WorldWideScience

Sample records for distributed parallel storage

  1. Globally distributed software defined storage (proposal)

    Science.gov (United States)

    Shevel, A.; Khoruzhnikov, S.; Grudinin, V.; Sadov, O.; Kairkanov, A.

    2017-10-01

    The volume of the coming data in HEP is growing. The volume of the data to be held for a long time is growing as well. Large volume of data - big data - is distributed around the planet. The methods, approaches how to organize and manage the globally distributed data storage are required. The distributed storage has several examples for personal needs like own-cloud.org, pydio.com, seafile.com, sparkleshare.org. For enterprise-level there is a number of systems: SWIFT - distributed storage systems (part of Openstack), CEPH and the like which are mostly object storage. When several data center’s resources are integrated, the organization of data links becomes very important issue especially if several parallel data links between data centers are used. The situation in data centers and in data links may vary each hour. All that means each part of distributed data storage has to be able to rearrange usage of data links and storage servers in each data center. In addition, for each customer of distributed storage different requirements could appear. The above topics are planned to be discussed in data storage proposal.

  2. Distributed Storage Algorithm for Geospatial Image Data Based on Data Access Patterns.

    Directory of Open Access Journals (Sweden)

    Shaoming Pan

    Full Text Available Declustering techniques are widely used in distributed environments to reduce query response time through parallel I/O by splitting large files into several small blocks and then distributing those blocks among multiple storage nodes. Unfortunately, however, many small geospatial image data files cannot be further split for distributed storage. In this paper, we propose a complete theoretical system for the distributed storage of small geospatial image data files based on mining the access patterns of geospatial image data using their historical access log information. First, an algorithm is developed to construct an access correlation matrix based on the analysis of the log information, which reveals the patterns of access to the geospatial image data. Then, a practical heuristic algorithm is developed to determine a reasonable solution based on the access correlation matrix. Finally, a number of comparative experiments are presented, demonstrating that our algorithm displays a higher total parallel access probability than those of other algorithms by approximately 10-15% and that the performance can be further improved by more than 20% by simultaneously applying a copy storage strategy. These experiments show that the algorithm can be applied in distributed environments to help realize parallel I/O and thereby improve system performance.

  3. Distributed Storage Algorithm for Geospatial Image Data Based on Data Access Patterns.

    Science.gov (United States)

    Pan, Shaoming; Li, Yongkai; Xu, Zhengquan; Chong, Yanwen

    2015-01-01

    Declustering techniques are widely used in distributed environments to reduce query response time through parallel I/O by splitting large files into several small blocks and then distributing those blocks among multiple storage nodes. Unfortunately, however, many small geospatial image data files cannot be further split for distributed storage. In this paper, we propose a complete theoretical system for the distributed storage of small geospatial image data files based on mining the access patterns of geospatial image data using their historical access log information. First, an algorithm is developed to construct an access correlation matrix based on the analysis of the log information, which reveals the patterns of access to the geospatial image data. Then, a practical heuristic algorithm is developed to determine a reasonable solution based on the access correlation matrix. Finally, a number of comparative experiments are presented, demonstrating that our algorithm displays a higher total parallel access probability than those of other algorithms by approximately 10-15% and that the performance can be further improved by more than 20% by simultaneously applying a copy storage strategy. These experiments show that the algorithm can be applied in distributed environments to help realize parallel I/O and thereby improve system performance.

  4. Study on parallel and distributed management of RS data based on spatial database

    Science.gov (United States)

    Chen, Yingbiao; Qian, Qinglan; Wu, Hongqiao; Liu, Shijin

    2009-10-01

    With the rapid development of current earth-observing technology, RS image data storage, management and information publication become a bottle-neck for its appliance and popularization. There are two prominent problems in RS image data storage and management system. First, background server hardly handle the heavy process of great capacity of RS data which stored at different nodes in a distributing environment. A tough burden has put on the background server. Second, there is no unique, standard and rational organization of Multi-sensor RS data for its storage and management. And lots of information is lost or not included at storage. Faced at the above two problems, the paper has put forward a framework for RS image data parallel and distributed management and storage system. This system aims at RS data information system based on parallel background server and a distributed data management system. Aiming at the above two goals, this paper has studied the following key techniques and elicited some revelatory conclusions. The paper has put forward a solid index of "Pyramid, Block, Layer, Epoch" according to the properties of RS image data. With the solid index mechanism, a rational organization for different resolution, different area, different band and different period of Multi-sensor RS image data is completed. In data storage, RS data is not divided into binary large objects to be stored at current relational database system, while it is reconstructed through the above solid index mechanism. A logical image database for the RS image data file is constructed. In system architecture, this paper has set up a framework based on a parallel server of several common computers. Under the framework, the background process is divided into two parts, the common WEB process and parallel process.

  5. Design and Implementation of Papyrus: Parallel Aggregate Persistent Storage

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Jungwon [ORNL; Sajjapongse, Kittisak [ORNL; Lee, Seyong [ORNL; Vetter, Jeffrey S [ORNL

    2017-01-01

    A surprising development in recently announced HPC platforms is the addition of, sometimes massive amounts of, persistent (nonvolatile) memory (NVM) in order to increase memory capacity and compensate for plateauing I/O capabilities. However, there are no portable and scalable programming interfaces using aggregate NVM effectively. This paper introduces Papyrus: a new software system built to exploit emerging capability of NVM in HPC architectures. Papyrus (or Parallel Aggregate Persistent -YRU- Storage) is a novel programming system that provides features for scalable, aggregate, persistent memory in an extreme-scale system for typical HPC usage scenarios. Papyrus mainly consists of Papyrus Virtual File System (VFS) and Papyrus Template Container Library (TCL). Papyrus VFS provides a uniform aggregate NVM storage image across diverse NVM architectures. It enables Papyrus TCL to provide a portable and scalable high-level container programming interface whose data elements are distributed across multiple NVM nodes without requiring the user to handle complex communication, synchronization, replication, and consistency model. We evaluate Papyrus on two HPC systems, including UTK Beacon and NERSC Cori, using real NVM storage devices.

  6. An Approach Using Parallel Architecture to Storage DICOM Images in Distributed File System

    International Nuclear Information System (INIS)

    Soares, Tiago S; Prado, Thiago C; Dantas, M A R; De Macedo, Douglas D J; Bauer, Michael A

    2012-01-01

    Telemedicine is a very important area in medical field that is expanding daily motivated by many researchers interested in improving medical applications. In Brazil was started in 2005, in the State of Santa Catarina has a developed server called the CyclopsDCMServer, which the purpose to embrace the HDF for the manipulation of medical images (DICOM) using a distributed file system. Since then, many researches were initiated in order to seek better performance. Our approach for this server represents an additional parallel implementation in I/O operations since HDF version 5 has an essential feature for our work which supports parallel I/O, based upon the MPI paradigm. Early experiments using four parallel nodes, provide good performance when compare to the serial HDF implemented in the CyclopsDCMServer.

  7. Distributed Parallel Architecture for "Big Data"

    Directory of Open Access Journals (Sweden)

    Catalin BOJA

    2012-01-01

    Full Text Available This paper is an extension to the "Distributed Parallel Architecture for Storing and Processing Large Datasets" paper presented at the WSEAS SEPADS’12 conference in Cambridge. In its original version the paper went over the benefits of using a distributed parallel architecture to store and process large datasets. This paper analyzes the problem of storing, processing and retrieving meaningful insight from petabytes of data. It provides a survey on current distributed and parallel data processing technologies and, based on them, will propose an architecture that can be used to solve the analyzed problem. In this version there is more emphasis put on distributed files systems and the ETL processes involved in a distributed environment.

  8. Optimal File-Distribution in Heterogeneous and Asymmetric Storage Networks

    Science.gov (United States)

    Langner, Tobias; Schindelhauer, Christian; Souza, Alexander

    We consider an optimisation problem which is motivated from storage virtualisation in the Internet. While storage networks make use of dedicated hardware to provide homogeneous bandwidth between servers and clients, in the Internet, connections between storage servers and clients are heterogeneous and often asymmetric with respect to upload and download. Thus, for a large file, the question arises how it should be fragmented and distributed among the servers to grant "optimal" access to the contents. We concentrate on the transfer time of a file, which is the time needed for one upload and a sequence of n downloads, using a set of m servers with heterogeneous bandwidths. We assume that fragments of the file can be transferred in parallel to and from multiple servers. This model yields a distribution problem that examines the question of how these fragments should be distributed onto those servers in order to minimise the transfer time. We present an algorithm, called FlowScaling, that finds an optimal solution within running time {O}(m log m). We formulate the distribution problem as a maximum flow problem, which involves a function that states whether a solution with a given transfer time bound exists. This function is then used with a scaling argument to determine an optimal solution within the claimed time complexity.

  9. Differences Between Distributed and Parallel Systems

    Energy Technology Data Exchange (ETDEWEB)

    Brightwell, R.; Maccabe, A.B.; Rissen, R.

    1998-10-01

    Distributed systems have been studied for twenty years and are now coming into wider use as fast networks and powerful workstations become more readily available. In many respects a massively parallel computer resembles a network of workstations and it is tempting to port a distributed operating system to such a machine. However, there are significant differences between these two environments and a parallel operating system is needed to get the best performance out of a massively parallel system. This report characterizes the differences between distributed systems, networks of workstations, and massively parallel systems and analyzes the impact of these differences on operating system design. In the second part of the report, we introduce Puma, an operating system specifically developed for massively parallel systems. We describe Puma portals, the basic building blocks for message passing paradigms implemented on top of Puma, and show how the differences observed in the first part of the report have influenced the design and implementation of Puma.

  10. Research on Multi - Person Parallel Modeling Method Based on Integrated Model Persistent Storage

    Science.gov (United States)

    Qu, MingCheng; Wu, XiangHu; Tao, YongChao; Liu, Ying

    2018-03-01

    This paper mainly studies the multi-person parallel modeling method based on the integrated model persistence storage. The integrated model refers to a set of MDDT modeling graphics system, which can carry out multi-angle, multi-level and multi-stage description of aerospace general embedded software. Persistent storage refers to converting the data model in memory into a storage model and converting the storage model into a data model in memory, where the data model refers to the object model and the storage model is a binary stream. And multi-person parallel modeling refers to the need for multi-person collaboration, the role of separation, and even real-time remote synchronization modeling.

  11. Current distribution characteristics of superconducting parallel circuits

    International Nuclear Information System (INIS)

    Mori, K.; Suzuki, Y.; Hara, N.; Kitamura, M.; Tominaka, T.

    1994-01-01

    In order to increase the current carrying capacity of the current path of the superconducting magnet system, the portion of parallel circuits such as insulated multi-strand cables or parallel persistent current switches (PCS) are made. In superconducting parallel circuits of an insulated multi-strand cable or a parallel persistent current switch (PCS), the current distribution during the current sweep, the persistent mode, and the quench process were investigated. In order to measure the current distribution, two methods were used. (1) Each strand was surrounded with a pure iron core with the air gap. In the air gap, a Hall probe was located. The accuracy of this method was deteriorated by the magnetic hysteresis of iron. (2) The Rogowski coil without iron was used for the current measurement of each path in a 4-parallel PCS. As a result, it was shown that the current distribution characteristics of a parallel PCS is very similar to that of an insulated multi-strand cable for the quench process

  12. Distributed Cooperative Current-Sharing Control of Parallel Chargers Using Feedback Linearization

    Directory of Open Access Journals (Sweden)

    Jiangang Liu

    2014-01-01

    Full Text Available We propose a distributed current-sharing scheme to address the output current imbalance problem for the parallel chargers in the energy storage type light rail vehicle system. By treating the parallel chargers as a group of agents with output information sharing through communication network, the current-sharing control problem is recast as the consensus tracking problem of multiagents. To facilitate the design, input-output feedback linearization is first applied to transform the nonidentical nonlinear charging system model into the first-order integrator. Then, a general saturation function is introduced to design the cooperative current-sharing control law which can guarantee the boundedness of the proposed control. The cooperative stability of the closed-loop system under fixed and dynamic communication topologies is rigorously proved with the aid of Lyapunov function and LaSalle invariant principle. Simulation using a multicharging test system further illustrates that the output currents of parallel chargers are balanced using the proposed control.

  13. Structured storage in ATLAS Distributed Data Management: use cases and experiences

    International Nuclear Information System (INIS)

    Lassnig, Mario; Garonne, Vincent; Beermann, Thomas; Dimitrov, Gancho; Canali, Luca; Molfetas, Angelos; Zang Donal; Azzurra Chinzer, Lisa

    2012-01-01

    The distributed data management system of the high-energy physics experiment ATLAS has a critical dependency on the Oracle Relational Database Management System. Recently however, the increased appearance of data warehouselike workload in the experiment has put considerable and increasing strain on the Oracle database. In particular, the analysis of archived data, and the aggregation of data for summary purposes has been especially demanding. For this reason, structured storage systems were evaluated to offload the Oracle database, and to handle processing of data in a non-transactional way. This includes distributed file systems like HDFS that support parallel execution of computational tasks on distributed data, as well as non-relational databases like HBase, Cassandra, or MongoDB. In this paper, the most important analysis and aggregation use cases of the data management system are presented, and how structured storage systems were established to process them.

  14. Locally Minimum Storage Regenerating Codes in Distributed Cloud Storage Systems

    Institute of Scientific and Technical Information of China (English)

    Jing Wang; Wei Luo; Wei Liang; Xiangyang Liu; Xiaodai Dong

    2017-01-01

    In distributed cloud storage sys-tems, inevitably there exist multiple node fail-ures at the same time. The existing methods of regenerating codes, including minimum storage regenerating (MSR) codes and mini-mum bandwidth regenerating (MBR) codes, are mainly to repair one single or several failed nodes, unable to meet the repair need of distributed cloud storage systems. In this paper, we present locally minimum storage re-generating (LMSR) codes to recover multiple failed nodes at the same time. Specifically, the nodes in distributed cloud storage systems are divided into multiple local groups, and in each local group (4, 2) or (5, 3) MSR codes are constructed. Moreover, the grouping method of storage nodes and the repairing process of failed nodes in local groups are studied. The-oretical analysis shows that LMSR codes can achieve the same storage overhead as MSR codes. Furthermore, we verify by means of simulation that, compared with MSR codes, LMSR codes can reduce the repair bandwidth and disk I/O overhead effectively.

  15. Autonomous operation of distributed storages in microgrids

    DEFF Research Database (Denmark)

    Loh, Poh Chiang; Chai, Yi Kai; Li, Ding

    2014-01-01

    Operation of distributed generators in microgrids has been widely discussed, but would not be fully autonomous if distributed energy storages are not considered. Storages are important since they provide energy buffering to load changes, energy levelling to source variations and ride-through enha......Operation of distributed generators in microgrids has been widely discussed, but would not be fully autonomous if distributed energy storages are not considered. Storages are important since they provide energy buffering to load changes, energy levelling to source variations and ride......-through enhancement to the microgrids. Recognising their importance, this study presents a scheme for sharing power among multiple distributed storages in coordination with the distributed sources and loads. The scheme prompts the storages to autonomously sense for local operating conditions, requesting for maximum...... to help with meeting the extra load demand. The described process takes place autonomously with energy eventually shared among the storages in proportion to their ratings. To test the concepts discussed, experiments have been performed with favourable results obtained for performance verification....

  16. Autonomous control of distributed storages in microgrids

    DEFF Research Database (Denmark)

    Loh, Poh Chiang; Blaabjerg, Frede

    2011-01-01

    Operation of distributed generators in microgrids has widely been discussed, but would not be fully autonomous, if distributed storages are not considered. Storages in general are important, since they provide energy buffering to load changes, energy leveling to source variations and ride......-through enhancement to the overall microgrids. Recognizing their importance, this paper presents a scheme for sharing power among multiple distributed storages, in coordination with the distributed sources and loads. The scheme prompts the storages to autonomously sense for system conditions, requesting for maximum...

  17. Automatic Management of Parallel and Distributed System Resources

    Science.gov (United States)

    Yan, Jerry; Ngai, Tin Fook; Lundstrom, Stephen F.

    1990-01-01

    Viewgraphs on automatic management of parallel and distributed system resources are presented. Topics covered include: parallel applications; intelligent management of multiprocessing systems; performance evaluation of parallel architecture; dynamic concurrent programs; compiler-directed system approach; lattice gaseous cellular automata; and sparse matrix Cholesky factorization.

  18. Intelligent Distributed Generation and Storage Units for DC Microgrids - A New Concept on Cooperative Control without Communications Beyond Droop Control

    DEFF Research Database (Denmark)

    Aldana, Nelson Leonardo Diaz; Dragicevic, Tomislav; Vasquez, Juan Carlos

    2014-01-01

    . Typically, voltage droop loops are used for interconnecting several different units in parallel to a microgrid. This paper proposes a new decentralized strategy based on fuzzy logic that ensures stored energy balance, for a low voltage DC microgrid with distributed battery energy storage systems......Low voltage DC microgrids have been widely used for supplying critical loads, such as data centers and remote communication stations. Consequently, it is important to ensure redundancy and enough energy capacity in order to support possible increments in load consumption. This is achieved by means...... of expansion of the energy storage system by adding extra distributed energy storage units. However, using distributed energy storage units adds more challenges in microgrids control, since stored energy should be balanced in order to avoid deep discharge or over-charge in one of the energy storage units...

  19. Methods and apparatus for capture and storage of semantic information with sub-files in a parallel computing system

    Science.gov (United States)

    Faibish, Sorin; Bent, John M; Tzelnic, Percy; Grider, Gary; Torres, Aaron

    2015-02-03

    Techniques are provided for storing files in a parallel computing system using sub-files with semantically meaningful boundaries. A method is provided for storing at least one file generated by a distributed application in a parallel computing system. The file comprises one or more of a complete file and a plurality of sub-files. The method comprises the steps of obtaining a user specification of semantic information related to the file; providing the semantic information as a data structure description to a data formatting library write function; and storing the semantic information related to the file with one or more of the sub-files in one or more storage nodes of the parallel computing system. The semantic information provides a description of data in the file. The sub-files can be replicated based on semantically meaningful boundaries.

  20. Storage, transmission and distribution of hydrogen

    Science.gov (United States)

    Kelley, J. H.; Hagler, R., Jr.

    1979-01-01

    Current practices and future requirements for the storage, transmission and distribution of hydrogen are reviewed in order to identify inadequacies to be corrected before hydrogen can achieve its full potential as a substitute for fossil fuels. Consideration is given to the storage of hydrogen in underground solution-mined salt caverns, portable high-pressure containers and dewars, pressure vessels and aquifers and as metal hydrides, hydrogen transmission in evacuated double-walled insulated containers and by pipeline, and distribution by truck and internal distribution networks. Areas for the improvement of these techniques are indicated, and these technological deficiencies, including materials development, low-cost storage and transmission methods, low-cost, long-life metal hydrides and novel methods for hydrogen storage, are presented as challenges for research and development.

  1. Distributed parallel messaging for multiprocessor systems

    Science.gov (United States)

    Chen, Dong; Heidelberger, Philip; Salapura, Valentina; Senger, Robert M; Steinmacher-Burrow, Burhard; Sugawara, Yutaka

    2013-06-04

    A method and apparatus for distributed parallel messaging in a parallel computing system. The apparatus includes, at each node of a multiprocessor network, multiple injection messaging engine units and reception messaging engine units, each implementing a DMA engine and each supporting both multiple packet injection into and multiple reception from a network, in parallel. The reception side of the messaging unit (MU) includes a switch interface enabling writing of data of a packet received from the network to the memory system. The transmission side of the messaging unit, includes switch interface for reading from the memory system when injecting packets into the network.

  2. Influence of Paralleling Dies and Paralleling Half-Bridges on Transient Current Distribution in Multichip Power Modules

    DEFF Research Database (Denmark)

    Li, Helong; Zhou, Wei; Wang, Xiongfei

    2018-01-01

    This paper addresses the transient current distribution in the multichip half-bridge power modules, where two types of paralleling connections with different current commutation mechanisms are considered: paralleling dies and paralleling half-bridges. It reveals that with paralleling dies, both t...

  3. Parallel paving: An algorithm for generating distributed, adaptive, all-quadrilateral meshes on parallel computers

    Energy Technology Data Exchange (ETDEWEB)

    Lober, R.R.; Tautges, T.J.; Vaughan, C.T.

    1997-03-01

    Paving is an automated mesh generation algorithm which produces all-quadrilateral elements. It can additionally generate these elements in varying sizes such that the resulting mesh adapts to a function distribution, such as an error function. While powerful, conventional paving is a very serial algorithm in its operation. Parallel paving is the extension of serial paving into parallel environments to perform the same meshing functions as conventional paving only on distributed, discretized models. This extension allows large, adaptive, parallel finite element simulations to take advantage of paving`s meshing capabilities for h-remap remeshing. A significantly modified version of the CUBIT mesh generation code has been developed to host the parallel paving algorithm and demonstrate its capabilities on both two dimensional and three dimensional surface geometries and compare the resulting parallel produced meshes to conventionally paved meshes for mesh quality and algorithm performance. Sandia`s {open_quotes}tiling{close_quotes} dynamic load balancing code has also been extended to work with the paving algorithm to retain parallel efficiency as subdomains undergo iterative mesh refinement.

  4. Parallel file system performances in fusion data storage

    International Nuclear Information System (INIS)

    Iannone, F.; Podda, S.; Bracco, G.; Manduchi, G.; Maslennikov, A.; Migliori, S.; Wolkersdorfer, K.

    2012-01-01

    High I/O flow rates, up to 10 GB/s, are required in large fusion Tokamak experiments like ITER where hundreds of nodes store simultaneously large amounts of data acquired during the plasma discharges. Typical network topologies such as linear arrays (systolic), rings, meshes (2-D arrays), tori (3-D arrays), trees, butterfly, hypercube in combination with high speed data transports like Infiniband or 10G-Ethernet, are the main areas in which the effort to overcome the so-called parallel I/O bottlenecks is most focused. The high I/O flow rates were modelled in an emulated testbed based on the parallel file systems such as Lustre and GPFS, commonly used in High Performance Computing. The test runs on High Performance Computing–For Fusion (8640 cores) and ENEA CRESCO (3392 cores) supercomputers. Message Passing Interface based applications were developed to emulate parallel I/O on Lustre and GPFS using data archival and access solutions like MDSPLUS and Universal Access Layer. These methods of data storage organization are widely diffused in nuclear fusion experiments and are being developed within the EFDA Integrated Tokamak Modelling – Task Force; the authors tried to evaluate their behaviour in a realistic emulation setup.

  5. Parallel file system performances in fusion data storage

    Energy Technology Data Exchange (ETDEWEB)

    Iannone, F., E-mail: francesco.iannone@enea.it [Associazione EURATOM-ENEA sulla Fusione, C.R.ENEA Frascati, via E.Fermi, 45 - 00044 Frascati, Rome (Italy); Podda, S.; Bracco, G. [ENEA Information Communication Tecnologies, Lungotevere Thaon di Revel, 76 - 00196 Rome (Italy); Manduchi, G. [Associazione EURATOM-ENEA sulla Fusione, Consorzio RFX, Corso Stati Uniti, 4 - 35127 Padua (Italy); Maslennikov, A. [CASPUR Inter-University Consortium for the Application of Super-Computing for Research, via dei Tizii, 6b - 00185 Rome (Italy); Migliori, S. [ENEA Information Communication Tecnologies, Lungotevere Thaon di Revel, 76 - 00196 Rome (Italy); Wolkersdorfer, K. [Juelich Supercomputing Centre-FZJ, D-52425 Juelich (Germany)

    2012-12-15

    High I/O flow rates, up to 10 GB/s, are required in large fusion Tokamak experiments like ITER where hundreds of nodes store simultaneously large amounts of data acquired during the plasma discharges. Typical network topologies such as linear arrays (systolic), rings, meshes (2-D arrays), tori (3-D arrays), trees, butterfly, hypercube in combination with high speed data transports like Infiniband or 10G-Ethernet, are the main areas in which the effort to overcome the so-called parallel I/O bottlenecks is most focused. The high I/O flow rates were modelled in an emulated testbed based on the parallel file systems such as Lustre and GPFS, commonly used in High Performance Computing. The test runs on High Performance Computing-For Fusion (8640 cores) and ENEA CRESCO (3392 cores) supercomputers. Message Passing Interface based applications were developed to emulate parallel I/O on Lustre and GPFS using data archival and access solutions like MDSPLUS and Universal Access Layer. These methods of data storage organization are widely diffused in nuclear fusion experiments and are being developed within the EFDA Integrated Tokamak Modelling - Task Force; the authors tried to evaluate their behaviour in a realistic emulation setup.

  6. Storage and distribution/Linear programming for storage operations

    Energy Technology Data Exchange (ETDEWEB)

    Coleman, D

    1978-07-15

    The techniques of linear programing to solve storage problems as applied in a tank farm tie-in with refinery throughput operation include: (1) the time-phased model which works on storage and refinery operations input parameters, e.g., production, distribution, cracking, etc., and is capable of representing product stockpiling in slack periods to meet future peak demands, and investigating alternative strategies such as exchange deals and purchase and leasing of additional storage, and (2) the Monte Carlo simulation method, which inputs parameters, e.g., arrival of crude products at refinery, tankage size, likely demand for products, etc., as probability distributions rather than single values, and is capable of showing the average utilization of facilities, potential bottlenecks, investment required to achieve an increase in utilization, and to enable the user to predict total investment, cash flow, and profit emanating from the original financing decision. The increasing use of computer techniques to solve refinery and storage problems is attributed to potential savings resulting from more effective planning, reduced computer costs, ease of access and more usable software. Diagrams.

  7. Rocket measurement of auroral partial parallel distribution functions

    Science.gov (United States)

    Lin, C.-A.

    1980-01-01

    The auroral partial parallel distribution functions are obtained by using the observed energy spectra of electrons. The experiment package was launched by a Nike-Tomahawk rocket from Poker Flat, Alaska over a bright auroral band and covered an altitude range of up to 180 km. Calculated partial distribution functions are presented with emphasis on their slopes. The implications of the slopes are discussed. It should be pointed out that the slope of the partial parallel distribution function obtained from one energy spectra will be changed by superposing another energy spectra on it.

  8. Parallel Harmony Search Based Distributed Energy Resource Optimization

    Energy Technology Data Exchange (ETDEWEB)

    Ceylan, Oguzhan [ORNL; Liu, Guodong [ORNL; Tomsovic, Kevin [University of Tennessee, Knoxville (UTK)

    2015-01-01

    This paper presents a harmony search based parallel optimization algorithm to minimize voltage deviations in three phase unbalanced electrical distribution systems and to maximize active power outputs of distributed energy resources (DR). The main contribution is to reduce the adverse impacts on voltage profile during a day as photovoltaics (PVs) output or electrical vehicles (EVs) charging changes throughout a day. The IEEE 123- bus distribution test system is modified by adding DRs and EVs under different load profiles. The simulation results show that by using parallel computing techniques, heuristic methods may be used as an alternative optimization tool in electrical power distribution systems operation.

  9. Parallel Breadth-First Search on Distributed Memory Systems

    Energy Technology Data Exchange (ETDEWEB)

    Computational Research Division; Buluc, Aydin; Madduri, Kamesh

    2011-04-15

    Data-intensive, graph-based computations are pervasive in several scientific applications, and are known to to be quite challenging to implement on distributed memory systems. In this work, we explore the design space of parallel algorithms for Breadth-First Search (BFS), a key subroutine in several graph algorithms. We present two highly-tuned par- allel approaches for BFS on large parallel systems: a level-synchronous strategy that relies on a simple vertex-based partitioning of the graph, and a two-dimensional sparse matrix- partitioning-based approach that mitigates parallel commu- nication overhead. For both approaches, we also present hybrid versions with intra-node multithreading. Our novel hybrid two-dimensional algorithm reduces communication times by up to a factor of 3.5, relative to a common vertex based approach. Our experimental study identifies execu- tion regimes in which these approaches will be competitive, and we demonstrate extremely high performance on lead- ing distributed-memory parallel systems. For instance, for a 40,000-core parallel execution on Hopper, an AMD Magny- Cours based system, we achieve a BFS performance rate of 17.8 billion edge visits per second on an undirected graph of 4.3 billion vertices and 68.7 billion edges with skewed degree distribution.

  10. A compositional reservoir simulator on distributed memory parallel computers

    International Nuclear Information System (INIS)

    Rame, M.; Delshad, M.

    1995-01-01

    This paper presents the application of distributed memory parallel computes to field scale reservoir simulations using a parallel version of UTCHEM, The University of Texas Chemical Flooding Simulator. The model is a general purpose highly vectorized chemical compositional simulator that can simulate a wide range of displacement processes at both field and laboratory scales. The original simulator was modified to run on both distributed memory parallel machines (Intel iPSC/960 and Delta, Connection Machine 5, Kendall Square 1 and 2, and CRAY T3D) and a cluster of workstations. A domain decomposition approach has been taken towards parallelization of the code. A portion of the discrete reservoir model is assigned to each processor by a set-up routine that attempts a data layout as even as possible from the load-balance standpoint. Each of these subdomains is extended so that data can be shared between adjacent processors for stencil computation. The added routines that make parallel execution possible are written in a modular fashion that makes the porting to new parallel platforms straight forward. Results of the distributed memory computing performance of Parallel simulator are presented for field scale applications such as tracer flood and polymer flood. A comparison of the wall-clock times for same problems on a vector supercomputer is also presented

  11. Distributed coordination of energy storage with distributed generators

    NARCIS (Netherlands)

    Yang, Tao; Wu, Di; Stoorvogel, Antonie Arij; Stoustrup, Jakob

    2016-01-01

    With a growing emphasis on energy efficiency and system flexibility, a great effort has been made recently in developing distributed energy resources (DER), including distributed generators and energy storage systems. This paper first formulates an optimal DER coordination problem considering

  12. Massively Parallel Finite Element Programming

    KAUST Repository

    Heister, Timo; Kronbichler, Martin; Bangerth, Wolfgang

    2010-01-01

    Today's large finite element simulations require parallel algorithms to scale on clusters with thousands or tens of thousands of processor cores. We present data structures and algorithms to take advantage of the power of high performance computers in generic finite element codes. Existing generic finite element libraries often restrict the parallelization to parallel linear algebra routines. This is a limiting factor when solving on more than a few hundreds of cores. We describe routines for distributed storage of all major components coupled with efficient, scalable algorithms. We give an overview of our effort to enable the modern and generic finite element library deal.II to take advantage of the power of large clusters. In particular, we describe the construction of a distributed mesh and develop algorithms to fully parallelize the finite element calculation. Numerical results demonstrate good scalability. © 2010 Springer-Verlag.

  13. Massively Parallel Finite Element Programming

    KAUST Repository

    Heister, Timo

    2010-01-01

    Today\\'s large finite element simulations require parallel algorithms to scale on clusters with thousands or tens of thousands of processor cores. We present data structures and algorithms to take advantage of the power of high performance computers in generic finite element codes. Existing generic finite element libraries often restrict the parallelization to parallel linear algebra routines. This is a limiting factor when solving on more than a few hundreds of cores. We describe routines for distributed storage of all major components coupled with efficient, scalable algorithms. We give an overview of our effort to enable the modern and generic finite element library deal.II to take advantage of the power of large clusters. In particular, we describe the construction of a distributed mesh and develop algorithms to fully parallelize the finite element calculation. Numerical results demonstrate good scalability. © 2010 Springer-Verlag.

  14. Paging memory from random access memory to backing storage in a parallel computer

    Science.gov (United States)

    Archer, Charles J; Blocksome, Michael A; Inglett, Todd A; Ratterman, Joseph D; Smith, Brian E

    2013-05-21

    Paging memory from random access memory (`RAM`) to backing storage in a parallel computer that includes a plurality of compute nodes, including: executing a data processing application on a virtual machine operating system in a virtual machine on a first compute node; providing, by a second compute node, backing storage for the contents of RAM on the first compute node; and swapping, by the virtual machine operating system in the virtual machine on the first compute node, a page of memory from RAM on the first compute node to the backing storage on the second compute node.

  15. Parallelization and automatic data distribution for nuclear reactor simulations

    Energy Technology Data Exchange (ETDEWEB)

    Liebrock, L.M. [Liebrock-Hicks Research, Calumet, MI (United States)

    1997-07-01

    Detailed attempts at realistic nuclear reactor simulations currently take many times real time to execute on high performance workstations. Even the fastest sequential machine can not run these simulations fast enough to ensure that the best corrective measure is used during a nuclear accident to prevent a minor malfunction from becoming a major catastrophe. Since sequential computers have nearly reached the speed of light barrier, these simulations will have to be run in parallel to make significant improvements in speed. In physical reactor plants, parallelism abounds. Fluids flow, controls change, and reactions occur in parallel with only adjacent components directly affecting each other. These do not occur in the sequentialized manner, with global instantaneous effects, that is often used in simulators. Development of parallel algorithms that more closely approximate the real-world operation of a reactor may, in addition to speeding up the simulations, actually improve the accuracy and reliability of the predictions generated. Three types of parallel architecture (shared memory machines, distributed memory multicomputers, and distributed networks) are briefly reviewed as targets for parallelization of nuclear reactor simulation. Various parallelization models (loop-based model, shared memory model, functional model, data parallel model, and a combined functional and data parallel model) are discussed along with their advantages and disadvantages for nuclear reactor simulation. A variety of tools are introduced for each of the models. Emphasis is placed on the data parallel model as the primary focus for two-phase flow simulation. Tools to support data parallel programming for multiple component applications and special parallelization considerations are also discussed.

  16. Parallelization and automatic data distribution for nuclear reactor simulations

    International Nuclear Information System (INIS)

    Liebrock, L.M.

    1997-01-01

    Detailed attempts at realistic nuclear reactor simulations currently take many times real time to execute on high performance workstations. Even the fastest sequential machine can not run these simulations fast enough to ensure that the best corrective measure is used during a nuclear accident to prevent a minor malfunction from becoming a major catastrophe. Since sequential computers have nearly reached the speed of light barrier, these simulations will have to be run in parallel to make significant improvements in speed. In physical reactor plants, parallelism abounds. Fluids flow, controls change, and reactions occur in parallel with only adjacent components directly affecting each other. These do not occur in the sequentialized manner, with global instantaneous effects, that is often used in simulators. Development of parallel algorithms that more closely approximate the real-world operation of a reactor may, in addition to speeding up the simulations, actually improve the accuracy and reliability of the predictions generated. Three types of parallel architecture (shared memory machines, distributed memory multicomputers, and distributed networks) are briefly reviewed as targets for parallelization of nuclear reactor simulation. Various parallelization models (loop-based model, shared memory model, functional model, data parallel model, and a combined functional and data parallel model) are discussed along with their advantages and disadvantages for nuclear reactor simulation. A variety of tools are introduced for each of the models. Emphasis is placed on the data parallel model as the primary focus for two-phase flow simulation. Tools to support data parallel programming for multiple component applications and special parallelization considerations are also discussed

  17. An Educational Tool for Interactive Parallel and Distributed Processing

    DEFF Research Database (Denmark)

    Pagliarini, Luigi; Lund, Henrik Hautop

    2011-01-01

    In this paper we try to describe how the Modular Interactive Tiles System (MITS) can be a valuable tool for introducing students to interactive parallel and distributed processing programming. This is done by providing an educational hands-on tool that allows a change of representation of the abs......In this paper we try to describe how the Modular Interactive Tiles System (MITS) can be a valuable tool for introducing students to interactive parallel and distributed processing programming. This is done by providing an educational hands-on tool that allows a change of representation...... of the abstract problems related to designing interactive parallel and distributed systems. Indeed, MITS seems to bring a series of goals into the education, such as parallel programming, distributedness, communication protocols, master dependency, software behavioral models, adaptive interactivity, feedback......, connectivity, topology, island modeling, user and multiuser interaction, which can hardly be found in other tools. Finally, we introduce the system of modular interactive tiles as a tool for easy, fast, and flexible hands-on exploration of these issues, and through examples show how to implement interactive...

  18. Distributed Cloud Storage Using Network Coding

    OpenAIRE

    Sipos, Marton A.; Fitzek, Frank; Roetter, Daniel Enrique Lucani; Pedersen, Morten Videbæk

    2014-01-01

    Distributed storage is usually considered within acloud provider to ensure availability and reliability of the data.However, the user is still directly dependent on the quality of asingle system. It is also entrusting the service provider with largeamounts of private data, which may be accessed by a successfulattack to that cloud system or even be inspected by governmentagencies in some countries. This paper advocates a generalframework for network coding enabled distributed storage overmulti...

  19. Lempel–Ziv Data Compression on Parallel and Distributed Systems

    Directory of Open Access Journals (Sweden)

    Sergio De Agostino

    2011-09-01

    Full Text Available We present a survey of results concerning Lempel–Ziv data compression on parallel and distributed systems, starting from the theoretical approach to parallel time complexity to conclude with the practical goal of designing distributed algorithms with low communication cost. Storer’s extension for image compression is also discussed.

  20. Parallel multi-join query optimization algorithm for distributed sensor network in the internet of things

    Science.gov (United States)

    Zheng, Yan

    2015-03-01

    Internet of things (IoT), focusing on providing users with information exchange and intelligent control, attracts a lot of attention of researchers from all over the world since the beginning of this century. IoT is consisted of large scale of sensor nodes and data processing units, and the most important features of IoT can be illustrated as energy confinement, efficient communication and high redundancy. With the sensor nodes increment, the communication efficiency and the available communication band width become bottle necks. Many research work is based on the instance which the number of joins is less. However, it is not proper to the increasing multi-join query in whole internet of things. To improve the communication efficiency between parallel units in the distributed sensor network, this paper proposed parallel query optimization algorithm based on distribution attributes cost graph. The storage information relations and the network communication cost are considered in this algorithm, and an optimized information changing rule is established. The experimental result shows that the algorithm has good performance, and it would effectively use the resource of each node in the distributed sensor network. Therefore, executive efficiency of multi-join query between different nodes could be improved.

  1. Transparent Distribution in a Storage Manager

    NARCIS (Netherlands)

    J.S. Karlsson; M.L. Kersten (Martin)

    1998-01-01

    textabstractScalable Distributed Data Structures (SDDSs) provide a self-managing and self-organizing data storage of potentially unbounded size. This stands in contrast to common distribution schemas deployed in conventional distributed DBMS. SDDSs, however, have mostly been used in synthetic

  2. Ceph, a distributed storage system for scientific computing

    CERN Multimedia

    CERN. Geneva

    2013-01-01

    Ceph is a distributed storage system designed to providing high performance and reliability at scales of up to thousands of storage nodes. The system is based on a distributed object storage layer call RADOS that provides durability, availability, efficient data distribution, and rich object semantics. This storage can be consumed directly via an object-based interface, or via file, block, or REST-based object services that are built on top of it. Clusters are composed of commodity components to provide a reliable storage service serving multiple use-cases. This seminar will cover the basic architecture of Ceph, with a focus on how each service can be consumed in a research and infrastructure environment. About the speaker Sage Weil, Founder and current CTO of Inktank Inc, is the creator of the Ceph project. He originally designed it as part of his PhD research in Storage Systems at the University of California, Santa Cruz. Since graduating, he has continued to refine the system with the goal of providi...

  3. Parallel and distributed processing: applications to power systems

    Energy Technology Data Exchange (ETDEWEB)

    Wu, Felix; Murphy, Liam [California Univ., Berkeley, CA (United States). Dept. of Electrical Engineering and Computer Sciences

    1994-12-31

    Applications of parallel and distributed processing to power systems problems are still in the early stages. Rapid progress in computing and communications promises a revolutionary increase in the capacity of distributed processing systems. In this paper, the state-of-the art in distributed processing technology and applications is reviewed and future trends are discussed. (author) 14 refs.,1 tab.

  4. Analysis of flow distribution instability in parallel thin rectangular multi-channel system

    Energy Technology Data Exchange (ETDEWEB)

    Xia, G.L. [School of Nuclear Science and Technology, Xi’an Jiaotong University, Xi’an City 710049 (China); Fundamental Science on Nuclear Safety and Simulation Technology Laboratory, Harbin Engineering University, Harbin City 150001 (China); Su, G.H., E-mail: ghsu@mail.xjtu.edu.cn [School of Nuclear Science and Technology, Xi’an Jiaotong University, Xi’an City 710049 (China); Peng, M.J. [Fundamental Science on Nuclear Safety and Simulation Technology Laboratory, Harbin Engineering University, Harbin City 150001 (China)

    2016-08-15

    Highlights: • Flow distribution instability in parallel thin rectangular multi-channel system is studied using RELAP5 codes. • Flow excursion may bring parallel heating channel into the density wave oscillations region. • Flow distribution instability is more likely to happen at low power/flow ratio conditions. • The increase of channel number will not affect the flow distribution instability boundary. • Asymmetry inlet throttling and heating will make system more unstable. - Abstract: The flow distribution instability in parallel thin rectangular multi-channel system has been researched in the present study. The research model of parallel channel system is established by using RELAP5/MOD3.4 codes. The transient process of flow distribution instability is studied at imposed inlet mass flow rate and imposed pressure drop conditions. The influence of heating power, mass flow rate, system pressure and channel number on flow distribution instability are analyzed. Furthermore, the flow distribution instability of parallel two-channel system under asymmetric inlet throttling and heating power is studied. The results show that, if multi-channel system operates at the negative slope region of channel ΔP–G curve, small disturbance in pressure drop will lead to flow redistribution between parallel channels. Flow excursion may bring the operating point of heating channel into the density-wave oscillations region, this will result in out-phase or in-phase flow oscillations. Flow distribution instability is more likely to happen at low power/flow ratio conditions, the stability of parallel channel system increases with system pressure, the channel number has a little effect on system stability, but the asymmetry inlet throttling or heating power will make the system more unstable.

  5. Modeling and optimization of parallel and distributed embedded systems

    CERN Document Server

    Munir, Arslan; Ranka, Sanjay

    2016-01-01

    This book introduces the state-of-the-art in research in parallel and distributed embedded systems, which have been enabled by developments in silicon technology, micro-electro-mechanical systems (MEMS), wireless communications, computer networking, and digital electronics. These systems have diverse applications in domains including military and defense, medical, automotive, and unmanned autonomous vehicles. The emphasis of the book is on the modeling and optimization of emerging parallel and distributed embedded systems in relation to the three key design metrics of performance, power and dependability.

  6. Event parallelism: Distributed memory parallel computing for high energy physics experiments

    International Nuclear Information System (INIS)

    Nash, T.

    1989-05-01

    This paper describes the present and expected future development of distributed memory parallel computers for high energy physics experiments. It covers the use of event parallel microprocessor farms, particularly at Fermilab, including both ACP multiprocessors and farms of MicroVAXES. These systems have proven very cost effective in the past. A case is made for moving to the more open environment of UNIX and RISC processors. The 2nd Generation ACP Multiprocessor System, which is based on powerful RISC systems, is described. Given the promise of still more extraordinary increases in processor performance, a new emphasis on point to point, rather than bussed, communication will be required. Developments in this direction are described. 6 figs

  7. Event parallelism: Distributed memory parallel computing for high energy physics experiments

    International Nuclear Information System (INIS)

    Nash, T.

    1989-01-01

    This paper describes the present and expected future development of distributed memory parallel computers for high energy physics experiments. It covers the use of event parallel microprocessor farms, particularly at Fermilab, including both ACP multiprocessors and farms of MicroVAXES. These systems have proven very cost effective in the past. A case is made for moving to the more open environment of UNIX and RISC processors. The 2nd Generation ACP Multiprocessor System, which is based on powerful RISC systems, is described. Given the promise of still more extraordinary increases in processor performance, a new emphasis on point to point, rather than bussed, communication will be required. Developments in this direction are described. (orig.)

  8. Event parallelism: Distributed memory parallel computing for high energy physics experiments

    Science.gov (United States)

    Nash, Thomas

    1989-12-01

    This paper describes the present and expected future development of distributed memory parallel computers for high energy physics experiments. It covers the use of event parallel microprocessor farms, particularly at Fermilab, including both ACP multiprocessors and farms of MicroVAXES. These systems have proven very cost effective in the past. A case is made for moving to the more open environment of UNIX and RISC processors. The 2nd Generation ACP Multiprocessor System, which is based on powerful RISC system, is described. Given the promise of still more extraordinary increases in processor performance, a new emphasis on point to point, rather than bussed, communication will be required. Developments in this direction are described.

  9. Parallel and distributed processing in power system simulation and control

    Energy Technology Data Exchange (ETDEWEB)

    Falcao, Djalma M [Universidade Federal, Rio de Janeiro, RJ (Brazil). Coordenacao dos Programas de Pos-graduacao de Engenharia

    1994-12-31

    Recent advances in computer technology will certainly have a great impact in the methodologies used in power system expansion and operational planning as well as in real-time control. Parallel and distributed processing are among the new technologies that present great potential for application in these areas. Parallel computers use multiple functional or processing units to speed up computation while distributed processing computer systems are collection of computers joined together by high speed communication networks having many objectives and advantages. The paper presents some ideas for the use of parallel and distributed processing in power system simulation and control. It also comments on some of the current research work in these topics and presents a summary of the work presently being developed at COPPE. (author) 53 refs., 2 figs.

  10. Distributed Large Data-Object Environments: End-to-End Performance Analysis of High Speed Distributed Storage Systems in Wide Area ATM Networks

    Science.gov (United States)

    Johnston, William; Tierney, Brian; Lee, Jason; Hoo, Gary; Thompson, Mary

    1996-01-01

    We have developed and deployed a distributed-parallel storage system (DPSS) in several high speed asynchronous transfer mode (ATM) wide area networks (WAN) testbeds to support several different types of data-intensive applications. Architecturally, the DPSS is a network striped disk array, but is fairly unique in that its implementation allows applications complete freedom to determine optimal data layout, replication and/or coding redundancy strategy, security policy, and dynamic reconfiguration. In conjunction with the DPSS, we have developed a 'top-to-bottom, end-to-end' performance monitoring and analysis methodology that has allowed us to characterize all aspects of the DPSS operating in high speed ATM networks. In particular, we have run a variety of performance monitoring experiments involving the DPSS in the MAGIC testbed, which is a large scale, high speed, ATM network and we describe our experience using the monitoring methodology to identify and correct problems that limit the performance of high speed distributed applications. Finally, the DPSS is part of an overall architecture for using high speed, WAN's for enabling the routine, location independent use of large data-objects. Since this is part of the motivation for a distributed storage system, we describe this architecture.

  11. A general purpose subroutine for fast fourier transform on a distributed memory parallel machine

    Science.gov (United States)

    Dubey, A.; Zubair, M.; Grosch, C. E.

    1992-01-01

    One issue which is central in developing a general purpose Fast Fourier Transform (FFT) subroutine on a distributed memory parallel machine is the data distribution. It is possible that different users would like to use the FFT routine with different data distributions. Thus, there is a need to design FFT schemes on distributed memory parallel machines which can support a variety of data distributions. An FFT implementation on a distributed memory parallel machine which works for a number of data distributions commonly encountered in scientific applications is presented. The problem of rearranging the data after computing the FFT is also addressed. The performance of the implementation on a distributed memory parallel machine Intel iPSC/860 is evaluated.

  12. Efficient implementation of multidimensional fast fourier transform on a distributed-memory parallel multi-node computer

    Science.gov (United States)

    Bhanot, Gyan V [Princeton, NJ; Chen, Dong [Croton-On-Hudson, NY; Gara, Alan G [Mount Kisco, NY; Giampapa, Mark E [Irvington, NY; Heidelberger, Philip [Cortlandt Manor, NY; Steinmacher-Burow, Burkhard D [Mount Kisco, NY; Vranas, Pavlos M [Bedford Hills, NY

    2012-01-10

    The present in invention is directed to a method, system and program storage device for efficiently implementing a multidimensional Fast Fourier Transform (FFT) of a multidimensional array comprising a plurality of elements initially distributed in a multi-node computer system comprising a plurality of nodes in communication over a network, comprising: distributing the plurality of elements of the array in a first dimension across the plurality of nodes of the computer system over the network to facilitate a first one-dimensional FFT; performing the first one-dimensional FFT on the elements of the array distributed at each node in the first dimension; re-distributing the one-dimensional FFT-transformed elements at each node in a second dimension via "all-to-all" distribution in random order across other nodes of the computer system over the network; and performing a second one-dimensional FFT on elements of the array re-distributed at each node in the second dimension, wherein the random order facilitates efficient utilization of the network thereby efficiently implementing the multidimensional FFT. The "all-to-all" re-distribution of array elements is further efficiently implemented in applications other than the multidimensional FFT on the distributed-memory parallel supercomputer.

  13. An educational tool for interactive parallel and distributed processing

    DEFF Research Database (Denmark)

    Pagliarini, Luigi; Lund, Henrik Hautop

    2012-01-01

    In this article we try to describe how the modular interactive tiles system (MITS) can be a valuable tool for introducing students to interactive parallel and distributed processing programming. This is done by providing a handson educational tool that allows a change in the representation...... of abstract problems related to designing interactive parallel and distributed systems. Indeed, the MITS seems to bring a series of goals into education, such as parallel programming, distributedness, communication protocols, master dependency, software behavioral models, adaptive interactivity, feedback......, connectivity, topology, island modeling, and user and multi-user interaction which can rarely be found in other tools. Finally, we introduce the system of modular interactive tiles as a tool for easy, fast, and flexible hands-on exploration of these issues, and through examples we show how to implement...

  14. Distributed parallel computing in stochastic modeling of groundwater systems.

    Science.gov (United States)

    Dong, Yanhui; Li, Guomin; Xu, Haizhen

    2013-03-01

    Stochastic modeling is a rapidly evolving, popular approach to the study of the uncertainty and heterogeneity of groundwater systems. However, the use of Monte Carlo-type simulations to solve practical groundwater problems often encounters computational bottlenecks that hinder the acquisition of meaningful results. To improve the computational efficiency, a system that combines stochastic model generation with MODFLOW-related programs and distributed parallel processing is investigated. The distributed computing framework, called the Java Parallel Processing Framework, is integrated into the system to allow the batch processing of stochastic models in distributed and parallel systems. As an example, the system is applied to the stochastic delineation of well capture zones in the Pinggu Basin in Beijing. Through the use of 50 processing threads on a cluster with 10 multicore nodes, the execution times of 500 realizations are reduced to 3% compared with those of a serial execution. Through this application, the system demonstrates its potential in solving difficult computational problems in practical stochastic modeling. © 2012, The Author(s). Groundwater © 2012, National Ground Water Association.

  15. A distributed storage system with dCache

    DEFF Research Database (Denmark)

    Behrmann, Gerd; Fuhrmann, Patrick; Grønager, Michael

    2008-01-01

    The LCG collaboration is encompassed by a number of Tier 1 centers. The Nordic LCG Tier 1, operated by NDGF, is in contrast to many other Tier 1 centers distributed over the Nordic countries. A distributed setup was chosen for both political and technical reasons, but also provides a number...... of unique challenges. dCache is well known and respected as a powerful distributed storage resource manager, and was chosen for implementing the storage aspects of the Nordic Tier 1. In contrast to classic dCache deployments, we deploy dCache over a WAN with limited bandwidth, high latency, frequent network...

  16. From parallel to distributed computing for reactive scattering calculations

    International Nuclear Information System (INIS)

    Lagana, A.; Gervasi, O.; Baraglia, R.

    1994-01-01

    Some reactive scattering codes have been ported on different innovative computer architectures ranging from massively parallel machines to clustered workstations. The porting has required a drastic restructuring of the codes to single out computationally decoupled cpu intensive subsections. The suitability of different theoretical approaches for parallel and distributed computing restructuring is discussed and the efficiency of related algorithms evaluated

  17. A strategy for load balancing in distributed storage systems

    CERN Multimedia

    CERN. Geneva

    2012-01-01

    Distributed storage systems are critical to the operation of the WLCG. These systems are not limited to fulfilling the long term storage requirements. They also serve data for computational analysis and other computational jobs. Distributed storage systems provide the ability to aggregate the storage and IO capacity of disks and tapes, but at the end of the day IO rate is still bound by the capabilities of the hardware, in particular the hard drives. Throughput of hard drives has increased dramatically over the decades, however for computational analysis IOPS is typically the limiting factor. To maximize return of investment, balancing IO load over available hardware is crucial. The task is made complicated by the common use of heterogeneous hardware and software environments that results from combining new and old hardware into a single storage system. This paper describes recent advances made in load balancing in the dCache distributed storage system. We describe a set of common requirements for load balan...

  18. A novel two-level dynamic parallel data scheme for large 3-D SN calculations

    International Nuclear Information System (INIS)

    Sjoden, G.E.; Shedlock, D.; Haghighat, A.; Yi, C.

    2005-01-01

    We introduce a new dynamic parallel memory optimization scheme for executing large scale 3-D discrete ordinates (Sn) simulations on distributed memory parallel computers. In order for parallel transport codes to be truly scalable, they must use parallel data storage, where only the variables that are locally computed are locally stored. Even with parallel data storage for the angular variables, cumulative storage requirements for large discrete ordinates calculations can be prohibitive. To address this problem, Memory Tuning has been implemented into the PENTRAN 3-D parallel discrete ordinates code as an optimized, two-level ('large' array, 'small' array) parallel data storage scheme. Memory Tuning can be described as the process of parallel data memory optimization. Memory Tuning dynamically minimizes the amount of required parallel data in allocated memory on each processor using a statistical sampling algorithm. This algorithm is based on the integral average and standard deviation of the number of fine meshes contained in each coarse mesh in the global problem. Because PENTRAN only stores the locally computed problem phase space, optimal two-level memory assignments can be unique on each node, depending upon the parallel decomposition used (hybrid combinations of angular, energy, or spatial). As demonstrated in the two large discrete ordinates models presented (a storage cask and an OECD MOX Benchmark), Memory Tuning can save a substantial amount of memory per parallel processor, allowing one to accomplish very large scale Sn computations. (authors)

  19. Prototyping and Simulating Parallel, Distributed Computations with VISA

    National Research Council Canada - National Science Library

    Demeure, Isabelle M; Nutt, Gary J

    1989-01-01

    ...] to support the design, prototyping, and simulation of parallel, distributed computations. In particular, VISA is meant to guide the choice of partitioning and communication strategies for such computations, based on their performance...

  20. The Design of Distributed Micro Grid Energy Storage System

    Science.gov (United States)

    Liang, Ya-feng; Wang, Yan-ping

    2018-03-01

    Distributed micro-grid runs in island mode, the energy storage system is the core to maintain the micro-grid stable operation. For the problems that it is poor to adjust at work and easy to cause the volatility of micro-grid caused by the existing energy storage structure of fixed connection. In this paper, an array type energy storage structure is proposed, and the array type energy storage system structure and working principle are analyzed. Finally, the array type energy storage structure model is established based on MATLAB, the simulation results show that the array type energy storage system has great flexibility, which can maximize the utilization of energy storage system, guarantee the reliable operation of distributed micro-grid and achieve the function of peak clipping and valley filling.

  1. Parallel and distributed processing in two SGBDS: A case study

    OpenAIRE

    Francisco Javier Moreno; Nataly Castrillón Charari; Camilo Taborda Zuluaga

    2017-01-01

    Context: One of the strategies for managing large volumes of data is distributed and parallel computing. Among the tools that allow applying these characteristics are some Data Base Management Systems (DBMS), such as Oracle, DB2, and SQL Server. Method: In this paper we present a case study where we evaluate the performance of an SQL query in two of these DBMS. The evaluation is done through various forms of data distribution in a computer network with different degrees of parallelism. ...

  2. Efficient implementation of a multidimensional fast fourier transform on a distributed-memory parallel multi-node computer

    Science.gov (United States)

    Bhanot, Gyan V [Princeton, NJ; Chen, Dong [Croton-On-Hudson, NY; Gara, Alan G [Mount Kisco, NY; Giampapa, Mark E [Irvington, NY; Heidelberger, Philip [Cortlandt Manor, NY; Steinmacher-Burow, Burkhard D [Mount Kisco, NY; Vranas, Pavlos M [Bedford Hills, NY

    2008-01-01

    The present in invention is directed to a method, system and program storage device for efficiently implementing a multidimensional Fast Fourier Transform (FFT) of a multidimensional array comprising a plurality of elements initially distributed in a multi-node computer system comprising a plurality of nodes in communication over a network, comprising: distributing the plurality of elements of the array in a first dimension across the plurality of nodes of the computer system over the network to facilitate a first one-dimensional FFT; performing the first one-dimensional FFT on the elements of the array distributed at each node in the first dimension; re-distributing the one-dimensional FFT-transformed elements at each node in a second dimension via "all-to-all" distribution in random order across other nodes of the computer system over the network; and performing a second one-dimensional FFT on elements of the array re-distributed at each node in the second dimension, wherein the random order facilitates efficient utilization of the network thereby efficiently implementing the multidimensional FFT. The "all-to-all" re-distribution of array elements is further efficiently implemented in applications other than the multidimensional FFT on the distributed-memory parallel supercomputer.

  3. Influence of Distributed Residential Energy Storage on Voltage in Rural Distribution Network and Capacity Configuration

    Science.gov (United States)

    Liu, Lu; Tong, Yibin; Zhao, Zhigang; Zhang, Xuefen

    2018-03-01

    Large-scale access of distributed residential photovoltaic (PV) in rural areas has solved the voltage problem to a certain extent. However, due to the intermittency of PV and the particularity of rural residents’ power load, the problem of low voltage in the evening peak remains to be resolved. This paper proposes to solve the problem by accessing residential energy storage. Firstly, the influence of access location and capacity of energy storage on voltage distribution in rural distribution network is analyzed. Secondly, the relation between the storage capacity and load capacity is deduced for four typical load and energy storage cases when the voltage deviation meets the demand. Finally, the optimal storage position and capacity are obtained by using PSO and power flow simulation.

  4. Optimal scheduling for distribution network with redox flow battery storage

    International Nuclear Information System (INIS)

    Hosseina, Majid; Bathaee, Seyed Mohammad Taghi

    2016-01-01

    Highlights: • A novel method for optimal scheduling of storages in radial network is presented. • Peak shaving and load leveling are the main objectives. • Vanadium redox flow battery is considered as the energy storage unit. • Real data is used for simulation. - Abstract: There are many advantages to utilize storages in electric power system. Peak shaving, load leveling, load frequency control, integration of renewable, energy trading and spinning reserve are the most important of them. Batteries, especially redox flow batteries, are one of the appropriate storages for utilization in distribution network. This paper presents a novel, heuristic and practical method for optimal scheduling in distribution network with flow battery storage. This heuristic method is more suitable for scheduling and operation of distribution networks which require installation of storages. Peak shaving and load leveling is considered as the main objective in this paper. Several indices are presented in this paper for determine the place of storages and also scheduling for optimal use of energy in them. Simulations of this paper are based on real information of distribution network substation that located in Semnan, Iran.

  5. Distributed Cooperative Control of Multi Flywheel Energy Storage System for Electrical Vehicle Fast Charging Stations

    DEFF Research Database (Denmark)

    Sun, Bo; Dragicevic, Tomislav; Quintero, Juan Carlos Vasquez

    2015-01-01

    Plug-in electrical vehicles will play a critical role in future smart grid and sudden connection of electrical vehicles chargers may cause huge power-peaks with high slew-rates on grid. In order to cope with this issue, this paper applies a distributed cooperative control for fast charging station...... with dedicated paralleled flywheel-based energy storage system. The distributed DC-bus signaling method is employed in the power coordination of grid and flywheel converters, and a distributed secondary controller generates DC voltage correction term to adjust the local voltage set-point through a dynamic...... consensus based voltage observer by communicating with its neighbors. The control system can realize the power balancing and DC voltage regulation with low reliance on communications. Finally, real-time hardware-in-the-loop results have been reported in order to verify the feasibility of proposed approach....

  6. A distributed storage system with dCache

    Science.gov (United States)

    Behrmann, G.; Fuhrmann, P.; Grønager, M.; Kleist, J.

    2008-07-01

    The LCG collaboration is encompassed by a number of Tier 1 centers. The Nordic LCG Tier 1, operated by NDGF, is in contrast to many other Tier 1 centers distributed over the Nordic countries. A distributed setup was chosen for both political and technical reasons, but also provides a number of unique challenges. dCache is well known and respected as a powerful distributed storage resource manager, and was chosen for implementing the storage aspects of the Nordic Tier 1. In contrast to classic dCache deployments, we deploy dCache over a WAN with limited bandwidth, high latency, frequent network failures, and spanning many administrative domains. These properties provide unique challenges, covering topics such as security, administration, maintenance, upgradability, reliability, and performance. Our initial focus has been on implementing the GFD.47 OGF recommendation (which introduced the GridFTP 2 protocol) in dCache and the Globus Toolkit. Compared to GridFTP 1, GridFTP 2 allows for more intelligent data flow between clients and storage pools, thus enabling more efficient use of our limited bandwidth.

  7. A distributed storage system with dCache

    International Nuclear Information System (INIS)

    Behrmann, G; Groenager, M; Fuhrmann, P; Kleist, J

    2008-01-01

    The LCG collaboration is encompassed by a number of Tier 1 centers. The Nordic LCG Tier 1, operated by NDGF, is in contrast to many other Tier 1 centers distributed over the Nordic countries. A distributed setup was chosen for both political and technical reasons, but also provides a number of unique challenges. dCache is well known and respected as a powerful distributed storage resource manager, and was chosen for implementing the storage aspects of the Nordic Tier 1. In contrast to classic dCache deployments, we deploy dCache over a WAN with limited bandwidth, high latency, frequent network failures, and spanning many administrative domains. These properties provide unique challenges, covering topics such as security, administration, maintenance, upgradability, reliability, and performance. Our initial focus has been on implementing the GFD.47 OGF recommendation (which introduced the GridFTP 2 protocol) in dCache and the Globus Toolkit. Compared to GridFTP 1, GridFTP 2 allows for more intelligent data flow between clients and storage pools, thus enabling more efficient use of our limited bandwidth

  8. Parallel interactive data analysis with PROOF

    International Nuclear Information System (INIS)

    Ballintijn, Maarten; Biskup, Marek; Brun, Rene; Canal, Philippe; Feichtinger, Derek; Ganis, Gerardo; Kickinger, Guenter; Peters, Andreas; Rademakers, Fons

    2006-01-01

    The Parallel ROOT Facility, PROOF, enables the analysis of much larger data sets on a shorter time scale. It exploits the inherent parallelism in data of uncorrelated events via a multi-tier architecture that optimizes I/O and CPU utilization in heterogeneous clusters with distributed storage. The system provides transparent and interactive access to gigabytes today. Being part of the ROOT framework PROOF inherits the benefits of a performant object storage system and a wealth of statistical and visualization tools. This paper describes the data analysis model of ROOT and the latest developments on closer integration of PROOF into that model and the ROOT user environment, e.g. support for PROOF-based browsing of trees stored remotely, and the popular TTree::Draw() interface. We also outline the ongoing developments aimed to improve the flexibility and user-friendliness of the system

  9. Web based parallel/distributed medical data mining using software agents

    Energy Technology Data Exchange (ETDEWEB)

    Kargupta, H.; Stafford, B.; Hamzaoglu, I.

    1997-12-31

    This paper describes an experimental parallel/distributed data mining system PADMA (PArallel Data Mining Agents) that uses software agents for local data accessing and analysis and a web based interface for interactive data visualization. It also presents the results of applying PADMA for detecting patterns in unstructured texts of postmortem reports and laboratory test data for Hepatitis C patients.

  10. Energy Storage and Distributed Energy Generation Project, Final Project Report

    Energy Technology Data Exchange (ETDEWEB)

    Schwank, Johannes; Mader, Jerry; Chen, Xiaoyin; Mi, Chris; Linic, Suljo; Sastry, Ann Marie; Stefanopoulou, Anna; Thompson, Levi; Varde, Keshav

    2008-03-31

    This report serves as a Final Report under the “Energy Storage and Distribution Energy Generation Project” carried out by the Transportation Energy Center (TEC) at the University of Michigan (UM). An interdisciplinary research team has been working on fundamental and applied research on: -distributed power generation and microgrids, -power electronics, and -advanced energy storage. The long-term objective of the project was to provide a framework for identifying fundamental research solutions to technology challenges of transmission and distribution, with special emphasis on distributed power generation, energy storage, control methodologies, and power electronics for microgrids, and to develop enabling technologies for novel energy storage and harvesting concepts that can be simulated, tested, and scaled up to provide relief for both underserved and overstressed portions of the Nation’s grid. TEC’s research is closely associated with Sections 5.0 and 6.0 of the DOE "Five-year Program Plan for FY2008 to FY2012 for Electric Transmission and Distribution Programs, August 2006.”

  11. Storage Solutions for Power Quality Problems in Cyprus Electricity Distribution Network

    Directory of Open Access Journals (Sweden)

    Andreas Poullikkas

    2014-01-01

    Full Text Available In this work, a prediction of the effects of introducing energy storage systems on the network stability of the distribution network of Cyprus and a comparison in terms of cost with a traditional solution is carried out. In particular, for solving possible overvoltage problems, several scenarios of storage units' installation are used and compared with the alternative solution of extra cable connection between the node with the lowest voltage and the node with the highest voltage of the distribution network. For the comparison, a case study of a typical LV distribution feeder in the power system of Cyprus is used. The results indicated that the performance indicator of each solution depends on the type, the size and the position of installation of the storage unit. Also, as more storage units are installed the better the performance indicator and the more attractive is the investment in storage units to solve power quality problems in the distribution network. In the case where the technical requirements in voltage limitations according to distribution regulations are satisfied with one storage unit, the installation of an additional storage unit will only increase the final cost. The best solution, however, still remains the alternative solution of extra cable connection between the node with the lowest voltage and the node with the highest voltage of the distribution network, due to the lower investment costs compared to that of the storage units.

  12. Distributed Parallel Endmember Extraction of Hyperspectral Data Based on Spark

    Directory of Open Access Journals (Sweden)

    Zebin Wu

    2016-01-01

    Full Text Available Due to the increasing dimensionality and volume of remotely sensed hyperspectral data, the development of acceleration techniques for massive hyperspectral image analysis approaches is a very important challenge. Cloud computing offers many possibilities of distributed processing of hyperspectral datasets. This paper proposes a novel distributed parallel endmember extraction method based on iterative error analysis that utilizes cloud computing principles to efficiently process massive hyperspectral data. The proposed method takes advantage of technologies including MapReduce programming model, Hadoop Distributed File System (HDFS, and Apache Spark to realize distributed parallel implementation for hyperspectral endmember extraction, which significantly accelerates the computation of hyperspectral processing and provides high throughput access to large hyperspectral data. The experimental results, which are obtained by extracting endmembers of hyperspectral datasets on a cloud computing platform built on a cluster, demonstrate the effectiveness and computational efficiency of the proposed method.

  13. Research on the Orientation and Application of Distributed Energy Storage in Energy Internet

    Science.gov (United States)

    Zeng, Ming; Zhou, Pengcheng; Li, Ran; Zhou, Jingjing; Chen, Tao; Li, Zhe

    2018-01-01

    Energy storage is indispensable resources to achieve a high proportion of new energy power consumption in electric power system. As an important support to energy Internet, energy storage system can achieve a variety of energy integration operation to ensure maximum energy efficiency. In this paper, firstly, the SWOT analysis method is used to express the internal and external advantages and disadvantages of distributed energy storage participating in the energy Internet. Secondly, the function orientation of distributed energy storage in energy Internet is studied, based on which the application modes of distributed energy storage in virtual power plant, community energy storage and auxiliary services are deeply studied. Finally, this paper puts forward the development strategy of distributed energy storage which is suitable for the development of China’s energy Internet, and summarizes and prospects the application of distributed energy storage system.

  14. The welfare effects of unbundling gas storage and distribution

    International Nuclear Information System (INIS)

    Breton, Michele; Kharbach, Mohammed

    2008-01-01

    We use a stylized gas system to study the use of access-to-gas storage in a seasonal model. In a duopoly setting, we find that welfare is higher under vertical integration and open access organization than under separate management of storage and distribution. This raises questions about recent regulatory reforms in the gas sectors in the US and Europe, supporting the separation of storage and merchant activities. In the absence of other justifying reasons such as encouraging competition by creating a level playing field, separating the management and accounting functions of storage activities from those of distribution may be a better option than real divestiture, on the basis of welfare arguments. (author)

  15. Scalable storage for a DBMS using transparent distribution

    NARCIS (Netherlands)

    J.S. Karlsson; M.L. Kersten (Martin)

    1997-01-01

    textabstractScalable Distributed Data Structures (SDDSs) provide a self-managing and self-organizing data storage of potentially unbounded size. This stands in contrast to common distribution schemas deployed in conventional distributed DBMS. SDDSs, however, have mostly been used in synthetic

  16. Parallel and distributed processing in two SGBDS: A case study

    Directory of Open Access Journals (Sweden)

    Francisco Javier Moreno

    2017-04-01

    Full Text Available Context: One of the strategies for managing large volumes of data is distributed and parallel computing. Among the tools that allow applying these characteristics are some Data Base Management Systems (DBMS, such as Oracle, DB2, and SQL Server. Method: In this paper we present a case study where we evaluate the performance of an SQL query in two of these DBMS. The evaluation is done through various forms of data distribution in a computer network with different degrees of parallelism. Results: The tests of the SQL query evidenced the performance differences between the two DBMS analyzed. However, more thorough testing and a wider variety of queries are needed. Conclusions: The differences in performance between the two DBMSs analyzed show that when evaluating this aspect, it is necessary to consider the particularities of each DBMS and the degree of parallelism of the queries.

  17. Hybrid Hydrogen and Mechanical Distributed Energy Storage

    Directory of Open Access Journals (Sweden)

    Stefano Ubertini

    2017-12-01

    Full Text Available Effective energy storage technologies represent one of the key elements to solving the growing challenges of electrical energy supply of the 21st century. Several energy storage systems are available, from ones that are technologically mature to others still at a research stage. Each technology has its inherent limitations that make its use economically or practically feasible only for specific applications. The present paper aims at integrating hydrogen generation into compressed air energy storage systems to avoid natural gas combustion or thermal energy storage. A proper design of such a hybrid storage system could provide high roundtrip efficiencies together with enhanced flexibility thanks to the possibility of providing additional energy outputs (heat, cooling, and hydrogen as a fuel, in a distributed energy storage framework. Such a system could be directly connected to the power grid at the distribution level to reduce power and energy intermittence problems related to renewable energy generation. Similarly, it could be located close to the user (e.g., office buildings, commercial centers, industrial plants, hospitals, etc.. Finally, it could be integrated in decentralized energy generation systems to reduce the peak electricity demand charges and energy costs, to increase power generation efficiency, to enhance the security of electrical energy supply, and to facilitate the market penetration of small renewable energy systems. Different configurations have been investigated (simple hybrid storage system, regenerate system, multistage system demonstrating the compressed air and hydrogen storage systems effectiveness in improving energy source flexibility and efficiency, and possibly in reducing the costs of energy supply. Round-trip efficiency up to 65% can be easily reached. The analysis is conducted through a mixed theoretical-numerical approach, which allows the definition of the most relevant physical parameters affecting the system

  18. A parallelization study of the general purpose Monte Carlo code MCNP4 on a distributed memory highly parallel computer

    International Nuclear Information System (INIS)

    Yamazaki, Takao; Fujisaki, Masahide; Okuda, Motoi; Takano, Makoto; Masukawa, Fumihiro; Naito, Yoshitaka

    1993-01-01

    The general purpose Monte Carlo code MCNP4 has been implemented on the Fujitsu AP1000 distributed memory highly parallel computer. Parallelization techniques developed and studied are reported. A shielding analysis function of the MCNP4 code is parallelized in this study. A technique to map a history to each processor dynamically and to map control process to a certain processor was applied. The efficiency of parallelized code is up to 80% for a typical practical problem with 512 processors. These results demonstrate the advantages of a highly parallel computer to the conventional computers in the field of shielding analysis by Monte Carlo method. (orig.)

  19. Experimental studies of current sharing in parallel driven Graetz bridge units for diurnal superconductive magnetic energy storage

    International Nuclear Information System (INIS)

    Kustom, R.L.; Akita, S.; Okada, H.; Skiles, J.

    1985-01-01

    Superconductive Magnetic Energy Storage (SMES) coils for diurnal load leveling and system peaking are envisioned to operate at hundreds of thousands of amperes and a few kilovolts. The interface between the SMES coil and the electric utility is envisioned to be Graetz bridges using SCR switches. Many parallel SCR switches or bridge units will have to operate in parallel because of the high operating current of the coil. Current balancing on parallel Graetz bridges driving a single 8-hy superconducting coil has been achieved on a laboratory model using delay-angle control with an LSI 11/2 microprocessor and external digital control hardware

  20. Design and Verification of Remote Sensing Image Data Center Storage Architecture Based on Hadoop

    Science.gov (United States)

    Tang, D.; Zhou, X.; Jing, Y.; Cong, W.; Li, C.

    2018-04-01

    The data center is a new concept of data processing and application proposed in recent years. It is a new method of processing technologies based on data, parallel computing, and compatibility with different hardware clusters. While optimizing the data storage management structure, it fully utilizes cluster resource computing nodes and improves the efficiency of data parallel application. This paper used mature Hadoop technology to build a large-scale distributed image management architecture for remote sensing imagery. Using MapReduce parallel processing technology, it called many computing nodes to process image storage blocks and pyramids in the background to improve the efficiency of image reading and application and sovled the need for concurrent multi-user high-speed access to remotely sensed data. It verified the rationality, reliability and superiority of the system design by testing the storage efficiency of different image data and multi-users and analyzing the distributed storage architecture to improve the application efficiency of remote sensing images through building an actual Hadoop service system.

  1. Battery Storage Systems as Grid-Balancing Measure in Low-Voltage Distribution Grids with Distributed Generation

    Directory of Open Access Journals (Sweden)

    Bernhard Faessler

    2017-12-01

    Full Text Available Due to the promoted integration of renewable sources, a further growth of strongly transient, distributed generation is expected. Thus, the existing electrical grid may reach its physical limits. To counteract this, and to fully exploit the viable potential of renewables, grid-balancing measures are crucial. In this work, battery storage systems are embedded in a grid simulation to evaluate their potential for grid balancing. The overall setup is based on a real, low-voltage distribution grid topology, real smart meter household load profiles, and real photovoltaics load data. An autonomous optimization routine, driven by a one-way communicated incentive, determines the prospective battery operation mode. Different battery positions and incentives are compared to evaluate their impact. The configurations incorporate a baseline simulation without storage, a single, central battery storage or multiple, distributed battery storages which together have the same power and capacity. The incentives address either market conditions, grid balancing, optimal photovoltaic utilization, load shifting, or self-consumption. Simulations show that grid-balancing incentives result in lowest peak-to-average power ratios, while maintaining negligible voltage changes in comparison to a reference case. Incentives reflecting market conditions for electricity generation, such as real-time pricing, negatively influence the power quality, especially with respect to the peak-to-average power ratio. A central, feed-in-tied storage performs better in terms of minimizing the voltage drop/rise and shows lower distribution losses, while distributed storages attached at nodes with electricity generation by photovoltaics achieve lower peak-to-average power ratios.

  2. 7 CFR 250.14 - Warehousing, distribution and storage of donated foods.

    Science.gov (United States)

    2010-01-01

    .... (iv) All initial data regarding the cost of the current warehousing and distribution system and the... 7 Agriculture 4 2010-01-01 2010-01-01 false Warehousing, distribution and storage of donated foods... General Operating Provisions § 250.14 Warehousing, distribution and storage of donated foods. (a...

  3. Hybrid shared/distributed parallelism for 3D characteristics transport solvers

    International Nuclear Information System (INIS)

    Dahmani, M.; Roy, R.

    2005-01-01

    In this paper, we will present a new hybrid parallel model for solving large-scale 3-dimensional neutron transport problems used in nuclear reactor simulations. Large heterogeneous reactor problems, like the ones that occurs when simulating Candu cores, have remained computationally intensive and impractical for routine applications on single-node or even vector computers. Based on the characteristics method, this new model is designed to solve the transport equation after distributing the calculation load on a network of shared memory multi-processors. The tracks are either generated on the fly at each characteristics sweep or stored in sequential files. The load balancing is taken into account by estimating the calculation load of tracks and by distributing batches of uniform load on each node of the network. Moreover, the communication overhead can be predicted after benchmarking the latency and bandwidth using appropriate network test suite. These models are useful for predicting the performance of the parallel applications and to analyze the scalability of the parallel systems. (authors)

  4. Geospatial Applications on Different Parallel and Distributed Systems in enviroGRIDS Project

    Science.gov (United States)

    Rodila, D.; Bacu, V.; Gorgan, D.

    2012-04-01

    The execution of Earth Science applications and services on parallel and distributed systems has become a necessity especially due to the large amounts of Geospatial data these applications require and the large geographical areas they cover. The parallelization of these applications comes to solve important performance issues and can spread from task parallelism to data parallelism as well. Parallel and distributed architectures such as Grid, Cloud, Multicore, etc. seem to offer the necessary functionalities to solve important problems in the Earth Science domain: storing, distribution, management, processing and security of Geospatial data, execution of complex processing through task and data parallelism, etc. A main goal of the FP7-funded project enviroGRIDS (Black Sea Catchment Observation and Assessment System supporting Sustainable Development) [1] is the development of a Spatial Data Infrastructure targeting this catchment region but also the development of standardized and specialized tools for storing, analyzing, processing and visualizing the Geospatial data concerning this area. For achieving these objectives, the enviroGRIDS deals with the execution of different Earth Science applications, such as hydrological models, Geospatial Web services standardized by the Open Geospatial Consortium (OGC) and others, on parallel and distributed architecture to maximize the obtained performance. This presentation analysis the integration and execution of Geospatial applications on different parallel and distributed architectures and the possibility of choosing among these architectures based on application characteristics and user requirements through a specialized component. Versions of the proposed platform have been used in enviroGRIDS project on different use cases such as: the execution of Geospatial Web services both on Web and Grid infrastructures [2] and the execution of SWAT hydrological models both on Grid and Multicore architectures [3]. The current

  5. Storing files in a parallel computing system based on user-specified parser function

    Science.gov (United States)

    Faibish, Sorin; Bent, John M; Tzelnic, Percy; Grider, Gary; Manzanares, Adam; Torres, Aaron

    2014-10-21

    Techniques are provided for storing files in a parallel computing system based on a user-specified parser function. A plurality of files generated by a distributed application in a parallel computing system are stored by obtaining a parser from the distributed application for processing the plurality of files prior to storage; and storing one or more of the plurality of files in one or more storage nodes of the parallel computing system based on the processing by the parser. The plurality of files comprise one or more of a plurality of complete files and a plurality of sub-files. The parser can optionally store only those files that satisfy one or more semantic requirements of the parser. The parser can also extract metadata from one or more of the files and the extracted metadata can be stored with one or more of the plurality of files and used for searching for files.

  6. Transport or Store? Synthesizing Flow-based Microfluidic Biochips using Distributed Channel Storage

    DEFF Research Database (Denmark)

    Liu, Chunfeng; Li, Bing; Yao, Hailong

    2017-01-01

    of the performance of biochips. In this paper, we propose the first architectural synthesis framework considering distributed storage constructed temporarily from transportation channels to cache fluid samples. Since distributed storage can be accessed more efficiently than a dedicated storage unit and channels can...... switch between the roles of transportation and storage easily, biochips with this distributed computing architecture can achieve a higher execution efficiency even with fewer resources. Experimental results confirm that the execution efficiency of a bioassay can be improved by up to 28% while the number...

  7. Stampi: a message passing library for distributed parallel computing. User's guide

    International Nuclear Information System (INIS)

    Imamura, Toshiyuki; Koide, Hiroshi; Takemiya, Hiroshi

    1998-11-01

    A new message passing library, Stampi, has been developed to realize a computation with different kind of parallel computers arbitrarily and making MPI (Message Passing Interface) as an unique interface for communication. Stampi is based on MPI2 specification. It realizes dynamic process creation to different machines and communication between spawned one within the scope of MPI semantics. Vender implemented MPI as a closed system in one parallel machine and did not support both functions; process creation and communication to external machines. Stampi supports both functions and enables us distributed parallel computing. Currently Stampi has been implemented on COMPACS (COMplex PArallel Computer System) introduced in CCSE, five parallel computers and one graphic workstation, and any communication on them can be processed on. (author)

  8. Optimal Operation of Energy Storage in Power Transmission and Distribution

    Science.gov (United States)

    Akhavan Hejazi, Seyed Hossein

    In this thesis, we investigate optimal operation of energy storage units in power transmission and distribution grids. At transmission level, we investigate the problem where an investor-owned independently-operated energy storage system seeks to offer energy and ancillary services in the day-ahead and real-time markets. We specifically consider the case where a significant portion of the power generated in the grid is from renewable energy resources and there exists significant uncertainty in system operation. In this regard, we formulate a stochastic programming framework to choose optimal energy and reserve bids for the storage units that takes into account the fluctuating nature of the market prices due to the randomness in the renewable power generation availability. At distribution level, we develop a comprehensive data set to model various stochastic factors on power distribution networks, with focus on networks that have high penetration of electric vehicle charging load and distributed renewable generation. Furthermore, we develop a data-driven stochastic model for energy storage operation at distribution level, where the distribution of nodal voltage and line power flow are modelled as stochastic functions of the energy storage unit's charge and discharge schedules. In particular, we develop new closed-form stochastic models for such key operational parameters in the system. Our approach is analytical and allows formulating tractable optimization problems. Yet, it does not involve any restricting assumption on the distribution of random parameters, hence, it results in accurate modeling of uncertainties. By considering the specific characteristics of random variables, such as their statistical dependencies and often irregularly-shaped probability distributions, we propose a non-parametric chance-constrained optimization approach to operate and plan energy storage units in power distribution girds. In the proposed stochastic optimization, we consider

  9. PARALLEL AND ADAPTIVE UNIFORM-DISTRIBUTED REGISTRATION METHOD FOR CHANG’E-1 LUNAR REMOTE SENSED IMAGERY

    Directory of Open Access Journals (Sweden)

    X. Ning

    2012-08-01

    To resolve the above-mentioned registration difficulties, a parallel and adaptive uniform-distributed registration method for CE-1 lunar remote sensed imagery is proposed in this paper. Based on 6 pairs of randomly selected images, both the standard SIFT algorithm and the parallel and adaptive uniform-distributed registration method were executed, the versatility and effectiveness were assessed. The experimental results indicate that: by applying the parallel and adaptive uniform-distributed registration method, the efficiency of CE-1 lunar remote sensed imagery registration were increased dramatically. Therefore, the proposed method in the paper could acquire uniform-distributed registration results more effectively, the registration difficulties including difficult to obtain results, time-consuming, non-uniform distribution could be successfully solved.

  10. Unified storage systems for distributed Tier-2 centres

    International Nuclear Information System (INIS)

    Cowan, G A; Stewart, G A; Elwell, A

    2008-01-01

    The start of data taking at the Large Hadron Collider will herald a new era in data volumes and distributed processing in particle physics. Data volumes of hundreds of Terabytes will be shipped to Tier-2 centres for analysis by the LHC experiments using the Worldwide LHC Computing Grid (WLCG). In many countries Tier-2 centres are distributed between a number of institutes, e.g., the geographically spread Tier-2s of GridPP in the UK. This presents a number of challenges for experiments to utilise these centres efficaciously, as CPU and storage resources may be subdivided and exposed in smaller units than the experiment would ideally want to work with. In addition, unhelpful mismatches between storage and CPU at the individual centres may be seen, which make efficient exploitation of a Tier-2's resources difficult. One method of addressing this is to unify the storage across a distributed Tier-2, presenting the centres' aggregated storage as a single system. This greatly simplifies data management for the VO, which then can access a greater amount of data across the Tier-2. However, such an approach will lead to scenarios where analysis jobs on one site's batch system must access data hosted on another site. We investigate this situation using the Glasgow and Edinburgh clusters, which are part of the ScotGrid distributed Tier-2. In particular we look at how to mitigate the problems associated with 'distant' data access and discuss the security implications of having LAN access protocols traverse the WAN between centres

  11. Monte Carlo photon transport on shared memory and distributed memory parallel processors

    International Nuclear Information System (INIS)

    Martin, W.R.; Wan, T.C.; Abdel-Rahman, T.S.; Mudge, T.N.; Miura, K.

    1987-01-01

    Parallelized Monte Carlo algorithms for analyzing photon transport in an inertially confined fusion (ICF) plasma are considered. Algorithms were developed for shared memory (vector and scalar) and distributed memory (scalar) parallel processors. The shared memory algorithm was implemented on the IBM 3090/400, and timing results are presented for dedicated runs with two, three, and four processors. Two alternative distributed memory algorithms (replication and dispatching) were implemented on a hypercube parallel processor (1 through 64 nodes). The replication algorithm yields essentially full efficiency for all cube sizes; with the 64-node configuration, the absolute performance is nearly the same as with the CRAY X-MP. The dispatching algorithm also yields efficiencies above 80% in a large simulation for the 64-processor configuration

  12. Resonance analysis in parallel voltage-controlled Distributed Generation inverters

    DEFF Research Database (Denmark)

    Wang, Xiongfei; Blaabjerg, Frede; Chen, Zhe

    2013-01-01

    Thanks to the fast responses of the inner voltage and current control loops, the dynamic behaviors of parallel voltage-controlled Distributed Generation (DG) inverters not only relies on the stability of load sharing among them, but subjects to the interactions between the voltage control loops...

  13. P3T+: A Performance Estimator for Distributed and Parallel Programs

    Directory of Open Access Journals (Sweden)

    T. Fahringer

    2000-01-01

    Full Text Available Developing distributed and parallel programs on today's multiprocessor architectures is still a challenging task. Particular distressing is the lack of effective performance tools that support the programmer in evaluating changes in code, problem and machine sizes, and target architectures. In this paper we introduce P3T+ which is a performance estimator for mostly regular HPF (High Performance Fortran programs but partially covers also message passing programs (MPI. P3T+ is unique by modeling programs, compiler code transformations, and parallel and distributed architectures. It computes at compile-time a variety of performance parameters including work distribution, number of transfers, amount of data transferred, transfer times, computation times, and number of cache misses. Several novel technologies are employed to compute these parameters: loop iteration spaces, array access patterns, and data distributions are modeled by employing highly effective symbolic analysis. Communication is estimated by simulating the behavior of a communication library used by the underlying compiler. Computation times are predicted through pre-measured kernels on every target architecture of interest. We carefully model most critical architecture specific factors such as cache lines sizes, number of cache lines available, startup times, message transfer time per byte, etc. P3T+ has been implemented and is closely integrated with the Vienna High Performance Compiler (VFC to support programmers develop parallel and distributed applications. Experimental results for realistic kernel codes taken from real-world applications are presented to demonstrate both accuracy and usefulness of P3T+.

  14. New adaptive differencing strategy in the PENTRAN 3-d parallel Sn code

    International Nuclear Information System (INIS)

    Sjoden, G.E.; Haghighat, A.

    1996-01-01

    It is known that three-dimensional (3-D) discrete ordinates (S n ) transport problems require an immense amount of storage and computational effort to solve. For this reason, parallel codes that offer a capability to completely decompose the angular, energy, and spatial domains among a distributed network of processors are required. One such code recently developed is PENTRAN, which iteratively solves 3-D multi-group, anisotropic S n problems on distributed-memory platforms, such as the IBM-SP2. Because large problems typically contain several different material zones with various properties, available differencing schemes should automatically adapt to the transport physics in each material zone. To minimize the memory and message-passing overhead required for massively parallel S n applications, available differencing schemes in an adaptive strategy should also offer reasonable accuracy and positivity, yet require only the zeroth spatial moment of the transport equation; differencing schemes based on higher spatial moments, in spite of their greater accuracy, require at least twice the amount of storage and communication cost for implementation in a massively parallel transport code. This paper discusses a new adaptive differencing strategy that uses increasingly accurate schemes with low parallel memory and communication overhead. This strategy, implemented in PENTRAN, includes a new scheme, exponential directional averaged (EDA) differencing

  15. Towards Integrating Distributed Energy Resources and Storage Devices in Smart Grid.

    Science.gov (United States)

    Xu, Guobin; Yu, Wei; Griffith, David; Golmie, Nada; Moulema, Paul

    2017-02-01

    Internet of Things (IoT) provides a generic infrastructure for different applications to integrate information communication techniques with physical components to achieve automatic data collection, transmission, exchange, and computation. The smart grid, as one of typical applications supported by IoT, denoted as a re-engineering and a modernization of the traditional power grid, aims to provide reliable, secure, and efficient energy transmission and distribution to consumers. How to effectively integrate distributed (renewable) energy resources and storage devices to satisfy the energy service requirements of users, while minimizing the power generation and transmission cost, remains a highly pressing challenge in the smart grid. To address this challenge and assess the effectiveness of integrating distributed energy resources and storage devices, in this paper we develop a theoretical framework to model and analyze three types of power grid systems: the power grid with only bulk energy generators, the power grid with distributed energy resources, and the power grid with both distributed energy resources and storage devices. Based on the metrics of the power cumulative cost and the service reliability to users, we formally model and analyze the impact of integrating distributed energy resources and storage devices in the power grid. We also use the concept of network calculus, which has been traditionally used for carrying out traffic engineering in computer networks, to derive the bounds of both power supply and user demand to achieve a high service reliability to users. Through an extensive performance evaluation, our data shows that integrating distributed energy resources conjointly with energy storage devices can reduce generation costs, smooth the curve of bulk power generation over time, reduce bulk power generation and power distribution losses, and provide a sustainable service reliability to users in the power grid.

  16. Parallel computation for distributed parameter system-from vector processors to Adena computer

    Energy Technology Data Exchange (ETDEWEB)

    Nogi, T

    1983-04-01

    Research on advanced parallel hardware and software architectures for very high-speed computation deserves and needs more support and attention to fulfil its promise. Novel architectures for parallel processing are being made ready. Architectures for parallel processing can be roughly divided into two groups. One is a vector processor in which a single central processing unit involves multiple vector-arithmetic registers. The other is a processor array in which slave processors are connected to a host processor to perform parallel computation. In this review, the concept and data structure of the Adena (alternating-direction edition nexus array) architecture, which is conformable to distributed-parameter simulation algorithms, are described. 5 references.

  17. Multilayer PV-storage Microgrids Algorithm for the Dispatch of Distributed Network

    Directory of Open Access Journals (Sweden)

    Yang Ping

    2016-01-01

    Full Text Available In recent years, due to the support of our country, PV-storage microgrid develops rapidly. However, the flexible network operation modes of PV-storage microgrid change flexibly and the operating characteristics with a large amout of sources is highly complicated. Based on the existing microgrid coordinate control methods, this paper proposes multilayer PV-storage microgrid algorithm for fitting dispatch of distributed network, which achieves maximum output of renewable energy when meeting the scheduling requirements of network, by building PV-storage microgrid type dynamic simulation system in a variety of conditions in PSCAD. Simulation results show that the heuristic algorithm proposed can achieve microgrid stable operation and satisfy the demands of the dispatch in distributed network.

  18. TME (Task Mapping Editor): tool for executing distributed parallel computing. TME user's manual

    International Nuclear Information System (INIS)

    Takemiya, Hiroshi; Yamagishi, Nobuhiro; Imamura, Toshiyuki

    2000-03-01

    At the Center for Promotion of Computational Science and Engineering, a software environment PPExe has been developed to support scientific computing on a parallel computer cluster (distributed parallel scientific computing). TME (Task Mapping Editor) is one of components of the PPExe and provides a visual programming environment for distributed parallel scientific computing. Users can specify data dependence among tasks (programs) visually as a data flow diagram and map these tasks onto computers interactively through GUI of TME. The specified tasks are processed by other components of PPExe such as Meta-scheduler, RIM (Resource Information Monitor), and EMS (Execution Management System) according to the execution order of these tasks determined by TME. In this report, we describe the usage of TME. (author)

  19. Distributed and parallel approach for handle and perform huge datasets

    Science.gov (United States)

    Konopko, Joanna

    2015-12-01

    Big Data refers to the dynamic, large and disparate volumes of data comes from many different sources (tools, machines, sensors, mobile devices) uncorrelated with each others. It requires new, innovative and scalable technology to collect, host and analytically process the vast amount of data. Proper architecture of the system that perform huge data sets is needed. In this paper, the comparison of distributed and parallel system architecture is presented on the example of MapReduce (MR) Hadoop platform and parallel database platform (DBMS). This paper also analyzes the problem of performing and handling valuable information from petabytes of data. The both paradigms: MapReduce and parallel DBMS are described and compared. The hybrid architecture approach is also proposed and could be used to solve the analyzed problem of storing and processing Big Data.

  20. On Network Coded Distributed Storage

    DEFF Research Database (Denmark)

    Cabrera Guerrero, Juan Alberto; Roetter, Daniel Enrique Lucani; Fitzek, Frank Hanns Paul

    2016-01-01

    systems typically rely on expensive infrastructure with centralized control to store, repair and access the data. This approach introduces a large delay for accessing and storing the data driven in part by a high RTT between users and the cloud. These characteristics are at odds with the massive increase......This paper focuses on distributed fog storage solutions, where a number of unreliable devices organize themselves in Peer-to-Peer (P2P) networks with the purpose to store reliably their data and that of other devices and/or local users and provide lower delay and higher throughput. Cloud storage...... of devices and generated data in coming years as well as the requirements of low latency in many applications. We focus on characterizing optimal solutions for maintaining data availability when nodes in the fog continuously leave the network. In contrast with state-of-the-art data repair formulations, which...

  1. Kemari: A Portable High Performance Fortran System for Distributed Memory Parallel Processors

    Directory of Open Access Journals (Sweden)

    T. Kamachi

    1997-01-01

    Full Text Available We have developed a compilation system which extends High Performance Fortran (HPF in various aspects. We support the parallelization of well-structured problems with loop distribution and alignment directives similar to HPF's data distribution directives. Such directives give both additional control to the user and simplify the compilation process. For the support of unstructured problems, we provide directives for dynamic data distribution through user-defined mappings. The compiler also allows integration of message-passing interface (MPI primitives. The system is part of a complete programming environment which also comprises a parallel debugger and a performance monitor and analyzer. After an overview of the compiler, we describe the language extensions and related compilation mechanisms in detail. Performance measurements demonstrate the compiler's applicability to a variety of application classes.

  2. Parallel Distributed Processing theory in the age of deep networks

    OpenAIRE

    Bowers, Jeffrey

    2017-01-01

    Parallel Distributed Processing (PDP) models in psychology are the precursors of deep networks used in computer science. However, only PDP models are associated with two core psychological claims, namely, that all knowledge is coded in a distributed format, and cognition is mediated by non-symbolic computations. These claims have long been debated within cognitive science, and recent work with deep networks speaks to this debate. Specifically, single-unit recordings show that deep networks le...

  3. Dynamic overset grid communication on distributed memory parallel processors

    Science.gov (United States)

    Barszcz, Eric; Weeratunga, Sisira K.; Meakin, Robert L.

    1993-01-01

    A parallel distributed memory implementation of intergrid communication for dynamic overset grids is presented. Included are discussions of various options considered during development. Results are presented comparing an Intel iPSC/860 to a single processor Cray Y-MP. Results for grids in relative motion show the iPSC/860 implementation to be faster than the Cray implementation.

  4. Community energy storage and distribution SCADA improvements

    International Nuclear Information System (INIS)

    Riggins, M.

    2010-01-01

    The mission of American Electric Power (AEP) is to sustain the real time balance of energy supply and demand. Approximately 2.5 percent of energy generated in the United States (USA) is stored as pumped hydro, compressed air, or in batteries and other devices. This power point presentation discussed the use of SCADA for improving community energy storage (CES) and distribution systems. CES is a distributed fleet of small energy units connected to the transformers in order to serve houses or small commercial loads. CES is operated as a fleet offering multi-megawatt (MW) multi-hour storage. The benefits of CES include backup power, flicker mitigation, and renewable integration. Benefits to the electricity grid include power factor correct, ancillary services, and load leveling at the substation level. SCADA is being used to determine when emergency load reductions are required or when emergency inspections on fans, oil pumps or other devices are needed. An outline of AEP's monitoring system installation plan was also included. tabs., figs.

  5. Global distribution of grid connected electrical energy storage systems

    Directory of Open Access Journals (Sweden)

    Katja Buss

    2016-06-01

    Full Text Available This article gives an overview of grid connected electrical energy storage systems worldwide, based on public available data. Technologies considered in this study are pumped hydroelectric energy storage (PHES, compressed air energy storage (CAES, sodium-sulfur batteries (NaS, lead-acid batteries, redox-flow batteries, nickel-cadmium batteries (NiCd and lithium-ion batteries. As the research indicates, the worldwide installed capacity of grid connected electrical energy storage systems is approximately 154 GW. This corresponds to a share of 5.5 % of the worldwide installed generation capacity. Furthermore, the article gives an overview of the historical development of installed and used storage systems worldwide. Subsequently, the focus is on each considered technology concerning the current storage size, number of plants and location. In summary it can be stated, PHES is the most commonly used technology worldwide, whereas electrochemical technologies are increasingly gaining in importance. Regarding the distribution of grid connected storage systems reveals the share of installed storage capacity is in Europe and Eastern Asia twice as high as in North America.

  6. Distributed and cloud computing from parallel processing to the Internet of Things

    CERN Document Server

    Hwang, Kai; Fox, Geoffrey C

    2012-01-01

    Distributed and Cloud Computing, named a 2012 Outstanding Academic Title by the American Library Association's Choice publication, explains how to create high-performance, scalable, reliable systems, exposing the design principles, architecture, and innovative applications of parallel, distributed, and cloud computing systems. Starting with an overview of modern distributed models, the book provides comprehensive coverage of distributed and cloud computing, including: Facilitating management, debugging, migration, and disaster recovery through virtualization Clustered systems for resear

  7. Storage of sparse files using parallel log-structured file system

    Science.gov (United States)

    Bent, John M.; Faibish, Sorin; Grider, Gary; Torres, Aaron

    2017-11-07

    A sparse file is stored without holes by storing a data portion of the sparse file using a parallel log-structured file system; and generating an index entry for the data portion, the index entry comprising a logical offset, physical offset and length of the data portion. The holes can be restored to the sparse file upon a reading of the sparse file. The data portion can be stored at a logical end of the sparse file. Additional storage efficiency can optionally be achieved by (i) detecting a write pattern for a plurality of the data portions and generating a single patterned index entry for the plurality of the patterned data portions; and/or (ii) storing the patterned index entries for a plurality of the sparse files in a single directory, wherein each entry in the single directory comprises an identifier of a corresponding sparse file.

  8. The parallel processing of EGS4 code on distributed memory scalar parallel computer:Intel Paragon XP/S15-256

    Energy Technology Data Exchange (ETDEWEB)

    Takemiya, Hiroshi; Ohta, Hirofumi; Honma, Ichirou

    1996-03-01

    The parallelization of Electro-Magnetic Cascade Monte Carlo Simulation Code, EGS4 on distributed memory scalar parallel computer: Intel Paragon XP/S15-256 is described. EGS4 has the feature that calculation time for one incident particle is quite different from each other because of the dynamic generation of secondary particles and different behavior of each particle. Granularity for parallel processing, parallel programming model and the algorithm of parallel random number generation are discussed and two kinds of method, each of which allocates particles dynamically or statically, are used for the purpose of realizing high speed parallel processing of this code. Among four problems chosen for performance evaluation, the speedup factors for three problems have been attained to nearly 100 times with 128 processor. It has been found that when both the calculation time for each incident particles and its dispersion are large, it is preferable to use dynamic particle allocation method which can average the load for each processor. And it has also been found that when they are small, it is preferable to use static particle allocation method which reduces the communication overhead. Moreover, it is pointed out that to get the result accurately, it is necessary to use double precision variables in EGS4 code. Finally, the workflow of program parallelization is analyzed and tools for program parallelization through the experience of the EGS4 parallelization are discussed. (author).

  9. The specification of Stampi, a message passing library for distributed parallel computing

    International Nuclear Information System (INIS)

    Imamura, Toshiyuki; Takemiya, Hiroshi; Koide, Hiroshi

    2000-03-01

    At CCSE, Center for Promotion of Computational Science and Engineering, a new message passing library for heterogeneous and distributed parallel computing has been developed, and it is called as Stampi. Stampi enables us to communicate between any combination of parallel computers as well as workstations. Currently, a Stampi system is constructed from Stampi library and Stampi/Java. It provides functions to connect a Stampi application with not only those on COMPACS, COMplex Parallel Computer System, but also applets which work on WWW browsers. This report summarizes the specifications of Stampi and details the development of its system. (author)

  10. Conflict free network coding for distributed storage networks

    KAUST Repository

    Al-Habob, Ahmed A.; Sorour, Sameh; Aboutorab, Neda; Sadeghi, Parastoo

    2015-01-01

    © 2015 IEEE. In this paper, we design a conflict free instantly decodable network coding (IDNC) solution for file download from distributed storage servers. Considering previously downloaded files at the clients from these servers as side

  11. Embodied and Distributed Parallel DJing.

    Science.gov (United States)

    Cappelen, Birgitta; Andersson, Anders-Petter

    2016-01-01

    Everyone has a right to take part in cultural events and activities, such as music performances and music making. Enforcing that right, within Universal Design, is often limited to a focus on physical access to public areas, hearing aids etc., or groups of persons with special needs performing in traditional ways. The latter might be people with disabilities, being musicians playing traditional instruments, or actors playing theatre. In this paper we focus on the innovative potential of including people with special needs, when creating new cultural activities. In our project RHYME our goal was to create health promoting activities for children with severe disabilities, by developing new musical and multimedia technologies. Because of the users' extreme demands and rich contribution, we ended up creating both a new genre of musical instruments and a new art form. We call this new art form Embodied and Distributed Parallel DJing, and the new genre of instruments for Empowering Multi-Sensorial Things.

  12. Three-dimensional magnetic field computation on a distributed memory parallel processor

    International Nuclear Information System (INIS)

    Barion, M.L.

    1990-01-01

    The analysis of three-dimensional magnetic fields by finite element methods frequently proves too onerous a task for the computing resource on which it is attempted. When non-linear and transient effects are included, it may become impossible to calculate the field distribution to sufficient resolution. One approach to this problem is to exploit the natural parallelism in the finite element method via parallel processing. This paper reports on an implementation of a finite element code for non-linear three-dimensional low-frequency magnetic field calculation on Intel's iPSC/2

  13. Enabling Requirements-Based Programming for Highly-Dependable Complex Parallel and Distributed Systems

    Science.gov (United States)

    Hinchey, Michael G.; Rash, James L.; Rouff, Christopher A.

    2005-01-01

    The manual application of formal methods in system specification has produced successes, but in the end, despite any claims and assertions by practitioners, there is no provable relationship between a manually derived system specification or formal model and the customer's original requirements. Complex parallel and distributed system present the worst case implications for today s dearth of viable approaches for achieving system dependability. No avenue other than formal methods constitutes a serious contender for resolving the problem, and so recognition of requirements-based programming has come at a critical juncture. We describe a new, NASA-developed automated requirement-based programming method that can be applied to certain classes of systems, including complex parallel and distributed systems, to achieve a high degree of dependability.

  14. Control of hybrid fuel cell/energy storage distributed generation system against voltage sag

    Energy Technology Data Exchange (ETDEWEB)

    Hajizadeh, Amin; Golkar, Masoud Aliakbar [Electrical Engineering Department, K.N. Toosi University of Technology, Seyedkhandan, Dr. Shariati Ave, P.O. Box 16315-1355, Tehran (Iran)

    2010-06-15

    Fuel cell (FC) and energy storage (ES) based hybrid distributed power generation systems appear to be very promising for satisfying high energy and high power requirements of power quality problems in distributed generation (DG) systems. In this study, design of control strategy for hybrid fuel cell/energy storage distributed power generation system during voltage sag has been presented. The proposed control strategy allows hybrid distributed generation system works properly when a voltage disturbance occurs in distribution system and hybrid system stays connected to the main grid. Hence, modeling, controller design, and simulation study of a hybrid distributed generation system are investigated. The physical model of the fuel cell stack, energy storage and the models of power conditioning units are described. Then the control design methodology for each component of the hybrid system is proposed. Simulation results are given to show the overall system performance including active power control and voltage sag ride-through capability of the hybrid distributed generation system. (author)

  15. AdiosStMan: Parallelizing Casacore Table Data System using Adaptive IO System

    Science.gov (United States)

    Wang, R.; Harris, C.; Wicenec, A.

    2016-07-01

    In this paper, we investigate the Casacore Table Data System (CTDS) used in the casacore and CASA libraries, and methods to parallelize it. CTDS provides a storage manager plugin mechanism for third-party developers to design and implement their own CTDS storage managers. Having this in mind, we looked into various storage backend techniques that can possibly enable parallel I/O for CTDS by implementing new storage managers. After carrying on benchmarks showing the excellent parallel I/O throughput of the Adaptive IO System (ADIOS), we implemented an ADIOS based parallel CTDS storage manager. We then applied the CASA MSTransform frequency split task to verify the ADIOS Storage Manager. We also ran a series of performance tests to examine the I/O throughput in a massively parallel scenario.

  16. Parallel file system with metadata distributed across partitioned key-value store c

    Science.gov (United States)

    Bent, John M.; Faibish, Sorin; Grider, Gary; Torres, Aaron

    2017-09-19

    Improved techniques are provided for storing metadata associated with a plurality of sub-files associated with a single shared file in a parallel file system. The shared file is generated by a plurality of applications executing on a plurality of compute nodes. A compute node implements a Parallel Log Structured File System (PLFS) library to store at least one portion of the shared file generated by an application executing on the compute node and metadata for the at least one portion of the shared file on one or more object storage servers. The compute node is also configured to implement a partitioned data store for storing a partition of the metadata for the shared file, wherein the partitioned data store communicates with partitioned data stores on other compute nodes using a message passing interface. The partitioned data store can be implemented, for example, using Multidimensional Data Hashing Indexing Middleware (MDHIM).

  17. Cost Benefit and Alternatives Analysis of Distribution Systems with Energy Storage Systems: Preprint

    Energy Technology Data Exchange (ETDEWEB)

    Harris, Tom; Nagarajan, Adarsh; Baggu, Murali; Bialek, Tom

    2017-06-27

    This paper explores monetized and non-monetized benefits from storage interconnected to distribution system through use cases illustrating potential applications for energy storage in California's electric utility system. This work supports SDG&E in its efforts to quantify, summarize, and compare the cost and benefit streams related to implementation and operation of energy storage on its distribution feeders. This effort develops the cost benefit and alternatives analysis platform, integrated with QSTS feeder simulation capability, and analyzed use cases to explore the cost-benefit of implementation and operation of energy storage for feeder support and market participation.

  18. Computational cost of isogeometric multi-frontal solvers on parallel distributed memory machines

    KAUST Repository

    Woźniak, Maciej

    2015-02-01

    This paper derives theoretical estimates of the computational cost for isogeometric multi-frontal direct solver executed on parallel distributed memory machines. We show theoretically that for the Cp-1 global continuity of the isogeometric solution, both the computational cost and the communication cost of a direct solver are of order O(log(N)p2) for the one dimensional (1D) case, O(Np2) for the two dimensional (2D) case, and O(N4/3p2) for the three dimensional (3D) case, where N is the number of degrees of freedom and p is the polynomial order of the B-spline basis functions. The theoretical estimates are verified by numerical experiments performed with three parallel multi-frontal direct solvers: MUMPS, PaStiX and SuperLU, available through PETIGA toolkit built on top of PETSc. Numerical results confirm these theoretical estimates both in terms of p and N. For a given problem size, the strong efficiency rapidly decreases as the number of processors increases, becoming about 20% for 256 processors for a 3D example with 1283 unknowns and linear B-splines with C0 global continuity, and 15% for a 3D example with 643 unknowns and quartic B-splines with C3 global continuity. At the same time, one cannot arbitrarily increase the problem size, since the memory required by higher order continuity spaces is large, quickly consuming all the available memory resources even in the parallel distributed memory version. Numerical results also suggest that the use of distributed parallel machines is highly beneficial when solving higher order continuity spaces, although the number of processors that one can efficiently employ is somehow limited.

  19. Adaptive data migration scheme with facilitator database and multi-tier distributed storage in LHD

    International Nuclear Information System (INIS)

    Nakanishi, Hideya; Masaki, Ohsuna; Mamoru, Kojima; Setsuo, Imazu; Miki, Nonomura; Kenji, Watanabe; Masayoshi, Moriya; Yoshio, Nagayama; Kazuo, Kawahata

    2008-01-01

    Recent 'data explosion' induces the demand for high flexibility of storage extension and data migration. The data amount of LHD plasma diagnostics has grown 4.6 times bigger than that of three years before. Frequent migration or replication between plenty of distributed storage becomes mandatory, and thus increases the human operational costs. To reduce them computationally, a new adaptive migration scheme has been developed on LHD's multi-tier distributed storage. So-called the HSM (Hierarchical Storage Management) software usually adopts a low-level cache mechanism or simple watermarks for triggering the data stage-in and out between two storage devices. However, the new scheme can deal with a number of distributed storage by the facilitator database that manages the whole data locations with their access histories and retrieval priorities. Not only the inter-tier migration but also the intra-tier replication and moving are even manageable so that it can be a big help in extending or replacing storage equipment. The access history of each data object is also utilized to optimize the volume size of fast and costly RAID, in addition to a normal cache effect for frequently retrieved data. The new scheme has been verified its effectiveness so that LHD multi-tier distributed storage and other next-generation experiments can obtain such the flexible expandability

  20. Academic training: From Evolution Theory to Parallel and Distributed Genetic Programming

    CERN Multimedia

    2007-01-01

    2006-2007 ACADEMIC TRAINING PROGRAMME LECTURE SERIES 15, 16 March From 11:00 to 12:00 - Main Auditorium, bldg. 500 From Evolution Theory to Parallel and Distributed Genetic Programming F. FERNANDEZ DE VEGA / Univ. of Extremadura, SP Lecture No. 1: From Evolution Theory to Evolutionary Computation Evolutionary computation is a subfield of artificial intelligence (more particularly computational intelligence) involving combinatorial optimization problems, which are based to some degree on the evolution of biological life in the natural world. In this tutorial we will review the source of inspiration for this metaheuristic and its capability for solving problems. We will show the main flavours within the field, and different problems that have been successfully solved employing this kind of techniques. Lecture No. 2: Parallel and Distributed Genetic Programming The successful application of Genetic Programming (GP, one of the available Evolutionary Algorithms) to optimization problems has encouraged an ...

  1. Storage Operation for Peak Shaving of Distributed PV and Wind Generation

    NARCIS (Netherlands)

    Nykamp, Stefan; Molderink, Albert; Hurink, Johann L.; Smit, Gerardus Johannes Maria

    2013-01-01

    The integration of fluctuating power generation based on renewable energy systems (RES-E) in distribution grids require grid reinforcement resulting from high feed-in peaks. Introducing storage assets can decrease these peaks. For this, storage technologies need to be chosen and dimensioned

  2. Implementation of Parallel Dynamic Simulation on Shared-Memory vs. Distributed-Memory Environments

    Energy Technology Data Exchange (ETDEWEB)

    Jin, Shuangshuang; Chen, Yousu; Wu, Di; Diao, Ruisheng; Huang, Zhenyu

    2015-12-09

    Power system dynamic simulation computes the system response to a sequence of large disturbance, such as sudden changes in generation or load, or a network short circuit followed by protective branch switching operation. It consists of a large set of differential and algebraic equations, which is computational intensive and challenging to solve using single-processor based dynamic simulation solution. High-performance computing (HPC) based parallel computing is a very promising technology to speed up the computation and facilitate the simulation process. This paper presents two different parallel implementations of power grid dynamic simulation using Open Multi-processing (OpenMP) on shared-memory platform, and Message Passing Interface (MPI) on distributed-memory clusters, respectively. The difference of the parallel simulation algorithms and architectures of the two HPC technologies are illustrated, and their performances for running parallel dynamic simulation are compared and demonstrated.

  3. Particle simulation on a distributed memory highly parallel processor

    International Nuclear Information System (INIS)

    Sato, Hiroyuki; Ikesaka, Morio

    1990-01-01

    This paper describes parallel molecular dynamics simulation of atoms governed by local force interaction. The space in the model is divided into cubic subspaces and mapped to the processor array of the CAP-256, a distributed memory, highly parallel processor developed at Fujitsu Labs. We developed a new technique to avoid redundant calculation of forces between atoms in different processors. Experiments showed the communication overhead was less than 5%, and the idle time due to load imbalance was less than 11% for two model problems which contain 11,532 and 46,128 argon atoms. From the software simulation, the CAP-II which is under development is estimated to be about 45 times faster than CAP-256 and will be able to run the same problem about 40 times faster than Fujitsu's M-380 mainframe when 256 processors are used. (author)

  4. Distributed analysis with PROOF in ATLAS collaboration

    International Nuclear Information System (INIS)

    Panitkin, S Y; Ernst, M; Ito, H; Maeno, T; Majewski, S; Rind, O; Tarrade, F; Wenaus, T; Ye, S; Benjamin, D; Montoya, G Carillo; Guan, W; Mellado, B; Xu, N; Cranmer, K; Shibata, A

    2010-01-01

    The Parallel ROOT Facility - PROOF is a distributed analysis system which allows to exploit inherent event level parallelism of high energy physics data. PROOF can be configured to work with centralized storage systems, but it is especially effective together with distributed local storage systems - like Xrootd, when data are distributed over computing nodes. It works efficiently on different types of hardware and scales well from a multi-core laptop to large computing farms. From that point of view it is well suited for both large central analysis facilities and Tier 3 type analysis farms. PROOF can be used in interactive or batch like regimes. The interactive regime allows the user to work with typically distributed data from the ROOT command prompt and get a real time feedback on analysis progress and intermediate results. We will discuss our experience with PROOF in the context of ATLAS Collaboration distributed analysis. In particular we will discuss PROOF performance in various analysis scenarios and in multi-user, multi-session environments. We will also describe PROOF integration with the ATLAS distributed data management system and prospects of running PROOF on geographically distributed analysis farms.

  5. Distributed analysis with PROOF in ATLAS collaboration

    Energy Technology Data Exchange (ETDEWEB)

    Panitkin, S Y; Ernst, M; Ito, H; Maeno, T; Majewski, S; Rind, O; Tarrade, F; Wenaus, T; Ye, S [Brookhaven National Laboratory, Upton, NY 11973 (United States); Benjamin, D [Duke University, Durham, NC 27708 (United States); Montoya, G Carillo; Guan, W; Mellado, B; Xu, N [University of Wisconsin-Madison, Madison, WI 53706 (United States); Cranmer, K; Shibata, A [New York University, New York, NY 10003 (United States)

    2010-04-01

    The Parallel ROOT Facility - PROOF is a distributed analysis system which allows to exploit inherent event level parallelism of high energy physics data. PROOF can be configured to work with centralized storage systems, but it is especially effective together with distributed local storage systems - like Xrootd, when data are distributed over computing nodes. It works efficiently on different types of hardware and scales well from a multi-core laptop to large computing farms. From that point of view it is well suited for both large central analysis facilities and Tier 3 type analysis farms. PROOF can be used in interactive or batch like regimes. The interactive regime allows the user to work with typically distributed data from the ROOT command prompt and get a real time feedback on analysis progress and intermediate results. We will discuss our experience with PROOF in the context of ATLAS Collaboration distributed analysis. In particular we will discuss PROOF performance in various analysis scenarios and in multi-user, multi-session environments. We will also describe PROOF integration with the ATLAS distributed data management system and prospects of running PROOF on geographically distributed analysis farms.

  6. Parallel hierarchical global illumination

    Energy Technology Data Exchange (ETDEWEB)

    Snell, Quinn O. [Iowa State Univ., Ames, IA (United States)

    1997-10-08

    Solving the global illumination problem is equivalent to determining the intensity of every wavelength of light in all directions at every point in a given scene. The complexity of the problem has led researchers to use approximation methods for solving the problem on serial computers. Rather than using an approximation method, such as backward ray tracing or radiosity, the authors have chosen to solve the Rendering Equation by direct simulation of light transport from the light sources. This paper presents an algorithm that solves the Rendering Equation to any desired accuracy, and can be run in parallel on distributed memory or shared memory computer systems with excellent scaling properties. It appears superior in both speed and physical correctness to recent published methods involving bidirectional ray tracing or hybrid treatments of diffuse and specular surfaces. Like progressive radiosity methods, it dynamically refines the geometry decomposition where required, but does so without the excessive storage requirements for ray histories. The algorithm, called Photon, produces a scene which converges to the global illumination solution. This amounts to a huge task for a 1997-vintage serial computer, but using the power of a parallel supercomputer significantly reduces the time required to generate a solution. Currently, Photon can be run on most parallel environments from a shared memory multiprocessor to a parallel supercomputer, as well as on clusters of heterogeneous workstations.

  7. Computational cost of isogeometric multi-frontal solvers on parallel distributed memory machines

    KAUST Repository

    Woźniak, Maciej; Paszyński, Maciej R.; Pardo, D.; Dalcin, Lisandro; Calo, Victor M.

    2015-01-01

    This paper derives theoretical estimates of the computational cost for isogeometric multi-frontal direct solver executed on parallel distributed memory machines. We show theoretically that for the Cp-1 global continuity of the isogeometric solution

  8. Current distribution in parallel paths of the coils of a 50 Hz prototype dipole magnet

    International Nuclear Information System (INIS)

    Otter, A.J.

    1995-06-01

    The prototype dipole made for TRIUMF's Kaon Factory proposal used coils with 12 parallel paths to reduce eddy current losses in the conductors. The ac current distribution in these paths was non-uniform due to different self and mutual inductances. Small differences in inductance can cause large circulating currents in the parallel windings. This paper describes the measurement of the inductances and shows an attempt to predict the current distribution for two alternative connection schemes. (author). 4 refs., 8 figs

  9. Parallelization of MCNP 4, a Monte Carlo neutron and photon transport code system, in highly parallel distributed memory type computer

    International Nuclear Information System (INIS)

    Masukawa, Fumihiro; Takano, Makoto; Naito, Yoshitaka; Yamazaki, Takao; Fujisaki, Masahide; Suzuki, Koichiro; Okuda, Motoi.

    1993-11-01

    In order to improve the accuracy and calculating speed of shielding analyses, MCNP 4, a Monte Carlo neutron and photon transport code system, has been parallelized and measured of its efficiency in the highly parallel distributed memory type computer, AP1000. The code has been analyzed statically and dynamically, then the suitable algorithm for parallelization has been determined for the shielding analysis functions of MCNP 4. This includes a strategy where a new history is assigned to the idling processor element dynamically during the execution. Furthermore, to avoid the congestion of communicative processing, the batch concept, processing multi-histories by a unit, has been introduced. By analyzing a sample cask problem with 2,000,000 histories by the AP1000 with 512 processor elements, the 82 % of parallelization efficiency is achieved, and the calculational speed has been estimated to be around 50 times as fast as that of FACOM M-780. (author)

  10. Large Survey Database: A Distributed Framework for Storage and Analysis of Large Datasets

    Science.gov (United States)

    Juric, Mario

    2011-01-01

    The Large Survey Database (LSD) is a Python framework and DBMS for distributed storage, cross-matching and querying of large survey catalogs (>10^9 rows, >1 TB). The primary driver behind its development is the analysis of Pan-STARRS PS1 data. It is specifically optimized for fast queries and parallel sweeps of positionally and temporally indexed datasets. It transparently scales to more than >10^2 nodes, and can be made to function in "shared nothing" architectures. An LSD database consists of a set of vertically and horizontally partitioned tables, physically stored as compressed HDF5 files. Vertically, we partition the tables into groups of related columns ('column groups'), storing together logically related data (e.g., astrometry, photometry). Horizontally, the tables are partitioned into partially overlapping ``cells'' by position in space (lon, lat) and time (t). This organization allows for fast lookups based on spatial and temporal coordinates, as well as data and task distribution. The design was inspired by the success of Google BigTable (Chang et al., 2006). Our programming model is a pipelined extension of MapReduce (Dean and Ghemawat, 2004). An SQL-like query language is used to access data. For complex tasks, map-reduce ``kernels'' that operate on query results on a per-cell basis can be written, with the framework taking care of scheduling and execution. The combination leverages users' familiarity with SQL, while offering a fully distributed computing environment. LSD adds little overhead compared to direct Python file I/O. In tests, we sweeped through 1.1 Grows of PanSTARRS+SDSS data (220GB) less than 15 minutes on a dual CPU machine. In a cluster environment, we achieved bandwidths of 17Gbits/sec (I/O limited). Based on current experience, we believe LSD should scale to be useful for analysis and storage of LSST-scale datasets. It can be downloaded from http://mwscience.net/lsd.

  11. Issues in developing parallel iterative algorithms for solving partial differential equations on a (transputer-based) distributed parallel computing system

    International Nuclear Information System (INIS)

    Rajagopalan, S.; Jethra, A.; Khare, A.N.; Ghodgaonkar, M.D.; Srivenkateshan, R.; Menon, S.V.G.

    1990-01-01

    Issues relating to implementing iterative procedures, for numerical solution of elliptic partial differential equations, on a distributed parallel computing system are discussed. Preliminary investigations show that a speed-up of about 3.85 is achievable on a four transputer pipeline network. (author). 2 figs., 3 a ppendixes., 7 refs

  12. Research on parallel algorithm for sequential pattern mining

    Science.gov (United States)

    Zhou, Lijuan; Qin, Bai; Wang, Yu; Hao, Zhongxiao

    2008-03-01

    Sequential pattern mining is the mining of frequent sequences related to time or other orders from the sequence database. Its initial motivation is to discover the laws of customer purchasing in a time section by finding the frequent sequences. In recent years, sequential pattern mining has become an important direction of data mining, and its application field has not been confined to the business database and has extended to new data sources such as Web and advanced science fields such as DNA analysis. The data of sequential pattern mining has characteristics as follows: mass data amount and distributed storage. Most existing sequential pattern mining algorithms haven't considered the above-mentioned characteristics synthetically. According to the traits mentioned above and combining the parallel theory, this paper puts forward a new distributed parallel algorithm SPP(Sequential Pattern Parallel). The algorithm abides by the principal of pattern reduction and utilizes the divide-and-conquer strategy for parallelization. The first parallel task is to construct frequent item sets applying frequent concept and search space partition theory and the second task is to structure frequent sequences using the depth-first search method at each processor. The algorithm only needs to access the database twice and doesn't generate the candidated sequences, which abates the access time and improves the mining efficiency. Based on the random data generation procedure and different information structure designed, this paper simulated the SPP algorithm in a concrete parallel environment and implemented the AprioriAll algorithm. The experiments demonstrate that compared with AprioriAll, the SPP algorithm had excellent speedup factor and efficiency.

  13. Experience with highly-parallel software for the storage system of the ATLAS Experiment at CERN

    CERN Document Server

    Colombo, T; The ATLAS collaboration

    2012-01-01

    The ATLAS experiment is observing proton-proton collisions delivered by the LHC accelerator. The ATLAS Trigger and Data Acquisition (TDAQ) system selects interesting events on-line in a three-level trigger system in order to store them at a budgeted rate of several hundred Hz. This paper focuses on the TDAQ data-logging system and in particular on the implementation and performance of a novel parallel software design. In this respect, the main challenge presented by the data-logging workload is the conflict between the largely parallel nature of the event processing, especially the recently introduced event compression, and the constraint of sequential file writing and checksum evaluation. This is further complicated by the necessity of operating in a fully data-driven mode, to cope with continuously evolving trigger and detector configurations. In this paper we report on the design of the new ATLAS on-line storage software. In particular we will discuss our development experience using recent concurrency-ori...

  14. The flow distribution in the parallel tubes of the cavity receiver under variable heat flux

    International Nuclear Information System (INIS)

    Hao, Yun; Wang, Yueshe; Hu, Tian

    2016-01-01

    Highlights: • An experimental loop is built to find the flow distribution in the parallel tubes. • With the concentration of heat flux, two-phase flow makes distribution more uneven. • The total flow rate is chosen appropriately for a wider heat flux distribution. • A suitable system pressure is essential for the optimization of flow distribution. - Abstract: As an optical component of tower solar thermal power station, the heliostat mirror reflects sunlight to one point of the heated surface in the solar cavity receiver, called as one-point focusing system. The radiation heat flux concentrated in the cavity receiver is always non-uniform temporally and spatially, which may lead to extremely local over-heat on the receiver evaporation panels. In this paper, an electrical heated evaporating experimental loop, including five parallel vertical tubes, is set up to evaluate the hydrodynamic characteristics of evaporation panels in a solar cavity receiver under various non-uniform heat flux. The influence of the heat flux concentration ratio, total flow rate, and system pressure on the flow distribution of parallel tubes is discussed. It is found that the flow distribution becomes significantly worse with the increase of heat flux and concentration ratio; and as the system pressure decreased, the flow distribution is improved. It is extremely important to obtain these interesting findings for the safe and stable operation of solar cavity receiver, and can also provide valuable references for the design and optimization of operating parameters solar tower power station system.

  15. A Communication-less Distributed Control Architecture for Islanded Microgrids with Renewable Generation and Storage

    DEFF Research Database (Denmark)

    Aldana, Nelson Leonardo Diaz; Quintero, Juan Carlos Vasquez; Guerrero, Josep M.

    2018-01-01

    proposes a coordination architecture for islanded ac microgrids, which considers the appropriate charge profiles for battery-based energy storage systems. The architecture is based on distributed decisionmaking mechanisms, which use only local measurements for determining the operation mode of each unit......For reliable operation of an islanded microgrid, at least one of its distributed resources should assume the responsibility of forming the off-grid power system. This responsibility is usually assumed by energy storage systems based on their capability of compensating the unbalance between...... independently. The coordination relies on a bus-signalling method, which enables the distributed units to have a global perception about the operation of the microgrid, without any communication infrastructure. The proposed architecture includes cooperative operation between distributed energy storage systems...

  16. RAID Unbound: Storage Fault Tolerance in a Distributed Environment

    Science.gov (United States)

    Ritchie, Brian

    1996-01-01

    Mirroring, data replication, backup, and more recently, redundant arrays of independent disks (RAID) are all technologies used to protect and ensure access to critical company data. A new set of problems has arisen as data becomes more and more geographically distributed. Each of the technologies listed above provides important benefits; but each has failed to adapt fully to the realities of distributed computing. The key to data high availability and protection is to take the technologies' strengths and 'virtualize' them across a distributed network. RAID and mirroring offer high data availability, which data replication and backup provide strong data protection. If we take these concepts at a very granular level (defining user, record, block, file, or directory types) and them liberate them from the physical subsystems with which they have traditionally been associated, we have the opportunity to create a highly scalable network wide storage fault tolerance. The network becomes the virtual storage space in which the traditional concepts of data high availability and protection are implemented without their corresponding physical constraints.

  17. Modelling distribution of evaporating CO2 in parallel minichannels

    DEFF Research Database (Denmark)

    Brix, Wiebke; Kærn, Martin Ryhl; Elmegaard, Brian

    2010-01-01

    The effects of airflow non-uniformity and uneven inlet qualities on the performance of a minichannel evaporator with parallel channels, using CO2 as refrigerant, are investigated numerically. For this purpose a one-dimensional discretised steady-state model was developed, applying well-known empi......The effects of airflow non-uniformity and uneven inlet qualities on the performance of a minichannel evaporator with parallel channels, using CO2 as refrigerant, are investigated numerically. For this purpose a one-dimensional discretised steady-state model was developed, applying well...... to maldistribution of the refrigerant and considerable capacity reduction of the evaporator. Uneven inlet ualities to the different channels show only minor effects on the refrigerant distribution and evaporator capacity as long as the channels are vertically oriented with CO2 flowing upwards. For horizontal...... channels capacity reductions are found for both non-uniform airflow and uneven inlet qualities. For horizontal minichannels the results are very similar to those obtained using R134a as refrigerant....

  18. Multiagent-Based Distributed State of Charge Balancing Control for Distributed Energy Storage Units in AC Microgrids

    DEFF Research Database (Denmark)

    Li, Chendan; Coelho, Ernane Antônio Alves; Dragicevic, Tomislav

    2017-01-01

    In this paper, a multiagent-based distributed control algorithm has been proposed to achieve state of charge (SoC) balance of distributed energy storage (DES) units in an ac microgrid. The proposal uses frequency scheduling instead of adaptive droop gain to regulate the active power. Each DES unit...

  19. Probabilistic tools for planning and operating power systems with distributed energy storage

    DEFF Research Database (Denmark)

    Klöckl, Bernd; Papaefthymiou, George; Pinson, Pierre

    2008-01-01

    Stochastic energy flows are an increasingly important phenomenon in today's power system planning and operation. They are – among other reasons – caused by large amounts of stochastic generation such as wind. The inclusion of energy storage devices, distributed in future systems (distributed energy...... owners are either the grid operators, the generation owners, or the energy traders. For the grid operators being the DES owners, storage operation will have to be integrated into the planning of the system, therefore multivariate nonparametric time series analysis and synthesis methods have to be applied...... to recorded data of stochastic energy resources. Together with suited storage models, the implications of DES on the planning of the system can then be assessed. For the producers or traders being the owners of the DES, the topic to be addressed is the real-time operation of each storage device in the power...

  20. A lightweight, flow-based toolkit for parallel and distributed bioinformatics pipelines

    Directory of Open Access Journals (Sweden)

    Cieślik Marcin

    2011-02-01

    Full Text Available Abstract Background Bioinformatic analyses typically proceed as chains of data-processing tasks. A pipeline, or 'workflow', is a well-defined protocol, with a specific structure defined by the topology of data-flow interdependencies, and a particular functionality arising from the data transformations applied at each step. In computer science, the dataflow programming (DFP paradigm defines software systems constructed in this manner, as networks of message-passing components. Thus, bioinformatic workflows can be naturally mapped onto DFP concepts. Results To enable the flexible creation and execution of bioinformatics dataflows, we have written a modular framework for parallel pipelines in Python ('PaPy'. A PaPy workflow is created from re-usable components connected by data-pipes into a directed acyclic graph, which together define nested higher-order map functions. The successive functional transformations of input data are evaluated on flexibly pooled compute resources, either local or remote. Input items are processed in batches of adjustable size, all flowing one to tune the trade-off between parallelism and lazy-evaluation (memory consumption. An add-on module ('NuBio' facilitates the creation of bioinformatics workflows by providing domain specific data-containers (e.g., for biomolecular sequences, alignments, structures and functionality (e.g., to parse/write standard file formats. Conclusions PaPy offers a modular framework for the creation and deployment of parallel and distributed data-processing workflows. Pipelines derive their functionality from user-written, data-coupled components, so PaPy also can be viewed as a lightweight toolkit for extensible, flow-based bioinformatics data-processing. The simplicity and flexibility of distributed PaPy pipelines may help users bridge the gap between traditional desktop/workstation and grid computing. PaPy is freely distributed as open-source Python code at http://muralab.org/PaPy, and

  1. Design of multiple sequence alignment algorithms on parallel, distributed memory supercomputers.

    Science.gov (United States)

    Church, Philip C; Goscinski, Andrzej; Holt, Kathryn; Inouye, Michael; Ghoting, Amol; Makarychev, Konstantin; Reumann, Matthias

    2011-01-01

    The challenge of comparing two or more genomes that have undergone recombination and substantial amounts of segmental loss and gain has recently been addressed for small numbers of genomes. However, datasets of hundreds of genomes are now common and their sizes will only increase in the future. Multiple sequence alignment of hundreds of genomes remains an intractable problem due to quadratic increases in compute time and memory footprint. To date, most alignment algorithms are designed for commodity clusters without parallelism. Hence, we propose the design of a multiple sequence alignment algorithm on massively parallel, distributed memory supercomputers to enable research into comparative genomics on large data sets. Following the methodology of the sequential progressiveMauve algorithm, we design data structures including sequences and sorted k-mer lists on the IBM Blue Gene/P supercomputer (BG/P). Preliminary results show that we can reduce the memory footprint so that we can potentially align over 250 bacterial genomes on a single BG/P compute node. We verify our results on a dataset of E.coli, Shigella and S.pneumoniae genomes. Our implementation returns results matching those of the original algorithm but in 1/2 the time and with 1/4 the memory footprint for scaffold building. In this study, we have laid the basis for multiple sequence alignment of large-scale datasets on a massively parallel, distributed memory supercomputer, thus enabling comparison of hundreds instead of a few genome sequences within reasonable time.

  2. Efficiency Analysis of the Parallel Implementation of the SIMPLE Algorithm on Multiprocessor Computers

    Science.gov (United States)

    Lashkin, S. V.; Kozelkov, A. S.; Yalozo, A. V.; Gerasimov, V. Yu.; Zelensky, D. K.

    2017-12-01

    This paper describes the details of the parallel implementation of the SIMPLE algorithm for numerical solution of the Navier-Stokes system of equations on arbitrary unstructured grids. The iteration schemes for the serial and parallel versions of the SIMPLE algorithm are implemented. In the description of the parallel implementation, special attention is paid to computational data exchange among processors under the condition of the grid model decomposition using fictitious cells. We discuss the specific features for the storage of distributed matrices and implementation of vector-matrix operations in parallel mode. It is shown that the proposed way of matrix storage reduces the number of interprocessor exchanges. A series of numerical experiments illustrates the effect of the multigrid SLAE solver tuning on the general efficiency of the algorithm; the tuning involves the types of the cycles used (V, W, and F), the number of iterations of a smoothing operator, and the number of cells for coarsening. Two ways (direct and indirect) of efficiency evaluation for parallelization of the numerical algorithm are demonstrated. The paper presents the results of solving some internal and external flow problems with the evaluation of parallelization efficiency by two algorithms. It is shown that the proposed parallel implementation enables efficient computations for the problems on a thousand processors. Based on the results obtained, some general recommendations are made for the optimal tuning of the multigrid solver, as well as for selecting the optimal number of cells per processor.

  3. Parallel peak pruning for scalable SMP contour tree computation

    Energy Technology Data Exchange (ETDEWEB)

    Carr, Hamish A. [Univ. of Leeds (United Kingdom); Weber, Gunther H. [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Univ. of California, Davis, CA (United States); Sewell, Christopher M. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Ahrens, James P. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2017-03-09

    As data sets grow to exascale, automated data analysis and visualisation are increasingly important, to intermediate human understanding and to reduce demands on disk storage via in situ analysis. Trends in architecture of high performance computing systems necessitate analysis algorithms to make effective use of combinations of massively multicore and distributed systems. One of the principal analytic tools is the contour tree, which analyses relationships between contours to identify features of more than local importance. Unfortunately, the predominant algorithms for computing the contour tree are explicitly serial, and founded on serial metaphors, which has limited the scalability of this form of analysis. While there is some work on distributed contour tree computation, and separately on hybrid GPU-CPU computation, there is no efficient algorithm with strong formal guarantees on performance allied with fast practical performance. Here in this paper, we report the first shared SMP algorithm for fully parallel contour tree computation, withfor-mal guarantees of O(lgnlgt) parallel steps and O(n lgn) work, and implementations with up to 10x parallel speed up in OpenMP and up to 50x speed up in NVIDIA Thrust.

  4. Study of Solid State Drives performance in PROOF distributed analysis system

    Science.gov (United States)

    Panitkin, S. Y.; Ernst, M.; Petkus, R.; Rind, O.; Wenaus, T.

    2010-04-01

    Solid State Drives (SSD) is a promising storage technology for High Energy Physics parallel analysis farms. Its combination of low random access time and relatively high read speed is very well suited for situations where multiple jobs concurrently access data located on the same drive. It also has lower energy consumption and higher vibration tolerance than Hard Disk Drive (HDD) which makes it an attractive choice in many applications raging from personal laptops to large analysis farms. The Parallel ROOT Facility - PROOF is a distributed analysis system which allows to exploit inherent event level parallelism of high energy physics data. PROOF is especially efficient together with distributed local storage systems like Xrootd, when data are distributed over computing nodes. In such an architecture the local disk subsystem I/O performance becomes a critical factor, especially when computing nodes use multi-core CPUs. We will discuss our experience with SSDs in PROOF environment. We will compare performance of HDD with SSD in I/O intensive analysis scenarios. In particular we will discuss PROOF system performance scaling with a number of simultaneously running analysis jobs.

  5. On the effectiveness of recoding-based repair in network coded distributed storage

    DEFF Research Database (Denmark)

    Sipos, Marton A.; Braun, Patrik J.; Roetter, Daniel Enrique Lucani

    2017-01-01

    High capacity storage systems distribute less across several storage devices (nodes) and apply an erasure code to meet availability and reliability requirements. Since devices can lose network connectivity or fail permanently, a dynamic repair mechanism must be put in place. In such cases a new r...

  6. Distributed system for parallel data processing of ECT signals for electromagnetic flaw detection in materials

    International Nuclear Information System (INIS)

    Guliashki, Vassil; Marinova, Galia

    2002-01-01

    The paper proposes a distributed system for parallel data processing of ECT signals for flaw detection in materials. The measured data are stored in files on a host computer, where a JAVA server is located. The host computer is connected through Internet to a set of client computers, distributed geographically. The data are distributed from the host computer by means of the JAVA server to the client computers according their requests. The software necessary for the data processing is installed on each client computer in advance. The organization of the data processing on many computers, working simultaneously in parallel, leads to great time reducing, especially in cases when huge amount of data should be processed in very short time. (Author)

  7. Mass-storage management for distributed image/video archives

    Science.gov (United States)

    Franchi, Santina; Guarda, Roberto; Prampolini, Franco

    1993-04-01

    The realization of image/video database requires a specific design for both database structures and mass storage management. This issue has addressed the project of the digital image/video database system that has been designed at IBM SEMEA Scientific & Technical Solution Center. Proper database structures have been defined to catalog image/video coding technique with the related parameters, and the description of image/video contents. User workstations and servers are distributed along a local area network. Image/video files are not managed directly by the DBMS server. Because of their wide size, they are stored outside the database on network devices. The database contains the pointers to the image/video files and the description of the storage devices. The system can use different kinds of storage media, organized in a hierarchical structure. Three levels of functions are available to manage the storage resources. The functions of the lower level provide media management. They allow it to catalog devices and to modify device status and device network location. The medium level manages image/video files on a physical basis. It manages file migration between high capacity media and low access time media. The functions of the upper level work on image/video file on a logical basis, as they archive, move and copy image/video data selected by user defined queries. These functions are used to support the implementation of a storage management strategy. The database information about characteristics of both storage devices and coding techniques are used by the third level functions to fit delivery/visualization requirements and to reduce archiving costs.

  8. Equalization Algorithm for Distributed Energy Storage Systems in Islanded AC Microgrids

    DEFF Research Database (Denmark)

    Aldana, Nelson Leonardo Diaz; Hernández, Adriana Carolina Luna; Quintero, Juan Carlos Vasquez

    2015-01-01

    This paper presents a centralized strategy for equalizing the state of charge of distributed energy storage systems in an islanded ac microgrid. The strategy is based on a simple algorithm denoted as equalization algorithm, which modifies the charge or discharge ratio on the time, for distributed...

  9. A portable implementation of ARPACK for distributed memory parallel architectures

    Energy Technology Data Exchange (ETDEWEB)

    Maschhoff, K.J.; Sorensen, D.C.

    1996-12-31

    ARPACK is a package of Fortran 77 subroutines which implement the Implicitly Restarted Arnoldi Method used for solving large sparse eigenvalue problems. A parallel implementation of ARPACK is presented which is portable across a wide range of distributed memory platforms and requires minimal changes to the serial code. The communication layers used for message passing are the Basic Linear Algebra Communication Subprograms (BLACS) developed for the ScaLAPACK project and Message Passing Interface(MPI).

  10. Parallel Distributed Processing at 25: Further Explorations in the Microstructure of Cognition

    Science.gov (United States)

    Rogers, Timothy T.; McClelland, James L.

    2014-01-01

    This paper introduces a special issue of "Cognitive Science" initiated on the 25th anniversary of the publication of "Parallel Distributed Processing" (PDP), a two-volume work that introduced the use of neural network models as vehicles for understanding cognition. The collection surveys the core commitments of the PDP…

  11. RAMA: A file system for massively parallel computers

    Science.gov (United States)

    Miller, Ethan L.; Katz, Randy H.

    1993-01-01

    This paper describes a file system design for massively parallel computers which makes very efficient use of a few disks per processor. This overcomes the traditional I/O bottleneck of massively parallel machines by storing the data on disks within the high-speed interconnection network. In addition, the file system, called RAMA, requires little inter-node synchronization, removing another common bottleneck in parallel processor file systems. Support for a large tertiary storage system can easily be integrated in lo the file system; in fact, RAMA runs most efficiently when tertiary storage is used.

  12. Distributed Memory Parallel Computing with SEAWAT

    Science.gov (United States)

    Verkaik, J.; Huizer, S.; van Engelen, J.; Oude Essink, G.; Ram, R.; Vuik, K.

    2017-12-01

    Fresh groundwater reserves in coastal aquifers are threatened by sea-level rise, extreme weather conditions, increasing urbanization and associated groundwater extraction rates. To counteract these threats, accurate high-resolution numerical models are required to optimize the management of these precious reserves. The major model drawbacks are long run times and large memory requirements, limiting the predictive power of these models. Distributed memory parallel computing is an efficient technique for reducing run times and memory requirements, where the problem is divided over multiple processor cores. A new Parallel Krylov Solver (PKS) for SEAWAT is presented. PKS has recently been applied to MODFLOW and includes Conjugate Gradient (CG) and Biconjugate Gradient Stabilized (BiCGSTAB) linear accelerators. Both accelerators are preconditioned by an overlapping additive Schwarz preconditioner in a way that: a) subdomains are partitioned using Recursive Coordinate Bisection (RCB) load balancing, b) each subdomain uses local memory only and communicates with other subdomains by Message Passing Interface (MPI) within the linear accelerator, c) it is fully integrated in SEAWAT. Within SEAWAT, the PKS-CG solver replaces the Preconditioned Conjugate Gradient (PCG) solver for solving the variable-density groundwater flow equation and the PKS-BiCGSTAB solver replaces the Generalized Conjugate Gradient (GCG) solver for solving the advection-diffusion equation. PKS supports the third-order Total Variation Diminishing (TVD) scheme for computing advection. Benchmarks were performed on the Dutch national supercomputer (https://userinfo.surfsara.nl/systems/cartesius) using up to 128 cores, for a synthetic 3D Henry model (100 million cells) and the real-life Sand Engine model ( 10 million cells). The Sand Engine model was used to investigate the potential effect of the long-term morphological evolution of a large sand replenishment and climate change on fresh groundwater resources

  13. Stampi: a message passing library for distributed parallel computing. User's guide, second edition

    International Nuclear Information System (INIS)

    Imamura, Toshiyuki; Koide, Hiroshi; Takemiya, Hiroshi

    2000-02-01

    A new message passing library, Stampi, has been developed to realize a computation with different kind of parallel computers arbitrarily and making MPI (Message Passing Interface) as an unique interface for communication. Stampi is based on the MPI2 specification, and it realizes dynamic process creation to different machines and communication between spawned one within the scope of MPI semantics. Main features of Stampi are summarized as follows: (i) an automatic switch function between external- and internal communications, (ii) a message routing/relaying with a routing module, (iii) a dynamic process creation, (iv) a support of two types of connection, Master/Slave and Client/Server, (v) a support of a communication with Java applets. Indeed vendors implemented MPI libraries as a closed system in one parallel machine or their systems, and did not support both functions; process creation and communication to external machines. Stampi supports both functions and enables us distributed parallel computing. Currently Stampi has been implemented on COMPACS (COMplex PArallel Computer System) introduced in CCSE, five parallel computers and one graphic workstation, moreover on eight kinds of parallel machines, totally fourteen systems. Stampi provides us MPI communication functionality on them. This report describes mainly the usage of Stampi. (author)

  14. Searchable Data Vault: Encrypted Queries in Secure Distributed Cloud Storage

    Directory of Open Access Journals (Sweden)

    Geong Sen Poh

    2017-05-01

    Full Text Available Cloud storage services allow users to efficiently outsource their documents anytime and anywhere. Such convenience, however, leads to privacy concerns. While storage providers may not read users’ documents, attackers may possibly gain access by exploiting vulnerabilities in the storage system. Documents may also be leaked by curious administrators. A simple solution is for the user to encrypt all documents before submitting them. This method, however, makes it impossible to efficiently search for documents as they are all encrypted. To resolve this problem, we propose a multi-server searchable symmetric encryption (SSE scheme and construct a system called the searchable data vault (SDV. A unique feature of the scheme is that it allows an encrypted document to be divided into blocks and distributed to different storage servers so that no single storage provider has a complete document. By incorporating the scheme, the SDV protects the privacy of documents while allowing for efficient private queries. It utilizes a web interface and a controller that manages user credentials, query indexes and submission of encrypted documents to cloud storage services. It is also the first system that enables a user to simultaneously outsource and privately query documents from a few cloud storage services. Our preliminary performance evaluation shows that this feature introduces acceptable computation overheads when compared to submitting documents directly to a cloud storage service.

  15. Using high performance interconnects in a distributed computing and mass storage environment

    International Nuclear Information System (INIS)

    Ernst, M.

    1994-01-01

    Detector Collaborations of the HERA Experiments typically involve more than 500 physicists from a few dozen institutes. These physicists require access to large amounts of data in a fully transparent manner. Important issues include Distributed Mass Storage Management Systems in a Distributed and Heterogeneous Computing Environment. At the very center of a distributed system, including tens of CPUs and network attached mass storage peripherals are the communication links. Today scientists are witnessing an integration of computing and communication technology with the open-quote network close-quote becoming the computer. This contribution reports on a centrally operated computing facility for the HERA Experiments at DESY, including Symmetric Multiprocessor Machines (84 Processors), presently more than 400 GByte of magnetic disk and 40 TB of automoted tape storage, tied together by a HIPPI open-quote network close-quote. Focussing on the High Performance Interconnect technology, details will be provided about the HIPPI based open-quote Backplane close-quote configured around a 20 Gigabit/s Multi Media Router and the performance and efficiency of the related computer interfaces

  16. Improving the analysis, storage and sharing of neuroimaging data using relational databases and distributed computing.

    Science.gov (United States)

    Hasson, Uri; Skipper, Jeremy I; Wilde, Michael J; Nusbaum, Howard C; Small, Steven L

    2008-01-15

    The increasingly complex research questions addressed by neuroimaging research impose substantial demands on computational infrastructures. These infrastructures need to support management of massive amounts of data in a way that affords rapid and precise data analysis, to allow collaborative research, and to achieve these aims securely and with minimum management overhead. Here we present an approach that overcomes many current limitations in data analysis and data sharing. This approach is based on open source database management systems that support complex data queries as an integral part of data analysis, flexible data sharing, and parallel and distributed data processing using cluster computing and Grid computing resources. We assess the strengths of these approaches as compared to current frameworks based on storage of binary or text files. We then describe in detail the implementation of such a system and provide a concrete description of how it was used to enable a complex analysis of fMRI time series data.

  17. Parallel evolution of storage roots in Morning Glories (Convolvulaceae)

    Science.gov (United States)

    Storage roots are an ecologically and agriculturally important plant trait. In morning glories, storage roots are well characterized in the crop species sweetpotato. Storage roots have evolved numerous times across the morning glory family. This study aims to understand whether this was through para...

  18. Parallel Distributed Processing Theory in the Age of Deep Networks.

    Science.gov (United States)

    Bowers, Jeffrey S

    2017-12-01

    Parallel distributed processing (PDP) models in psychology are the precursors of deep networks used in computer science. However, only PDP models are associated with two core psychological claims, namely that all knowledge is coded in a distributed format and cognition is mediated by non-symbolic computations. These claims have long been debated in cognitive science, and recent work with deep networks speaks to this debate. Specifically, single-unit recordings show that deep networks learn units that respond selectively to meaningful categories, and researchers are finding that deep networks need to be supplemented with symbolic systems to perform some tasks. Given the close links between PDP and deep networks, it is surprising that research with deep networks is challenging PDP theory. Copyright © 2017. Published by Elsevier Ltd.

  19. Implementation and performance of parallelized elegant

    International Nuclear Information System (INIS)

    Wang, Y.; Borland, M.

    2008-01-01

    The program elegant is widely used for design and modeling of linacs for free-electron lasers and energy recovery linacs, as well as storage rings and other applications. As part of a multi-year effort, we have parallelized many aspects of the code, including single-particle dynamics, wakefields, and coherent synchrotron radiation. We report on the approach used for gradual parallelization, which proved very beneficial in getting parallel features into the hands of users quickly. We also report details of parallelization of collective effects. Finally, we discuss performance of the parallelized code in various applications.

  20. POSIX and Object Distributed Storage Systems Performance Comparison Studies With Real-Life Scenarios in an Experimental Data Taking Context Leveraging OpenStack Swift & Ceph

    Science.gov (United States)

    Poat, M. D.; Lauret, J.; Betts, W.

    2015-12-01

    The STAR online computing infrastructure has become an intensive dynamic system used for first-hand data collection and analysis resulting in a dense collection of data output. As we have transitioned to our current state, inefficient, limited storage systems have become an impediment to fast feedback to online shift crews. Motivation for a centrally accessible, scalable and redundant distributed storage system had become a necessity in this environment. OpenStack Swift Object Storage and Ceph Object Storage are two eye-opening technologies as community use and development have led to success elsewhere. In this contribution, OpenStack Swift and Ceph have been put to the test with single and parallel I/O tests, emulating real world scenarios for data processing and workflows. The Ceph file system storage, offering a POSIX compliant file system mounted similarly to an NFS share was of particular interest as it aligned with our requirements and was retained as our solution. I/O performance tests were run against the Ceph POSIX file system and have presented surprising results indicating true potential for fast I/O and reliability. STAR'S online compute farm historical use has been for job submission and first hand data analysis. The goal of reusing the online compute farm to maintain a storage cluster and job submission will be an efficient use of the current infrastructure.

  1. Design of a chemical batch plant : a study of dedicated parallel lines with intermediate storage and the plant performance

    OpenAIRE

    Verbiest, Floor; Cornelissens, Trijntje; Springael, Johan

    2016-01-01

    Abstract: Production plants worldwide face huge challenges in satisfying high service levels and outperforming competition. These challenges require appropriate strategic decisions on plant design and production strategies. In this paper, we focus on multiproduct chemical batch plants, which are typically equipped with multiple production lines and intermediate storage tanks. First we extend the existing MI(N) LP design models with the concept of parallel production lines, and optimise the as...

  2. Integration experiences and performance studies of A COTS parallel archive systems

    Energy Technology Data Exchange (ETDEWEB)

    Chen, Hsing-bung [Los Alamos National Laboratory; Scott, Cody [Los Alamos National Laboratory; Grider, Bary [Los Alamos National Laboratory; Torres, Aaron [Los Alamos National Laboratory; Turley, Milton [Los Alamos National Laboratory; Sanchez, Kathy [Los Alamos National Laboratory; Bremer, John [Los Alamos National Laboratory

    2010-01-01

    Current and future Archive Storage Systems have been asked to (a) scale to very high bandwidths, (b) scale in metadata performance, (c) support policy-based hierarchical storage management capability, (d) scale in supporting changing needs of very large data sets, (e) support standard interface, and (f) utilize commercial-off-the-shelf(COTS) hardware. Parallel file systems have been asked to do the same thing but at one or more orders of magnitude faster in performance. Archive systems continue to move closer to file systems in their design due to the need for speed and bandwidth, especially metadata searching speeds such as more caching and less robust semantics. Currently the number of extreme highly scalable parallel archive solutions is very small especially those that will move a single large striped parallel disk file onto many tapes in parallel. We believe that a hybrid storage approach of using COTS components and innovative software technology can bring new capabilities into a production environment for the HPC community much faster than the approach of creating and maintaining a complete end-to-end unique parallel archive software solution. In this paper, we relay our experience of integrating a global parallel file system and a standard backup/archive product with a very small amount of additional code to provide a scalable, parallel archive. Our solution has a high degree of overlap with current parallel archive products including (a) doing parallel movement to/from tape for a single large parallel file, (b) hierarchical storage management, (c) ILM features, (d) high volume (non-single parallel file) archives for backup/archive/content management, and (e) leveraging all free file movement tools in Linux such as copy, move, ls, tar, etc. We have successfully applied our working COTS Parallel Archive System to the current world's first petaflop/s computing system, LANL's Roadrunner, and demonstrated its capability to address requirements of

  3. Integration experiments and performance studies of a COTS parallel archive system

    Energy Technology Data Exchange (ETDEWEB)

    Chen, Hsing-bung [Los Alamos National Laboratory; Scott, Cody [Los Alamos National Laboratory; Grider, Gary [Los Alamos National Laboratory; Torres, Aaron [Los Alamos National Laboratory; Turley, Milton [Los Alamos National Laboratory; Sanchez, Kathy [Los Alamos National Laboratory; Bremer, John [Los Alamos National Laboratory

    2010-06-16

    Current and future Archive Storage Systems have been asked to (a) scale to very high bandwidths, (b) scale in metadata performance, (c) support policy-based hierarchical storage management capability, (d) scale in supporting changing needs of very large data sets, (e) support standard interface, and (f) utilize commercial-off-the-shelf (COTS) hardware. Parallel file systems have been asked to do the same thing but at one or more orders of magnitude faster in performance. Archive systems continue to move closer to file systems in their design due to the need for speed and bandwidth, especially metadata searching speeds such as more caching and less robust semantics. Currently the number of extreme highly scalable parallel archive solutions is very small especially those that will move a single large striped parallel disk file onto many tapes in parallel. We believe that a hybrid storage approach of using COTS components and innovative software technology can bring new capabilities into a production environment for the HPC community much faster than the approach of creating and maintaining a complete end-to-end unique parallel archive software solution. In this paper, we relay our experience of integrating a global parallel file system and a standard backup/archive product with a very small amount of additional code to provide a scalable, parallel archive. Our solution has a high degree of overlap with current parallel archive products including (a) doing parallel movement to/from tape for a single large parallel file, (b) hierarchical storage management, (c) ILM features, (d) high volume (non-single parallel file) archives for backup/archive/content management, and (e) leveraging all free file movement tools in Linux such as copy, move, Is, tar, etc. We have successfully applied our working COTS Parallel Archive System to the current world's first petafiop/s computing system, LANL's Roadrunner machine, and demonstrated its capability to address

  4. Template based parallel checkpointing in a massively parallel computer system

    Science.gov (United States)

    Archer, Charles Jens [Rochester, MN; Inglett, Todd Alan [Rochester, MN

    2009-01-13

    A method and apparatus for a template based parallel checkpoint save for a massively parallel super computer system using a parallel variation of the rsync protocol, and network broadcast. In preferred embodiments, the checkpoint data for each node is compared to a template checkpoint file that resides in the storage and that was previously produced. Embodiments herein greatly decrease the amount of data that must be transmitted and stored for faster checkpointing and increased efficiency of the computer system. Embodiments are directed to a parallel computer system with nodes arranged in a cluster with a high speed interconnect that can perform broadcast communication. The checkpoint contains a set of actual small data blocks with their corresponding checksums from all nodes in the system. The data blocks may be compressed using conventional non-lossy data compression algorithms to further reduce the overall checkpoint size.

  5. Studies of electron collisions with polyatomic molecules using distributed-memory parallel computers

    International Nuclear Information System (INIS)

    Winstead, C.; Hipes, P.G.; Lima, M.A.P.; McKoy, V.

    1991-01-01

    Elastic electron scattering cross sections from 5--30 eV are reported for the molecules C 2 H 4 , C 2 H 6 , C 3 H 8 , Si 2 H 6 , and GeH 4 , obtained using an implementation of the Schwinger multichannel method for distributed-memory parallel computer architectures. These results, obtained within the static-exchange approximation, are in generally good agreement with the available experimental data. These calculations demonstrate the potential of highly parallel computation in the study of collisions between low-energy electrons and polyatomic gases. The computational methodology discussed is also directly applicable to the calculation of elastic cross sections at higher levels of approximation (target polarization) and of electronic excitation cross sections

  6. Parallelization Issues and Particle-In Codes.

    Science.gov (United States)

    Elster, Anne Cathrine

    1994-01-01

    "Everything should be made as simple as possible, but not simpler." Albert Einstein. The field of parallel scientific computing has concentrated on parallelization of individual modules such as matrix solvers and factorizers. However, many applications involve several interacting modules. Our analyses of a particle-in-cell code modeling charged particles in an electric field, show that these accompanying dependencies affect data partitioning and lead to new parallelization strategies concerning processor, memory and cache utilization. Our test-bed, a KSR1, is a distributed memory machine with a globally shared addressing space. However, most of the new methods presented hold generally for hierarchical and/or distributed memory systems. We introduce a novel approach that uses dual pointers on the local particle arrays to keep the particle locations automatically partially sorted. Complexity and performance analyses with accompanying KSR benchmarks, have been included for both this scheme and for the traditional replicated grids approach. The latter approach maintains load-balance with respect to particles. However, our results demonstrate it fails to scale properly for problems with large grids (say, greater than 128-by-128) running on as few as 15 KSR nodes, since the extra storage and computation time associated with adding the grid copies, becomes significant. Our grid partitioning scheme, although harder to implement, does not need to replicate the whole grid. Consequently, it scales well for large problems on highly parallel systems. It may, however, require load balancing schemes for non-uniform particle distributions. Our dual pointer approach may facilitate this through dynamically partitioned grids. We also introduce hierarchical data structures that store neighboring grid-points within the same cache -line by reordering the grid indexing. This alignment produces a 25% savings in cache-hits for a 4-by-4 cache. A consideration of the input data's effect on

  7. PDDP, A Data Parallel Programming Model

    Directory of Open Access Journals (Sweden)

    Karen H. Warren

    1996-01-01

    Full Text Available PDDP, the parallel data distribution preprocessor, is a data parallel programming model for distributed memory parallel computers. PDDP implements high-performance Fortran-compatible data distribution directives and parallelism expressed by the use of Fortran 90 array syntax, the FORALL statement, and the WHERE construct. Distributed data objects belong to a global name space; other data objects are treated as local and replicated on each processor. PDDP allows the user to program in a shared memory style and generates codes that are portable to a variety of parallel machines. For interprocessor communication, PDDP uses the fastest communication primitives on each platform.

  8. Characterization of water distribution in bread during storage using magnetic resonance imaging.

    Science.gov (United States)

    Lodi, Alessia; Abduljalil, Amir M; Vodovotz, Yael

    2007-12-01

    A soy bread of fully acceptable quality and containing 49% soy ingredients (with or without 5% almond powder) has been recently developed in our laboratory. An investigation on water distribution and mobility, as probed by proton signal intensity and T2 magnetic resonance images, during storage was designed to examine possible relations between water states and hindered staling rate upon soy or soy-almond addition. Water proton distribution throughout soy-containing loaves was found to be very homogeneous in fresh breads with and without almond, with minimal water migration occurring during prolonged storage. In contrast, traditional wheat bread displayed an inhomogeneous water proton population that tended to change (with higher moisture migration towards the outer perimeter of the slice) during storage. Similar results were found for water mobility throughout the loaves, as depicted in T2 images. On intensity images of all considered bread varieties, the outer perimeter corresponding to the crust exhibited lower signal intensity due to decreased water content. Higher T2 values were found in the crust of soy breads with and without almond, which were attributed to lipids. The results indicated that the addition of soy to bread improved the homogeneous distribution of water molecules, which may hinder the staling rate of soy-containing breads. However, incorporation of almond had little effect on the water proton distribution or mobility of soy breads.

  9. A distributed charge storage with GeO2 nanodots

    International Nuclear Information System (INIS)

    Chang, T.C.; Yan, S.T.; Hsu, C.H.; Tang, M.T.; Lee, J.F.; Tai, Y.H.; Liu, P.T.; Sze, S.M.

    2004-01-01

    In this study, a distributed charge storage with GeO 2 nanodots is demonstrated. The mean size and aerial density of the nanodots embedded in SiO 2 are estimated to be about 5.5 nm and 4.3x10 11 cm -2 , respectively. The composition of the dots is also confirmed to be GeO 2 by x-ray absorption near-edge structure analyses. A significant memory effect is observed through the electrical measurements. Under the low voltage operation of 5 V, the memory window is estimated to ∼0.45 V. Also, a physical model is proposed to demonstrate the charge storage effect through the interfacial traps of GeO 2 nanodots

  10. Second International Workshop on Software Engineering and Code Design in Parallel Meteorological and Oceanographic Applications

    Science.gov (United States)

    OKeefe, Matthew (Editor); Kerr, Christopher L. (Editor)

    1998-01-01

    This report contains the abstracts and technical papers from the Second International Workshop on Software Engineering and Code Design in Parallel Meteorological and Oceanographic Applications, held June 15-18, 1998, in Scottsdale, Arizona. The purpose of the workshop is to bring together software developers in meteorology and oceanography to discuss software engineering and code design issues for parallel architectures, including Massively Parallel Processors (MPP's), Parallel Vector Processors (PVP's), Symmetric Multi-Processors (SMP's), Distributed Shared Memory (DSM) multi-processors, and clusters. Issues to be discussed include: (1) code architectures for current parallel models, including basic data structures, storage allocation, variable naming conventions, coding rules and styles, i/o and pre/post-processing of data; (2) designing modular code; (3) load balancing and domain decomposition; (4) techniques that exploit parallelism efficiently yet hide the machine-related details from the programmer; (5) tools for making the programmer more productive; and (6) the proliferation of programming models (F--, OpenMP, MPI, and HPF).

  11. Distributed Demand Side Management with Battery Storage for Smart Home Energy Scheduling

    Directory of Open Access Journals (Sweden)

    Omowunmi Mary Longe

    2017-01-01

    Full Text Available The role of Demand Side Management (DSM with Distributed Energy Storage (DES has been gaining attention in recent studies due to the impact of the latter on energy management in the smart grid. In this work, an Energy Scheduling and Distributed Storage (ESDS algorithm is proposed to be installed into the smart meters of Time-of-Use (TOU pricing consumers possessing in-home energy storage devices. Source of energy supply to the smart home appliances was optimized between the utility grid and the DES device depending on energy tariff and consumer demand satisfaction information. This is to minimize consumer energy expenditure and maximize demand satisfaction simultaneously. The ESDS algorithm was found to offer consumer-friendly and utility-friendly enhancements to the DSM program such as energy, financial, and investment savings, reduced/eliminated consumer dissatisfaction even at peak periods, Peak-to-Average-Ratio (PAR demand reduction, grid energy sustainability, socio-economic benefits, and other associated benefits such as environmental-friendliness.

  12. Carbon footprint of apple and pear : orchards, storage and distribution

    OpenAIRE

    Figueiredo, F.; Castanheira, E.G.; Feliciano, M.; Rodrigues, M.A.; Peres, A.; Maia, F.; Ramos, A.; Carneiro, J.P.; Coroama, V.C.; Freire, F.

    2013-01-01

    Apple and pear represent 51% of fresh fruit orchards in Portugal. This paper presents a life-cycle (LC) greenhouse gas (GHG) assessment (so-called carbon footprint) of 3 apple and 1 pear Portuguese production systems. An LC model and inventory were implemented, encompassing the farm stage (cultivation of fruit trees in orchards), storage and distribution (transport to retail). The functional unit considered in this study was 1 kg of distributed fruit (at retail). Four different LC inventories...

  13. Advanced Energy Storage Management in Distribution Network

    Energy Technology Data Exchange (ETDEWEB)

    Liu, Guodong [ORNL; Ceylan, Oguzhan [ORNL; Xiao, Bailu [ORNL; Starke, Michael R [ORNL; Ollis, T Ben [ORNL; King, Daniel J [ORNL; Irminger, Philip [ORNL; Tomsovic, Kevin [University of Tennessee, Knoxville (UTK)

    2016-01-01

    With increasing penetration of distributed generation (DG) in the distribution networks (DN), the secure and optimal operation of DN has become an important concern. In this paper, an iterative mixed integer quadratic constrained quadratic programming model to optimize the operation of a three phase unbalanced distribution system with high penetration of Photovoltaic (PV) panels, DG and energy storage (ES) is developed. The proposed model minimizes not only the operating cost, including fuel cost and purchasing cost, but also voltage deviations and power loss. The optimization model is based on the linearized sensitivity coefficients between state variables (e.g., node voltages) and control variables (e.g., real and reactive power injections of DG and ES). To avoid slow convergence when close to the optimum, a golden search method is introduced to control the step size and accelerate the convergence. The proposed algorithm is demonstrated on modified IEEE 13 nodes test feeders with multiple PV panels, DG and ES. Numerical simulation results validate the proposed algorithm. Various scenarios of system configuration are studied and some critical findings are concluded.

  14. Parallel algorithms for islanded microgrid with photovoltaic and energy storage systems planning optimization problem: Material selection and quantity demand optimization

    Science.gov (United States)

    Cao, Yang; Liu, Chun; Huang, Yuehui; Wang, Tieqiang; Sun, Chenjun; Yuan, Yue; Zhang, Xinsong; Wu, Shuyun

    2017-02-01

    With the development of roof photovoltaic power (PV) generation technology and the increasingly urgent need to improve supply reliability levels in remote areas, islanded microgrid with photovoltaic and energy storage systems (IMPE) is developing rapidly. The high costs of photovoltaic panel material and energy storage battery material have become the primary factors that hinder the development of IMPE. The advantages and disadvantages of different types of photovoltaic panel materials and energy storage battery materials are analyzed in this paper, and guidance is provided on material selection for IMPE planners. The time sequential simulation method is applied to optimize material demands of the IMPE. The model is solved by parallel algorithms that are provided by a commercial solver named CPLEX. Finally, to verify the model, an actual IMPE is selected as a case system. Simulation results on the case system indicate that the optimization model and corresponding algorithm is feasible. Guidance for material selection and quantity demand for IMPEs in remote areas is provided by this method.

  15. Parallel nanostructuring of GeSbTe film with particle mask

    Energy Technology Data Exchange (ETDEWEB)

    Wang, Z.B.; Hong, M.H.; Wang, Q.F.; Chong, T.C. [Data Storage Institute, DSI Building, 5 Engineering Drive 1, 117608, Singapore (Singapore); Department of Electrical and Computer Engineering, National University of Singapore, 119260, Singapore (Singapore); Luk' yanchuk, B.S.; Huang, S.M.; Shi, L.P. [Data Storage Institute, DSI Building, 5 Engineering Drive 1, 117608, Singapore (Singapore)

    2004-09-01

    Parallel nanostructuring of a GeSbTe film may significantly improve the recording performance in data storage. In this paper, a method that permits direct and massively parallel nanopatterning of the substrate surface by laser irradiation is investigated. Polystyrene spherical particles were deposited on the surface in a monolayer array by self-assembly. The array was then irradiated with a 248-nm KrF laser. A sub-micron nanodent array can be obtained after single-pulse irradiation. These nanodents change their shapes at different laser energies. The optical near-field distribution around the particles was calculated according to the exact solution of the light-scattering problem. The influence of the presence of the substrate on the optical near field was also studied. The mechanisms for the generation of the nanodent structures are discussed. (orig.)

  16. libstable: Fast, Parallel, and High-Precision Computation of α-Stable Distributions in R, C/C++, and MATLAB

    Directory of Open Access Journals (Sweden)

    Javier Royuela-del-Val

    2017-06-01

    Full Text Available α-stable distributions are a family of well-known probability distributions. However, the lack of closed analytical expressions hinders their application. Currently, several tools have been developed to numerically evaluate their density and distribution functions or to estimate their parameters, but available solutions either do not reach sufficient precision on their evaluations or are excessively slow for practical purposes. Moreover, they do not take full advantage of the parallel processing capabilities of current multi-core machines. Other solutions work only on a subset of the α-stable parameter space. In this paper we present an R package and a C/C++ library with a MATLAB front-end that permit parallelized, fast and high precision evaluation of density, distribution and quantile functions, as well as random variable generation and parameter estimation of α-stable distributions in their whole parameter space. The described library can be easily integrated into third party developments.

  17. Parallel Hybrid Vehicle Optimal Storage System

    Science.gov (United States)

    Bloomfield, Aaron P.

    2009-01-01

    A paper reports the results of a Hybrid Diesel Vehicle Project focused on a parallel hybrid configuration suitable for diesel-powered, medium-sized, commercial vehicles commonly used for parcel delivery and shuttle buses, as the missions of these types of vehicles require frequent stops. During these stops, electric hybridization can effectively recover the vehicle's kinetic energy during the deceleration, store it onboard, and then use that energy to assist in the subsequent acceleration.

  18. DCODE: A Distributed Column-Oriented Database Engine for Big Data Analytics

    OpenAIRE

    Liu, Yanchen; Cao, Fang; Mortazavi, Masood; Chen, Mengmeng; Yan, Ning; Ku, Chi; Adnaik, Aniket; Morgan, Stephen; Shi, Guangyu; Wang, Yuhu; Fang, Fan

    2015-01-01

    Part 10: Big Data and Text Mining; International audience; We propose a novel Distributed Column-Oriented Database Engine (DCODE) for efficient analytic query processing that combines advantages of both column storage and parallel processing. In DCODE, we enhance an existing open-source columnar database engine by adding the capability for handling queries over a cluster. Specifically, we studied parallel query execution and optimization techniques such as horizontal partitioning, exchange op...

  19. Plane-wave electronic structure calculations on a parallel supercomputer

    International Nuclear Information System (INIS)

    Nelson, J.S.; Plimpton, S.J.; Sears, M.P.

    1993-01-01

    The development of iterative solutions of Schrodinger's equation in a plane-wave (pw) basis over the last several years has coincided with great advances in the computational power available for performing the calculations. These dual developments have enabled many new and interesting condensed matter phenomena to be studied from a first-principles approach. The authors present a detailed description of the implementation on a parallel supercomputer (hypercube) of the first-order equation-of-motion solution to Schrodinger's equation, using plane-wave basis functions and ab initio separable pseudopotentials. By distributing the plane-waves across the processors of the hypercube many of the computations can be performed in parallel, resulting in decreases in the overall computation time relative to conventional vector supercomputers. This partitioning also provides ample memory for large Fast Fourier Transform (FFT) meshes and the storage of plane-wave coefficients for many hundreds of energy bands. The usefulness of the parallel techniques is demonstrated by benchmark timings for both the FFT's and iterations of the self-consistent solution of Schrodinger's equation for different sized Si unit cells of up to 512 atoms

  20. Adaptive Dynamic Process Scheduling on Distributed Memory Parallel Computers

    Directory of Open Access Journals (Sweden)

    Wei Shu

    1994-01-01

    Full Text Available One of the challenges in programming distributed memory parallel machines is deciding how to allocate work to processors. This problem is particularly important for computations with unpredictable dynamic behaviors or irregular structures. We present a scheme for dynamic scheduling of medium-grained processes that is useful in this context. The adaptive contracting within neighborhood (ACWN is a dynamic, distributed, load-dependent, and scalable scheme. It deals with dynamic and unpredictable creation of processes and adapts to different systems. The scheme is described and contrasted with two other schemes that have been proposed in this context, namely the randomized allocation and the gradient model. The performance of the three schemes on an Intel iPSC/2 hypercube is presented and analyzed. The experimental results show that even though the ACWN algorithm incurs somewhat larger overhead than the randomized allocation, it achieves better performance in most cases due to its adaptiveness. Its feature of quickly spreading the work helps it outperform the gradient model in performance and scalability.

  1. Electromagnetic ion-cyclotron instability in the presence of a parallel electric field with general loss-cone distribution function - particle aspect analysis

    Directory of Open Access Journals (Sweden)

    G. Ahirwar

    2006-08-01

    Full Text Available The effect of parallel electric field on the growth rate, parallel and perpendicular resonant energy and marginal stability of the electromagnetic ion-cyclotron (EMIC wave with general loss-cone distribution function in a low β homogeneous plasma is investigated by particle aspect approach. The effect of the steepness of the loss-cone distribution is investigated on the electromagnetic ion-cyclotron wave. The whole plasma is considered to consist of resonant and non-resonant particles. It is assumed that resonant particles participate in the energy exchange with the wave, whereas non-resonant particles support the oscillatory motion of the wave. The wave is assumed to propagate parallel to the static magnetic field. The effect of the parallel electric field with the general distribution function is to control the growth rate of the EMIC waves, whereas the effect of steep loss-cone distribution is to enhance the growth rate and perpendicular heating of the ions. This study is relevant to the analysis of ion conics in the presence of an EMIC wave in the auroral acceleration region of the Earth's magnetoplasma.

  2. First experiences with large SAN storage and Linux

    International Nuclear Information System (INIS)

    Wezel, Jos van; Marten, Holger; Verstege, Bernhard; Jaeger, Axel

    2004-01-01

    The use of a storage area network (SAN) with Linux opens possibilities for scalable and affordable large data storage and poses a new challenge for cluster computing. The GridKa center uses a commercial parallel file system to create a highly available high-speed data storage using a combination of Fibre Channel (SAN) and Ethernet (LAN) to optimize between data throughput and costs. This article describes the design, implementation and optimizations of the GridKa storage solution which will offer over 400 TB online storage for 600 nodes. Presented are some throughput measurements of one of the largest Linux-based parallel storage systems in the world

  3. A parallel model for SQL astronomical databases based on solid state storage. Application to the Gaia Archive PostgreSQL database

    Science.gov (United States)

    González-Núñez, J.; Gutiérrez-Sánchez, R.; Salgado, J.; Segovia, J. C.; Merín, B.; Aguado-Agelet, F.

    2017-07-01

    Query planning and optimisation algorithms in most popular relational databases were developed at the times hard disk drives were the only storage technology available. The advent of higher parallel random access capacity devices, such as solid state disks, opens up the way for intra-machine parallel computing over large datasets. We describe a two phase parallel model for the implementation of heavy analytical processes in single instance PostgreSQL astronomical databases. This model is particularised to fulfil two frequent astronomical problems, density maps and crossmatch computation with Quad Tree Cube (Q3C) indexes. They are implemented as part of the relational databases infrastructure for the Gaia Archive and performance is assessed. Improvement of a factor 28.40 in comparison to sequential execution is observed in the reference implementation for a histogram computation. Speedup ratios of 3.7 and 4.0 are attained for the reference positional crossmatches considered. We observe large performance enhancements over sequential execution for both CPU and disk access intensive computations, suggesting these methods might be useful with the growing data volumes in Astronomy.

  4. Discovery of resources using MADM approaches for parallel and distributed computing

    Directory of Open Access Journals (Sweden)

    Mandeep Kaur

    2017-06-01

    Full Text Available Grid, a form of parallel and distributed computing, allows the sharing of data and computational resources among its users from various geographical locations. The grid resources are diverse in terms of their underlying attributes. The majority of the state-of-the-art resource discovery techniques rely on the static resource attributes during resource selection. However, the matching resources based on the static resource attributes may not be the most appropriate resources for the execution of user applications because they may have heavy job loads, less storage space or less working memory (RAM. Hence, there is a need to consider the current state of the resources in order to find the most suitable resources. In this paper, we have proposed a two-phased multi-attribute decision making (MADM approach for discovery of grid resources by using P2P formalism. The proposed approach considers multiple resource attributes for decision making of resource selection and provides the best suitable resource(s to grid users. The first phase describes a mechanism to discover all matching resources and applies SAW method to shortlist the top ranked resources, which are communicated to the requesting super-peer. The second phase of our proposed methodology applies integrated MADM approach (AHP enriched PROMETHEE-II on the list of selected resources received from different super-peers. The pairwise comparison of the resources with respect to their attributes is made and the rank of each resource is determined. The top ranked resource is then communicated to the grid user by the grid scheduler. Our proposed methodology enables the grid scheduler to allocate the most suitable resource to the user application and also reduces the search complexity by filtering out the less suitable resources during resource discovery.

  5. PRISMA/DB: A Parallel Main-Memory Relational DBMS

    NARCIS (Netherlands)

    Apers, Peter M.G.; Flokstra, Jan; van den Berg, Carel A.; Grefen, P.W.P.J.; Wilschut, A.N.; Kersten, Martin L.; van den Berg, C.A.

    1992-01-01

    PRISMA/DB, a full-fledged parallel, main memory relational database management system (DBMS) is described. PRISMA/DB's high performance is obtained by the use of parallelism for query processing and main memory storage of the entire database. A flexible architecture for experimenting with

  6. Lithium ion storage between graphenes

    Directory of Open Access Journals (Sweden)

    Chan Yue

    2011-01-01

    Full Text Available Abstract In this article, we investigate the storage of lithium ions between two parallel graphene sheets using the continuous approximation and the 6-12 Lennard-Jones potential. The continuous approximation assumes that the carbon atoms can be replaced by a uniform distribution across the surface of the graphene sheets so that the total interaction potential can be approximated by performing surface integrations. The number of ion layers determines the major storage characteristics of the battery, and our results show three distinct ionic configurations, namely single, double, and triple ion forming layers between graphenes. The number densities of lithium ions between the two graphenes are estimated from existing semi-empirical molecular orbital calculations, and the graphene sheets giving rise to the triple ion layers admit the largest storage capacity at all temperatures, followed by a marginal decrease of storage capacity for the case of double ion layers. These two configurations exceed the maximum theoretical storage capacity of graphite. Further, on taking into account the charge-discharge property, the double ion layers are the most preferable choice for enhanced lithium storage. Although the single ion layer provides the least charge storage, it turns out to be the most stable configuration at all temperatures. One application of the present study is for the design of future high energy density alkali batteries using graphene sheets as anodes for which an analytical formulation might greatly facilitate rapid computational results.

  7. Estimating electricity storage power rating and discharge duration for utility transmission and distribution deferral :a study for the DOE energy storage program.

    Energy Technology Data Exchange (ETDEWEB)

    Eyer, James M. (Distributed Utility Associates, Livermore, CA); Butler, Paul Charles; Iannucci, Joseph J., Jr. (,.Distributed Utility Associates, Livermore, CA)

    2005-11-01

    This report describes a methodology for estimating the power and energy capacities for electricity energy storage systems that can be used to defer costly upgrades to fully overloaded, or nearly overloaded, transmission and distribution (T&D) nodes. This ''sizing'' methodology may be used to estimate the amount of storage needed so that T&D upgrades may be deferred for one year. The same methodology can also be used to estimate the characteristics of storage needed for subsequent years of deferral.

  8. Parallel SN algorithms in shared- and distributed-memory environments

    International Nuclear Information System (INIS)

    Haghighat, Alireza; Hunter, Melissa A.; Mattis, Ronald E.

    1995-01-01

    Different 2-D spatial domain partitioning Sn transport theory algorithms have been developed on the basis of the Block-Jacobi iterative scheme. These algorithms have been incorporated into TWOTRAN-II, and tested on a shared-memory CRAY Y-MP C90 and a distributed-memory IBM SP1. For a series of fixed source r-z geometry homogeneous problems, parallel efficiencies in a range of 50-90% are achieved on the C90 with 6 processors, and lower values (20-60%) are obtained on the SP1. It is demonstrated that better performance is attainable if one addresses issues such as convergence rate, load-balancing, and granularity for both architectures, as well as message passing (network bandwidth and latency) for SP1. (author). 17 refs, 4 figs

  9. Tuning HDF5 subfiling performance on parallel file systems

    Energy Technology Data Exchange (ETDEWEB)

    Byna, Suren [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Chaarawi, Mohamad [Intel Corp. (United States); Koziol, Quincey [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Mainzer, John [The HDF Group (United States); Willmore, Frank [The HDF Group (United States)

    2017-05-12

    Subfiling is a technique used on parallel file systems to reduce locking and contention issues when multiple compute nodes interact with the same storage target node. Subfiling provides a compromise between the single shared file approach that instigates the lock contention problems on parallel file systems and having one file per process, which results in generating a massive and unmanageable number of files. In this paper, we evaluate and tune the performance of recently implemented subfiling feature in HDF5. In specific, we explain the implementation strategy of subfiling feature in HDF5, provide examples of using the feature, and evaluate and tune parallel I/O performance of this feature with parallel file systems of the Cray XC40 system at NERSC (Cori) that include a burst buffer storage and a Lustre disk-based storage. We also evaluate I/O performance on the Cray XC30 system, Edison, at NERSC. Our results show performance benefits of 1.2X to 6X performance advantage with subfiling compared to writing a single shared HDF5 file. We present our exploration of configurations, such as the number of subfiles and the number of Lustre storage targets to storing files, as optimization parameters to obtain superior I/O performance. Based on this exploration, we discuss recommendations for achieving good I/O performance as well as limitations with using the subfiling feature.

  10. Electric utility transmission and distribution upgrade deferral benefits from modular electricity storage : a study for the DOE Energy Storage Systems Program.

    Energy Technology Data Exchange (ETDEWEB)

    Eyer, James M. (Distributed Utility Associates, Inc., Livermore, CA)

    2009-06-01

    The work documented in this report was undertaken as part of an ongoing investigation of innovative and potentially attractive value propositions for electricity storage by the United States Department of Energy (DOE) and Sandia National Laboratories (SNL) Electricity Storage Systems (ESS) Program. This study characterizes one especially attractive value proposition for modular electricity storage (MES): electric utility transmission and distribution (T&D) upgrade deferral. The T&D deferral benefit is characterized in detail. Also presented is a generalized framework for estimating the benefit. Other important and complementary (to T&D deferral) elements of possible value propositions involving MES are also characterized.

  11. Research on Distributed PV Storage Virtual Synchronous Generator System and Its Static Frequency Characteristic Analysis

    Directory of Open Access Journals (Sweden)

    Xiangwu Yan

    2018-03-01

    Full Text Available The increasing penetration rate of grid connected renewable energy power generation reduces the primary frequency regulation capability of the system and poses a challenge to the security and stability of the power grid. In this paper, a distributed photovoltaic (PV storage virtual synchronous generator system is constructed, which realizes the external characteristics of synchronous generator/motor. For this kind of input/output bidirectional devices (e.g., renewable power generation/storage combined systems, pumped storage power stations, battery energy storage systems, and vehicle-to-grid electric vehicles, a synthesis analysis method for system power-frequency considering source-load static frequency characteristics (S-L analysis method is proposed in order to depict the system’s power balance dynamic adjustment process visually. Simultaneously, an inertia matching method is proposed to solve the problem of inertia matching in the power grid. Through the simulation experiment in MATLAB, the feasibility of the distributed PV storage synchronous virtual machine system is verified as well as the effectiveness of S-L analysis method and inertia matching method.

  12. Performance evaluation of parallel electric field tunnel field-effect transistor by a distributed-element circuit model

    Science.gov (United States)

    Morita, Yukinori; Mori, Takahiro; Migita, Shinji; Mizubayashi, Wataru; Tanabe, Akihito; Fukuda, Koichi; Matsukawa, Takashi; Endo, Kazuhiko; O'uchi, Shin-ichi; Liu, Yongxun; Masahara, Meishoku; Ota, Hiroyuki

    2014-12-01

    The performance of parallel electric field tunnel field-effect transistors (TFETs), in which band-to-band tunneling (BTBT) was initiated in-line to the gate electric field was evaluated. The TFET was fabricated by inserting an epitaxially-grown parallel-plate tunnel capacitor between heavily doped source wells and gate insulators. Analysis using a distributed-element circuit model indicated there should be a limit of the drain current caused by the self-voltage-drop effect in the ultrathin channel layer.

  13. From evolution theory to parallel and distributed genetic

    CERN Multimedia

    CERN. Geneva

    2007-01-01

    Lecture #1: From Evolution Theory to Evolutionary Computation. Evolutionary computation is a subfield of artificial intelligence (more particularly computational intelligence) involving combinatorial optimization problems, which are based to some degree on the evolution of biological life in the natural world. In this tutorial we will review the source of inspiration for this metaheuristic and its capability for solving problems. We will show the main flavours within the field, and different problems that have been successfully solved employing this kind of techniques. Lecture #2: Parallel and Distributed Genetic Programming. The successful application of Genetic Programming (GP, one of the available Evolutionary Algorithms) to optimization problems has encouraged an increasing number of researchers to apply these techniques to a large set of problems. Given the difficulty of some problems, much effort has been applied to improving the efficiency of GP during the last few years. Among the available proposals,...

  14. Subgrid Parameterization of the Soil Moisture Storage Capacity for a Distributed Rainfall-Runoff Model

    Directory of Open Access Journals (Sweden)

    Weijian Guo

    2015-05-01

    Full Text Available Spatial variability plays an important role in nonlinear hydrologic processes. Due to the limitation of computational efficiency and data resolution, subgrid variability is usually assumed to be uniform for most grid-based rainfall-runoff models, which leads to the scale-dependence of model performances. In this paper, the scale effect on the Grid-Xinanjiang model was examined. The bias of the estimation of precipitation, runoff, evapotranspiration and soil moisture at the different grid scales, along with the scale-dependence of the effective parameters, highlights the importance of well representing the subgrid variability. This paper presents a subgrid parameterization method to incorporate the subgrid variability of the soil storage capacity, which is a key variable that controls runoff generation and partitioning in the Grid-Xinanjiang model. In light of the similar spatial pattern and physical basis, the soil storage capacity is correlated with the topographic index, whose spatial distribution can more readily be measured. A beta distribution is introduced to represent the spatial distribution of the soil storage capacity within the grid. The results derived from the Yanduhe Basin show that the proposed subgrid parameterization method can effectively correct the watershed soil storage capacity curve. Compared to the original Grid-Xinanjiang model, the model performances are quite consistent at the different grid scales when the subgrid variability is incorporated. This subgrid parameterization method reduces the recalibration necessity when the Digital Elevation Model (DEM resolution is changed. Moreover, it improves the potential for the application of the distributed model in the ungauged basin.

  15. Optimal sizing and operation of energy storage systems considering long term assessment

    Directory of Open Access Journals (Sweden)

    Gerardo Guerra

    2018-01-01

    Full Text Available This paper proposes a procedure for estimating the optimal sizing of Photovoltaic Generators and Energy Storage units when they are operated from the utility’s perspective. The goal is to explore the potential improvement on the overall operating conditions of the distribution system to which the Generators and Storage units will be connected. Optimization is conducted by means of a General Parallel Genetic Algorithm that seeks to maximize the technical benefits for the distribution system. The paper proposes an operation strategy for Energy Storage units based on the daily variation of load and generation; the operation strategy is optimized for an evaluation period of one year using hourly power curves. The construction of the yearly Storage operation curve results in a high-dimension optimization problem; as a result, different day-classification methods are applied in order to reduce the dimension of the optimization. Results show that the proposed approach is capable of producing significant improvements in system operating conditions and that the best performance is obtained when the day-classification is based on the similarity among daily power curves.

  16. Faster Parallel Traversal of Scale Free Graphs at Extreme Scale with Vertex Delegates

    KAUST Repository

    Pearce, Roger

    2014-11-01

    © 2014 IEEE. At extreme scale, irregularities in the structure of scale-free graphs such as social network graphs limit our ability to analyze these important and growing datasets. A key challenge is the presence of high-degree vertices (hubs), that leads to parallel workload and storage imbalances. The imbalances occur because existing partitioning techniques are not able to effectively partition high-degree vertices. We present techniques to distribute storage, computation, and communication of hubs for extreme scale graphs in distributed memory supercomputers. To balance the hub processing workload, we distribute hub data structures and related computation among a set of delegates. The delegates coordinate using highly optimized, yet portable, asynchronous broadcast and reduction operations. We demonstrate scalability of our new algorithmic technique using Breadth-First Search (BFS), Single Source Shortest Path (SSSP), K-Core Decomposition, and Page-Rank on synthetically generated scale-free graphs. Our results show excellent scalability on large scale-free graphs up to 131K cores of the IBM BG/P, and outperform the best known Graph500 performance on BG/P Intrepid by 15%

  17. Faster Parallel Traversal of Scale Free Graphs at Extreme Scale with Vertex Delegates

    KAUST Repository

    Pearce, Roger; Gokhale, Maya; Amato, Nancy M.

    2014-01-01

    © 2014 IEEE. At extreme scale, irregularities in the structure of scale-free graphs such as social network graphs limit our ability to analyze these important and growing datasets. A key challenge is the presence of high-degree vertices (hubs), that leads to parallel workload and storage imbalances. The imbalances occur because existing partitioning techniques are not able to effectively partition high-degree vertices. We present techniques to distribute storage, computation, and communication of hubs for extreme scale graphs in distributed memory supercomputers. To balance the hub processing workload, we distribute hub data structures and related computation among a set of delegates. The delegates coordinate using highly optimized, yet portable, asynchronous broadcast and reduction operations. We demonstrate scalability of our new algorithmic technique using Breadth-First Search (BFS), Single Source Shortest Path (SSSP), K-Core Decomposition, and Page-Rank on synthetically generated scale-free graphs. Our results show excellent scalability on large scale-free graphs up to 131K cores of the IBM BG/P, and outperform the best known Graph500 performance on BG/P Intrepid by 15%

  18. Distributed Control of Battery Energy Storage Systems for Voltage Regulation in Distribution Networks with High PV Penetration

    DEFF Research Database (Denmark)

    Zeraati, Mehdi; Golshan, Mohamad Esmaeil Hamedani; Guerrero, Josep M.

    2018-01-01

    of their installed capacity whereas the second one modifies the BESs performance in terms of their state of charge (SoC) to prevent the excessive saturation or depletion of batteries. The proposed controller enables the effective use of storage capacity in different conditions. Finally, the simulation results based...... issues of distribution networks. In this paper, the battery energy storage (BES) systems are used in order to solve the voltage rise during the peak PV generation as well as the voltage drop while meeting the peak load. A coordinated control strategy is proposed to regulate the charge/discharge of BESs...

  19. Implementation and Performance Evaluation of Distributed Cloud Storage Solutions using Random Linear Network Coding

    DEFF Research Database (Denmark)

    Fitzek, Frank; Toth, Tamas; Szabados, Áron

    2014-01-01

    This paper advocates the use of random linear network coding for storage in distributed clouds in order to reduce storage and traffic costs in dynamic settings, i.e. when adding and removing numerous storage devices/clouds on-the-fly and when the number of reachable clouds is limited. We introduce...... various network coding approaches that trade-off reliability, storage and traffic costs, and system complexity relying on probabilistic recoding for cloud regeneration. We compare these approaches with other approaches based on data replication and Reed-Solomon codes. A simulator has been developed...... to carry out a thorough performance evaluation of the various approaches when relying on different system settings, e.g., finite fields, and network/storage conditions, e.g., storage space used per cloud, limited network use, and limited recoding capabilities. In contrast to standard coding approaches, our...

  20. Distributed generation, storage, demand response and energy efficiency as alternatives to grid capacity enhancement

    International Nuclear Information System (INIS)

    Poudineh, Rahmatallah; Jamasb, Tooraj

    2014-01-01

    The need for investment in capital intensive electricity networks is on the rise in many countries. A major advantage of distributed resources is their potential for deferring investments in distribution network capacity. However, utilizing the full benefits of these resources requires addressing several technical, economic and regulatory challenges. A significant barrier pertains to the lack of an efficient market mechanism that enables this concept and also is consistent with business model of distribution companies under an unbundled power sector paradigm. This paper proposes a market-oriented approach termed as “contract for deferral scheme” (CDS). The scheme outlines how an economically efficient portfolio of distributed generation, storage, demand response and energy efficiency can be integrated as network resources to reduce the need for grid capacity and defer demand driven network investments. - Highlights: • The paper explores a practical framework for smart electricity distribution grids. • The aim is to defer large capital investments in the network by utilizing and incentivising distributed generation, demand response, energy efficiency and storage as network resources. • The paper discusses a possible new market model that enables integration of distributed resources as alternative to grid capacity enhancement

  1. High-performance mass storage system for workstations

    Science.gov (United States)

    Chiang, T.; Tang, Y.; Gupta, L.; Cooperman, S.

    1993-01-01

    Reduced Instruction Set Computer (RISC) workstations and Personnel Computers (PC) are very popular tools for office automation, command and control, scientific analysis, database management, and many other applications. However, when using Input/Output (I/O) intensive applications, the RISC workstations and PC's are often overburdened with the tasks of collecting, staging, storing, and distributing data. Also, by using standard high-performance peripherals and storage devices, the I/O function can still be a common bottleneck process. Therefore, the high-performance mass storage system, developed by Loral AeroSys' Independent Research and Development (IR&D) engineers, can offload a RISC workstation of I/O related functions and provide high-performance I/O functions and external interfaces. The high-performance mass storage system has the capabilities to ingest high-speed real-time data, perform signal or image processing, and stage, archive, and distribute the data. This mass storage system uses a hierarchical storage structure, thus reducing the total data storage cost, while maintaining high-I/O performance. The high-performance mass storage system is a network of low-cost parallel processors and storage devices. The nodes in the network have special I/O functions such as: SCSI controller, Ethernet controller, gateway controller, RS232 controller, IEEE488 controller, and digital/analog converter. The nodes are interconnected through high-speed direct memory access links to form a network. The topology of the network is easily reconfigurable to maximize system throughput for various applications. This high-performance mass storage system takes advantage of a 'busless' architecture for maximum expandability. The mass storage system consists of magnetic disks, a WORM optical disk jukebox, and an 8mm helical scan tape to form a hierarchical storage structure. Commonly used files are kept in the magnetic disk for fast retrieval. The optical disks are used as archive

  2. Efficient Sustainable Operation Mechanism of Distributed Desktop Integration Storage Based on Virtualization with Ubiquitous Computing

    Directory of Open Access Journals (Sweden)

    Hyun-Woo Kim

    2015-06-01

    Full Text Available Following the rapid growth of ubiquitous computing, many jobs that were previously manual have now been automated. This automation has increased the amount of time available for leisure; diverse services are now being developed for this leisure time. In addition, the development of small and portable devices like smartphones, diverse Internet services can be used regardless of time and place. Studies regarding diverse virtualization are currently in progress. These studies aim to determine ways to efficiently store and process the big data generated by the multitude of devices and services in use. One topic of such studies is desktop storage virtualization, which integrates distributed desktop resources and provides these resources to users to integrate into distributed legacy desktops via virtualization. In the case of desktop storage virtualization, high availability of virtualization is necessary and important for providing reliability to users. Studies regarding hierarchical structures and resource integration are currently in progress. These studies aim to create efficient data distribution and storage for distributed desktops based on resource integration environments. However, studies regarding efficient responses to server faults occurring in desktop-based resource integration environments have been insufficient. This paper proposes a mechanism for the sustainable operation of desktop storage (SODS for high operational availability. It allows for the easy addition and removal of desktops in desktop-based integration environments. It also activates alternative servers when a fault occurs within a system.

  3. Distributed Database Storage Solution in Java

    OpenAIRE

    Funck, Johan

    2010-01-01

    Car sales companies have in the last couple of years discovered that there is a big market in storing their customer's summer and winter tires for a small fee. For the customers it is very convenient to get rid of the all known storage problem with season tires. Burlin Motor Umeå is one of these companies and they are offering seasonal storage and change of tires in autumn and spring as well as washing of tires.The main problem for this kind of storage is how to make the storage easy to overv...

  4. Energy storage reinforces competitive business practices

    International Nuclear Information System (INIS)

    Makansi, J.

    1994-01-01

    This article examines how the ability to ''store'' electricity can pay handsome dividends in a competitive environment. Priorities change when industries are deregulated. Indeed, new priorities are being established for electric generation--low cost, efficiency, product distinction for marketing purposes, etc. are all more critical today. Perhaps not so obvious is the fundamental role of energy storage in a fully competitive marketplace. In fact, rarely do a technology development and a changing business climate play off against each other so nicely. Consider the function of the emerging electricity broker, or power marketer. Imagine the premium that broker could command with access to a large increment of electricity--purchased at a low price--and supplied at a moment's notice for a substantially higher price. Storage of electricity would mean that the investment in excess available generation capacity to supply so-called peak demand could be avoided. It also means that electricity could be brokered like other commodities--that is purchased, stockpiled, and sold to reflect market conditions across a wider geographical region and time spain. Benefits accrue to transmission and distribution, in addition to generation. Energy storage helps to manage the increasing stress placed on the grid as a result of intermittent sources of power and large numbers of cogenerators and small power producers. On the customer side, any ratepayer large or small could, theoretically, play the spot market in electric supply with a reserve to tap in emergencies. For a parallel in other deregulated markets, recall how storage has become an important factor in natural-gas contracting. Quality of electricity also can be improved by applying storage to stabilize the grid, especially along the distribution system at substations. And the opening of vast markets for electricity consumption, such as electric vehicles, depends in large measure on electric storage

  5. Rapid optimization of tension distribution for cable-driven parallel manipulators with redundant cables

    Science.gov (United States)

    Ouyang, Bo; Shang, Weiwei

    2016-03-01

    The solution of tension distributions is infinite for cable-driven parallel manipulators(CDPMs) with redundant cables. A rapid optimization method for determining the optimal tension distribution is presented. The new optimization method is primarily based on the geometry properties of a polyhedron and convex analysis. The computational efficiency of the optimization method is improved by the designed projection algorithm, and a fast algorithm is proposed to determine which two of the lines are intersected at the optimal point. Moreover, a method for avoiding the operating point on the lower tension limit is developed. Simulation experiments are implemented on a six degree-of-freedom(6-DOF) CDPM with eight cables, and the results indicate that the new method is one order of magnitude faster than the standard simplex method. The optimal distribution of tension distribution is thus rapidly established on real-time by the proposed method.

  6. Semantically Secure Symmetric Encryption with Error Correction for Distributed Storage

    Directory of Open Access Journals (Sweden)

    Juha Partala

    2017-01-01

    Full Text Available A distributed storage system (DSS is a fundamental building block in many distributed applications. It applies linear network coding to achieve an optimal tradeoff between storage and repair bandwidth when node failures occur. Additively homomorphic encryption is compatible with linear network coding. The homomorphic property ensures that a linear combination of ciphertext messages decrypts to the same linear combination of the corresponding plaintext messages. In this paper, we construct a linearly homomorphic symmetric encryption scheme that is designed for a DSS. Our proposal provides simultaneous encryption and error correction by applying linear error correcting codes. We show its IND-CPA security for a limited number of messages based on binary Goppa codes and the following assumption: when dividing a scrambled generator matrix G^ into two parts G1^ and G2^, it is infeasible to distinguish G2^ from random and to find a statistical connection between G1^ and G2^. Our infeasibility assumptions are closely related to those underlying the McEliece public key cryptosystem but are considerably weaker. We believe that the proposed problem has independent cryptographic interest.

  7. A brief summary on formalizing parallel tensor distributions redistributions and algorithm derivations.

    Energy Technology Data Exchange (ETDEWEB)

    Schatz, Martin D. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Kolda, Tamara G. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); van de Geijn, Robert [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2015-09-01

    Large-scale datasets in computational chemistry typically require distributed-memory parallel methods to perform a special operation known as tensor contraction. Tensors are multidimensional arrays, and a tensor contraction is akin to matrix multiplication with special types of permutations. Creating an efficient algorithm and optimized im- plementation in this domain is complex, tedious, and error-prone. To address this, we develop a notation to express data distributions so that we can apply use automated methods to find optimized implementations for tensor contractions. We consider the spin-adapted coupled cluster singles and doubles method from computational chemistry and use our methodology to produce an efficient implementation. Experiments per- formed on the IBM Blue Gene/Q and Cray XC30 demonstrate impact both improved performance and reduced memory consumption.

  8. Sparse distributed memory overview

    Science.gov (United States)

    Raugh, Mike

    1990-01-01

    The Sparse Distributed Memory (SDM) project is investigating the theory and applications of massively parallel computing architecture, called sparse distributed memory, that will support the storage and retrieval of sensory and motor patterns characteristic of autonomous systems. The immediate objectives of the project are centered in studies of the memory itself and in the use of the memory to solve problems in speech, vision, and robotics. Investigation of methods for encoding sensory data is an important part of the research. Examples of NASA missions that may benefit from this work are Space Station, planetary rovers, and solar exploration. Sparse distributed memory offers promising technology for systems that must learn through experience and be capable of adapting to new circumstances, and for operating any large complex system requiring automatic monitoring and control. Sparse distributed memory is a massively parallel architecture motivated by efforts to understand how the human brain works. Sparse distributed memory is an associative memory, able to retrieve information from cues that only partially match patterns stored in the memory. It is able to store long temporal sequences derived from the behavior of a complex system, such as progressive records of the system's sensory data and correlated records of the system's motor controls.

  9. The STAPL Parallel Graph Library

    KAUST Repository

    Harshvardhan,; Fidel, Adam; Amato, Nancy M.; Rauchwerger, Lawrence

    2013-01-01

    This paper describes the stapl Parallel Graph Library, a high-level framework that abstracts the user from data-distribution and parallelism details and allows them to concentrate on parallel graph algorithm development. It includes a customizable

  10. PRELIMINARY STUDY ON APPLICATION OF MAX PLUS ALGEBRA IN DISTRIBUTED STORAGE SYSTEM THROUGH NETWORK CODING

    Directory of Open Access Journals (Sweden)

    Agus Maman Abadi

    2016-04-01

    Full Text Available The increasing need in techniques of storing big data presents a new challenge. One way to address this challenge is the use of distributed storage systems. One strategy that implemented in distributed data storage systems is the use of Erasure Code which applied to network coding. The code used in this technique is based on the algebraic structure which is called as vector space. Some studies have also been carried out to create code that is based on other algebraic structures such as module.  In this study, we are going to try to set up a code based on the algebraic structure which is a generalization of the module that is semimodule by utilizing the max operations and sum operations at max plus algebra. The results of this study indicate that the max operation and the addition operation on max plus algebra cannot be used to establish a semimodule code, but by modifying the operation "+" as "min", we get a code based on semimodule. Keywords:   code, distributed storage systems, network coding, semimodule, max plus algebra

  11. Real - time Optimization of Distributed Energy Storage System Operation Strategy Based on Peak Load Shifting

    Science.gov (United States)

    Wang, Qian; Lu, Guangqi; Li, Xiaoyu; Zhang, Yichi; Yun, Zejian; Bian, Di

    2018-01-01

    To take advantage of the energy storage system (ESS) sufficiently, the factors that the service life of the distributed energy storage system (DESS) and the load should be considered when establishing optimization model. To reduce the complexity of the load shifting of DESS in the solution procedure, the loss coefficient and the equal capacity ratio distribution principle were adopted in this paper. Firstly, the model was established considering the constraint conditions of the cycles, depth, power of the charge-discharge of the ESS, the typical daily load curves, as well. Then, dynamic programming method was used to real-time solve the model in which the difference of power Δs, the real-time revised energy storage capacity Sk and the permission error of depth of charge-discharge were introduced to optimize the solution process. The simulation results show that the optimized results was achieved when the load shifting in the load variance was not considered which means the charge-discharge of the energy storage system was not executed. In the meantime, the service life of the ESS would increase.

  12. Distributed energy storage systems on the basis of electric-vehicle fleets

    Science.gov (United States)

    Zhuk, A. Z.; Buzoverov, E. A.; Sheindlin, A. E.

    2015-01-01

    Several power technologies directed to solving the problem of covering nonuniform loads in power systems are developed at the Joint Institute of High Temperatures, Russian Academy of Sciences (JIHT RAS). One direction of investigations is the use of storage batteries of electric vehicles to compensate load peaks in the power system (V2G—vehicle-to-grid technology). The efficiency of energy storage systems based on electric vehicles with traditional energy-saving technologies is compared in the article by means of performing computations. The comparison is performed by the minimum-cost criterion for the peak energy supply to the system. Computations show that the distributed storage systems based on fleets of electric cars are efficient economically with their usage regime to 1 h/day. In contrast to traditional methods, the prime cost of regulation of the loads in the power system based on V2G technology is independent of the duration of the load compensation period (the duration of the consumption peak).

  13. Temperature distributions in trapezoidal built in storage solar water heaters with/without phase change materials

    International Nuclear Information System (INIS)

    Tarhan, Sefa; Sari, Ahmet; Yardim, M. Hakan

    2006-01-01

    Built in storage solar water heaters (BSSWHs) have been recognized for their more compact constructions and faster solar gain than conventional solar water heaters, however, their water temperatures quickly go down during the cooling period. A trapezoidal BSSWH without PCM storage unit was used as the control heater (reference) to investigate the effect of two differently configured PCM storage units on the temperature distributions in water tanks. In the first design, myristic acid was filled into the PCM storage tank, which also served as an absorbing plate. In the second design, lauric acid was filled into the PCM storage tank, which also served as a baffle plate. The water temperature changes were followed by five thermocouples placed evenly and longitudinally into each of the three BSSWHs. The effects of the PCMs on the water temperature distributions depended on the configuration of the PCM storage unit and the longitudinal position in the water tanks. The use of lauric acid lowered the values of the peak temperatures by 15% compared to the control heater at the upper portion of the water tanks because of the low melting temperature of lauric acid, but it did not have any consistent effect on the retention of the water temperatures during the cooling period. The ability of the myristic acid storage unit to retain the water temperatures got more remarkable, especially at the middle portion of the water tank. The myristic acid storage increased the dip temperatures by approximately 8.8% compared to the control heater. In conclusion, lauric acid storage can be used to stabilize the water temperature during the day time, while the myristic acid storage unit can be used as a thermal barrier against heat loss during the night time because of its relatively high melting temperature and low heat conduction coefficient in its solid phase. The experimental results have also indicated that the thermal characteristics of the PCM and the configuration of the PCM storage

  14. Pattern-Driven Automatic Parallelization

    Directory of Open Access Journals (Sweden)

    Christoph W. Kessler

    1996-01-01

    Full Text Available This article describes a knowledge-based system for automatic parallelization of a wide class of sequential numerical codes operating on vectors and dense matrices, and for execution on distributed memory message-passing multiprocessors. Its main feature is a fast and powerful pattern recognition tool that locally identifies frequently occurring computations and programming concepts in the source code. This tool also works for dusty deck codes that have been "encrypted" by former machine-specific code transformations. Successful pattern recognition guides sophisticated code transformations including local algorithm replacement such that the parallelized code need not emerge from the sequential program structure by just parallelizing the loops. It allows access to an expert's knowledge on useful parallel algorithms, available machine-specific library routines, and powerful program transformations. The partially restored program semantics also supports local array alignment, distribution, and redistribution, and allows for faster and more exact prediction of the performance of the parallelized target code than is usually possible.

  15. Real-time modeling and simulation of distribution feeder and distributed resources

    Science.gov (United States)

    Singh, Pawan

    The analysis of the electrical system dates back to the days when analog network analyzers were used. With the advent of digital computers, many programs were written for power-flow and short circuit analysis for the improvement of the electrical system. Real-time computer simulations can answer many what-if scenarios in the existing or the proposed power system. In this thesis, the standard IEEE 13-Node distribution feeder is developed and validated on a real-time platform OPAL-RT. The concept and the challenges of the real-time simulation are studied and addressed. Distributed energy resources include some of the commonly used distributed generation and storage devices like diesel engine, solar photovoltaic array, and battery storage system are modeled and simulated on a real-time platform. A microgrid encompasses a portion of an electric power distribution which is located downstream of the distribution substation. Normally, the microgrid operates in paralleled mode with the grid; however, scheduled or forced isolation can take place. In such conditions, the microgrid must have the ability to operate stably and autonomously. The microgrid can operate in grid connected and islanded mode, both the operating modes are studied in the last chapter. Towards the end, a simple microgrid controller modeled and simulated on the real-time platform is developed for energy management and protection for the microgrid.

  16. Exergy efficient production, storage and distribution of solar energy

    Energy Technology Data Exchange (ETDEWEB)

    Sandnes, Bjoernar

    2003-07-01

    There are two main themes in this thesis. 1) Exergy efficient utilization of solar energy, where the introduction of alternative technologies such as photovoltaic/thermal collectors and phase change energy storage in a low temperature solar system is investigated. 2) The possibility of storing thermal energy in supercooled liquids is investigated. The introductory chapters introduce the concept of exergy, and focus on the use of solar heat as an inherently low quality source for covering low quality demands associated with space heating and hot water. The different stages of solar energy production, storage, and distribution of heat is discussed, with emphasis on exergy relevant issues. With the low temperature solar heating system as background, the introduction of some additional technologies that are investigated. A section of this thesis presents a study of a small scale PV/T collector as a possible component in a low temperature system. In another section the instrumentation that has been built for studies of full-size PV and thermal systems is described, and the possibility of using the PV unit outputs as parameters for controlling the thermal system operation is briefly discussed. It is suggested that the design of the PV/T unit in terms of whether priority should be given to electricity or heat production should be based on how consumption of high quality auxiliary energy is minimized, and not on adding up the combined exergy which is being produced. Solar combisystems require larger heat storage capacities compared to the more common solar hot water systems. Increased volumetric heat storage capacity can be achieved by latent heat storage systems where thermal energy is stored as heat of fusion in phase change materials (PCMs). A section presents a study where spherically encapsulated PCM is incorporated in a solar heat store. Solar combisystems are often complex, and have a relatively large number of interacting components. Another section describes a

  17. A parallel solver for huge dense linear systems

    Science.gov (United States)

    Badia, J. M.; Movilla, J. L.; Climente, J. I.; Castillo, M.; Marqués, M.; Mayo, R.; Quintana-Ortí, E. S.; Planelles, J.

    2011-11-01

    : Linux/Unix Has the code been vectorized or parallelized?: Yes, includes MPI primitives. RAM: Tested for up to 190 GB Classification: 6.5 External routines: MPI ( http://www.mpi-forum.org/), BLAS ( http://www.netlib.org/blas/), PLAPACK ( http://www.cs.utexas.edu/~plapack/), POOCLAPACK ( ftp://ftp.cs.utexas.edu/pub/rvdg/PLAPACK/pooclapack.ps) (code for PLAPACK and POOCLAPACK is included in the distribution). Catalogue identifier of previous version: AEHU_v1_0 Journal reference of previous version: Comput. Phys. Comm. 182 (2011) 533 Does the new version supersede the previous version?: Yes Nature of problem: Huge scale dense systems of linear equations, Ax=B, beyond standard LAPACK capabilities. Solution method: The linear systems are solved by means of parallelized routines based on the LU factorization, using efficient secondary storage algorithms when the available main memory is insufficient. Reasons for new version: In many applications we need to guarantee a high accuracy in the solution of very large linear systems and we can do it by using double-precision arithmetic. Summary of revisions: Version 1.1 Can be used to solve linear systems using double-precision arithmetic. New version of the initialization routine. The user can choose the kind of arithmetic and the values of several parameters of the environment. Running time: About 5 hours to solve a system with more than 200 000 equations and more than 10 000 right-hand side vectors using double-precision arithmetic on an eight-node commodity cluster with a total of 64 Intel cores.

  18. Design and simulation of parallel and distributed architectures for images processing

    International Nuclear Information System (INIS)

    Pirson, Alain

    1990-01-01

    The exploitation of visual information requires special computers. The diversity of operations and the Computing power involved bring about structures founded on the concepts of concurrency and distributed processing. This work identifies a vision computer with an association of dedicated intelligent entities, exchanging messages according to the model of parallelism introduced by the language Occam. It puts forward an architecture of the 'enriched processor network' type. It consists of a classical multiprocessor structure where each node is provided with specific devices. These devices perform processing tasks as well as inter-nodes dialogues. Such an architecture benefits from the homogeneity of multiprocessor networks and the power of dedicated resources. Its implementation corresponds to that of a distributed structure, tasks being allocated to each Computing element. This approach culminates in an original architecture called ATILA. This modular structure is based on a transputer network supplied with vision dedicated co-processors and powerful communication devices. (author) [fr

  19. Multi-layer distributed storage of LHD plasma diagnostic database

    International Nuclear Information System (INIS)

    Nakanishi, Hideya; Kojima, Mamoru; Ohsuna, Masaki; Nonomura, Miki; Imazu, Setsuo; Nagayama, Yoshio

    2006-01-01

    At the end of LHD experimental campaign in 2003, the amount of whole plasma diagnostics raw data had reached 3.16 GB in a long-pulse experiment. This is a new world record in fusion plasma experiments, far beyond the previous value of 1.5 GB/shot. The total size of the LHD diagnostic data is about 21.6 TB for the whole six years of experiments, and it continues to grow at an increasing rate. The LHD diagnostic database and storage system, i.e. the LABCOM system, has a completely distributed architecture to be sufficiently flexible and easily expandable to maintain integrity of the total amount of data. It has three categories of the storage layer: OODBMS volumes in data acquisition servers, RAID servers, and mass storage systems, such as MO jukeboxes and DVD-R changers. These are equally accessible through the network. By data migration between them, they can be considered a virtual OODB extension area. Their data contents have been listed in a 'facilitator' PostgreSQL RDBMS, which contains about 6.2 million entries, and informs the optimized priority to clients requesting data. Using the 'glib' compression for all of the binary data and applying the three-tier application model for the OODB data transfer/retrieval, an optimized OODB read-out rate of 1.7 MB/s and effective client access speed of 3-25 MB/s have been achieved. As a result, the LABCOM data system has succeeded in combination of the use of RDBMS, OODBMS, RAID, and MSS to enable a virtual and always expandable storage volume, simultaneously with rapid data access. (author)

  20. A Parallel Restoration for Black Start of Microgrids Considering Characteristics of Distributed Generations

    Directory of Open Access Journals (Sweden)

    Jing Wang

    2017-12-01

    Full Text Available The black start capability is vital for microgrids, which can potentially improve the reliability of the power grid. This paper proposes a black start strategy for microgrids based on a parallel restoration strategy. Considering the characteristics of distributed generations (DGs, an evaluation model, which is used to assess the black start capability of DGs, is established by adopting the variation coefficient method. Thus, the DGs with good black start capability, which are selected by a diversity sequence method, are restored first in parallel under the constraints of DGs and network. During the selection process of recovery paths, line weight and node importance degree are proposed under the consideration of the node topological importance and the load importance as well as the backbone network restoration time. Therefore, the whole optimization of the reconstructed network is realized. Finally, the simulation results verify the feasibility and effectiveness of the strategy.

  1. Adaptive Distributed Data Structure Management for Parallel CFD Applications

    KAUST Repository

    Frisch, Jerome

    2013-09-01

    Computational fluid dynamics (CFD) simulations require a lot of computing resources in terms of CPU time and memory in order to compute with a reasonable physical accuracy. If only uniformly refined domains are applied, the amount of computing cells is growing rather fast if a certain small resolution is physically required. This can be remedied by applying adaptively refined grids. Unfortunately, due to the adaptive refinement procedures, errors are introduced which have to be taken into account. This paper is focussing on implementation details of the applied adaptive data structure management and a qualitative analysis of the introduced errors by analysing a Poisson problem on the given data structure, which has to be solved in every time step of a CFD analysis. Furthermore an adaptive CFD benchmark example is computed, showing the benefits of an adaptive refinement as well as measurements of parallel data distribution and performance. © 2013 IEEE.

  2. Parallel computing by Monte Carlo codes MVP/GMVP

    International Nuclear Information System (INIS)

    Nagaya, Yasunobu; Nakagawa, Masayuki; Mori, Takamasa

    2001-01-01

    General-purpose Monte Carlo codes MVP/GMVP are well-vectorized and thus enable us to perform high-speed Monte Carlo calculations. In order to achieve more speedups, we parallelized the codes on the different types of parallel computing platforms or by using a standard parallelization library MPI. The platforms used for benchmark calculations are a distributed-memory vector-parallel computer Fujitsu VPP500, a distributed-memory massively parallel computer Intel paragon and a distributed-memory scalar-parallel computer Hitachi SR2201, IBM SP2. As mentioned generally, linear speedup could be obtained for large-scale problems but parallelization efficiency decreased as the batch size per a processing element(PE) was smaller. It was also found that the statistical uncertainty for assembly powers was less than 0.1% by the PWR full-core calculation with more than 10 million histories and it took about 1.5 hours by massively parallel computing. (author)

  3. Parallel computing: numerics, applications, and trends

    National Research Council Canada - National Science Library

    Trobec, Roman; Vajteršic, Marián; Zinterhof, Peter

    2009-01-01

    ... and/or distributed systems. The contributions to this book are focused on topics most concerned in the trends of today's parallel computing. These range from parallel algorithmics, programming, tools, network computing to future parallel computing. Particular attention is paid to parallel numerics: linear algebra, differential equations, numerica...

  4. Research on charging and discharging control strategy for electric vehicles as distributed energy storage devices

    Science.gov (United States)

    Zhang, Min; Yang, Feng; Zhang, Dongqing; Tang, Pengcheng

    2018-02-01

    A large number of electric vehicles are connected to the family micro grid will affect the operation safety of the power grid and the quality of power. Considering the factors of family micro grid price and electric vehicle as a distributed energy storage device, a two stage optimization model is established, and the improved discrete binary particle swarm optimization algorithm is used to optimize the parameters in the model. The proposed control strategy of electric vehicle charging and discharging is of practical significance for the rational control of electric vehicle as a distributed energy storage device and electric vehicle participating in the peak load regulation of power consumption.

  5. Parallel grid generation algorithm for distributed memory computers

    Science.gov (United States)

    Moitra, Stuti; Moitra, Anutosh

    1994-01-01

    A parallel grid-generation algorithm and its implementation on the Intel iPSC/860 computer are described. The grid-generation scheme is based on an algebraic formulation of homotopic relations. Methods for utilizing the inherent parallelism of the grid-generation scheme are described, and implementation of multiple levELs of parallelism on multiple instruction multiple data machines are indicated. The algorithm is capable of providing near orthogonality and spacing control at solid boundaries while requiring minimal interprocessor communications. Results obtained on the Intel hypercube for a blended wing-body configuration are used to demonstrate the effectiveness of the algorithm. Fortran implementations bAsed on the native programming model of the iPSC/860 computer and the Express system of software tools are reported. Computational gains in execution time speed-up ratios are given.

  6. The socio-technical transition of distributed electricity storage into future networks—System value and stakeholder views

    International Nuclear Information System (INIS)

    Grünewald, Philipp H.; Cockerill, Timothy T.; Contestabile, Marcello; Pearson, Peter J.G.

    2012-01-01

    Whole system models for the GB electricity system suggest that distributed electricity storage has the potential to significantly reduce the system integration cost for future system scenarios. From a policy perspective, this poses the question why this value should not be realised within existing market structures. Opinion among stakeholders is divided. Some believe that storage deployment constitutes a ‘special case’ in need of policy support. Others insist that markets can provide the necessary platform to negotiate contracts, which reward storage operators for the range of services they could provide. This paper seeks to inform this debate with a process of stakeholder engagement using a perspective informed by socio-technical transition literatures. This approach allows the identification of tensions among actors in the electricity system and of possibilities for co-evolution in the deployment of storage technologies during a transition towards a low carbon electricity system. It also draws attention to policy-related challenges of technology lock-in and path dependency resulting from poor alignment of incumbent regimes with the requirements for distributed electricity storage. - Highlights: ► Electricity storage is poorly aligned with existing regimes in the electricity system. ► Stakeholders perceive electricity storage as “somebody else's problem”. ► Combining stakeholder views and transition theory provides new insight. ► Transition from network to operational benefits poses regulatory challenge. ► Value aggregation made difficult due to institutional barriers.

  7. Research on high-performance mass storage system

    International Nuclear Information System (INIS)

    Cheng Yaodong; Wang Lu; Huang Qiulan; Zheng Wei

    2010-01-01

    With the enlargement of scientific experiments, more and more data will be produced, which brings great challenge to storage system. Large storage capacity and high data access performance are both important to Mass storage system. This paper firstly reviews some kinds of popular storage systems including network storage system, SAN-based sharing system, WAN File system, object-based parallel file system, hierarchical storage system and cloud storage systems. Then some key technologies are presented. Finally, this paper takes BES storage system as an example and introduces its requirements, architecture and operation results. (authors)

  8. A program system for ab initio MO calculations on vector and parallel processing machines. Pt. 3

    International Nuclear Information System (INIS)

    Wiest, R.; Demuynck, J.; Benard, M.; Rohmer, M.M.; Ernenwein, R.

    1991-01-01

    This series of three papers presents a program system for ab initio molecular orbital calculations on vector and parallel computers. Part III is devoted to the four-index transformation on a molecular orbital basis of size NMO of the file of two-electorn integrals (pqparallelrs) generated by a contracted Gaussian set of size NATO (number of atomic orbitals). A fast Yoshimine algorithm first sorts the (pqparallelrs) integrals with respect to index pq only. This file of half-sorted integrals labelled by their rs-index can be processed without further modification to generate either the transformed integrals or the supermatrix elements. The large memory available on the CRAY-2 hase made possible to implement the transformation algorithm proposed by Bender in 1972, which requires a core-storage allocation varying as (NATO) 3 . Two versions of Bender's algorithm are included in the present program. The first version is an in-core version, where the complete file of accumulated contributions to transformed integrals in stored and updated in central memory. This version has been parallelized by distributing over a limited number of logical tasks the NATO steps corresponding to the scanning of the most external loop. The second version is an out-of-core version, in which twin files are alternatively used as input and output for the accumulated contributions to transformed integrals. This version is not parallel. The choice of one or another version and (for version 1) the determination of the number of tasks depends upon the balance between the available and the requested amounts of storage. The storage management and the choice of the proper version are carried out automatically using dynamic storage allocation. Both versions are vectorized and take advantage of the molecular symmetry. (orig.)

  9. The STAPL Parallel Graph Library

    KAUST Repository

    Harshvardhan,

    2013-01-01

    This paper describes the stapl Parallel Graph Library, a high-level framework that abstracts the user from data-distribution and parallelism details and allows them to concentrate on parallel graph algorithm development. It includes a customizable distributed graph container and a collection of commonly used parallel graph algorithms. The library introduces pGraph pViews that separate algorithm design from the container implementation. It supports three graph processing algorithmic paradigms, level-synchronous, asynchronous and coarse-grained, and provides common graph algorithms based on them. Experimental results demonstrate improved scalability in performance and data size over existing graph libraries on more than 16,000 cores and on internet-scale graphs containing over 16 billion vertices and 250 billion edges. © Springer-Verlag Berlin Heidelberg 2013.

  10. Methods and apparatus for multi-resolution replication of files in a parallel computing system using semantic information

    Science.gov (United States)

    Faibish, Sorin; Bent, John M.; Tzelnic, Percy; Grider, Gary; Torres, Aaron

    2015-10-20

    Techniques are provided for storing files in a parallel computing system using different resolutions. A method is provided for storing at least one file generated by a distributed application in a parallel computing system. The file comprises one or more of a complete file and a sub-file. The method comprises the steps of obtaining semantic information related to the file; generating a plurality of replicas of the file with different resolutions based on the semantic information; and storing the file and the plurality of replicas of the file in one or more storage nodes of the parallel computing system. The different resolutions comprise, for example, a variable number of bits and/or a different sub-set of data elements from the file. A plurality of the sub-files can be merged to reproduce the file.

  11. Redox Flow Batteries, Hydrogen and Distributed Storage.

    Science.gov (United States)

    Dennison, C R; Vrubel, Heron; Amstutz, Véronique; Peljo, Pekka; Toghill, Kathryn E; Girault, Hubert H

    2015-01-01

    Social, economic, and political pressures are causing a shift in the global energy mix, with a preference toward renewable energy sources. In order to realize widespread implementation of these resources, large-scale storage of renewable energy is needed. Among the proposed energy storage technologies, redox flow batteries offer many unique advantages. The primary limitation of these systems, however, is their limited energy density which necessitates very large installations. In order to enhance the energy storage capacity of these systems, we have developed a unique dual-circuit architecture which enables two levels of energy storage; first in the conventional electrolyte, and then through the formation of hydrogen. Moreover, we have begun a pilot-scale demonstration project to investigate the scalability and technical readiness of this approach. This combination of conventional energy storage and hydrogen production is well aligned with the current trajectory of modern energy and mobility infrastructure. The combination of these two means of energy storage enables the possibility of an energy economy dominated by renewable resources.

  12. Parallel compression of data chunks of a shared data object using a log-structured file system

    Science.gov (United States)

    Bent, John M.; Faibish, Sorin; Grider, Gary

    2016-10-25

    Techniques are provided for parallel compression of data chunks being written to a shared object. A client executing on a compute node or a burst buffer node in a parallel computing system stores a data chunk generated by the parallel computing system to a shared data object on a storage node by compressing the data chunk; and providing the data compressed data chunk to the storage node that stores the shared object. The client and storage node may employ Log-Structured File techniques. The compressed data chunk can be de-compressed by the client when the data chunk is read. A storage node stores a data chunk as part of a shared object by receiving a compressed version of the data chunk from a compute node; and storing the compressed version of the data chunk to the shared data object on the storage node.

  13. Parallel checksumming of data chunks of a shared data object using a log-structured file system

    Science.gov (United States)

    Bent, John M.; Faibish, Sorin; Grider, Gary

    2016-09-06

    Checksum values are generated and used to verify the data integrity. A client executing in a parallel computing system stores a data chunk to a shared data object on a storage node in the parallel computing system. The client determines a checksum value for the data chunk; and provides the checksum value with the data chunk to the storage node that stores the shared object. The data chunk can be stored on the storage node with the corresponding checksum value as part of the shared object. The storage node may be part of a Parallel Log-Structured File System (PLFS), and the client may comprise, for example, a Log-Structured File System client on a compute node or burst buffer. The checksum value can be evaluated when the data chunk is read from the storage node to verify the integrity of the data that is read.

  14. Comprehensive Monitoring for Heterogeneous Geographically Distributed Storage

    Energy Technology Data Exchange (ETDEWEB)

    Ratnikova, N. [Fermilab; Karavakis, E. [CERN; Lammel, S. [Fermilab; Wildish, T. [Princeton U.

    2015-12-23

    Storage capacity at CMS Tier-1 and Tier-2 sites reached over 100 Petabytes in 2014, and will be substantially increased during Run 2 data taking. The allocation of storage for the individual users analysis data, which is not accounted as a centrally managed storage space, will be increased to up to 40%. For comprehensive tracking and monitoring of the storage utilization across all participating sites, CMS developed a space monitoring system, which provides a central view of the geographically dispersed heterogeneous storage systems. The first prototype was deployed at pilot sites in summer 2014, and has been substantially reworked since then. In this paper we discuss the functionality and our experience of system deployment and operation on the full CMS scale.

  15. Coordinated Control of Distributed and Bulk Energy Storage for Alleviation of Post-Contingency Overloads

    Directory of Open Access Journals (Sweden)

    Yunfeng Wen

    2014-03-01

    Full Text Available This paper presents a novel corrective control strategy that can effectively coordinate distributed and bulk energy storage to relieve post-contingency overloads. Immediately following a contingency, distributed batteries are implemented to provide fast corrective actions to reduce power flows below their short-term emergency ratings. During the long-term period, Pumped Hydro Storage units work in pumping or generation mode to aid conventional generating units keep line flows below the normal ratings. This problem is formulated as a multi-stage Corrective Security-constrained OPF (CSCOPF. An algorithm based on Benders decomposition was proposed to find the optimal base case solution and seek feasible corrective actions to handle all contingencies. Case studies based on a modified RTS-96 system demonstrate the performance and effectiveness of the proposed control strategy.

  16. Installation of the first Distributed Energy Storage System (DESS) at American Electric Power (AEP).

    Energy Technology Data Exchange (ETDEWEB)

    Nourai, Ali (American Electric Power Company, Columbus, OH)

    2007-06-01

    AEP studied the direct and indirect benefits, strengths, and weaknesses of distributed energy storage systems (DESS) and chose to transform its entire utility grid into a system that achieves optimal integration of both central and distributed energy assets. To that end, AEP installed the first NAS battery-based, energy storage system in North America. After one year of operation and testing, AEP has concluded that, although the initial costs of DESS are greater than conventional power solutions, the net benefits justify the AEP decision to create a grid of DESS with intelligent monitoring, communications, and control, in order to enable the utility grid of the future. This report details the site selection, construction, benefits and lessons learned of the first installation, at Chemical Station in North Charleston, WV.

  17. Design and analysis of electrical energy storage demonstration projects on UK distribution networks

    International Nuclear Information System (INIS)

    Lyons, P.F.; Wade, N.S.; Jiang, T.; Taylor, P.C.; Hashiesh, F.; Michel, M.; Miller, D.

    2015-01-01

    Highlights: • Results of an EES system demonstration project carried out in the UK. • Approaches to the design of trials for EES and observation on their application. • A formalised methodology for analysis of smart grids trials. • Validated models of energy storage. • Capability of EES to connect larger quantities of heat pumps and PV is evaluated. - Abstract: The UK government’s CO 2 emissions targets will require electrification of much of the country’s infrastructure with low carbon technologies such as photovoltaic panels, electric vehicles and heat pumps. The large scale proliferation of these technologies will necessitate major changes to the planning and operation of distribution networks. Distribution network operators are trialling electrical energy storage (EES) across their networks to increase their understanding of the contribution that it can make to enable the expected paradigm shift in generation and consumption of electricity. In order to evaluate a range of applications for EES, including voltage control and power flow management, installations have taken place at various distribution network locations and voltage levels. This article reports on trial design approaches and their application to a UK trial of an EES system to ensure broad applicability of the results. Results from these trials of an EES system, low carbon technologies and trial distribution networks are used to develop validated power system models. These models are used to evaluate, using a formalised methodology, the impact that EES could have on the design and operation of future distribution networks

  18. Parallelization of quantum molecular dynamics simulation code

    International Nuclear Information System (INIS)

    Kato, Kaori; Kunugi, Tomoaki; Shibahara, Masahiko; Kotake, Susumu

    1998-02-01

    A quantum molecular dynamics simulation code has been developed for the analysis of the thermalization of photon energies in the molecule or materials in Kansai Research Establishment. The simulation code is parallelized for both Scalar massively parallel computer (Intel Paragon XP/S75) and Vector parallel computer (Fujitsu VPP300/12). Scalable speed-up has been obtained with a distribution to processor units by division of particle group in both parallel computers. As a result of distribution to processor units not only by particle group but also by the particles calculation that is constructed with fine calculations, highly parallelization performance is achieved in Intel Paragon XP/S75. (author)

  19. Simulation models developed for voltage control in a distribution network using energy storage systems for PV penetration

    DEFF Research Database (Denmark)

    Mihet-Popa, Lucian; Bindner, Henrik W.

    2013-01-01

    This paper presents the development of simulation models for DER components in a distribution network, with focus on voltage controllers using energy storage systems for PV penetration. The Vanadium Redox Battery (VRB) system model, used as an energy storage system, was implemented in MATLAB....../Simulink and DIgSILENT PowerFactory, based on the efficiency of different components-such as: cell stacks, electrolytes, pumps and power converters, whilst power losses were also taken into account. The simulation results have been validated against measurements using experimental facility of a distributed power...

  20. The simplified spherical harmonics (SPL) methodology with space and moment decomposition in parallel environments

    International Nuclear Information System (INIS)

    Gianluca, Longoni; Alireza, Haghighat

    2003-01-01

    In recent years, the SP L (simplified spherical harmonics) equations have received renewed interest for the simulation of nuclear systems. We have derived the SP L equations starting from the even-parity form of the S N equations. The SP L equations form a system of (L+1)/2 second order partial differential equations that can be solved with standard iterative techniques such as the Conjugate Gradient (CG). We discretized the SP L equations with the finite-volume approach in a 3-D Cartesian space. We developed a new 3-D general code, Pensp L (Parallel Environment Neutral-particle SP L ). Pensp L solves both fixed source and criticality eigenvalue problems. In order to optimize the memory management, we implemented a Compressed Diagonal Storage (CDS) to store the SP L matrices. Pensp L includes parallel algorithms for space and moment domain decomposition. The computational load is distributed on different processors, using a mapping function, which maps the 3-D Cartesian space and moments onto processors. The code is written in Fortran 90 using the Message Passing Interface (MPI) libraries for the parallel implementation of the algorithm. The code has been tested on the Pcpen cluster and the parallel performance has been assessed in terms of speed-up and parallel efficiency. (author)

  1. Optimal Sizing and Placement of Battery Energy Storage in Distribution System Based on Solar Size for Voltage Regulation

    Energy Technology Data Exchange (ETDEWEB)

    Nazaripouya, Hamidreza [Univ. of California, Los Angeles, CA (United States); Wang, Yubo [Univ. of California, Los Angeles, CA (United States); Chu, Peter [Univ. of California, Los Angeles, CA (United States); Pota, Hemanshu R. [Univ. of California, Los Angeles, CA (United States); Gadh, Rajit [Univ. of California, Los Angeles, CA (United States)

    2016-07-26

    This paper proposes a new strategy to achieve voltage regulation in distributed power systems in the presence of solar energy sources and battery storage systems. The goal is to find the minimum size of battery storage and its corresponding location in the network based on the size and place of the integrated solar generation. The proposed method formulates the problem by employing the network impedance matrix to obtain an analytical solution instead of using a recursive algorithm such as power flow. The required modifications for modeling the slack and PV buses (generator buses) are utilized to increase the accuracy of the approach. The use of reactive power control to regulate the voltage regulation is not always an optimal solution as in distribution systems R/X is large. In this paper the minimum size and the best place of battery storage is achieved by optimizing the amount of both active and reactive power exchanged by battery storage and its gridtie inverter (GTI) based on the network topology and R/X ratios in the distribution system. Simulation results for the IEEE 14-bus system verify the effectiveness of the proposed approach.

  2. Scheduling of distributed energy storage for passive contribution in the imbalance settlement system of the Netherlands

    NARCIS (Netherlands)

    Lampropoulos, I.; Garoufalis, P.; Kling, W.L.

    2014-01-01

    The subject addressed in this paper is the intra-hour scheduling of a distributed battery energy storage system (BESS). The case study is about a lithium-ion BESS integrated in a low voltage (LV) distribution system with residential customers and photovoltaic (PV) generation in the Netherlands. The

  3. New superconducting coil configuration for energy storage

    International Nuclear Information System (INIS)

    Tokorabet, M.; Mailfert, A.; Colteu, A.

    1998-01-01

    Energy storage using superconducting coils involves the problem of electromagnetic field pollution outside the considered system. Different configurations are widely studied: the torus, the alone solenoid and multiple parallel solenoids enclosed in one container. A new configuration which minimizes the external pollution is studied in this paper. The theoretical system is composed of two spherical distributions of the current which are concentric. The analytical study uses solution of Laplace equations. Parametric study covers energy, flux density and geometrical data. The second study concerns the numerical approach of this design using coaxial solenoids. A comparison between this new system and the known systems is presented as a conclusion. (orig.)

  4. Parallel discrete ordinates algorithms on distributed and common memory systems

    International Nuclear Information System (INIS)

    Wienke, B.R.; Hiromoto, R.E.; Brickner, R.G.

    1987-01-01

    The S/sub n/ algorithm employs iterative techniques in solving the linear Boltzmann equation. These methods, both ordered and chaotic, were compared on both the Denelcor HEP and the Intel hypercube. Strategies are linked to the organization and accessibility of memory (common memory versus distributed memory architectures), with common concern for acquisition of global information. Apart from this, the inherent parallelism of the algorithm maps directly onto the two architectures. Results comparing execution times, speedup, and efficiency are based on a representative 16-group (full upscatter and downscatter) sample problem. Calculations were performed on both the Los Alamos National Laboratory (LANL) Denelcor HEP and the LANL Intel hypercube. The Denelcor HEP is a 64-bit multi-instruction, multidate MIMD machine consisting of up to 16 process execution modules (PEMs), each capable of executing 64 processes concurrently. Each PEM can cooperate on a job, or run several unrelated jobs, and share a common global memory through a crossbar switch. The Intel hypercube, on the other hand, is a distributed memory system composed of 128 processing elements, each with its own local memory. Processing elements are connected in a nearest-neighbor hypercube configuration and sharing of data among processors requires execution of explicit message-passing constructs

  5. Distribution of Evaporating CO2 in Parallel Microchannels

    DEFF Research Database (Denmark)

    Brix, Wiebke; Elmegaard, Brian

    2008-01-01

    The impact on the heat exchanger performance due to maldistribution of evaporating CO2 in parallel channels is investigated numerically. A 1D steady state simulation model of a microchannel evaporator is built using correlations from the literature to calculate frictional pressure drop and heat...... transfer coefficients. For two channels in parallel two different cases of maldistribution are studied. Firstly, the impact of a non-uniform air flow is considered, and secondly the impact of maldistribution of the two phases in the inlet manifold is investigated. The results for both cases are compared...

  6. Wide-area-distributed storage system for a multimedia database

    Science.gov (United States)

    Ueno, Masahiro; Kinoshita, Shigechika; Kuriki, Makato; Murata, Setsuko; Iwatsu, Shigetaro

    1998-12-01

    We have developed a wide-area-distribution storage system for multimedia databases, which minimizes the possibility of simultaneous failure of multiple disks in the event of a major disaster. It features a RAID system, whose member disks are spatially distributed over a wide area. Each node has a device, which includes the controller of the RAID and the controller of the member disks controlled by other nodes. The devices in the node are connected to a computer, using fiber optic cables and communicate using fiber-channel technology. Any computer at a node can utilize multiple devices connected by optical fibers as a single 'virtual disk.' The advantage of this system structure is that devices and fiber optic cables are shared by the computers. In this report, we first described our proposed system, and a prototype was used for testing. We then discussed its performance; i.e., how to read and write throughputs are affected by data-access delay, the RAID level, and queuing.

  7. Distributed demand-side management optimisation for multi-residential users with energy production and storage strategies

    Directory of Open Access Journals (Sweden)

    Emmanuel Chifuel Manasseh

    2014-12-01

    Full Text Available This study considers load control in a multi-residential setup where energy scheduler (ES devices installed in smart meters are employed for demand-side management (DSM. Several residential end-users share the same energy source and each residential user has non-adjustable loads and adjustable loads. In addition, residential users may have storage devices and renewable energy sources such as wind turbines or solar as well as dispatchable generators. The ES devices exchange information automatically by executing an iterative distributed algorithm to locate the optimal energy schedule for each end-user. This will reduce the total energy cost and the peak-to-average ratio (PAR in energy demand in the electric power distribution. Users possessing storage devices and dispatchable generators strategically utilise their resources to minimise the total energy cost together with the PAR. Simulation results are provided to evaluate the performance of the proposed game theoretic-based distributed DSM technique.

  8. Optimal Stochastic Management of Distributed Energy Storage Embedded with Wind Farms

    OpenAIRE

    Yanchi, Xiao; Vargas, Bruce; Hamdi, Mohammd

    2018-01-01

    Increasing wind turbines (WT) penetration and low carbon demand can potentially lead to two different flow peaks, generation and load, within distribution networks. This will not only constrain WT penetration but also pose serious threats to network reliability. This paper proposes energy storage (ES) to reduce system congestion cost caused by the two peaks by sending cost-reflective economic signals to affect ES operation in responding to network conditions. Firstly, a new charging and disch...

  9. An Efficient MapReduce-Based Parallel Clustering Algorithm for Distributed Traffic Subarea Division

    Directory of Open Access Journals (Sweden)

    Dawen Xia

    2015-01-01

    Full Text Available Traffic subarea division is vital for traffic system management and traffic network analysis in intelligent transportation systems (ITSs. Since existing methods may not be suitable for big traffic data processing, this paper presents a MapReduce-based Parallel Three-Phase K-Means (Par3PKM algorithm for solving traffic subarea division problem on a widely adopted Hadoop distributed computing platform. Specifically, we first modify the distance metric and initialization strategy of K-Means and then employ a MapReduce paradigm to redesign the optimized K-Means algorithm for parallel clustering of large-scale taxi trajectories. Moreover, we propose a boundary identifying method to connect the borders of clustering results for each cluster. Finally, we divide traffic subarea of Beijing based on real-world trajectory data sets generated by 12,000 taxis in a period of one month using the proposed approach. Experimental evaluation results indicate that when compared with K-Means, Par2PK-Means, and ParCLARA, Par3PKM achieves higher efficiency, more accuracy, and better scalability and can effectively divide traffic subarea with big taxi trajectory data.

  10. Fuzzy-driven energy storage system for mitigating voltage unbalance factor on distribution network with photovoltaic system

    Science.gov (United States)

    Wong, Jianhui; Lim, Yun Seng; Morris, Stella; Morris, Ezra; Chua, Kein Huat

    2017-04-01

    The amount of small-scaled renewable energy sources is anticipated to increase on the low-voltage distribution networks for the improvement of energy efficiency and reduction of greenhouse gas emission. The growth of the PV systems on the low-voltage distribution networks can create voltage unbalance, voltage rise, and reverse-power flow. Usually these issues happen with little fluctuation. However, it tends to fluctuate severely as Malaysia is a region with low clear sky index. A large amount of clouds often passes over the country, hence making the solar irradiance to be highly scattered. Therefore, the PV power output fluctuates substantially. These issues can lead to the malfunction of the electronic based equipment, reduction in the network efficiency and improper operation of the power protection system. At the current practice, the amount of PV system installed on the distribution network is constraint by the utility company. As a result, this can limit the reduction of carbon footprint. Therefore, energy storage system is proposed as a solution for these power quality issues. To ensure an effective operation of the distribution network with PV system, a fuzzy control system is developed and implemented to govern the operation of an energy storage system. The fuzzy driven energy storage system is able to mitigate the fluctuating voltage rise and voltage unbalance on the electrical grid by actively manipulates the flow of real power between the grid and the batteries. To verify the effectiveness of the proposed fuzzy driven energy storage system, an experimental network integrated with 7.2kWp PV system was setup. Several case studies are performed to evaluate the response of the proposed solution to mitigate voltage rises, voltage unbalance and reduce the amount of reverse power flow under highly intermittent PV power output.

  11. Cooperative storage of shared files in a parallel computing system with dynamic block size

    Science.gov (United States)

    Bent, John M.; Faibish, Sorin; Grider, Gary

    2015-11-10

    Improved techniques are provided for parallel writing of data to a shared object in a parallel computing system. A method is provided for storing data generated by a plurality of parallel processes to a shared object in a parallel computing system. The method is performed by at least one of the processes and comprises: dynamically determining a block size for storing the data; exchanging a determined amount of the data with at least one additional process to achieve a block of the data having the dynamically determined block size; and writing the block of the data having the dynamically determined block size to a file system. The determined block size comprises, e.g., a total amount of the data to be stored divided by the number of parallel processes. The file system comprises, for example, a log structured virtual parallel file system, such as a Parallel Log-Structured File System (PLFS).

  12. Expressing Parallelism with ROOT

    Energy Technology Data Exchange (ETDEWEB)

    Piparo, D. [CERN; Tejedor, E. [CERN; Guiraud, E. [CERN; Ganis, G. [CERN; Mato, P. [CERN; Moneta, L. [CERN; Valls Pla, X. [CERN; Canal, P. [Fermilab

    2017-11-22

    The need for processing the ever-increasing amount of data generated by the LHC experiments in a more efficient way has motivated ROOT to further develop its support for parallelism. Such support is being tackled both for shared-memory and distributed-memory environments. The incarnations of the aforementioned parallelism are multi-threading, multi-processing and cluster-wide executions. In the area of multi-threading, we discuss the new implicit parallelism and related interfaces, as well as the new building blocks to safely operate with ROOT objects in a multi-threaded environment. Regarding multi-processing, we review the new MultiProc framework, comparing it with similar tools (e.g. multiprocessing module in Python). Finally, as an alternative to PROOF for cluster-wide executions, we introduce the efforts on integrating ROOT with state-of-the-art distributed data processing technologies like Spark, both in terms of programming model and runtime design (with EOS as one of the main components). For all the levels of parallelism, we discuss, based on real-life examples and measurements, how our proposals can increase the productivity of scientists.

  13. Expressing Parallelism with ROOT

    Science.gov (United States)

    Piparo, D.; Tejedor, E.; Guiraud, E.; Ganis, G.; Mato, P.; Moneta, L.; Valls Pla, X.; Canal, P.

    2017-10-01

    The need for processing the ever-increasing amount of data generated by the LHC experiments in a more efficient way has motivated ROOT to further develop its support for parallelism. Such support is being tackled both for shared-memory and distributed-memory environments. The incarnations of the aforementioned parallelism are multi-threading, multi-processing and cluster-wide executions. In the area of multi-threading, we discuss the new implicit parallelism and related interfaces, as well as the new building blocks to safely operate with ROOT objects in a multi-threaded environment. Regarding multi-processing, we review the new MultiProc framework, comparing it with similar tools (e.g. multiprocessing module in Python). Finally, as an alternative to PROOF for cluster-wide executions, we introduce the efforts on integrating ROOT with state-of-the-art distributed data processing technologies like Spark, both in terms of programming model and runtime design (with EOS as one of the main components). For all the levels of parallelism, we discuss, based on real-life examples and measurements, how our proposals can increase the productivity of scientists.

  14. Parallel-Vector Algorithm For Rapid Structural Anlysis

    Science.gov (United States)

    Agarwal, Tarun R.; Nguyen, Duc T.; Storaasli, Olaf O.

    1993-01-01

    New algorithm developed to overcome deficiency of skyline storage scheme by use of variable-band storage scheme. Exploits both parallel and vector capabilities of modern high-performance computers. Gives engineers and designers opportunity to include more design variables and constraints during optimization of structures. Enables use of more refined finite-element meshes to obtain improved understanding of complex behaviors of aerospace structures leading to better, safer designs. Not only attractive for current supercomputers but also for next generation of shared-memory supercomputers.

  15. Second derivative parallel block backward differentiation type ...

    African Journals Online (AJOL)

    Second derivative parallel block backward differentiation type formulas for Stiff ODEs. ... Log in or Register to get access to full text downloads. ... and the methods are inherently parallel and can be distributed over parallel processors. They are ...

  16. Application of the distributed genetic algorithm for in-core fuel optimization problems under parallel computational environment

    International Nuclear Information System (INIS)

    Yamamoto, Akio; Hashimoto, Hiroshi

    2002-01-01

    The distributed genetic algorithm (DGA) is applied for loading pattern optimization problems of the pressurized water reactors. A basic concept of DGA follows that of the conventional genetic algorithm (GA). However, DGA equally distributes candidates of solutions (i.e. loading patterns) to several independent ''islands'' and evolves them in each island. Communications between islands, i.e. migrations of some candidates between islands are performed with a certain period. Since candidates of solutions independently evolve in each island while accepting different genes of migrants, premature convergence in the conventional GA can be prevented. Because many candidate loading patterns should be evaluated in GA or DGA, the parallelization is efficient to reduce turn around time. Parallel efficiency of DGA was measured using our optimization code and good efficiency was attained even in a heterogeneous cluster environment due to dynamic distribution of the calculation load. The optimization code is based on the client/server architecture with the TCP/IP native socket and a client (optimization) module and calculation server modules communicate the objects of loading patterns each other. Throughout the sensitivity study on optimization parameters of DGA, a suitable set of the parameters for a test problem was identified. Finally, optimization capability of DGA and the conventional GA was compared in the test problem and DGA provided better optimization results than the conventional GA. (author)

  17. Kmerind: A Flexible Parallel Library for K-mer Indexing of Biological Sequences on Distributed Memory Systems.

    Science.gov (United States)

    Pan, Tony; Flick, Patrick; Jain, Chirag; Liu, Yongchao; Aluru, Srinivas

    2017-10-09

    Counting and indexing fixed length substrings, or k-mers, in biological sequences is a key step in many bioinformatics tasks including genome alignment and mapping, genome assembly, and error correction. While advances in next generation sequencing technologies have dramatically reduced the cost and improved latency and throughput, few bioinformatics tools can efficiently process the datasets at the current generation rate of 1.8 terabases every 3 days. We present Kmerind, a high performance parallel k-mer indexing library for distributed memory environments. The Kmerind library provides a set of simple and consistent APIs with sequential semantics and parallel implementations that are designed to be flexible and extensible. Kmerind's k-mer counter performs similarly or better than the best existing k-mer counting tools even on shared memory systems. In a distributed memory environment, Kmerind counts k-mers in a 120 GB sequence read dataset in less than 13 seconds on 1024 Xeon CPU cores, and fully indexes their positions in approximately 17 seconds. Querying for 1% of the k-mers in these indices can be completed in 0.23 seconds and 28 seconds, respectively. Kmerind is the first k-mer indexing library for distributed memory environments, and the first extensible library for general k-mer indexing and counting. Kmerind is available at https://github.com/ParBLiSS/kmerind.

  18. Model Predictive Control for Distributed Microgrid Battery Energy Storage Systems

    DEFF Research Database (Denmark)

    Morstyn, Thomas; Hredzak, Branislav; Aguilera, Ricardo P.

    2018-01-01

    , and converter current constraints to be addressed. In addition, nonlinear variations in the charge and discharge efficiencies of lithium ion batteries are analyzed and included in the control strategy. Real-time digital simulations were carried out for an islanded microgrid based on the IEEE 13 bus prototypical......This brief proposes a new convex model predictive control (MPC) strategy for dynamic optimal power flow between battery energy storage (ES) systems distributed in an ac microgrid. The proposed control strategy uses a new problem formulation, based on a linear $d$ – $q$ reference frame voltage...... feeder, with distributed battery ES systems and intermittent photovoltaic generation. It is shown that the proposed control strategy approaches the performance of a strategy based on nonconvex optimization, while reducing the required computation time by a factor of 1000, making it suitable for a real...

  19. State-of-Charge Balance Using Adaptive Droop Control for Distributed Energy Storage Systems in DC MicroGrid Applications

    DEFF Research Database (Denmark)

    Lu, Xiaonan; Sun, Kai; Guerrero, Josep M.

    2014-01-01

    This paper presents the coordinated control of distributed energy storage systems (DESSs) in DC micro-grids. In order to balance the state-of-charge (SoC) of each energy storage unit (ESU), an SoC-based adaptive droop control method is proposed. In this decentralized control method, the droop...

  20. Coordinated Control Scheme of Battery Energy Storage System (BESS) and Distributed Generations (DGs) for Electric Distribution Grid Operation

    DEFF Research Database (Denmark)

    Cha, Seung-Tae; Zhao, Haoran; Wu, Qiuwei

    2012-01-01

    into the islanding operation mode, while the centralized joint load frequency control (CJLFC) utilizing DGs handles the secondary frequency regulation. The BESS with the associated controllers has been modelled in Real-time digital simulator (RTDS) in order to identify the improvement of the frequency and voltage......This paper describes a coordinated control scheme of battery energy storage system (BESS) and distributed generations (DGs) for electric distribution grid operation. The BESS is designed to stabilize frequency and voltages as a primary control after the electric distribution system enters...... response. The modified IEEE 9-bus system, which is comprised of several DG units, wind power plant and the BESS, has been employed to illustrate the performance of the proposed coordinated flexible control scheme using RTDS in order to verify its practical efficacy....

  1. Parallel phase model : a programming model for high-end parallel machines with manycores.

    Energy Technology Data Exchange (ETDEWEB)

    Wu, Junfeng (Syracuse University, Syracuse, NY); Wen, Zhaofang; Heroux, Michael Allen; Brightwell, Ronald Brian

    2009-04-01

    This paper presents a parallel programming model, Parallel Phase Model (PPM), for next-generation high-end parallel machines based on a distributed memory architecture consisting of a networked cluster of nodes with a large number of cores on each node. PPM has a unified high-level programming abstraction that facilitates the design and implementation of parallel algorithms to exploit both the parallelism of the many cores and the parallelism at the cluster level. The programming abstraction will be suitable for expressing both fine-grained and coarse-grained parallelism. It includes a few high-level parallel programming language constructs that can be added as an extension to an existing (sequential or parallel) programming language such as C; and the implementation of PPM also includes a light-weight runtime library that runs on top of an existing network communication software layer (e.g. MPI). Design philosophy of PPM and details of the programming abstraction are also presented. Several unstructured applications that inherently require high-volume random fine-grained data accesses have been implemented in PPM with very promising results.

  2. PETASCALE DATA STORAGE INSTITUTE (PDSI) Final Report

    Energy Technology Data Exchange (ETDEWEB)

    Gibson, Garth [Carnegie Mellon University

    2012-11-26

    Petascale computing infrastructures for scientific discovery make petascale demands on information storage capacity, performance, concurrency, reliability, availability, and manageability. The Petascale Data Storage Institute focuses on the data storage problems found in petascale scientific computing environments, with special attention to community issues such as interoperability, community buy-in, and shared tools. The Petascale Data Storage Institute is a collaboration between researchers at Carnegie Mellon University, National Energy Research Scientific Computing Center, Pacific Northwest National Laboratory, Oak Ridge National Laboratory, Sandia National Laboratory, Los Alamos National Laboratory, University of Michigan, and the University of California at Santa Cruz. Because the Institute focuses on low level files systems and storage systems, its role in improving SciDAC systems was one of supporting application middleware such as data management and system-level performance tuning. In retrospect, the Petascale Data Storage Institute’s most innovative and impactful contribution is the Parallel Log-structured File System (PLFS). Published in SC09, PLFS is middleware that operates in MPI-IO or embedded in FUSE for non-MPI applications. Its function is to decouple concurrently written files into a per-process log file, whose impact (the contents of the single file that the parallel application was concurrently writing) is determined on later reading, rather than during its writing. PLFS is transparent to the parallel application, offering a POSIX or MPI-IO interface, and it shows an order of magnitude speedup to the Chombo benchmark and two orders of magnitude to the FLASH benchmark. Moreover, LANL production applications see speedups of 5X to 28X, so PLFS has been put into production at LANL. Originally conceived and prototyped in a PDSI collaboration between LANL and CMU, it has grown to engage many other PDSI institutes, international partners like AWE

  3. A new method, with application, for analysis of the impacts on flood risk of widely distributed enhanced hillslope storage

    Directory of Open Access Journals (Sweden)

    P. Metcalfe

    2018-04-01

    Full Text Available Enhanced hillslope storage is utilised in natural flood management in order to retain overland storm run-off and to reduce connectivity between fast surface flow pathways and the channel. Examples include excavated ponds, deepened or bunded accumulation areas, and gullies and ephemeral channels blocked with wooden barriers or debris dams. The performance of large, distributed networks of such measures is poorly understood. Extensive schemes can potentially retain large quantities of run-off, but there are indications that much of their effectiveness can be attributed to desynchronisation of sub-catchment flood waves. Inappropriately sited measures may therefore increase, rather than mitigate, flood risk. Fully distributed hydrodynamic models have been applied in limited studies but introduce significant computational complexity. The longer run times of such models also restrict their use for uncertainty estimation or evaluation of the many potential configurations and storm sequences that may influence the timings and magnitudes of flood waves.Here a simplified overland flow-routing module and semi-distributed representation of enhanced hillslope storage is developed. It is applied to the headwaters of a large rural catchment in Cumbria, UK, where the use of an extensive network of storage features is proposed as a flood mitigation strategy. The models were run within a Monte Carlo framework against data for a 2-month period of extreme flood events that caused significant damage in areas downstream. Acceptable realisations and likelihood weightings were identified using the GLUE uncertainty estimation framework. Behavioural realisations were rerun against the catchment model modified with the addition of the hillslope storage. Three different drainage rate parameters were applied across the network of hillslope storage. The study demonstrates that schemes comprising widely distributed hillslope storage can be modelled effectively within such a

  4. A new method, with application, for analysis of the impacts on flood risk of widely distributed enhanced hillslope storage

    Science.gov (United States)

    Metcalfe, Peter; Beven, Keith; Hankin, Barry; Lamb, Rob

    2018-04-01

    Enhanced hillslope storage is utilised in natural flood management in order to retain overland storm run-off and to reduce connectivity between fast surface flow pathways and the channel. Examples include excavated ponds, deepened or bunded accumulation areas, and gullies and ephemeral channels blocked with wooden barriers or debris dams. The performance of large, distributed networks of such measures is poorly understood. Extensive schemes can potentially retain large quantities of run-off, but there are indications that much of their effectiveness can be attributed to desynchronisation of sub-catchment flood waves. Inappropriately sited measures may therefore increase, rather than mitigate, flood risk. Fully distributed hydrodynamic models have been applied in limited studies but introduce significant computational complexity. The longer run times of such models also restrict their use for uncertainty estimation or evaluation of the many potential configurations and storm sequences that may influence the timings and magnitudes of flood waves. Here a simplified overland flow-routing module and semi-distributed representation of enhanced hillslope storage is developed. It is applied to the headwaters of a large rural catchment in Cumbria, UK, where the use of an extensive network of storage features is proposed as a flood mitigation strategy. The models were run within a Monte Carlo framework against data for a 2-month period of extreme flood events that caused significant damage in areas downstream. Acceptable realisations and likelihood weightings were identified using the GLUE uncertainty estimation framework. Behavioural realisations were rerun against the catchment model modified with the addition of the hillslope storage. Three different drainage rate parameters were applied across the network of hillslope storage. The study demonstrates that schemes comprising widely distributed hillslope storage can be modelled effectively within such a reduced complexity

  5. Mapping robust parallel multigrid algorithms to scalable memory architectures

    Science.gov (United States)

    Overman, Andrea; Vanrosendale, John

    1993-01-01

    The convergence rate of standard multigrid algorithms degenerates on problems with stretched grids or anisotropic operators. The usual cure for this is the use of line or plane relaxation. However, multigrid algorithms based on line and plane relaxation have limited and awkward parallelism and are quite difficult to map effectively to highly parallel architectures. Newer multigrid algorithms that overcome anisotropy through the use of multiple coarse grids rather than relaxation are better suited to massively parallel architectures because they require only simple point-relaxation smoothers. In this paper, we look at the parallel implementation of a V-cycle multiple semicoarsened grid (MSG) algorithm on distributed-memory architectures such as the Intel iPSC/860 and Paragon computers. The MSG algorithms provide two levels of parallelism: parallelism within the relaxation or interpolation on each grid and across the grids on each multigrid level. Both levels of parallelism must be exploited to map these algorithms effectively to parallel architectures. This paper describes a mapping of an MSG algorithm to distributed-memory architectures that demonstrates how both levels of parallelism can be exploited. The result is a robust and effective multigrid algorithm for distributed-memory machines.

  6. Power Flow Distribution Strategy for Improved Power Electronics Energy Efficiency in Battery Storage Systems: Development and Implementation in a Utility-Scale System

    Directory of Open Access Journals (Sweden)

    Michael Schimpe

    2018-03-01

    Full Text Available Utility-scale battery storage systems typically consist of multiple smaller units contributing to the overall power dispatch of the system. Herein, the power distribution among these units is analyzed and optimized to operate the system with increased energy efficiency. To improve the real-life storage operation, a holistic system model for battery storage systems has been developed that enables a calculation of the energy efficiency. A utility-scale Second-Life battery storage system with a capacity of 3.3 MWh/3 MW is operated and evaluated in this work. The system is in operation for the provision of primary control reserve in combination with intraday trading for controlling the battery state of charge. The simulation model is parameterized with the system data. Results show that losses in power electronics dominate. An operational strategy improving the energy efficiency through an optimized power flow distribution within the storage system is developed. The power flow distribution strategy is based on the reduction of the power electronics losses at no-load/partial-load by minimizing their in-operation time. The simulation derived power flow distribution strategy is implemented in the real-life storage system. Field-test measurements and analysis prove the functionality of the power flow distribution strategy and reveal the reduction of the energy throughput of the units by 7%, as well as a significant reduction of energy losses in the units by 24%. The cost savings for electricity over the system’s lifetime are approximated to 4.4% of its investment cost.

  7. A Distributed Model Predictive Control approach for the integration of flexible loads, storage and renewables

    DEFF Research Database (Denmark)

    Ferrarini, Luca; Mantovani, Giancarlo; Costanzo, Giuseppe Tommaso

    2014-01-01

    This paper presents an innovative solution based on distributed model predictive controllers to integrate the control and management of energy consumption, energy storage, PV and wind generation at customer side. The overall goal is to enable an advanced prosumer to autoproduce part of the energy...... he needs with renewable sources and, at the same time, to optimally exploit the thermal and electrical storages, to trade off its comfort requirements with different pricing schemes (including real-time pricing), and apply optimal control techniques rather than sub-optimal heuristics....

  8. Energy Storage Systems

    Science.gov (United States)

    Elliott, David

    2017-07-01

    As renewable energy use expands there will be a need to develop ways to balance its variability. Storage is one of the options. Presently the main emphasis is for systems storing electrical power in advanced batteries (many of them derivatives of parallel developments in the electric vehicle field), as well as via liquid air storage, compressed air storage, super-capacitors and flywheels, and, the leader so far, pumped hydro reservoirs. In addition, new systems are emerging for hydrogen generation and storage, feeding fuel cell power production. Heat (and cold) is also a storage medium and some systems exploit thermal effects as part of wider energy management activity. Some of the more exotic ones even try to use gravity on a large scale. This short book looks at all the options, their potentials and their limits. There are no clear winners, with some being suited to short-term balancing and others to longer-term storage. The eventual mix adopted will be shaped by the pattern of development of other balancing measures, including smart-grid demand management and super-grid imports and exports.

  9. Parallel Monte Carlo reactor neutronics

    International Nuclear Information System (INIS)

    Blomquist, R.N.; Brown, F.B.

    1994-01-01

    The issues affecting implementation of parallel algorithms for large-scale engineering Monte Carlo neutron transport simulations are discussed. For nuclear reactor calculations, these include load balancing, recoding effort, reproducibility, domain decomposition techniques, I/O minimization, and strategies for different parallel architectures. Two codes were parallelized and tested for performance. The architectures employed include SIMD, MIMD-distributed memory, and workstation network with uneven interactive load. Speedups linear with the number of nodes were achieved

  10. Multi-objective problem of the modified distributed parallel machine and assembly scheduling problem (MDPMASP) with eligibility constraints

    Science.gov (United States)

    Amallynda, I.; Santosa, B.

    2017-11-01

    This paper proposes a new generalization of the distributed parallel machine and assembly scheduling problem (DPMASP) with eligibility constraints referred to as the modified distributed parallel machine and assembly scheduling problem (MDPMASP) with eligibility constraints. Within this generalization, we assume that there are a set non-identical factories or production lines, each one with a set unrelated parallel machine with different speeds in processing them disposed to a single assembly machine in series. A set of different products that are manufactured through an assembly program of a set of components (jobs) according to the requested demand. Each product requires several kinds of jobs with different sizes. Beside that we also consider to the multi-objective problem (MOP) of minimizing mean flow time and the number of tardy products simultaneously. This is known to be NP-Hard problem, is important to practice, as the former criterions to reflect the customer's demand and manufacturer's perspective. This is a realistic and complex problem with wide range of possible solutions, we propose four simple heuristics and two metaheuristics to solve it. Various parameters of the proposed metaheuristic algorithms are discussed and calibrated by means of Taguchi technique. All proposed algorithms are tested by Matlab software. Our computational experiments indicate that the proposed problem and fourth proposed algorithms are able to be implemented and can be used to solve moderately-sized instances, and giving efficient solutions, which are close to optimum in most cases.

  11. Parallel and Distributed Systems for Probabilistic Reasoning

    Science.gov (United States)

    2012-12-01

    Ranganathan "et"al...typically a random permutation over the vertices. Advances by Elidan et al. [2006] and Ranganathan et al. [2007] have focused on dynamic asynchronous...Wildfire algorithm shown in Alg. 3.6 is a direct parallelization of the algorithm proposed by [ Ranganathan et al., 2007]. The Wildfire algorithm

  12. Multi-objective optimization of water quality, pumps operation, and storage sizing of water distribution systems.

    Science.gov (United States)

    Kurek, Wojciech; Ostfeld, Avi

    2013-01-30

    A multi-objective methodology utilizing the Strength Pareto Evolutionary Algorithm (SPEA2) linked to EPANET for trading-off pumping costs, water quality, and tanks sizing of water distribution systems is developed and demonstrated. The model integrates variable speed pumps for modeling the pumps operation, two water quality objectives (one based on chlorine disinfectant concentrations and one on water age), and tanks sizing cost which are assumed to vary with location and diameter. The water distribution system is subject to extended period simulations, variable energy tariffs, Kirchhoff's laws 1 and 2 for continuity of flow and pressure, tanks water level closure constraints, and storage-reliability requirements. EPANET Example 3 is employed for demonstrating the methodology on two multi-objective models, which differ in the imposed water quality objective (i.e., either with disinfectant or water age considerations). Three-fold Pareto optimal fronts are presented. Sensitivity analysis on the storage-reliability constraint, its influence on pumping cost, water quality, and tank sizing are explored. The contribution of this study is in tailoring design (tank sizing), pumps operational costs, water quality of two types, and reliability through residual storage requirements, in a single multi-objective framework. The model was found to be stable in generating multi-objective three-fold Pareto fronts, while producing explainable engineering outcomes. The model can be used as a decision tool for both pumps operation, water quality, required storage for reliability considerations, and tank sizing decision-making. Copyright © 2012 Elsevier Ltd. All rights reserved.

  13. Fuzzy droop control loops adjustment for stored energy balance in distributed energy storage system

    DEFF Research Database (Denmark)

    Aldana, Nelson Leonardo Diaz; Wu, Dan; Dragicevic, Tomislav

    2015-01-01

    system, in order to smooth the variations at the prime energy generator. In this paper, a decentralized strategy based on fuzzy logic is proposed in order to balance the state of charge of distributed energy storage systems in lowvoltage three phase AC microgrid. The proposed method weights the action...

  14. Monitored Retrievable Storage conceptual system study: metal storage casks

    International Nuclear Information System (INIS)

    Unterzuber, R.; Cross, T.E.; Krasicki, B.R.

    1983-08-01

    A description of the metal cask storage facility concept is presented with the operations required to handle the spent fuel or high-level wastes and transuranic wastes. A generic Receiving and Handling Facility, provided by PNL, has been used for this study. Modifications to the storage delivery side of the handling facility, necessary to couple the Receiving and Handling Facility with the storage facility, are described. The equipment and support facilities needed for the storage facility are also described. Two separate storage facilities are presented herein: one for all spent fuel storage, and one for storage of high-level waste (HLW) and transuranic waste (TRU). Each facility is described for the capacities and rates defined by PNL in the Concept Technical Performance Criteria and Base Assumptions (see Table 1.3-1). Estimates of costs and time-distributions of expenditures have been developed to construct, operate, and decommission the conceptual MRS facilities in mid-1983 dollars, for the base cases given using the cost categories and percentages provided by PNL. Cost estimates and time-distributions of expenditures have also been developed to expand the facility throughput rate from 1800 MTU to 3000 MTU, and to expand the facility storage capacity from 15,000 MTU to 72,00 MTU. The life cycle cost of the facility for the bounding cases of all spent fuel and all HLW and TRU, using the time-distributions of costs developed above and assuming a two percent per year discount rate, are also presented. 3 references, 16 figures, 18 tables

  15. Probabilistic Model for Listeria monocytogenes Growth during Distribution, Retail Storage, and Domestic Storage of Pasteurized Milk ▿

    Science.gov (United States)

    Koutsoumanis, Konstantinos; Pavlis, Athanasios; Nychas, George-John E.; Xanthiakos, Konstantinos

    2010-01-01

    A survey on the time-temperature conditions of pasteurized milk in Greece during transportation to retail, retail storage, and domestic storage and handling was performed. The data derived from the survey were described with appropriate probability distributions and introduced into a growth model of Listeria monocytogenes in pasteurized milk which was appropriately modified for taking into account strain variability. Based on the above components, a probabilistic model was applied to evaluate the growth of L. monocytogenes during the chill chain of pasteurized milk using a Monte Carlo simulation. The model predicted that, in 44.8% of the milk cartons released in the market, the pathogen will grow until the time of consumption. For these products the estimated mean total growth of L. monocytogenes during transportation, retail storage, and domestic storage was 0.93 log CFU, with 95th and 99th percentiles of 2.68 and 4.01 log CFU, respectively. Although based on EU regulation 2073/2005 pasteurized milk produced in Greece belongs to the category of products that do not allow the growth of L. monocytogenes due to a shelf life (defined by law) of 5 days, the above results show that this shelf life limit cannot prevent L. monocytogenes from growing under the current chill chain conditions. The predicted percentage of milk cartons—initially contaminated with 1 cell/1-liter carton—in which the pathogen exceeds the safety criterion of 100 cells/ml at the time of consumption was 0.14%. The probabilistic model was used for an importance analysis of the chill chain factors, using rank order correlation, while selected intervention and shelf life increase scenarios were evaluated. The results showed that simple interventions, such as excluding the door shelf from the domestic storage of pasteurized milk, can effectively reduce the growth of the pathogen. The door shelf was found to be the warmest position in domestic refrigerators, and it was most frequently used by the

  16. Parallel Wavefront Analysis for a 4D Interferometer

    Science.gov (United States)

    Rao, Shanti R.

    2011-01-01

    This software provides a programming interface for automating data collection with a PhaseCam interferometer from 4D Technology, and distributing the image-processing algorithm across a cluster of general-purpose computers. Multiple instances of 4Sight (4D Technology s proprietary software) run on a networked cluster of computers. Each connects to a single server (the controller) and waits for instructions. The controller directs the interferometer to several images, then assigns each image to a different computer for processing. When the image processing is finished, the server directs one of the computers to collate and combine the processed images, saving the resulting measurement in a file on a disk. The available software captures approximately 100 images and analyzes them immediately. This software separates the capture and analysis processes, so that analysis can be done at a different time and faster by running the algorithm in parallel across several processors. The PhaseCam family of interferometers can measure an optical system in milliseconds, but it takes many seconds to process the data so that it is usable. In characterizing an adaptive optics system, like the next generation of astronomical observatories, thousands of measurements are required, and the processing time quickly becomes excessive. A programming interface distributes data processing for a PhaseCam interferometer across a Windows computing cluster. A scriptable controller program coordinates data acquisition from the interferometer, storage on networked hard disks, and parallel processing. Idle time of the interferometer is minimized. This architecture is implemented in Python and JavaScript, and may be altered to fit a customer s needs.

  17. A program system for ab initio MO calculations on vector and parallel processing machines. Pt. 1

    International Nuclear Information System (INIS)

    Ernenwein, R.; Rohmer, M.M.; Benard, M.

    1990-01-01

    We present a program system for ab initio molecular orbital calculations on vector and parallel computers. The present article is devoted to the computation of one- and two-electron integrals over contracted Gaussian basis sets involving s-, p-, d- and f-type functions. The McMurchie and Davidson (MMD) algorithm has been implemented and parallelized by distributing over a limited number of logical tasks the calculation of the 55 relevant classes of integrals. All sections of the MMD algorithm have been efficiently vectorized, leading to a scalar/vector ratio of 5.8. Different algorithms are proposed and compared for an optimal vectorization of the contraction of the 'intermediate integrals' generated by the MMD formalism. Advantage is taken of the dynamic storage allocation for tuning the length of the vector loops (i.e. the size of the vectorization buffer) as a function of (i) the total memory available for the job, (ii) the number of logical tasks defined by the user (≤13), and (iii) the storage requested by each specific class of integrals. Test calculations carried out on a CRAY-2 computer show that the average number of finite integrals computed over a (s, p, d, f) CGTO basis set is about 1180000 per second and per processor. The combination of vectorization and parallelism on this 4-processor machine reduces the CPU time by a factor larger than 20 with respect to the scalar and sequential performance. (orig.)

  18. Parallel processing of genomics data

    Science.gov (United States)

    Agapito, Giuseppe; Guzzi, Pietro Hiram; Cannataro, Mario

    2016-10-01

    The availability of high-throughput experimental platforms for the analysis of biological samples, such as mass spectrometry, microarrays and Next Generation Sequencing, have made possible to analyze a whole genome in a single experiment. Such platforms produce an enormous volume of data per single experiment, thus the analysis of this enormous flow of data poses several challenges in term of data storage, preprocessing, and analysis. To face those issues, efficient, possibly parallel, bioinformatics software needs to be used to preprocess and analyze data, for instance to highlight genetic variation associated with complex diseases. In this paper we present a parallel algorithm for the parallel preprocessing and statistical analysis of genomics data, able to face high dimension of data and resulting in good response time. The proposed system is able to find statistically significant biological markers able to discriminate classes of patients that respond to drugs in different ways. Experiments performed on real and synthetic genomic datasets show good speed-up and scalability.

  19. Massively Parallel Polar Decomposition on Distributed-Memory Systems

    KAUST Repository

    Ltaief, Hatem

    2018-01-01

    We present a high-performance implementation of the Polar Decomposition (PD) on distributed-memory systems. Building upon on the QR-based Dynamically Weighted Halley (QDWH) algorithm, the key idea lies in finding the best rational approximation for the scalar sign function, which also corresponds to the polar factor for symmetric matrices, to further accelerate the QDWH convergence. Based on the Zolotarev rational functions—introduced by Zolotarev (ZOLO) in 1877— this new PD algorithm ZOLO-PD converges within two iterations even for ill-conditioned matrices, instead of the original six iterations needed for QDWH. ZOLO-PD uses the property of Zolotarev functions that optimality is maintained when two functions are composed in an appropriate manner. The resulting ZOLO-PD has a convergence rate up to seventeen, in contrast to the cubic convergence rate for QDWH. This comes at the price of higher arithmetic costs and memory footprint. These extra floating-point operations can, however, be processed in an embarrassingly parallel fashion. We demonstrate performance using up to 102, 400 cores on two supercomputers. We demonstrate that, in the presence of a large number of processing units, ZOLO-PD is able to outperform QDWH by up to 2.3X speedup, especially in situations where QDWH runs out of work, for instance, in the strong scaling mode of operation.

  20. High-speed parallel solution of the neutron diffusion equation with the hierarchical domain decomposition boundary element method incorporating parallel communications

    International Nuclear Information System (INIS)

    Tsuji, Masashi; Chiba, Gou

    2000-01-01

    A hierarchical domain decomposition boundary element method (HDD-BEM) for solving the multiregion neutron diffusion equation (NDE) has been fully parallelized, both for numerical computations and for data communications, to accomplish a high parallel efficiency on distributed memory message passing parallel computers. Data exchanges between node processors that are repeated during iteration processes of HDD-BEM are implemented, without any intervention of the host processor that was used to supervise parallel processing in the conventional parallelized HDD-BEM (P-HDD-BEM). Thus, the parallel processing can be executed with only cooperative operations of node processors. The communication overhead was even the dominant time consuming part in the conventional P-HDD-BEM, and the parallelization efficiency decreased steeply with the increase of the number of processors. With the parallel data communication, the efficiency is affected only by the number of boundary elements assigned to decomposed subregions, and the communication overhead can be drastically reduced. This feature can be particularly advantageous in the analysis of three-dimensional problems where a large number of processors are required. The proposed P-HDD-BEM offers a promising solution to the deterioration problem of parallel efficiency and opens a new path to parallel computations of NDEs on distributed memory message passing parallel computers. (author)

  1. Scaling up machine learning: parallel and distributed approaches

    National Research Council Canada - National Science Library

    Bekkerman, Ron; Bilenko, Mikhail; Langford, John

    2012-01-01

    .... Demand for parallelizing learning algorithms is highly task-specific: in some settings it is driven by the enormous dataset sizes, in others by model complexity or by real-time performance requirements...

  2. Building a parallel file system simulator

    International Nuclear Information System (INIS)

    Molina-Estolano, E; Maltzahn, C; Brandt, S A; Bent, J

    2009-01-01

    Parallel file systems are gaining in popularity in high-end computing centers as well as commercial data centers. High-end computing systems are expected to scale exponentially and to pose new challenges to their storage scalability in terms of cost and power. To address these challenges scientists and file system designers will need a thorough understanding of the design space of parallel file systems. Yet there exist few systematic studies of parallel file system behavior at petabyte- and exabyte scale. An important reason is the significant cost of getting access to large-scale hardware to test parallel file systems. To contribute to this understanding we are building a parallel file system simulator that can simulate parallel file systems at very large scale. Our goal is to simulate petabyte-scale parallel file systems on a small cluster or even a single machine in reasonable time and fidelity. With this simulator, file system experts will be able to tune existing file systems for specific workloads, scientists and file system deployment engineers will be able to better communicate workload requirements, file system designers and researchers will be able to try out design alternatives and innovations at scale, and instructors will be able to study very large-scale parallel file system behavior in the class room. In this paper we describe our approach and provide preliminary results that are encouraging both in terms of fidelity and simulation scalability.

  3. Hierarchical predictive control scheme for distributed energy storage integrated with residential demand and photovoltaic generation

    NARCIS (Netherlands)

    Lampropoulos, I.; Garoufalis, P.; van den Bosch, P.P.J.; Kling, W.L.

    2015-01-01

    A hierarchical control scheme is defined for the energy management of a battery energy storage system which is integrated in a low-voltage distribution grid with residential customers and photovoltaic installations. The scope is the economic optimisation of the integrated system by employing

  4. On-line event reconstruction using a parallel in-memory data base

    OpenAIRE

    Argante, E; Van der Stok, P D V; Willers, Ian Malcolm

    1995-01-01

    PORS is a system designed for on-line event reconstruction in high energy physics (HEP) experiments. It uses the CPREAD reconstruction program. Central to the system is a parallel in-memory database which is used as communication medium between parallel workers. A farming control structure is implemented with PORS in a natural way. The database provides structured storage of data with a short life time. PORS serves as a case study for the construction of a methodology on how to apply parallel...

  5. Out of core, out of mind: Practical parallel I/O

    Energy Technology Data Exchange (ETDEWEB)

    Womble, D.E.; Greenberg, D.S.; Riesen, R.E.; Wheat, S.R.

    1993-11-01

    Parallel computers are becoming more powerful and more complex in response to the demand for computing power by scientists and engineers. Inevitably, new and more complex I/O systems will be developed for these systems. In particular we believe that the I/O system must provide the programmer with the ability to explcitly manage storage (despite the trend toward complex parallel file systems and caching schemes). One method of doing so is to have a partitioned secondary storage in which each processor owns a logical disk. Along with operating system enhancements which allow overheads such as buffer copying to be avoided and libraries to support optimal remapping of data, this sort of I/O system meets the needs of high performance computing.

  6. [Storages and distributed patterns of soil organic carbon and total nitrogen during the succession of artificial sand-binding vegetation in arid desert ecosystem].

    Science.gov (United States)

    Jia, Xiao-Hong; Li, Xin-Rong; Zhou, Yu-Yan; Li, Yuan-Shou

    2012-03-01

    Soil carbon pool acts as the largest one of carbon pools in the terrestrial ecosystem. The storages and distributed patterns of soil organic carbon (SOC) and total nitrogen (TN) evaluated accurately are helpful to predict the feedback between the terrestrial ecosystem and climate changes. Based on the data about bulk density, content of SOC and TN at 0-100 cm soil profile, the density of SOC and TN at the temporal (chronosequence of artificial vegetation) and spatial (vertical) distributed patterns have been estimated. The results indicated that storages of SOC and TN at 0-100 cm depth increased with the chronosequence of artificial vegetation. The storages of SOC and TN showed the same tendency with the succession time of artificial vegetation. Storages of SOC and TN significantly increased at the early stage of banding sand by artificially vegetation ( 25 a). The variation of storages mainly occurred in the 0-20 cm depth. The storages decreased with the soil vertical depth. At the early stage of banding sand, increase in storage included every depth (0-100 cm). Whereas, at the later stage, increase in storage at 0-20 cm depth was main, and increase in the 20-100 cm was inconspicuous. The accumulation of storage at the shallow soil depth was more notability with the succession of artificial vegetation. The distributed pattern of storage in SOC and TN has been confirmed in arid desert regions below 200 mm annual precipitation. This was beneficial to understand the carbon cycle and to predict the feedback relationship between desert ecosystem and climate changes.

  7. Secure Storage Architectures

    Energy Technology Data Exchange (ETDEWEB)

    Aderholdt, Ferrol [Tennessee Technological University; Caldwell, Blake A [ORNL; Hicks, Susan Elaine [ORNL; Koch, Scott M [ORNL; Naughton, III, Thomas J [ORNL; Pogge, James R [Tennessee Technological University; Scott, Stephen L [Tennessee Technological University; Shipman, Galen M [ORNL; Sorrillo, Lawrence [ORNL

    2015-01-01

    The purpose of this report is to clarify the challenges associated with storage for secure enclaves. The major focus areas for the report are: - review of relevant parallel filesystem technologies to identify assets and gaps; - review of filesystem isolation/protection mechanisms, to include native filesystem capabilities and auxiliary/layered techniques; - definition of storage architectures that can be used for customizable compute enclaves (i.e., clarification of use-cases that must be supported for shared storage scenarios); - investigate vendor products related to secure storage. This study provides technical details on the storage and filesystem used for HPC with particular attention on elements that contribute to creating secure storage. We outline the pieces for a a shared storage architecture that balances protection and performance by leveraging the isolation capabilities available in filesystems and virtualization technologies to maintain the integrity of the data. Key Points: There are a few existing and in-progress protection features in Lustre related to secure storage, which are discussed in (Chapter 3.1). These include authentication capabilities like GSSAPI/Kerberos and the in-progress work for GSSAPI/Host-keys. The GPFS filesystem provides native support for encryption, which is not directly available in Lustre. Additionally, GPFS includes authentication/authorization mechanisms for inter-cluster sharing of filesystems (Chapter 3.2). The limitations of key importance for secure storage/filesystems are: (i) restricting sub-tree mounts for parallel filesystem (which is not directly supported in Lustre or GPFS), and (ii) segregation of hosts on the storage network and practical complications with dynamic additions to the storage network, e.g., LNET. A challenge for VM based use cases will be to provide efficient IO forwarding of the parallel filessytem from the host to the guest (VM). There are promising options like para-virtualized filesystems to

  8. Federating Distributed Storage For Clouds In ATLAS

    CERN Document Server

    Berghaus, Frank; The ATLAS collaboration

    2017-01-01

    Input data for applications that run in cloud computing centres can be stored at distant repositories, often with multiple copies of the popular data stored at many sites. Locating and retrieving the remote data can be challenging, and we believe that federating the storage can address this problem. A federation would locate the closest copy of the data currently on the basis of GeoIP information. Currently we are using the DynaFed data federation software solution developed by CERN IT. DynaFed supports several industry standards for connection protocols like Amazon's S3, Microsofts Azure, as well as WebDav and HTTP. Protocol dependent authentication is hidden from the user by using their X509 certificate. We have setup an instance of DynaFed and integrated it into the ATLAS Data Distribution Management system. We report on the challenges faced during the installation and integration. We have tested ATLAS analysis jobs submitted by the PanDA production system and we report on our first experiences with its op...

  9. Welfare effects of unbundling gas storage and distribution

    International Nuclear Information System (INIS)

    Breton, M.; Montreal Univ., PQ; Kharbach, M.

    2006-01-01

    The creation of inventories and stockpiles can help to reduce price and production fluctuations. This paper presented the results of a simulation of market architectures using a 2 period model. The aim of the paper was to provide insights on the merits of a gas unbundling policy recently adopted in many gas markets. In terms of market architecture, it was first assumed that one of the gas firms owned the storage facility and was mandated to give a second firm access to it. It was then assumed that an independent third firm was responsible for the storage activity so that the other 2 firms competed in the downstream gas market and bought storage services from the independent firm. High and low price periods in a single year were considered. The first architecture assumed an Open Access framework which introduced a Stackelberg competition component in the downstream market through the storage participation in the final goods offering. The second architecture assumed an independent storage activity, and a Cournot component was present. Seasonal storage facilities were filled during the low price period and emptied during the high price period. Results of the simulation indicated that total welfare and consumer welfare were maximized in the case of the integrated firm owning the storage facilities and operating in the downstream market. Success was attributed to the Open Access framework and the Stackelberg competition component in the downstream market. The bundled architecture led to higher consumer surplus than the unbundled architecture. It was concluded that regulatory reforms in North American and European gas sectors that foster separating storage and merchant activities can not be justified based on welfare arguments. 12 refs., 1 tab

  10. Vitamin D3 in Pigs: Distribution, Storage and Turnover under Various Input Conditions

    DEFF Research Database (Denmark)

    Burild, Anders

    Vitamin D3 is important for the mineralization of the skeleton to prevent the deficiency diseases rickets and osteoporosis, and to maintain a healthy skeleton throughout life. Vitamin D3 is synthesized in the skin after exposure to the sun. Due to the low angle of the sun during wintertime at high...... latitudes, no or only a negligible amount of vitamin D3 is synthesized and the body needs to rely on its storages of vitamin D3, or dietary vitamin D3 in the form of vitamin D3 and 25-hydroxyvitamin D3. The information of the size of the storages of vitamin D3 in humans is sparse, but very low levels...... of vitamin D3 is found in tissues from animals fed physiologically relevant doses of vitamin D3. The natural synthesis of vitamin D3 might, however, influence on the storages of vitamin D3. The different inherent properties of the two forms of vitamin D3 might also affect the tissue distribution of vitamin D...

  11. Running ATLAS workloads within massively parallel distributed applications using Athena Multi-Process framework (AthenaMP)

    CERN Document Server

    Calafiura, Paolo; The ATLAS collaboration; Seuster, Rolf; Tsulaia, Vakhtang; van Gemmeren, Peter

    2015-01-01

    AthenaMP is a multi-process version of the ATLAS reconstruction and data analysis framework Athena. By leveraging Linux fork and copy-on-write, it allows the sharing of memory pages between event processors running on the same compute node with little to no change in the application code. Originally targeted to optimize the memory footprint of reconstruction jobs, AthenaMP has demonstrated that it can reduce the memory usage of certain confugurations of ATLAS production jobs by a factor of 2. AthenaMP has also evolved to become the parallel event-processing core of the recently developed ATLAS infrastructure for fine-grained event processing (Event Service) which allows to run AthenaMP inside massively parallel distributed applications on hundreds of compute nodes simultaneously. We present the architecture of AthenaMP, various strategies implemented by AthenaMP for scheduling workload to worker processes (for example: Shared Event Queue and Shared Distributor of Event Tokens) and the usage of AthenaMP in the...

  12. Running ATLAS workloads within massively parallel distributed applications using Athena Multi-Process framework (AthenaMP)

    CERN Document Server

    Calafiura, Paolo; Seuster, Rolf; Tsulaia, Vakhtang; van Gemmeren, Peter

    2015-01-01

    AthenaMP is a multi-process version of the ATLAS reconstruction, simulation and data analysis framework Athena. By leveraging Linux fork and copy-on-write, it allows for sharing of memory pages between event processors running on the same compute node with little to no change in the application code. Originally targeted to optimize the memory footprint of reconstruction jobs, AthenaMP has demonstrated that it can reduce the memory usage of certain configurations of ATLAS production jobs by a factor of 2. AthenaMP has also evolved to become the parallel event-processing core of the recently developed ATLAS infrastructure for fine-grained event processing (Event Service) which allows to run AthenaMP inside massively parallel distributed applications on hundreds of compute nodes simultaneously. We present the architecture of AthenaMP, various strategies implemented by AthenaMP for scheduling workload to worker processes (for example: Shared Event Queue and Shared Distributor of Event Tokens) and the usage of Ath...

  13. The numerical parallel computing of photon transport

    International Nuclear Information System (INIS)

    Huang Qingnan; Liang Xiaoguang; Zhang Lifa

    1998-12-01

    The parallel computing of photon transport is investigated, the parallel algorithm and the parallelization of programs on parallel computers both with shared memory and with distributed memory are discussed. By analyzing the inherent law of the mathematics and physics model of photon transport according to the structure feature of parallel computers, using the strategy of 'to divide and conquer', adjusting the algorithm structure of the program, dissolving the data relationship, finding parallel liable ingredients and creating large grain parallel subtasks, the sequential computing of photon transport into is efficiently transformed into parallel and vector computing. The program was run on various HP parallel computers such as the HY-1 (PVP), the Challenge (SMP) and the YH-3 (MPP) and very good parallel speedup has been gotten

  14. Energy storage management system with distributed wireless sensors

    Science.gov (United States)

    Farmer, Joseph C.; Bandhauer, Todd M.

    2015-12-08

    An energy storage system having a multiple different types of energy storage and conversion devices. Each device is equipped with one or more sensors and RFID tags to communicate sensor information wirelessly to a central electronic management system, which is used to control the operation of each device. Each device can have multiple RFID tags and sensor types. Several energy storage and conversion devices can be combined.

  15. Adding Data Management Services to Parallel File Systems

    Energy Technology Data Exchange (ETDEWEB)

    Brandt, Scott [Univ. of California, Santa Cruz, CA (United States)

    2015-03-04

    -based ecosystem; (3) common optimizations, e.g., indexing and caching, are readily supported across several file formats, avoiding effort duplication; and (4) performance improves significantly, as data processing is integrated more tightly with data storage. Our key contributions are: SciHadoop which explores changes to MapReduce assumption by taking advantage of semantics of structured data while preserving MapReduce’s failure and resource management; DataMods which extends common abstractions of parallel file systems so they become programmable such that they can be extended to natively support a variety of data models and can be hooked into emerging distributed runtimes such as Stanford’s Legion; and Miso which combines Hadoop and relational data warehousing to minimize time to insight, taking into account the overhead of ingesting data into data warehousing.

  16. Increase in the number of distributed power generation installations in electricity distribution grids - Storage technologies; Zunahme der dezentralen Energieerzeugungsanlagen in elektrischen Verteilnetzen: Grundlagen der Speicher

    Energy Technology Data Exchange (ETDEWEB)

    Luechinger, P.

    2003-07-01

    This is the fifth part of a ten-part final report for the Swiss Federal Office of Energy (SFOE) on a project that looked into potential problems relating to the Swiss electricity distribution grid with respect to the increasing number of distributed power generation facilities being put into service. The identification of special conditions for the grid's operation and future development that take increasing decentralised power production into account are discussed. The results of the project activities encompass the analysis and evaluation of various problem areas associated with planning and management of the grid during normal operation and periods of malfunction, as well as required modifications to safety systems and grid configurations. This fourth appendix to the main report describes six ways of storing electricity, including accumulators, super caps, super-conducting magnetic and flywheel energy storage units. The accumulator technologies discussed include lead-acid, nickel-cadmium and sodium-sulphur batteries. Each of these types of power storage technologies is briefly described. The characteristics of these various types of storage are compared.

  17. Spatial distribution and variability of carbon storage in different sympodial bamboo species in China.

    Science.gov (United States)

    Teng, Jiangnan; Xiang, Tingting; Huang, Zhangting; Wu, Jiasen; Jiang, Peikun; Meng, Cifu; Li, Yongfu; Fuhrmann, Jeffry J

    2016-03-01

    Selection of tree species is potentially an important management decision for increasing carbon storage in forest ecosystems. This study investigated and compared spatial distribution and variability of carbon storage in 8 sympodial bamboo species in China. The results of this study showed that average carbon densities (CDs) in the different organs decreased in the order: culms (0.4754 g g(-1)) > below-ground (0.4701 g g(-1)) > branches (0.4662 g g(-1)) > leaves (0.4420 g g(-1)). Spatial distribution of carbon storage (CS) on an area basis in the biomass of 8 sympodial bamboo species was in the order: culms (17.4-77.1%) > below-ground (10.6-71.7%) > branches (3.8-11.6%) > leaves (0.9-5.1%). Total CSs in the sympodial bamboo ecosystems ranged from 103.6 Mg C ha(-1) in Bambusa textilis McClure stand to 194.2 Mg C ha(-1) in Dendrocalamus giganteus Munro stand. Spatial distribution of CSs in 8 sympodial bamboo ecosystems decreased in the order: soil (68.0-83.5%) > vegetation (16.8-31.1%) > litter (0.3-1.7%). Total current CS and biomass carbon sequestration rate in the sympodial bamboo stands studied in China is 93.184 × 10(6) Mg C ha(-1) and 8.573 × 10(6) Mg C yr(-1), respectively. The sympodial bamboos had a greater CSs and higher carbon sequestration rates relative to other bamboo species. Sympodial bamboos can play an important role in improving climate and economy in the widely cultivated areas of the world. Copyright © 2015 Elsevier Ltd. All rights reserved.

  18. Parallelization characteristics of the DeCART code

    International Nuclear Information System (INIS)

    Cho, J. Y.; Joo, H. G.; Kim, H. Y.; Lee, C. C.; Chang, M. H.; Zee, S. Q.

    2003-12-01

    This report is to describe the parallelization characteristics of the DeCART code and also examine its parallel performance. Parallel computing algorithms are implemented to DeCART to reduce the tremendous computational burden and memory requirement involved in the three-dimensional whole core transport calculation. In the parallelization of the DeCART code, the axial domain decomposition is first realized by using MPI (Message Passing Interface), and then the azimuthal angle domain decomposition by using either MPI or OpenMP. When using the MPI for both the axial and the angle domain decomposition, the concept of MPI grouping is employed for convenient communication in each communication world. For the parallel computation, most of all the computing modules except for the thermal hydraulic module are parallelized. These parallelized computing modules include the MOC ray tracing, CMFD, NEM, region-wise cross section preparation and cell homogenization modules. For the distributed allocation, most of all the MOC and CMFD/NEM variables are allocated only for the assigned planes, which reduces the required memory by a ratio of the number of the assigned planes to the number of all planes. The parallel performance of the DeCART code is evaluated by solving two problems, a rodded variation of the C5G7 MOX three-dimensional benchmark problem and a simplified three-dimensional SMART PWR core problem. In the aspect of parallel performance, the DeCART code shows a good speedup of about 40.1 and 22.4 in the ray tracing module and about 37.3 and 20.2 in the total computing time when using 48 CPUs on the IBM Regatta and 24 CPUs on the LINUX cluster, respectively. In the comparison between the MPI and OpenMP, OpenMP shows a somewhat better performance than MPI. Therefore, it is concluded that the first priority in the parallel computation of the DeCART code is in the axial domain decomposition by using MPI, and then in the angular domain using OpenMP, and finally the angular

  19. Applying wind turbines and battery storage to defer Orcas Power and Light Company distribution circuit upgrades

    International Nuclear Information System (INIS)

    Zaininger, H.W.; Barnes, P.R.

    1997-03-01

    The purpose of this study is to conduct a detailed assessment of the Orcas Power and Light Company (OPALCO) system to determine the potential for deferring the costly upgrade of the 25-kV Lopez- Eastsound circuit, by the application of a MW-scale wind farm and battery storage facilities as appropriate. Local wind resource data has been collected over the past year and used to determine MW-scale wind farm performance. This hourly wind farm performance data is used with measured hourly Eastsound load data, and recent OPALCO distribution system expansion plans and cost projections in performing this detailed benefit-cost assessment. The OPALCO distribution circuit expansion project and assumptions are described. MW-scale wind farm performance results are given. The economic benefit-cost results for the wind farm and battery storage applications on the OPALCO system using OPALCO system design criteria and cost assumptions are reported. A recalculation is presented of the benefit-cost results for similar potential wind farm and battery storage applications on other utility systems with higher marginal energy and demand costs. Conclusions and recommendations are presented. costs. Conclusions and recommendations are presented

  20. Parallelization for first principles electronic state calculation program

    International Nuclear Information System (INIS)

    Watanabe, Hiroshi; Oguchi, Tamio.

    1997-03-01

    In this report we study the parallelization for First principles electronic state calculation program. The target machines are NEC SX-4 for shared memory type parallelization and FUJITSU VPP300 for distributed memory type parallelization. The features of each parallel machine are surveyed, and the parallelization methods suitable for each are proposed. It is shown that 1.60 times acceleration is achieved with 2 CPU parallelization by SX-4 and 4.97 times acceleration is achieved with 12 PE parallelization by VPP 300. (author)

  1. Parallel Implicit Algorithms for CFD

    Science.gov (United States)

    Keyes, David E.

    1998-01-01

    The main goal of this project was efficient distributed parallel and workstation cluster implementations of Newton-Krylov-Schwarz (NKS) solvers for implicit Computational Fluid Dynamics (CFD.) "Newton" refers to a quadratically convergent nonlinear iteration using gradient information based on the true residual, "Krylov" to an inner linear iteration that accesses the Jacobian matrix only through highly parallelizable sparse matrix-vector products, and "Schwarz" to a domain decomposition form of preconditioning the inner Krylov iterations with primarily neighbor-only exchange of data between the processors. Prior experience has established that Newton-Krylov methods are competitive solvers in the CFD context and that Krylov-Schwarz methods port well to distributed memory computers. The combination of the techniques into Newton-Krylov-Schwarz was implemented on 2D and 3D unstructured Euler codes on the parallel testbeds that used to be at LaRC and on several other parallel computers operated by other agencies or made available by the vendors. Early implementations were made directly in Massively Parallel Integration (MPI) with parallel solvers we adapted from legacy NASA codes and enhanced for full NKS functionality. Later implementations were made in the framework of the PETSC library from Argonne National Laboratory, which now includes pseudo-transient continuation Newton-Krylov-Schwarz solver capability (as a result of demands we made upon PETSC during our early porting experiences). A secondary project pursued with funding from this contract was parallel implicit solvers in acoustics, specifically in the Helmholtz formulation. A 2D acoustic inverse problem has been solved in parallel within the PETSC framework.

  2. Parallel log structured file system collective buffering to achieve a compact representation of scientific and/or dimensional data

    Science.gov (United States)

    Grider, Gary A.; Poole, Stephen W.

    2015-09-01

    Collective buffering and data pattern solutions are provided for storage, retrieval, and/or analysis of data in a collective parallel processing environment. For example, a method can be provided for data storage in a collective parallel processing environment. The method comprises receiving data to be written for a plurality of collective processes within a collective parallel processing environment, extracting a data pattern for the data to be written for the plurality of collective processes, generating a representation describing the data pattern, and saving the data and the representation.

  3. Distributed Energy Resources On-Site Optimization for Commercial Buildings with Electric and Thermal Storage Technologies

    International Nuclear Information System (INIS)

    Lacommare, Kristina S H; Stadler, Michael; Aki, Hirohisa; Firestone, Ryan; Lai, Judy; Marnay, Chris; Siddiqui, Afzal

    2008-01-01

    The addition of storage technologies such as flow batteries, conventional batteries, and heat storage can improve the economic as well as environmental attractiveness of on-site generation (e.g., PV, fuel cells, reciprocating engines or microturbines operating with or without CHP) and contribute to enhanced demand response. In order to examine the impact of storage technologies on demand response and carbon emissions, a microgrid's distributed energy resources (DER) adoption problem is formulated as a mixed-integer linear program that has the minimization of annual energy costs as its objective function. By implementing this approach in the General Algebraic Modeling System (GAMS), the problem is solved for a given test year at representative customer sites, such as schools and nursing homes, to obtain not only the level of technology investment, but also the optimal hourly operating schedules. This paper focuses on analysis of storage technologies in DER optimization on a building level, with example applications for commercial buildings. Preliminary analysis indicates that storage technologies respond effectively to time-varying electricity prices, i.e., by charging batteries during periods of low electricity prices and discharging them during peak hours. The results also indicate that storage technologies significantly alter the residual load profile, which can contribute to lower carbon emissions depending on the test site, its load profile, and its adopted DER technologies

  4. Optimization of Comb-Drive Actuators [Nanopositioners for probe-based data storage and musical MEMS

    NARCIS (Netherlands)

    Engelen, Johannes Bernardus Charles

    2011-01-01

    The era of infinite storage seems near. To reach it, data storage capabilities need to grow, and new storage technologies must be developed.This thesis studies one aspect of one of the emergent storage technologies: optimizing electrostatic combdrive actuation for a parallel probe-based data storage

  5. Parallel computing of physical maps--a comparative study in SIMD and MIMD parallelism.

    Science.gov (United States)

    Bhandarkar, S M; Chirravuri, S; Arnold, J

    1996-01-01

    Ordering clones from a genomic library into physical maps of whole chromosomes presents a central computational problem in genetics. Chromosome reconstruction via clone ordering is usually isomorphic to the NP-complete Optimal Linear Arrangement problem. Parallel SIMD and MIMD algorithms for simulated annealing based on Markov chain distribution are proposed and applied to the problem of chromosome reconstruction via clone ordering. Perturbation methods and problem-specific annealing heuristics are proposed and described. The SIMD algorithms are implemented on a 2048 processor MasPar MP-2 system which is an SIMD 2-D toroidal mesh architecture whereas the MIMD algorithms are implemented on an 8 processor Intel iPSC/860 which is an MIMD hypercube architecture. A comparative analysis of the various SIMD and MIMD algorithms is presented in which the convergence, speedup, and scalability characteristics of the various algorithms are analyzed and discussed. On a fine-grained, massively parallel SIMD architecture with a low synchronization overhead such as the MasPar MP-2, a parallel simulated annealing algorithm based on multiple periodically interacting searches performs the best. For a coarse-grained MIMD architecture with high synchronization overhead such as the Intel iPSC/860, a parallel simulated annealing algorithm based on multiple independent searches yields the best results. In either case, distribution of clonal data across multiple processors is shown to exacerbate the tendency of the parallel simulated annealing algorithm to get trapped in a local optimum.

  6. Static and Dynamic Stability Analysis of Distributed Energy Resources Components with Storage Devices and Loads for Smart Grids

    DEFF Research Database (Denmark)

    Mihet-Popa, Lucian; Groza, V.

    2011-01-01

    of the Smart Grids (SGs). A SG can operate interconnected to the main distribution grid or in islanded mode. This paper presents experimental tests for static and dynamic stability analysis carried out in a dedicated laboratory for research in distributed control and smart grid with a high share of renewable......The distributed energy resources (DER) contains several technologies, such as diesel engines, small wind turbines, photovoltaic inverters, etc. The control of DER components with storage devices and (controllable) loads, such as batteries, capacitors, dump loads, are central to the concept...... energy production. Moreover to point out, on a laboratory scale, the coupling between DR and storage and to effectively compensate wind fluctuations a number of tests have been done. In order to find out the parameters of various types of DER components for dynamic simulation models a number of tests...

  7. Energy Storage Characteristic Analysis of Voltage Sags Compensation for UPQC Based on MMC for Medium Voltage Distribution System

    Directory of Open Access Journals (Sweden)

    Yongchun Yang

    2018-04-01

    Full Text Available The modular multilevel converter (MMC, as a new type of voltage source converter, is increasingly used because it is a distributed storage system. There are many advantages of using the topological structure of the MMC on a unified power quality controller (UPQC, and voltage sag mitigation is an important use of the MMC energy storage system for the power quality compensation process. In this paper, based on the analysis of the topology of the MMC, the essence of energy conversion in a UPQC of voltage sag compensation is analyzed; then, the energy storage characteristics are calculated and analyzed to determine the performance index of voltage sag compensation; in addition, the simulation method is used to verify the voltage sag compensation characteristics of the UPQC; finally, an industrial prototype of the UPQC based on an MMC for 10 kV of medium voltage distribution network has been developed, and the basic functions of UPQC have been tested.

  8. Distributed energy systems with wind power and energy storage

    Energy Technology Data Exchange (ETDEWEB)

    Korpaas, Magnus

    2004-07-01

    The topic of this thesis is the study of energy storage systems operating with wind power plants. The motivation for applying energy storage in this context is that wind power generation is intermittent and generally difficult to predict, and that good wind energy resources are often found in areas with limited grid capacity. Moreover, energy storage in the form of hydrogen makes it possible to provide clean fuel for transportation. The aim of this work has been to evaluate how local energy storage systems should be designed and operated in order to increase the penetration and value of wind power in the power system. Optimization models and sequential and probabilistic simulation models have been developed for this purpose. Chapter 3 presents a sequential simulation model of a general wind hydrogen energy system. Electrolytic hydrogen is used either as a fuel for transportation or for power generation in a stationary fuel cell. The model is useful for evaluating how hydrogen storage can increase the penetration of wind power in areas with limited or no transmission capacity to the main grid. The simulation model is combined with a cost model in order to study how component sizing and choice of operation strategy influence the performance and economics of the wind-hydrogen system. If the stored hydrogen is not used as a separate product, but merely as electrical energy storage, it should be evaluated against other and more energy efficient storage options such as pumped hydro and redox flow cells. A probabilistic model of a grid-connected wind power plant with a general energy storage unit is presented in chapter 4. The energy storage unit is applied for smoothing wind power fluctuations by providing a firm power output to the grid over a specific period. The method described in the chapter is based on the statistical properties of the wind speed and a general representation of the wind energy conversion system and the energy storage unit. This method allows us to

  9. Hierarchical storage of large volume of multidector CT data using distributed servers

    Science.gov (United States)

    Ratib, Osman; Rosset, Antoine; Heuberger, Joris; Bandon, David

    2006-03-01

    Multidector scanners and hybrid multimodality scanners have the ability to generate large number of high-resolution images resulting in very large data sets. In most cases, these datasets are generated for the sole purpose of generating secondary processed images and 3D rendered images as well as oblique and curved multiplanar reformatted images. It is therefore not essential to archive the original images after they have been processed. We have developed an architecture of distributed archive servers for temporary storage of large image datasets for 3D rendering and image processing without the need for long term storage in PACS archive. With the relatively low cost of storage devices it is possible to configure these servers to hold several months or even years of data, long enough for allowing subsequent re-processing if required by specific clinical situations. We tested the latest generation of RAID servers provided by Apple computers with a capacity of 5 TBytes. We implemented a peer-to-peer data access software based on our Open-Source image management software called OsiriX, allowing remote workstations to directly access DICOM image files located on the server through a new technology called "bonjour". This architecture offers a seamless integration of multiple servers and workstations without the need for central database or complex workflow management tools. It allows efficient access to image data from multiple workstation for image analysis and visualization without the need for image data transfer. It provides a convenient alternative to centralized PACS architecture while avoiding complex and time-consuming data transfer and storage.

  10. The effect of the flow direction inside the header on two-phase flow distribution in parallel vertical channels

    International Nuclear Information System (INIS)

    Marchitto, A.; Fossa, M.; Guglielmini, G.

    2012-01-01

    Uniform fluid distribution is essential for efficient operation of chemical-processing equipment such as contactors, reactors, mixers, burners and in most refrigeration equipment, where two phases are acting together. To obtain optimum distribution, proper consideration must be given to flow behaviour in the distributor, flow conditions upstream and downstream of the distributor, and the distribution requirements (fluid or phase) of the equipment. Even though the principles of single phase distribution have been well developed for more than three decades, they are frequently not taken in the right account by equipment designers when a mixture is present, and a significant fraction of process equipment consequently suffers from maldistribution. The experimental investigation presented in this paper is aimed at understanding the main mechanisms which drive the flow distribution inside a two-phase horizontal header in order to design improved distributors and to optimise the flow distribution inside compact heat exchanger. Experimentation was devoted to establish the influence of the inlet conditions and of the channel/distributor geometry on the phase/mass distribution into parallel vertical channels. The study is carried out with air–water mixtures and it is based on the measurement of component flow rates in individual channels and on pressure drops across the distributor. The effects of the operating conditions, the header geometry and the inlet port nozzle were investigated in the ranges of liquid and gas superficial velocities of 0.2–1.2 and 1.5–16.5 m/s, respectively. In order to control the main flow direction inside the header, different fitting devices were tested; the insertion of a co-axial, multi-hole distributor inside the header has confirmed the possibility of greatly improving the liquid and gas flow distribution by the proper selection of position, diameter and number of the flow openings between the supplying distributor and the system of

  11. Transactive-Market-Based Operation of Distributed Electrical Energy Storage with Grid Constraints

    Directory of Open Access Journals (Sweden)

    M. Nazif Faqiry

    2017-11-01

    Full Text Available In a transactive energy market, distributed energy resources (DERs such as dispatchable distributed generators (DGs, electrical energy storages (EESs, distribution-scale load aggregators (LAs, and renewable energy sources (RESs have to earn their share of supply or demand through a bidding process. In such a market, the distribution system operator (DSO may optimally schedule these resources, first in a forward market, i.e., day-ahead, and in a real-time market later on, while maintaining a reliable and economic distribution grid. In this paper, an efficient day-ahead scheduling of these resources, in the presence of interaction with wholesale market at the locational marginal price (LMP, is studied. Due to inclusion of EES units with integer constraints, a detailed mixed integer linear programming (MILP formulation that incorporates simplified DistFlow equations to account for grid constraints is proposed. Convex quadratic line and transformer apparent power flow constraints have been linearized using an outer approximation. The proposed model schedules DERs based on distribution locational marginal price (DLMP, which is obtained as the Lagrange multiplier of the real power balance constraint at each distribution bus while maintaining physical grid constraints such as line limits, transformer limits, and bus voltage magnitudes. Case studies are performed on a modified IEEE 13-bus system with high DER penetration. Simulation results show the validity and efficiency of the proposed model.

  12. Distributed Renewable Generation and Storage System Sizing Based on Smart Dispatch of Microgrids

    Directory of Open Access Journals (Sweden)

    Raji Atia

    2016-03-01

    Full Text Available This paper considers the contribution of independent owners (IOs operating within microgrids (MGs toward green power generation in deregulated energy markets. An optimization scheme is introduced for sizing distributed renewable generation (DRG and a distributed energy storage system (DESS based on a novel energy management system (EMS that accounts for demand response (DR, DESS dispatch and performance degradation, dynamic pricing environments, power distribution loss and irregular renewable generation. The proposed EMS utilizes an iterative Newton-Raphson linear programming algorithm that schedules resources in order to minimize the objective function, to deal with the complicated nonlinear nature of the problem and to enable efficient long-term assessments. The EMS is used to evaluate candidate solutions that are generated by a genetic algorithm (GA to determine the optimal combination of DRG and DESS. A case study for IEEE 34-bus distribution MG in Okinawa, Japan, is used for testing the algorithm and analyzing the potential for IO/MG investments and their strategies.

  13. The simplified spherical harmonics (SP{sub L}) methodology with space and moment decomposition in parallel environments

    Energy Technology Data Exchange (ETDEWEB)

    Gianluca, Longoni; Alireza, Haghighat [Florida University, Nuclear and Radiological Engineering Department, Gainesville, FL (United States)

    2003-07-01

    In recent years, the SP{sub L} (simplified spherical harmonics) equations have received renewed interest for the simulation of nuclear systems. We have derived the SP{sub L} equations starting from the even-parity form of the S{sub N} equations. The SP{sub L} equations form a system of (L+1)/2 second order partial differential equations that can be solved with standard iterative techniques such as the Conjugate Gradient (CG). We discretized the SP{sub L} equations with the finite-volume approach in a 3-D Cartesian space. We developed a new 3-D general code, Pensp{sub L} (Parallel Environment Neutral-particle SP{sub L}). Pensp{sub L} solves both fixed source and criticality eigenvalue problems. In order to optimize the memory management, we implemented a Compressed Diagonal Storage (CDS) to store the SP{sub L} matrices. Pensp{sub L} includes parallel algorithms for space and moment domain decomposition. The computational load is distributed on different processors, using a mapping function, which maps the 3-D Cartesian space and moments onto processors. The code is written in Fortran 90 using the Message Passing Interface (MPI) libraries for the parallel implementation of the algorithm. The code has been tested on the Pcpen cluster and the parallel performance has been assessed in terms of speed-up and parallel efficiency. (author)

  14. Integrating renewables in distribution grids. Storage, regulation and the interaction of different stakeholders in future grids

    Energy Technology Data Exchange (ETDEWEB)

    Nykamp, S.

    2013-10-18

    In recent years, the transition of the power supply chain towards a sustainable system based on 'green' electricity generation out of renewable energy sources (RES-E) has become a main challenge for grid operators and further stakeholders in the power system. To enable the evaluation of new concepts for the integration of RES-E, first the feed-in characteristics of photovoltaic, wind and biomass generators located in a distribution grid area and based on numerous measured feed-in data are studied in this thesis. The achieved insights from the feed-in profiles can be used for the dimensioning of grid assets. Furthermore, the results are useful for the evaluation of congestion management or for the dimensioning of storage assets in distribution grids. The latter aspect is analyzed in detail such that suitable storage characteristics for an introduction in the grid are determined. An economic approach is presented to derive break-even points for storage assets as a substitute to conventional reinforcements. For a case study from a real world low voltage grid with reinforcement needs, these break-even points are determined and the main influencing parameters are evaluated. A further important question in this context concerns the role DSOs (distribution system operators) may play with the operation of decentralized storage assets since several stakeholders may be interested in using the flexibility provided by these assets. This unclear responsibility also applies to the steering of adjustable consumption devices such as electric heat pumps or electric cars. For decentralized storage assets as well as heat pump appliances, optimal operation modes based on the optimization objectives for a DSO and a trader are derived. It is shown based on real world data that choosing a 'copperplate' scenario is not only technically insufficient for a global balance of the consumption and generation. It may even be harmful for the society from a welfare economic

  15. Optimal placement, sizing, and daily charge/discharge of battery energy storage in low voltage distribution network with high photovoltaic penetration

    DEFF Research Database (Denmark)

    Jannesar, Mohammad Rasol; Sedighi, Alireza; Savaghebi, Mehdi

    2018-01-01

    when photovoltaic penetration is increased in low voltage distribution network. Local battery energy storage system can mitigate these disadvantages and as a result, improve the system operation. For this purpose, battery energy storage system is charged when production of photovoltaic is more than...... consumers’ demands and discharged when consumers’ demands are increased. Since the price of battery energy storage system is high, economic, environmental, and technical objectives should be considered together for its placement and sizing. In this paper, optimal placement, sizing, and daily (24 h) charge......Proper installation of rooftop photovoltaic generation in distribution networks can improve voltage profile, reduce energy losses, and enhance the reliability. But, on the other hand, some problems regarding harmonic distortion, voltage magnitude, reverse power flow, and energy losses can arise...

  16. Angular distribution of Pionization particles from experiments on CERN-storage rings

    CERN Document Server

    Jabs, A

    1972-01-01

    The angular distribution of secondary charged particles arising from high energy interaction in the CERN storage rings has been measured. The measurements near the centre-of-mass energies (20-50) GeV have been taken in the angle range of 20-90 degrees relative to the recoil axis. A complex formula is suggested to explain the variations described experimentally. Good agreement between the theoretical and experimental data is found. Simplifications are possible in the formula which allow the two-centre emission structure to be explained. It is shown that the angular dependence cannot be explained solely from rotary impulse behaviour. (9 refs).

  17. Impact of Nitrification on the Formation of N-Nitrosamines and Halogenated Disinfection Byproducts within Distribution System Storage Facilities.

    Science.gov (United States)

    Zeng, Teng; Mitch, William A

    2016-03-15

    Distribution system storage facilities are a critical, yet often overlooked, component of the urban water infrastructure. This study showed elevated concentrations of N-nitrosodimethylamine (NDMA), total N-nitrosamines (TONO), regulated trihalomethanes (THMs) and haloacetic acids (HAAs), 1,1-dichloropropanone (1,1-DCP), trichloroacetaldehyde (TCAL), haloacetonitriles (HANs), and haloacetamides (HAMs) in waters with ongoing nitrification as compared to non-nitrifying waters in storage facilities within five different chloraminated drinking water distribution systems. The concentrations of NDMA, TONO, HANs, and HAMs in the nitrifying waters further increased upon application of simulated distribution system chloramination. The addition of a nitrifying biofilm sample collected from a nitrifying facility to its non-nitrifying influent water led to increases in N-nitrosamine and halogenated DBP formation, suggesting the release of precursors from nitrifying biofilms. Periodic treatment of two nitrifying facilities with breakpoint chlorination (BPC) temporarily suppressed nitrification and reduced precursor levels for N-nitrosamines, HANs, and HAMs, as reflected by lower concentrations of these DBPs measured after re-establishment of a chloramine residual within the facilities than prior to the BPC treatment. However, BPC promoted the formation of halogenated DBPs while a free chlorine residual was maintained. Strategies that minimize application of free chlorine while preventing nitrification are needed to control DBP precursor release in storage facilities.

  18. Management issues for high performance storage systems

    Energy Technology Data Exchange (ETDEWEB)

    Louis, S. [Lawrence Livermore National Lab., CA (United States); Burris, R. [Oak Ridge National Lab., TN (United States)

    1995-03-01

    Managing distributed high-performance storage systems is complex and, although sharing common ground with traditional network and systems management, presents unique storage-related issues. Integration technologies and frameworks exist to help manage distributed network and system environments. Industry-driven consortia provide open forums where vendors and users cooperate to leverage solutions. But these new approaches to open management fall short addressing the needs of scalable, distributed storage. We discuss the motivation and requirements for storage system management (SSM) capabilities and describe how SSM manages distributed servers and storage resource objects in the High Performance Storage System (HPSS), a new storage facility for data-intensive applications and large-scale computing. Modem storage systems, such as HPSS, require many SSM capabilities, including server and resource configuration control, performance monitoring, quality of service, flexible policies, file migration, file repacking, accounting, and quotas. We present results of initial HPSS SSM development including design decisions and implementation trade-offs. We conclude with plans for follow-on work and provide storage-related recommendations for vendors and standards groups seeking enterprise-wide management solutions.

  19. Active and reactive power support of MV distribution systems using battery energy storage

    DEFF Research Database (Denmark)

    Wang, Jiawei; Hashemi Toghroljerdi, Seyedmostafa; You, Shi

    2017-01-01

    shaving and voltage support service from the perspective of Distribution System Operators (DSOs). An active power support algorithm is implemented and the effects of various load profiles as well as different Photovoltaic (PV) penetration scenarios on the operation of BESS and the optimal BESS converter......Adoption of Battery Energy Storage Systems (BESSs) for provision of grid services is increasing. This paper investigates the applications of BESS for the grid upgrade deferral and voltage support of Medium Voltage (MV) distribution systems. A BESS is modelled in Matlab/Simulink to perform peak load...... size for peak load shaving are investigated. The BESS annual lifetime degradation is also estimated using a rainflow counting algorithm. A reactive power support algorithm embedded with Q-U droop control is proposed in order to reduce the voltage drop in a part of 10 kV distribution network of Nordhavn...

  20. Data storage as a service

    OpenAIRE

    Tomšič, Jan

    2016-01-01

    The purpose of the thesis was comparison of interfaces to network attached file systems and object storage. The thesis describes network file system and mounting procedure in Linux operating system. Object storage and distributed storage systems are explained with examples of usage. Amazon S3 is an example of object store with access trough REST interface. Ceph, a system for distributed object storage, is explained in detail, and a Ceph cluster was deployed for the purpose of this thesis. Cep...

  1. Numerical simulation of ambient flow and thermal distributions in a spent fuel storage cask array

    International Nuclear Information System (INIS)

    Michener, T.; Trent, D.S.; Guttmann, J.; Bajwa, C.

    2001-01-01

    At the request of the U.S. Nuclear Regulatory Commission (USNRC), the staff at the Pacific Northwest National Laboratory (PNNL) analyzed the thermal performance of the Utah Private Fuel Storage (PFS) using the TEMPEST computational fluid dynamics software. A three-dimensional section of the PFS with a total of 20 casks was modeled to estimate the ambient flow and temperature distributions surrounding the casks. The purpose of this analysis was to compute the cask inlet vent air temperature to be used for boundary conditions in a detailed analysis of an individual Holtec Hi-Storm 100 cask using the COBRA-SFS (Spent Fuel Storage) thermal hydraulic computer software. (author)

  2. Coordinated operation of a neighborhood of smart households comprising electric vehicles, energy storage and distributed generation

    NARCIS (Netherlands)

    Paterakis, N.G.; Erdinc, O.; Pappi, I.N.; Bakirtzis, A.G.; Catalao, J.P.S.

    2016-01-01

    In this paper, the optimal operation of a neighborhood of smart households in terms of minimizing the total energy procurement cost is analyzed. Each household may comprise several assets such as electric vehicles, controllable appliances, energy storage and distributed generation. Bi-directional

  3. Research in Parallel Algorithms and Software for Computational Aerosciences

    Science.gov (United States)

    Domel, Neal D.

    1996-01-01

    Phase 1 is complete for the development of a computational fluid dynamics CFD) parallel code with automatic grid generation and adaptation for the Euler analysis of flow over complex geometries. SPLITFLOW, an unstructured Cartesian grid code developed at Lockheed Martin Tactical Aircraft Systems, has been modified for a distributed memory/massively parallel computing environment. The parallel code is operational on an SGI network, Cray J90 and C90 vector machines, SGI Power Challenge, and Cray T3D and IBM SP2 massively parallel machines. Parallel Virtual Machine (PVM) is the message passing protocol for portability to various architectures. A domain decomposition technique was developed which enforces dynamic load balancing to improve solution speed and memory requirements. A host/node algorithm distributes the tasks. The solver parallelizes very well, and scales with the number of processors. Partially parallelized and non-parallelized tasks consume most of the wall clock time in a very fine grain environment. Timing comparisons on a Cray C90 demonstrate that Parallel SPLITFLOW runs 2.4 times faster on 8 processors than its non-parallel counterpart autotasked over 8 processors.

  4. Power Flow Calculation for Weakly Meshed Distribution Networks with Multiple DGs Based on Generalized Chain-table Storage Structure

    DEFF Research Database (Denmark)

    Chen, Shuheng; Hu, Weihao; Chen, Zhe

    2014-01-01

    Based on generalized chain-table storage structure (GCTSS), a novel power flow method is proposed, which can be used to solve the power flow of weakly meshed distribution networks with multiple distributed generators (DGs). GCTSS is designed based on chain-table technology and its target is to de......Based on generalized chain-table storage structure (GCTSS), a novel power flow method is proposed, which can be used to solve the power flow of weakly meshed distribution networks with multiple distributed generators (DGs). GCTSS is designed based on chain-table technology and its target...... is to describe the topology of radial distribution networks with a clear logic and a small memory size. The strategies of compensating the equivalent currents of break-point branches and the reactive power outputs of PV-type DGs are presented on the basis of superposition theorem. Their formulations...... are simplified to be the final multi-variable linear functions. Furthermore, an accelerating factor is applied to the outer-layer reactive power compensation for improving the convergence procedure. Finally, the proposed power flow method is performed in program language VC++ 6.0, and numerical tests have been...

  5. Modelling transient temperature distribution for injecting hot water through a well to an aquifer thermal energy storage system

    Science.gov (United States)

    Yang, Shaw-Yang; Yeh, Hund-Der; Li, Kuang-Yi

    2010-10-01

    Heat storage systems are usually used to store waste heat and solar energy. In this study, a mathematical model is developed to predict both the steady-state and transient temperature distributions of an aquifer thermal energy storage (ATES) system after hot water is injected through a well into a confined aquifer. The ATES has a confined aquifer bounded by aquicludes with different thermomechanical properties and geothermal gradients along the depth. Consider that the heat is transferred by conduction and forced convection within the aquifer and by conduction within the aquicludes. The dimensionless semi-analytical solutions of temperature distributions of the ATES system are developed using Laplace and Fourier transforms and their corresponding time-domain results are evaluated numerically by the modified Crump method. The steady-state solution is obtained from the transient solution through the final-value theorem. The effect of the heat transfer coefficient on aquiclude temperature distribution is appreciable only near the outer boundaries of the aquicludes. The present solutions are useful for estimating the temperature distribution of heat injection and the aquifer thermal capacity of ATES systems.

  6. Design considerations for parallel graphics libraries

    Science.gov (United States)

    Crockett, Thomas W.

    1994-01-01

    Applications which run on parallel supercomputers are often characterized by massive datasets. Converting these vast collections of numbers to visual form has proven to be a powerful aid to comprehension. For a variety of reasons, it may be desirable to provide this visual feedback at runtime. One way to accomplish this is to exploit the available parallelism to perform graphics operations in place. In order to do this, we need appropriate parallel rendering algorithms and library interfaces. This paper provides a tutorial introduction to some of the issues which arise in designing parallel graphics libraries and their underlying rendering algorithms. The focus is on polygon rendering for distributed memory message-passing systems. We illustrate our discussion with examples from PGL, a parallel graphics library which has been developed on the Intel family of parallel systems.

  7. Shared Variable Oriented Parallel Precompiler for SPMD Model

    Institute of Scientific and Technical Information of China (English)

    1995-01-01

    For the moment,commercial parallel computer systems with distributed memory architecture are usually provided with parallel FORTRAN or parallel C compliers,which are just traditional sequential FORTRAN or C compilers expanded with communication statements.Programmers suffer from writing parallel programs with communication statements. The Shared Variable Oriented Parallel Precompiler (SVOPP) proposed in this paper can automatically generate appropriate communication statements based on shared variables for SPMD(Single Program Multiple Data) computation model and greatly ease the parallel programming with high communication efficiency.The core function of parallel C precompiler has been successfully verified on a transputer-based parallel computer.Its prominent performance shows that SVOPP is probably a break-through in parallel programming technique.

  8. Solar energy thermalization and storage device

    Science.gov (United States)

    McClelland, J.F.

    A passive solar thermalization and thermal energy storage assembly which is visually transparent is described. The assembly consists of two substantial parallel, transparent wall members mounted in a rectangular support frame to form a liquid-tight chamber. A semitransparent thermalization plate is located in the chamber, substantially paralled to and about equidistant from the transparent wall members to thermalize solar radiation which is stored in a transparent thermal energy storage liquid which fills the chamber. A number of the devices, as modules, can be stacked together to construct a visually transparent, thermal storage wall for passive solar-heated buildings.

  9. A Parallel Distributed-Memory Particle Method Enables Acquisition-Rate Segmentation of Large Fluorescence Microscopy Images.

    Science.gov (United States)

    Afshar, Yaser; Sbalzarini, Ivo F

    2016-01-01

    Modern fluorescence microscopy modalities, such as light-sheet microscopy, are capable of acquiring large three-dimensional images at high data rate. This creates a bottleneck in computational processing and analysis of the acquired images, as the rate of acquisition outpaces the speed of processing. Moreover, images can be so large that they do not fit the main memory of a single computer. We address both issues by developing a distributed parallel algorithm for segmentation of large fluorescence microscopy images. The method is based on the versatile Discrete Region Competition algorithm, which has previously proven useful in microscopy image segmentation. The present distributed implementation decomposes the input image into smaller sub-images that are distributed across multiple computers. Using network communication, the computers orchestrate the collectively solving of the global segmentation problem. This not only enables segmentation of large images (we test images of up to 10(10) pixels), but also accelerates segmentation to match the time scale of image acquisition. Such acquisition-rate image segmentation is a prerequisite for the smart microscopes of the future and enables online data compression and interactive experiments.

  10. PSHED: a simplified approach to developing parallel programs

    International Nuclear Information System (INIS)

    Mahajan, S.M.; Ramesh, K.; Rajesh, K.; Somani, A.; Goel, M.

    1992-01-01

    This paper presents a simplified approach in the forms of a tree structured computational model for parallel application programs. An attempt is made to provide a standard user interface to execute programs on BARC Parallel Processing System (BPPS), a scalable distributed memory multiprocessor. The interface package called PSHED provides a basic framework for representing and executing parallel programs on different parallel architectures. The PSHED package incorporates concepts from a broad range of previous research in programming environments and parallel computations. (author). 6 refs

  11. Scaling up machine learning: parallel and distributed approaches

    National Research Council Canada - National Science Library

    Bekkerman, Ron; Bilenko, Mikhail; Langford, John

    2012-01-01

    ... presented in the book cover a range of parallelization platforms from FPGAs and GPUs to multi-core systems and commodity clusters; concurrent programming frameworks that include CUDA, MPI, MapReduce, and DryadLINQ; and various learning settings: supervised, unsupervised, semi-supervised, and online learning. Extensive coverage of parallelizat...

  12. A Screen Space GPGPU Surface LIC Algorithm for Distributed Memory Data Parallel Sort Last Rendering Infrastructures

    Energy Technology Data Exchange (ETDEWEB)

    Loring, Burlen; Karimabadi, Homa; Rortershteyn, Vadim

    2014-07-01

    The surface line integral convolution(LIC) visualization technique produces dense visualization of vector fields on arbitrary surfaces. We present a screen space surface LIC algorithm for use in distributed memory data parallel sort last rendering infrastructures. The motivations for our work are to support analysis of datasets that are too large to fit in the main memory of a single computer and compatibility with prevalent parallel scientific visualization tools such as ParaView and VisIt. By working in screen space using OpenGL we can leverage the computational power of GPUs when they are available and run without them when they are not. We address efficiency and performance issues that arise from the transformation of data from physical to screen space by selecting an alternate screen space domain decomposition. We analyze the algorithm's scaling behavior with and without GPUs on two high performance computing systems using data from turbulent plasma simulations.

  13. PAPIRUS, a parallel computing framework for sensitivity analysis, uncertainty propagation, and estimation of parameter distribution

    International Nuclear Information System (INIS)

    Heo, Jaeseok; Kim, Kyung Doo

    2015-01-01

    Highlights: • We developed an interface between an engineering simulation code and statistical analysis software. • Multiple packages of the sensitivity analysis, uncertainty quantification, and parameter estimation algorithms are implemented in the framework. • Parallel computing algorithms are also implemented in the framework to solve multiple computational problems simultaneously. - Abstract: This paper introduces a statistical data analysis toolkit, PAPIRUS, designed to perform the model calibration, uncertainty propagation, Chi-square linearity test, and sensitivity analysis for both linear and nonlinear problems. The PAPIRUS was developed by implementing multiple packages of methodologies, and building an interface between an engineering simulation code and the statistical analysis algorithms. A parallel computing framework is implemented in the PAPIRUS with multiple computing resources and proper communications between the server and the clients of each processor. It was shown that even though a large amount of data is considered for the engineering calculation, the distributions of the model parameters and the calculation results can be quantified accurately with significant reductions in computational effort. A general description about the PAPIRUS with a graphical user interface is presented in Section 2. Sections 2.1–2.5 present the methodologies of data assimilation, uncertainty propagation, Chi-square linearity test, and sensitivity analysis implemented in the toolkit with some results obtained by each module of the software. Parallel computing algorithms adopted in the framework to solve multiple computational problems simultaneously are also summarized in the paper

  14. PAPIRUS, a parallel computing framework for sensitivity analysis, uncertainty propagation, and estimation of parameter distribution

    Energy Technology Data Exchange (ETDEWEB)

    Heo, Jaeseok, E-mail: jheo@kaeri.re.kr; Kim, Kyung Doo, E-mail: kdkim@kaeri.re.kr

    2015-10-15

    Highlights: • We developed an interface between an engineering simulation code and statistical analysis software. • Multiple packages of the sensitivity analysis, uncertainty quantification, and parameter estimation algorithms are implemented in the framework. • Parallel computing algorithms are also implemented in the framework to solve multiple computational problems simultaneously. - Abstract: This paper introduces a statistical data analysis toolkit, PAPIRUS, designed to perform the model calibration, uncertainty propagation, Chi-square linearity test, and sensitivity analysis for both linear and nonlinear problems. The PAPIRUS was developed by implementing multiple packages of methodologies, and building an interface between an engineering simulation code and the statistical analysis algorithms. A parallel computing framework is implemented in the PAPIRUS with multiple computing resources and proper communications between the server and the clients of each processor. It was shown that even though a large amount of data is considered for the engineering calculation, the distributions of the model parameters and the calculation results can be quantified accurately with significant reductions in computational effort. A general description about the PAPIRUS with a graphical user interface is presented in Section 2. Sections 2.1–2.5 present the methodologies of data assimilation, uncertainty propagation, Chi-square linearity test, and sensitivity analysis implemented in the toolkit with some results obtained by each module of the software. Parallel computing algorithms adopted in the framework to solve multiple computational problems simultaneously are also summarized in the paper.

  15. Testing and evaluating storage technology to build a distributed Tier1 for SuperB in Italy

    International Nuclear Information System (INIS)

    Pardi, S; Delprete, D; Russo, G; Fella, A; Corvo, M; Bianchi, F; Ciaschini, V; Giacomini, F; Simone, A Di; Donvito, G; Santeramo, B; Gianoli, A; Luppi, E; Manzali, M; Tomassetti, L; Longo, S; Stroili, R; Luitz, S; Perez, A; Rama, M

    2012-01-01

    The SuperB asymmetric energy e + e −- collider and detector to be built at the newly founded Nicola Cabibbo Lab will provide a uniquely sensitive probe of New Physics in the flavor sector of the Standard Model. Studying minute effects in the heavy quark and heavy lepton sectors requires a data sample of 75 ab −-1 and a luminosity target of 10 36 cm −-2 s −-1 . This luminosity translate in the requirement of storing more than 50 PByte of additional data each year, making SuperB an interesting challenge to the data management infrastructure, both at site level as at Wide Area Network level. A new Tier1, distributed among 3 or 4 sites in the south of Italy, is planned as part of the SuperB computing infrastructure. Data storage is a relevant topic whose development affects the way to configure and setup storage infrastructure both in local computing cluster and in a distributed paradigm. In this work we report the test on the software for data distribution and data replica focusing on the experiences made with Hadoop and GlusterFS.

  16. Algorithm and Implementation of Distributed ESN Using Spark Framework and Parallel PSO

    Directory of Open Access Journals (Sweden)

    Kehe Wu

    2017-04-01

    Full Text Available The echo state network (ESN employs a huge reservoir with sparsely and randomly connected internal nodes and only trains the output weights, which avoids the suboptimal problem, exploding and vanishing gradients, high complexity and other disadvantages faced by traditional recurrent neural network (RNN training. In light of the outstanding adaption to nonlinear dynamical systems, ESN has been applied into a wide range of applications. However, in the era of Big Data, with an enormous amount of data being generated continuously every day, the data are often distributed and stored in real applications, and thus the centralized ESN training process is prone to being technologically unsuitable. In order to achieve the requirement of Big Data applications in the real world, in this study we propose an algorithm and its implementation for distributed ESN training. The mentioned algorithm is based on the parallel particle swarm optimization (P-PSO technique and the implementation uses Spark, a famous large-scale data processing framework. Four extremely large-scale datasets, including artificial benchmarks, real-world data and image data, are adopted to verify our framework on a stretchable platform. Experimental results indicate that the proposed work is accurate in the era of Big Data, regarding speed, accuracy and generalization capabilities.

  17. Communication complexity of distributed computing and a parallel algorithm for polynomial roots

    International Nuclear Information System (INIS)

    Tiwari, P.

    1986-01-01

    The first part of this thesis begins with a discussion of the minimum communication requirements in some distributed networks. The main result is a general technique for determining lower bounds on the communication complexity of problems on various distributed computer networks. This general technique is derived by simulating the general network by a linear array and then using a lower bound on the communication complexity of the problem on the linear array. Applications of this technique yield nontrivial optimal or near-optimal lower bounds on the communication complexity of distinctness, ranking, uniqueness, merging, and triangle detection on a ring, a mesh, and a complete binary tree of processors. A technique similar to the one used in proving the above results, yields interesting graph theoretic results concerning decomposition of a graph into complete bipartite subgraphs. The second part of the this is devoted to the design of a fast parallel algorithm for determining all roots of a polynomial. Given a polynomial rho(z) of degree n with m bit integer coefficients and an integer μ, the author considers the problem of determining all its roots with error less than 2/sup -μ/. It is shown that this problem is in the class NC if rho(z) has all real roots

  18. Redesign Electricity Market for the Next Generation Power System of Renewable Energy and Distributed Storage Technologies

    DEFF Research Database (Denmark)

    Feng, Donghan; Xu, Zhao; Østergaard, Jacob

    2010-01-01

    This paper proposes a stochastic time-series based method to simulate the volatility of intermittent renewable generation and distributed storage devices along timeline. The proposed method can calculate the optimal timeline for different electricity markets and power systems. In practice......, the proposed method is potentially useful for designing market rules and evaluating different design options. Following works is underway on application and simulation of proposed method using the realistic distribution system of Bornholm Island in Denmark....

  19. Final Report: Migration Mechanisms for Large-scale Parallel Applications

    Energy Technology Data Exchange (ETDEWEB)

    Jason Nieh

    2009-10-30

    software environments, (4) network checkpoint-restart and migration of distributed and parallel applications, (5) a utility computing infrastructure for mobile desktop cloud computing based on process checkpoint-restart and migration functionality, (6) a process migration security architecture for protecting applications and infrastructure from denial-of-service attacks, and (7) a checkpoint-restart mobile computing system using portable storage devices.

  20. Optimizing Capacities of Distributed Generation and Energy Storage in a Small Autonomous Power System Considering Uncertainty in Renewables

    Directory of Open Access Journals (Sweden)

    Ying-Yi Hong

    2015-03-01

    Full Text Available This paper explores real power generation planning, considering distributed generation resources and energy storage in a small standalone power system. On account of the Kyoto Protocol and Copenhagen Accord, wind and photovoltaic (PV powers are considered as clean and renewable energies. In this study, a genetic algorithm (GA was used to determine the optimal capacities of wind-turbine-generators, PV, diesel generators and energy storage in a small standalone power system. The investment costs (installation, unit and maintenance costs of the distributed generation resources and energy storage and the cost of fuel for the diesel generators were minimized while the reliability requirement and CO2 emission limit were fulfilled. The renewable sources and loads were modeled by random variables because of their uncertainties. The equality and inequality constraints in the genetic algorithms were treated by cumulant effects and cumulative probability of random variables, respectively. The IEEE reliability data for an 8760 h load profile with a 150 kW peak load were used to demonstrate the applicability of the proposed method.

  1. The use of cold storage in household refrigerators

    International Nuclear Information System (INIS)

    Michael, Y.C.; Yang, K.T.

    1991-01-01

    A simulated household refrigerator was built that utilizes latent Thermal Energy Storage (TES) to reduce the on-off frequency of the compressor. It was found that the on-off frequency was reduced by a factor of six and that the energy consumption was less for a certain range of refrigerated space temperatures. A new evaporator design was used to accommodate the Phase Change Material (PCM). The evaporator consisted of four thin plates filled with PCM which were connected in parallel. The temperature distribution inside the PCM was also measured and the results are discussed. The PCM used was an aqueous solution of propylene glycol. In this paper freezing characteristics as well as some estimate of its latent heat are also discussed

  2. Scalable cloud without dedicated storage

    Science.gov (United States)

    Batkovich, D. V.; Kompaniets, M. V.; Zarochentsev, A. K.

    2015-05-01

    We present a prototype of a scalable computing cloud. It is intended to be deployed on the basis of a cluster without the separate dedicated storage. The dedicated storage is replaced by the distributed software storage. In addition, all cluster nodes are used both as computing nodes and as storage nodes. This solution increases utilization of the cluster resources as well as improves fault tolerance and performance of the distributed storage. Another advantage of this solution is high scalability with a relatively low initial and maintenance cost. The solution is built on the basis of the open source components like OpenStack, CEPH, etc.

  3. EFFICIENT LIDAR POINT CLOUD DATA MANAGING AND PROCESSING IN A HADOOP-BASED DISTRIBUTED FRAMEWORK

    Directory of Open Access Journals (Sweden)

    C. Wang

    2017-10-01

    Full Text Available Light Detection and Ranging (LiDAR is one of the most promising technologies in surveying and mapping,city management, forestry, object recognition, computer vision engineer and others. However, it is challenging to efficiently storage, query and analyze the high-resolution 3D LiDAR data due to its volume and complexity. In order to improve the productivity of Lidar data processing, this study proposes a Hadoop-based framework to efficiently manage and process LiDAR data in a distributed and parallel manner, which takes advantage of Hadoop’s storage and computing ability. At the same time, the Point Cloud Library (PCL, an open-source project for 2D/3D image and point cloud processing, is integrated with HDFS and MapReduce to conduct the Lidar data analysis algorithms provided by PCL in a parallel fashion. The experiment results show that the proposed framework can efficiently manage and process big LiDAR data.

  4. Efficient LIDAR Point Cloud Data Managing and Processing in a Hadoop-Based Distributed Framework

    Science.gov (United States)

    Wang, C.; Hu, F.; Sha, D.; Han, X.

    2017-10-01

    Light Detection and Ranging (LiDAR) is one of the most promising technologies in surveying and mapping city management, forestry, object recognition, computer vision engineer and others. However, it is challenging to efficiently storage, query and analyze the high-resolution 3D LiDAR data due to its volume and complexity. In order to improve the productivity of Lidar data processing, this study proposes a Hadoop-based framework to efficiently manage and process LiDAR data in a distributed and parallel manner, which takes advantage of Hadoop's storage and computing ability. At the same time, the Point Cloud Library (PCL), an open-source project for 2D/3D image and point cloud processing, is integrated with HDFS and MapReduce to conduct the Lidar data analysis algorithms provided by PCL in a parallel fashion. The experiment results show that the proposed framework can efficiently manage and process big LiDAR data.

  5. Advanced parallel processing with supercomputer architectures

    International Nuclear Information System (INIS)

    Hwang, K.

    1987-01-01

    This paper investigates advanced parallel processing techniques and innovative hardware/software architectures that can be applied to boost the performance of supercomputers. Critical issues on architectural choices, parallel languages, compiling techniques, resource management, concurrency control, programming environment, parallel algorithms, and performance enhancement methods are examined and the best answers are presented. The authors cover advanced processing techniques suitable for supercomputers, high-end mainframes, minisupers, and array processors. The coverage emphasizes vectorization, multitasking, multiprocessing, and distributed computing. In order to achieve these operation modes, parallel languages, smart compilers, synchronization mechanisms, load balancing methods, mapping parallel algorithms, operating system functions, application library, and multidiscipline interactions are investigated to ensure high performance. At the end, they assess the potentials of optical and neural technologies for developing future supercomputers

  6. Stabilization and control of tie-line power flow of microgrid including wind generation by distributed energy storage

    Energy Technology Data Exchange (ETDEWEB)

    Molina, M.G.; Mercado, P.E. [CONICET, Instituto de Energia Electrica, Universidad Nacional de San Juan, Av. Libertador San Martin Oeste 1109, J5400ARL San Juan (Argentina)

    2010-06-15

    High penetration of wind generation in electrical microgrids causes fluctuations of tie-line power flow and significantly affects the power system operation. This can lead to severe problems, such as system frequency oscillations, and/or violations of power lines capability. With proper control, a distribution static synchronous compensator (DSTATCOM) integrated with superconducting magnetic energy storage (SMES) is able to significantly enhance the dynamic security of the power system. This paper proposes the use of a SMES system in combination with a DSTATCOM as effective distributed energy storage (DES) for stabilization and control of the tie-line power flow of microgrids incorporating wind generation. A new detailed model of the integrated DSTATCOM-SMES device is derived and a novel three-level control scheme is designed. The dynamic performance of the proposed control schemes is fully validated using MATLAB/Simulink. (author)

  7. A Framework for Evaluating Economic Impacts of Rooftop PV Systems with or without Energy Storage on Thai Distribution Utilities and Ratepayers

    Science.gov (United States)

    Chaianong, A.; Bangviwat, A.; Menke, C.

    2017-07-01

    Driven by decreasing PV and energy storage prices, increasing electricity costs and policy supports from Thai government (self-consumption era), rooftop PV and energy storage systems are going to be deployed in the country rapidly that may disrupt existing business models structure of Thai distribution utilities due to revenue erosion and lost earnings opportunities. The retail rates that directly affect ratepayers (non-solar customers) are expected to increase. This paper focuses on a framework for evaluating impacts of PV with and without energy storage systems on Thai distribution utilities and ratepayers by using cost-benefit analysis (CBA). Prior to calculation of cost/benefit components, changes in energy sales need to be addressed. Government policies for the support of PV generation will also help in accelerating the rooftop PV installation. Benefit components include avoided costs due to transmission losses and deferring distribution capacity with appropriate PV penetration level, while cost components consist of losses in revenue, program costs, integration costs and unrecovered fixed costs. It is necessary for Thailand to compare total costs and total benefits of rooftop PV and energy storage systems in order to adopt policy supports and mitigation approaches, such as business model innovation and regulatory reform, effectively.

  8. Low latency network and distributed storage for next generation HPC systems: the ExaNeSt project

    Science.gov (United States)

    Ammendola, R.; Biagioni, A.; Cretaro, P.; Frezza, O.; Lo Cicero, F.; Lonardo, A.; Martinelli, M.; Paolucci, P. S.; Pastorelli, E.; Pisani, F.; Simula, F.; Vicini, P.; Navaridas, J.; Chaix, F.; Chrysos, N.; Katevenis, M.; Papaeustathiou, V.

    2017-10-01

    With processor architecture evolution, the HPC market has undergone a paradigm shift. The adoption of low-cost, Linux-based clusters extended the reach of HPC from its roots in modelling and simulation of complex physical systems to a broader range of industries, from biotechnology, cloud computing, computer analytics and big data challenges to manufacturing sectors. In this perspective, the near future HPC systems can be envisioned as composed of millions of low-power computing cores, densely packed — meaning cooling by appropriate technology — with a tightly interconnected, low latency and high performance network and equipped with a distributed storage architecture. Each of these features — dense packing, distributed storage and high performance interconnect — represents a challenge, made all the harder by the need to solve them at the same time. These challenges lie as stumbling blocks along the road towards Exascale-class systems; the ExaNeSt project acknowledges them and tasks itself with investigating ways around them.

  9. GPU Parallel Bundle Block Adjustment

    Directory of Open Access Journals (Sweden)

    ZHENG Maoteng

    2017-09-01

    Full Text Available To deal with massive data in photogrammetry, we introduce the GPU parallel computing technology. The preconditioned conjugate gradient and inexact Newton method are also applied to decrease the iteration times while solving the normal equation. A brand new workflow of bundle adjustment is developed to utilize GPU parallel computing technology. Our method can avoid the storage and inversion of the big normal matrix, and compute the normal matrix in real time. The proposed method can not only largely decrease the memory requirement of normal matrix, but also largely improve the efficiency of bundle adjustment. It also achieves the same accuracy as the conventional method. Preliminary experiment results show that the bundle adjustment of a dataset with about 4500 images and 9 million image points can be done in only 1.5 minutes while achieving sub-pixel accuracy.

  10. Application of battery-based storage systems in household-demand smoothening in electricity-distribution grids

    International Nuclear Information System (INIS)

    Purvins, Arturs; Papaioannou, Ioulia T.; Debarberis, Luigi

    2013-01-01

    Highlights: ► Battery system application in demand smoothening in distribution grids is analysed. ► Five European countries are studied with and without high photovoltaic deployment. ► A sensitivity analysis for different battery system parameters is performed. ► A simple battery system management is sufficient for low demand smoothening. ► More elaborate management is required for high demand smoothening. - Abstract: This article analyses in technical terms the application of battery-based storage systems for household-demand smoothening in electricity-distribution grids. The analysis includes case studies of Denmark, Portugal, Greece, France and Italy. A high penetration of photovoltaic systems in distribution grids is considered as an additional scenario. A sensitivity analysis is performed in order to examine the smoothening effect of daily demand profiles for different configurations of the battery system. In general, battery-storage systems with low rated power and low battery capacity can smooth the demand sufficiently with the aid of a simple management process. For example, with 1 kW of peak demand, a 30–45% decrease in the variability of the daily demand profile can be achieved with a battery system of 0.1 kW rated power and up to 0.6 kW h battery capacity. However, further smoothening requires higher battery-system capacity and power. In this case, more elaborate management is also needed to use the battery system efficiently.

  11. Computationally efficient implementation of combustion chemistry in parallel PDF calculations

    International Nuclear Information System (INIS)

    Lu Liuyan; Lantz, Steven R.; Ren Zhuyin; Pope, Stephen B.

    2009-01-01

    In parallel calculations of combustion processes with realistic chemistry, the serial in situ adaptive tabulation (ISAT) algorithm [S.B. Pope, Computationally efficient implementation of combustion chemistry using in situ adaptive tabulation, Combustion Theory and Modelling, 1 (1997) 41-63; L. Lu, S.B. Pope, An improved algorithm for in situ adaptive tabulation, Journal of Computational Physics 228 (2009) 361-386] substantially speeds up the chemistry calculations on each processor. To improve the parallel efficiency of large ensembles of such calculations in parallel computations, in this work, the ISAT algorithm is extended to the multi-processor environment, with the aim of minimizing the wall clock time required for the whole ensemble. Parallel ISAT strategies are developed by combining the existing serial ISAT algorithm with different distribution strategies, namely purely local processing (PLP), uniformly random distribution (URAN), and preferential distribution (PREF). The distribution strategies enable the queued load redistribution of chemistry calculations among processors using message passing. They are implemented in the software x2f m pi, which is a Fortran 95 library for facilitating many parallel evaluations of a general vector function. The relative performance of the parallel ISAT strategies is investigated in different computational regimes via the PDF calculations of multiple partially stirred reactors burning methane/air mixtures. The results show that the performance of ISAT with a fixed distribution strategy strongly depends on certain computational regimes, based on how much memory is available and how much overlap exists between tabulated information on different processors. No one fixed strategy consistently achieves good performance in all the regimes. Therefore, an adaptive distribution strategy, which blends PLP, URAN and PREF, is devised and implemented. It yields consistently good performance in all regimes. In the adaptive parallel

  12. Synchronization Techniques in Parallel Discrete Event Simulation

    OpenAIRE

    Lindén, Jonatan

    2018-01-01

    Discrete event simulation is an important tool for evaluating system models in many fields of science and engineering. To improve the performance of large-scale discrete event simulations, several techniques to parallelize discrete event simulation have been developed. In parallel discrete event simulation, the work of a single discrete event simulation is distributed over multiple processing elements. A key challenge in parallel discrete event simulation is to ensure that causally dependent ...

  13. Performance Analysis of Parallel Mathematical Subroutine library PARCEL

    International Nuclear Information System (INIS)

    Yamada, Susumu; Shimizu, Futoshi; Kobayashi, Kenichi; Kaburaki, Hideo; Kishida, Norio

    2000-01-01

    The parallel mathematical subroutine library PARCEL (Parallel Computing Elements) has been developed by Japan Atomic Energy Research Institute for easy use of typical parallelized mathematical codes in any application problems on distributed parallel computers. The PARCEL includes routines for linear equations, eigenvalue problems, pseudo-random number generation, and fast Fourier transforms. It is shown that the results of performance for linear equations routines exhibit good parallelization efficiency on vector, as well as scalar, parallel computers. A comparison of the efficiency results with the PETSc (Portable Extensible Tool kit for Scientific Computations) library has been reported. (author)

  14. Cloud object store for archive storage of high performance computing data using decoupling middleware

    Science.gov (United States)

    Bent, John M.; Faibish, Sorin; Grider, Gary

    2015-06-30

    Cloud object storage is enabled for archived data, such as checkpoints and results, of high performance computing applications using a middleware process. A plurality of archived files, such as checkpoint files and results, generated by a plurality of processes in a parallel computing system are stored by obtaining the plurality of archived files from the parallel computing system; converting the plurality of archived files to objects using a log structured file system middleware process; and providing the objects for storage in a cloud object storage system. The plurality of processes may run, for example, on a plurality of compute nodes. The log structured file system middleware process may be embodied, for example, as a Parallel Log-Structured File System (PLFS). The log structured file system middleware process optionally executes on a burst buffer node.

  15. Parallel simulated annealing algorithms for cell placement on hypercube multiprocessors

    Science.gov (United States)

    Banerjee, Prithviraj; Jones, Mark Howard; Sargent, Jeff S.

    1990-01-01

    Two parallel algorithms for standard cell placement using simulated annealing are developed to run on distributed-memory message-passing hypercube multiprocessors. The cells can be mapped in a two-dimensional area of a chip onto processors in an n-dimensional hypercube in two ways, such that both small and large cell exchange and displacement moves can be applied. The computation of the cost function in parallel among all the processors in the hypercube is described, along with a distributed data structure that needs to be stored in the hypercube to support the parallel cost evaluation. A novel tree broadcasting strategy is used extensively for updating cell locations in the parallel environment. A dynamic parallel annealing schedule estimates the errors due to interacting parallel moves and adapts the rate of synchronization automatically. Two novel approaches in controlling error in parallel algorithms are described: heuristic cell coloring and adaptive sequence control.

  16. MMC with parallel-connected MOSFETs as an alternative to wide bandgap converters for LVDC distribution networks

    Directory of Open Access Journals (Sweden)

    Yanni Zhong

    2017-03-01

    Full Text Available Low-voltage direct-current (LVDC networks offer improved conductor utilisation on existing infrastructure and reduced conversion stages, which can lead to a simpler and more efficient distribution network. However, LVDC networks must continue to support AC loads, requiring efficient, low-distortion DC–AC converters. Additionally, increasing numbers of DC loads on the LVAC network require controlled, low-distortion, unity power factor AC-DC converters with large capacity, and bi-directional capability. An AC–DC/DC–AC converter design is therefore proposed in this study to minimise conversion loss and maximise power quality. Comparative analysis is performed for a conventional IGBT two-level converter, a SiC MOSFET two-level converter, a Si MOSFET modular multi-level converter (MMC and a GaN HEMT MMC, in terms of power loss, reliability, fault tolerance, converter cost and heatsink size. The analysis indicates that the five-level MMC with parallel-connected Si MOSFETs is an efficient, cost-effective converter for low-voltage converter applications. MMC converters suffer negligible switching loss, which enables reduced device switching without loss penalty from increased harmonics and filtering. Optimal extent of parallel-connection for MOSFETs in an MMC is investigated. Experimental results are presented to show the reduction in device stress and electromagnetic interference generating transients through the use of reduced switching and device parallel-connection.

  17. Parallel thermal radiation transport in two dimensions

    International Nuclear Information System (INIS)

    Smedley-Stevenson, R.P.; Ball, S.R.

    2003-01-01

    This paper describes the distributed memory parallel implementation of a deterministic thermal radiation transport algorithm in a 2-dimensional ALE hydrodynamics code. The parallel algorithm consists of a variety of components which are combined in order to produce a state of the art computational capability, capable of solving large thermal radiation transport problems using Blue-Oak, the 3 Tera-Flop MPP (massive parallel processors) computing facility at AWE (United Kingdom). Particular aspects of the parallel algorithm are described together with examples of the performance on some challenging applications. (author)

  18. Parallel thermal radiation transport in two dimensions

    Energy Technology Data Exchange (ETDEWEB)

    Smedley-Stevenson, R.P.; Ball, S.R. [AWE Aldermaston (United Kingdom)

    2003-07-01

    This paper describes the distributed memory parallel implementation of a deterministic thermal radiation transport algorithm in a 2-dimensional ALE hydrodynamics code. The parallel algorithm consists of a variety of components which are combined in order to produce a state of the art computational capability, capable of solving large thermal radiation transport problems using Blue-Oak, the 3 Tera-Flop MPP (massive parallel processors) computing facility at AWE (United Kingdom). Particular aspects of the parallel algorithm are described together with examples of the performance on some challenging applications. (author)

  19. Shredder: GPU-Accelerated Incremental Storage and Computation

    OpenAIRE

    Bhatotia, Pramod; Rodrigues, Rodrigo; Verma, Akshat

    2012-01-01

    Redundancy elimination using data deduplication and incremental data processing has emerged as an important technique to minimize storage and computation requirements in data center computing. In this paper, we present the design, implementation and evaluation of Shredder, a high performance content-based chunking framework for supporting incremental storage and computation systems. Shredder exploits the massively parallel processing power of GPUs to overcome the CPU bottlenecks of content-ba...

  20. Modeling storage and demand management in power distribution grids

    International Nuclear Information System (INIS)

    Schroeder, Andreas

    2011-01-01

    Grahical abstract: The model informs an optimal investment sizing decision as regards specific 'smart grid' applications such as storage facilities and meters enabling load control. Results indicate that central storage facilities are a more promising option for generation cost reductions as compared to demand management. Highlights: → Stochastic versus deterministic model increases investment efficiency up to 5%. → Deterministic model under-estimates value of load control and storage. → Battery storage is beneficial at investment cost below 850 EUR/MW h. → Demand management equipment is not beneficial at cost beyond 200 EUR. → The stylized 10 kV grid constitutes no shortage factor. -- Abstract: Storage devices and demand control may constitute beneficial tools to optimize electricity generation with a large share of intermittent resources through inter-temporal substitution of load. This paper quantifies the related cost reductions in a simulation model of a simplified stylized medium-voltage grid (10 kV) under uncertain demand and wind output. Benders Decomposition Method is applied to create a two-stage stochastic optimization program. The model informs an optimal investment sizing decision as regards specific 'smart' applications such as storage facilities and meters enabling load control. Model results indicate that central storage facilities are a more promising option for generation cost reductions as compared to demand management. Grid extensions are not appropriate in any of the scenarios. A sensitivity analysis is applied with respect to the market penetration of uncoordinated Plug-In Electric Vehicles which are found to strongly encourage investment into load control equipment for 'smart' charging and slightly improve the case for central storage devices.

  1. A Hybrid RES Distributed Generation System for Autonomous Islands: A DER-CAM and Storage-Based Economic and Optimal Dispatch Analysis

    DEFF Research Database (Denmark)

    Michalitsakos, Panagiotis; Mihet-Popa, Lucian; Xydis, George

    2017-01-01

    -CAM (Distributed Energy Resources Customer Adoption Model) decision support tool was used for the multi-objective analysis conducted, which proposes a set of optimal solutions defining the appropriate Distributed Generation (DG) technologies, the capacities of storage and other technologies and the optimal......The possibility of replacing the existing autonomous thermal power plants by Distributed Energy Resources (DER) based on renewable energy sources (RES), along with the appropriate energy storage technologies in order to deal with the major problems that autonomous islands usually face...... was investigated. A case study of a small Greek island, which is confronted by various energy and water shortages, was studied for assessing the feasibility of DER deployment. The main objectives investigated were cost minimization, CO2 emissions minimization and DER reliability maximization. The DER...

  2. Conflict free network coding for distributed storage networks

    KAUST Repository

    Al-Habob, Ahmed A.

    2015-06-01

    © 2015 IEEE. In this paper, we design a conflict free instantly decodable network coding (IDNC) solution for file download from distributed storage servers. Considering previously downloaded files at the clients from these servers as side information, IDNC can speed up the current download process. However, transmission conflicts can occur since multiple servers can simultaneously send IDNC combinations of files to the same client, which can tune to only one of them at a time. To avoid such conflicts and design more efficient coded download patterns, we propose a dual conflict IDNC graph model, which extends the conventional IDNC graph model in order to guarantee conflict free server transmissions to each of the clients. We then formulate the download time minimization problem as a stochastic shortest path problem whose action space is defined by the independent sets of this new graph. Given the intractability of the solution, we design a channel-aware heuristic algorithm and show that it achieves a considerable reduction in the file download time, compared to applying the conventional IDNC approach separately at each of the servers.

  3. Balanced, parallel operation of flashlamps

    International Nuclear Information System (INIS)

    Carder, B.M.; Merritt, B.T.

    1979-01-01

    A new energy store, the Compensated Pulsed Alternator (CPA), promises to be a cost effective substitute for capacitors to drive flashlamps that pump large Nd:glass lasers. Because the CPA is large and discrete, it will be necessary that it drive many parallel flashlamp circuits, presenting a problem in equal current distribution. Current division to +- 20% between parallel flashlamps has been achieved, but this is marginal for laser pumping. A method is presented here that provides equal current sharing to about 1%, and it includes fused protection against short circuit faults. The method was tested with eight parallel circuits, including both open-circuit and short-circuit fault tests

  4. One click dataset transfer: toward efficient coupling of distributed storage resources and CPUs

    Czech Academy of Sciences Publication Activity Database

    Zerola, Michal; Lauret, J.; Barták, R.; Šumbera, Michal

    2012-01-01

    Roč. 368, 012022 (2012), s. 1-10 ISSN 1742-6588. [14th International Workshop on Advanced Computing and Analysis Techniques in Physics Research (ACAT). Uxbridge, 05.09.2011-09.09.2011] R&D Projects: GA MŠk LC07048; GA MŠk LA09013 Institutional support: RVO:61389005 Keywords : distributed storage * Grid computing * dataset transfer Subject RIV: BG - Nuclear, Atomic and Molecular Physics, Colliders http://iopscience.iop.org/1742-6596/368/1/012022/pdf/1742-6596_368_1_012022.pdf

  5. PERFORMANCE ANALYSIS BETWEEN EXPLICIT SCHEDULING AND IMPLICIT SCHEDULING OF PARALLEL ARRAY-BASED DOMAIN DECOMPOSITION USING OPENMP

    Directory of Open Access Journals (Sweden)

    MOHAMMED FAIZ ABOALMAALY

    2014-10-01

    Full Text Available With the continuous revolution of multicore architecture, several parallel programming platforms have been introduced in order to pave the way for fast and efficient development of parallel algorithms. Back into its categories, parallel computing can be done through two forms: Data-Level Parallelism (DLP or Task-Level Parallelism (TLP. The former can be done by the distribution of data among the available processing elements while the latter is based on executing independent tasks concurrently. Most of the parallel programming platforms have built-in techniques to distribute the data among processors, these techniques are technically known as automatic distribution (scheduling. However, due to their wide range of purposes, variation of data types, amount of distributed data, possibility of extra computational overhead and other hardware-dependent factors, manual distribution could achieve better outcomes in terms of performance when compared to the automatic distribution. In this paper, this assumption is investigated by conducting a comparison between automatic and our newly proposed manual distribution of data among threads in parallel. Empirical results of matrix addition and matrix multiplication show a considerable performance gain when manual distribution is applied against automatic distribution.

  6. Minimization of distribution system losses by exploiting storage and anticipating market-driven behaviour of wind power producers

    NARCIS (Netherlands)

    Farrokhseresht, M.; Paterakis, N.G.; Gibescu, M.; Slootweg, J.G.

    2017-01-01

    This paper presents a stochastic bi-level optimization model to determine the optimal dispatch of energy storage systems controlled directly by the distribution system operator (DSO) in order to achieve minimization of active power losses, taking into account the profit-driven participation of

  7. Voltage Scheduling Droop Control for State-of-Charge Balance of Distributed Energy Storage in DC Microgrids

    DEFF Research Database (Denmark)

    Li, Chendan; Dragicevic, Tomislav; Aldana, Nelson Leonardo Diaz

    2014-01-01

    Due to higher power quality, lower conversion loss, and more DC loads, there has been an increasing awareness on DC microgrid. Previous emphasis has been on equal power sharing among different units in the DC microgrid, while overlooking the coordination of the energy storage units to maintain...... the State-of-Charge balance. In this paper, a new droop method based on voltage scheduling for State-of-Charge balance is proposed to keep the SoC balance for the energy storage units. The proposed method has the advantage of avoiding the stability problem existed in traditional methods based on droop gain...... scheduling. Simulation experiment is taken in Matlab on a DC microgrid with two distributed energy storage units. The simulation results show that the proposed method has successfully achieved SoC balance during the load changes while maintaining the DC bus voltage within the allowable range....

  8. Energy storage

    CERN Document Server

    Brunet, Yves

    2013-01-01

    Energy storage examines different applications such as electric power generation, transmission and distribution systems, pulsed systems, transportation, buildings and mobile applications. For each of these applications, proper energy storage technologies are foreseen, with their advantages, disadvantages and limits. As electricity cannot be stored cheaply in large quantities, energy has to be stored in another form (chemical, thermal, electromagnetic, mechanical) and then converted back into electric power and/or energy using conversion systems. Most of the storage technologies are examined: b

  9. High resolution isotope data and ensemble modelling reveal ecohydrological controls on catchment storage-discharge relationships and flux travel time distributions

    Science.gov (United States)

    Soulsby, C.; Kuppel, S.; Smith, A.; Tetzlaff, D.

    2017-12-01

    The dynamics of water storage in a catchment provides a fundamental insight into the interlinkages between input and output fluxes, and how these are affected by environmental change. Such dynamics also mediate, and help us understand, the fundamental difference of the rapid celerity of the rainfall-runoff (minutes to hours) response of catchments and the much slower velocity of water particles (months to decades) as they are transported through catchment systems. In this contribution we report an intensive, long-term (>10year), multi-scale isotope study in the Scottish Highlands that has sought to better understand these issues. We have integrated empirical data collection with diverse modelling approaches to quantify the dynamics and residence times of storage in different compartments of the hydrological system (vegetation canopies, soils, ground waters etc.) and their relationship between the magnitude and travel time distributions of output fluxes (stream flow, transpiration and evaporation). Use of conceptual, physically-based and probabilistic modelling approaches give broadly consistent perspectives on the storage-discharge relationships and the preferential selection of younger waters in runoff, evaporation and transpiration; while older waters predominate in groundwater. The work also highlighted the importance role vegetation plays in regulating fluxes in evaporation and transpiration and how this contributes to the differential ageing of water in mobile and bulk waters in the soil compartment. A separate case study shows how land use change can affect storage distributions in a catchment and radically change travel time distributions in output fluxes.

  10. Molecular simulation workflows as parallel algorithms: the execution engine of Copernicus, a distributed high-performance computing platform.

    Science.gov (United States)

    Pronk, Sander; Pouya, Iman; Lundborg, Magnus; Rotskoff, Grant; Wesén, Björn; Kasson, Peter M; Lindahl, Erik

    2015-06-09

    Computational chemistry and other simulation fields are critically dependent on computing resources, but few problems scale efficiently to the hundreds of thousands of processors available in current supercomputers-particularly for molecular dynamics. This has turned into a bottleneck as new hardware generations primarily provide more processing units rather than making individual units much faster, which simulation applications are addressing by increasingly focusing on sampling with algorithms such as free-energy perturbation, Markov state modeling, metadynamics, or milestoning. All these rely on combining results from multiple simulations into a single observation. They are potentially powerful approaches that aim to predict experimental observables directly, but this comes at the expense of added complexity in selecting sampling strategies and keeping track of dozens to thousands of simulations and their dependencies. Here, we describe how the distributed execution framework Copernicus allows the expression of such algorithms in generic workflows: dataflow programs. Because dataflow algorithms explicitly state dependencies of each constituent part, algorithms only need to be described on conceptual level, after which the execution is maximally parallel. The fully automated execution facilitates the optimization of these algorithms with adaptive sampling, where undersampled regions are automatically detected and targeted without user intervention. We show how several such algorithms can be formulated for computational chemistry problems, and how they are executed efficiently with many loosely coupled simulations using either distributed or parallel resources with Copernicus.

  11. Same-source parallel implementation of the PSU/NCAR MM5

    Energy Technology Data Exchange (ETDEWEB)

    Michalakes, J.

    1997-12-31

    The Pennsylvania State/National Center for Atmospheric Research Mesoscale Model is a limited-area model of atmospheric systems, now in its fifth generation, MM5. Designed and maintained for vector and shared-memory parallel architectures, the official version of MM5 does not run on message-passing distributed memory (DM) parallel computers. The authors describe a same-source parallel implementation of the PSU/NCAR MM5 using FLIC, the Fortran Loop and Index Converter. The resulting source is nearly line-for-line identical with the original source code. The result is an efficient distributed memory parallel option to MM5 that can be seamlessly integrated into the official version.

  12. Loss optimization in distribution networks with distributed generation

    DEFF Research Database (Denmark)

    Pokhrel, Basanta Raj; Nainar, Karthikeyan; Bak-Jensen, Birgitte

    2017-01-01

    This paper presents a novel power loss minimization approach in distribution grids considering network reconfiguration, distributed generation and storage installation. Identification of optimum configuration in such scenario is one of the main challenges faced by distribution system operators...... in highly active distribution grids. This issue is tackled by formulating a hybrid loss optimization problem and solved using the Interior Point Method. Sensitivity analysis is used to identify the optimum location of storage units. Different scenarios of reconfiguration, storage and distributed generation...... penetration are created to test the proposed algorithm. It is tested in a benchmark medium voltage network to show the effectiveness and performance of the algorithm. Results obtained are found to be encouraging for radial distribution system. It shows that we can reduce the power loss by more than 30% using...

  13. A Parallel Distributed-Memory Particle Method Enables Acquisition-Rate Segmentation of Large Fluorescence Microscopy Images

    Science.gov (United States)

    Afshar, Yaser; Sbalzarini, Ivo F.

    2016-01-01

    Modern fluorescence microscopy modalities, such as light-sheet microscopy, are capable of acquiring large three-dimensional images at high data rate. This creates a bottleneck in computational processing and analysis of the acquired images, as the rate of acquisition outpaces the speed of processing. Moreover, images can be so large that they do not fit the main memory of a single computer. We address both issues by developing a distributed parallel algorithm for segmentation of large fluorescence microscopy images. The method is based on the versatile Discrete Region Competition algorithm, which has previously proven useful in microscopy image segmentation. The present distributed implementation decomposes the input image into smaller sub-images that are distributed across multiple computers. Using network communication, the computers orchestrate the collectively solving of the global segmentation problem. This not only enables segmentation of large images (we test images of up to 1010 pixels), but also accelerates segmentation to match the time scale of image acquisition. Such acquisition-rate image segmentation is a prerequisite for the smart microscopes of the future and enables online data compression and interactive experiments. PMID:27046144

  14. A Parallel Distributed-Memory Particle Method Enables Acquisition-Rate Segmentation of Large Fluorescence Microscopy Images.

    Directory of Open Access Journals (Sweden)

    Yaser Afshar

    Full Text Available Modern fluorescence microscopy modalities, such as light-sheet microscopy, are capable of acquiring large three-dimensional images at high data rate. This creates a bottleneck in computational processing and analysis of the acquired images, as the rate of acquisition outpaces the speed of processing. Moreover, images can be so large that they do not fit the main memory of a single computer. We address both issues by developing a distributed parallel algorithm for segmentation of large fluorescence microscopy images. The method is based on the versatile Discrete Region Competition algorithm, which has previously proven useful in microscopy image segmentation. The present distributed implementation decomposes the input image into smaller sub-images that are distributed across multiple computers. Using network communication, the computers orchestrate the collectively solving of the global segmentation problem. This not only enables segmentation of large images (we test images of up to 10(10 pixels, but also accelerates segmentation to match the time scale of image acquisition. Such acquisition-rate image segmentation is a prerequisite for the smart microscopes of the future and enables online data compression and interactive experiments.

  15. Resilient data staging through MxN distributed transactions.

    Energy Technology Data Exchange (ETDEWEB)

    Schwan, Karsten (Georgia Institute of Technology, Atlanta, GA); Oldfield, Ron A.; Lofstead, Gerald Fredrick, II; Dayal, Jai (Georgia Institute of Technology, Atlanta, GA)

    2011-11-01

    Scientific computing-driven discoveries are frequently driven from workflows that use persistent storage as a staging area for data between operations. With the bad and progressively worse bandwidth vs. data size issues as we continue towards exascale, eliminating persistent storage through techniques like data staging will both enable these workflows to continue online, but also enable more interactive workflows reducing the time to scientific discoveries. Data staging has shown to be an effective way for applications running on high-end computing platforms to offload expensive I/O operations and to manage the tremendous amounts of data they produce. This data staging approach, however, lacks the ACID style guarantees traditional straight-to-disk methods provide. Distributed transactions are a proven way to add ACID properties to data movements, however distributed transactions follow 1xN data movement semantics, where our highly parallel HPC environments employ MxN data movement semantics. In this paper we present a novel protocol that extends distributed transaction terminology to include MxN semantics which allows our data staging areas to benefit from ACID properties. We show that with our protocol we can provide resilient data staging with a limited performance penalty over current data staging implementations.

  16. Integrated load distribution and production planning in series-parallel multi-state systems with failure rate depending on load

    International Nuclear Information System (INIS)

    Nourelfath, Mustapha; Yalaoui, Farouk

    2012-01-01

    A production system containing a set of machines (also called components) arranged according to a series-parallel configuration is addressed. A set of products must be produced in lots on this production system during a specified finite planning horizon. This paper presents a method for integrating load distribution decisions, and tactical production planning considering the costs of capacity change and the costs of unused capacity. The objective is to minimize the sum of capacity change costs, unused capacity costs, setup costs, holding costs, backorder costs, and production costs. The main constraints consist in satisfying the demand for all products over the entire horizon, and in not exceeding available repair resource. The production series-parallel system is modeled as a multi-state system with binary-state components. The proposed model takes into account the dependence of machines' failure rates on their load. Universal generating function technique can be used in the optimization algorithm for evaluating the expected system production rate in each period. We show how the formulated problem can be solved by comparing the results of several multi-product lot-sizing problems with capacity associated costs. The importance of integrating load distribution decisions and production planning is illustrated through numerical examples.

  17. A parallel 2-opt algorithm for the traveling salesman problem

    NARCIS (Netherlands)

    Verhoeven, M.G.A.; Aarts, E.H.L.; Swinkels, P.C.J.

    1995-01-01

    We present a scalable parallel local search algorithm based on data parallelism. The concept of distributed neighborhood structures is introduced, and applied to the Traveling Salesman Problem (TSP). Our parallel local search algorithm finds the same quality solutions as the classical 2-opt

  18. runjags: An R Package Providing Interface Utilities, Model Templates, Parallel Computing Methods and Additional Distributions for MCMC Models in JAGS

    Directory of Open Access Journals (Sweden)

    Matthew J. Denwood

    2016-07-01

    Full Text Available The runjags package provides a set of interface functions to facilitate running Markov chain Monte Carlo models in JAGS from within R. Automated calculation of appropriate convergence and sample length diagnostics, user-friendly access to commonly used graphical outputs and summary statistics, and parallelized methods of running JAGS are provided. Template model specifications can be generated using a standard lme4-style formula interface to assist users less familiar with the BUGS syntax. Automated simulation study functions are implemented to facilitate model performance assessment, as well as drop-k type cross-validation studies, using high performance computing clusters such as those provided by parallel. A module extension for JAGS is also included within runjags, providing the Pareto family of distributions and a series of minimally-informative priors including the DuMouchel and half-Cauchy priors. This paper outlines the primary functions of this package, and gives an illustration of a simulation study to assess the sensitivity of two equivalent model formulations to different prior distributions.

  19. MSAProbs-MPI: parallel multiple sequence aligner for distributed-memory systems.

    Science.gov (United States)

    González-Domínguez, Jorge; Liu, Yongchao; Touriño, Juan; Schmidt, Bertil

    2016-12-15

    MSAProbs is a state-of-the-art protein multiple sequence alignment tool based on hidden Markov models. It can achieve high alignment accuracy at the expense of relatively long runtimes for large-scale input datasets. In this work we present MSAProbs-MPI, a distributed-memory parallel version of the multithreaded MSAProbs tool that is able to reduce runtimes by exploiting the compute capabilities of common multicore CPU clusters. Our performance evaluation on a cluster with 32 nodes (each containing two Intel Haswell processors) shows reductions in execution time of over one order of magnitude for typical input datasets. Furthermore, MSAProbs-MPI using eight nodes is faster than the GPU-accelerated QuickProbs running on a Tesla K20. Another strong point is that MSAProbs-MPI can deal with large datasets for which MSAProbs and QuickProbs might fail due to time and memory constraints, respectively. Source code in C ++ and MPI running on Linux systems as well as a reference manual are available at http://msaprobs.sourceforge.net CONTACT: jgonzalezd@udc.esSupplementary information: Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  20. [Development and evaluation of the medical imaging distribution system with dynamic web application and clustering technology].

    Science.gov (United States)

    Yokohama, Noriya; Tsuchimoto, Tadashi; Oishi, Masamichi; Itou, Katsuya

    2007-01-20

    It has been noted that the downtime of medical informatics systems is often long. Many systems encounter downtimes of hours or even days, which can have a critical effect on daily operations. Such systems remain especially weak in the areas of database and medical imaging data. The scheme design shows the three-layer architecture of the system: application, database, and storage layers. The application layer uses the DICOM protocol (Digital Imaging and Communication in Medicine) and HTTP (Hyper Text Transport Protocol) with AJAX (Asynchronous JavaScript+XML). The database is designed to decentralize in parallel using cluster technology. Consequently, restoration of the database can be done not only with ease but also with improved retrieval speed. In the storage layer, a network RAID (Redundant Array of Independent Disks) system, it is possible to construct exabyte-scale parallel file systems that exploit storage spread. Development and evaluation of the test-bed has been successful in medical information data backup and recovery in a network environment. This paper presents a schematic design of the new medical informatics system that can be accommodated from a recovery and the dynamic Web application for medical imaging distribution using AJAX.

  1. Parallel implementation of the PHOENIX generalized stellar atmosphere program. II. Wavelength parallelization

    International Nuclear Information System (INIS)

    Baron, E.; Hauschildt, Peter H.

    1998-01-01

    We describe an important addition to the parallel implementation of our generalized nonlocal thermodynamic equilibrium (NLTE) stellar atmosphere and radiative transfer computer program PHOENIX. In a previous paper in this series we described data and task parallel algorithms we have developed for radiative transfer, spectral line opacity, and NLTE opacity and rate calculations. These algorithms divided the work spatially or by spectral lines, that is, distributing the radial zones, individual spectral lines, or characteristic rays among different processors and employ, in addition, task parallelism for logically independent functions (such as atomic and molecular line opacities). For finite, monotonic velocity fields, the radiative transfer equation is an initial value problem in wavelength, and hence each wavelength point depends upon the previous one. However, for sophisticated NLTE models of both static and moving atmospheres needed to accurately describe, e.g., novae and supernovae, the number of wavelength points is very large (200,000 - 300,000) and hence parallelization over wavelength can lead both to considerable speedup in calculation time and the ability to make use of the aggregate memory available on massively parallel supercomputers. Here, we describe an implementation of a pipelined design for the wavelength parallelization of PHOENIX, where the necessary data from the processor working on a previous wavelength point is sent to the processor working on the succeeding wavelength point as soon as it is known. Our implementation uses a MIMD design based on a relatively small number of standard message passing interface (MPI) library calls and is fully portable between serial and parallel computers. copyright 1998 The American Astronomical Society

  2. Storing files in a parallel computing system using list-based index to identify replica files

    Science.gov (United States)

    Faibish, Sorin; Bent, John M.; Tzelnic, Percy; Zhang, Zhenhua; Grider, Gary

    2015-07-21

    Improved techniques are provided for storing files in a parallel computing system using a list-based index to identify file replicas. A file and at least one replica of the file are stored in one or more storage nodes of the parallel computing system. An index for the file comprises at least one list comprising a pointer to a storage location of the file and a storage location of the at least one replica of the file. The file comprises one or more of a complete file and one or more sub-files. The index may also comprise a checksum value for one or more of the file and the replica(s) of the file. The checksum value can be evaluated to validate the file and/or the file replica(s). A query can be processed using the list.

  3. An improved power flow method based on extended chain-table storage structure for distribution network with PV nodes

    DEFF Research Database (Denmark)

    Chen, Shuheng; Wang, Xiongfei; Su, Chi

    2014-01-01

    with a reduced memory size. The voltage error of each PV node is adjusted by a reactive power adjusting strategy. The adjusting strategy is based on a multi-variable linear function with an accelerating factor. Finally, this new improved power flow method is realized by the software system developed in VC......Based on an extended chain-table storage structure, an improved power flow method is presented, which can be applied to a distribution network with multi PV nodes. The extended chain-table storage structure is designed on the basis of address-pointer technology describing the radial topology...... and the corresponding case study has been done. The experimental data and the further analysis have proved that this method can calculate the power flow of a distribution network with multi PV nodes precisely and fast. © 2014 IEEE....

  4. Storage of cold and thermal neutrons with perfect crystals at the pulsed source

    International Nuclear Information System (INIS)

    Jericha, E.

    1996-12-01

    The possibility of storing cold neutrons by sequential Bragg reflections between two parallel perfect crystal plates in backscattering geometry has been implemented as the parasitic instrument VESTA at the pulsed neutron source ISIS. Filling the neutrons into and releasing them from the storage cavity is accomplished by applying a short-pulsed magnetic field at the crystal plates. The method takes advantage of the conservation of the axial component of the neutron wave vector after Bragg reflection and its Zeeman shift in a magnetic field. The setup at ISIS is presented where a monochromatic neutron beam with wavelength 6.27 A and 2.9 x 10 4 n/scm 2 flux is taken out of the neutron guide leading to the IRIS backscattering spectrometer by a pyrolytic graphite crystal monochromator. The longest storage period obtained with the setup was 2.655 s which corresponds to 1574 consecutive Bragg reflections and a distance traveled of 1675 n. The measurements are analyzed by heuristic methods developed for neutron storage experiments. The apparatus is seen as a passive resonator system and characteristics like stored neutron intensity, the efficiency of the storage process, the probability to remain in the system, the mirror reflectivity, the dispersion of the stored distribution, the penetration depth of a neutron into a crystal mirror and the figure of merit of the resonator system are discussed. Monte Carlo simulations of the extracted beam and of the stored neutron distribution were performed to deepen the understanding of the experimental results. (author)

  5. Parallel, Asynchronous Executive (PAX): System concepts, facilities, and architecture

    Science.gov (United States)

    Jones, W. H.

    1983-01-01

    The Parallel, Asynchronous Executive (PAX) is a software operating system simulation that allows many computers to work on a single problem at the same time. PAX is currently implemented on a UNIVAC 1100/42 computer system. Independent UNIVAC runstreams are used to simulate independent computers. Data are shared among independent UNIVAC runstreams through shared mass-storage files. PAX has achieved the following: (1) applied several computing processes simultaneously to a single, logically unified problem; (2) resolved most parallel processor conflicts by careful work assignment; (3) resolved by means of worker requests to PAX all conflicts not resolved by work assignment; (4) provided fault isolation and recovery mechanisms to meet the problems of an actual parallel, asynchronous processing machine. Additionally, one real-life problem has been constructed for the PAX environment. This is CASPER, a collection of aerodynamic and structural dynamic problem simulation routines. CASPER is not discussed in this report except to provide examples of parallel-processing techniques.

  6. Scalable Parallel Distributed Coprocessor System for Graph Searching Problems with Massive Data

    Directory of Open Access Journals (Sweden)

    Wanrong Huang

    2017-01-01

    Full Text Available The Internet applications, such as network searching, electronic commerce, and modern medical applications, produce and process massive data. Considerable data parallelism exists in computation processes of data-intensive applications. A traversal algorithm, breadth-first search (BFS, is fundamental in many graph processing applications and metrics when a graph grows in scale. A variety of scientific programming methods have been proposed for accelerating and parallelizing BFS because of the poor temporal and spatial locality caused by inherent irregular memory access patterns. However, new parallel hardware could provide better improvement for scientific methods. To address small-world graph problems, we propose a scalable and novel field-programmable gate array-based heterogeneous multicore system for scientific programming. The core is multithread for streaming processing. And the communication network InfiniBand is adopted for scalability. We design a binary search algorithm to address mapping to unify all processor addresses. Within the limits permitted by the Graph500 test bench after 1D parallel hybrid BFS algorithm testing, our 8-core and 8-thread-per-core system achieved superior performance and efficiency compared with the prior work under the same degree of parallelism. Our system is efficient not as a special acceleration unit but as a processor platform that deals with graph searching applications.

  7. Is Monte Carlo embarrassingly parallel?

    Energy Technology Data Exchange (ETDEWEB)

    Hoogenboom, J. E. [Delft Univ. of Technology, Mekelweg 15, 2629 JB Delft (Netherlands); Delft Nuclear Consultancy, IJsselzoom 2, 2902 LB Capelle aan den IJssel (Netherlands)

    2012-07-01

    Monte Carlo is often stated as being embarrassingly parallel. However, running a Monte Carlo calculation, especially a reactor criticality calculation, in parallel using tens of processors shows a serious limitation in speedup and the execution time may even increase beyond a certain number of processors. In this paper the main causes of the loss of efficiency when using many processors are analyzed using a simple Monte Carlo program for criticality. The basic mechanism for parallel execution is MPI. One of the bottlenecks turn out to be the rendez-vous points in the parallel calculation used for synchronization and exchange of data between processors. This happens at least at the end of each cycle for fission source generation in order to collect the full fission source distribution for the next cycle and to estimate the effective multiplication factor, which is not only part of the requested results, but also input to the next cycle for population control. Basic improvements to overcome this limitation are suggested and tested. Also other time losses in the parallel calculation are identified. Moreover, the threading mechanism, which allows the parallel execution of tasks based on shared memory using OpenMP, is analyzed in detail. Recommendations are given to get the maximum efficiency out of a parallel Monte Carlo calculation. (authors)

  8. Is Monte Carlo embarrassingly parallel?

    International Nuclear Information System (INIS)

    Hoogenboom, J. E.

    2012-01-01

    Monte Carlo is often stated as being embarrassingly parallel. However, running a Monte Carlo calculation, especially a reactor criticality calculation, in parallel using tens of processors shows a serious limitation in speedup and the execution time may even increase beyond a certain number of processors. In this paper the main causes of the loss of efficiency when using many processors are analyzed using a simple Monte Carlo program for criticality. The basic mechanism for parallel execution is MPI. One of the bottlenecks turn out to be the rendez-vous points in the parallel calculation used for synchronization and exchange of data between processors. This happens at least at the end of each cycle for fission source generation in order to collect the full fission source distribution for the next cycle and to estimate the effective multiplication factor, which is not only part of the requested results, but also input to the next cycle for population control. Basic improvements to overcome this limitation are suggested and tested. Also other time losses in the parallel calculation are identified. Moreover, the threading mechanism, which allows the parallel execution of tasks based on shared memory using OpenMP, is analyzed in detail. Recommendations are given to get the maximum efficiency out of a parallel Monte Carlo calculation. (authors)

  9. Pell-Sim - dynamic model for forecasting storage and distribution of wood pellets

    International Nuclear Information System (INIS)

    Vinterbaeck, Johan

    2004-01-01

    This study examined the system of wood pellet distribution to residential consumers. The distribution cost for a residential pellet consumer typically represents 30% of the per tonne price and of this share, the inventory cost could be more than 50%. Important administrative activities in physical distribution are forecasting demand and inventory control. One way to improve distribution systems would be to optimise inventory management for pellet distributors. The aim of this study was to propose improvements in pellet distribution management by using tools from systems analysis. The ultimate goal was to present an optimised storage level curve adapted to the mid-Swedish community of Avesta. An internal model for optimising inventory management, Pell-Sim, was constructed, composed of two integrated parts: a simulation unit to forecast residential wood pellet demand and a spreadsheet unit with inventory-related functions. Daily outdoor temperatures basically regulated the simulation unit. An order point system was chosen for reordering. The residential customers of a distribution company were divided into two groups, delivery and collecting customers, which were statistically treated separately. When collecting and delivery customer input inventories were normally distributed in the intervals from 0 to 3500 kg and 6500 kg, respectively, their annual means of total delivery were both about 7000 kg/customer, which was the desired and empirical level. The expected pellet customer orders were negatively correlated to mean daily temperatures, lagging behind about 1 month. Sensitivity analyses showed that monthly results for ordered quantity and total cost were particularly sensitive to ordering and carrying costs. The Pell-Sim programme can easily be adapted for distributors in other geographical regions. (Author)

  10. Pell-Sim - dynamic model for forecasting storage and distribution of wood pellets

    Energy Technology Data Exchange (ETDEWEB)

    Vinterbaeck, Johan [Swedish Univ. of Agricultural Sciences, Dept. of Forest Management and Products, Uppsala (Sweden)

    2004-12-01

    This study examined the system of wood pellet distribution to residential consumers. The distribution cost for a residential pellet consumer typically represents 30% of the per tonne price and of this share, the inventory cost could be more than 50%. Important administrative activities in physical distribution are forecasting demand and inventory control. One way to improve distribution systems would be to optimise inventory management for pellet distributors. The aim of this study was to propose improvements in pellet distribution management by using tools from systems analysis. The ultimate goal was to present an optimised storage level curve adapted to the mid-Swedish community of Avesta. An internal model for optimising inventory management, Pell-Sim, was constructed, composed of two integrated parts: a simulation unit to forecast residential wood pellet demand and a spreadsheet unit with inventory-related functions. Daily outdoor temperatures basically regulated the simulation unit. An order point system was chosen for reordering. The residential customers of a distribution company were divided into two groups, delivery and collecting customers, which were statistically treated separately. When collecting and delivery customer input inventories were normally distributed in the intervals from 0 to 3500 kg and 6500 kg, respectively, their annual means of total delivery were both about 7000 kg/customer, which was the desired and empirical level. The expected pellet customer orders were negatively correlated to mean daily temperatures, lagging behind about 1 month. Sensitivity analyses showed that monthly results for ordered quantity and total cost were particularly sensitive to ordering and carrying costs. The Pell-Sim programme can easily be adapted for distributors in other geographical regions. (Author)

  11. Implementations of BLAST for parallel computers.

    Science.gov (United States)

    Jülich, A

    1995-02-01

    The BLAST sequence comparison programs have been ported to a variety of parallel computers-the shared memory machine Cray Y-MP 8/864 and the distributed memory architectures Intel iPSC/860 and nCUBE. Additionally, the programs were ported to run on workstation clusters. We explain the parallelization techniques and consider the pros and cons of these methods. The BLAST programs are very well suited for parallelization for a moderate number of processors. We illustrate our results using the program blastp as an example. As input data for blastp, a 799 residue protein query sequence and the protein database PIR were used.

  12. Incorporating solid state drives into distributed storage systems

    OpenAIRE

    Wacha, Rosie

    2012-01-01

    Big data stores are becoming increasingly important in a variety of domains including scientific computing, internet applications, and business applications. For price and performance reasons, such storage is comprised of magnetic hard drives. To achieve the necessary degree of performance and reliability, the drives are configured into storage subsystems based on RAID (Redundant Array of Independent Disks). Because of their mechanical nature, hard drives are relatively power-hungry and slow ...

  13. Cache-aware data structure model for parallelism and dynamic load balancing

    International Nuclear Information System (INIS)

    Sridi, Marwa

    2016-01-01

    This PhD thesis is dedicated to the implementation of innovative parallel methods in the framework of fast transient fluid-structure dynamics. It improves existing methods within EUROPLEXUS software, in order to optimize the shared memory parallel strategy, complementary to the original distributed memory approach, brought together into a global hybrid strategy for clusters of multi-core nodes. Starting from a sound analysis of the state of the art concerning data structuring techniques correlated to the hierarchic memory organization of current multi-processor architectures, the proposed work introduces an approach suitable for an explicit time integration (i.e. with no linear system to solve at each step). A data structure of type 'Structure of arrays' is conserved for the global data storage, providing flexibility and efficiency for current operations on kinematics fields (displacement, velocity and acceleration). On the contrary, in the particular case of elementary operations (for internal forces generic computations, as well as fluxes computations between cell faces for fluid models), particularly time consuming but localized in the program, a temporary data structure of type 'Array of structures' is used instead, to force an efficient filling of the cache memory and increase the performance of the resolution, for both serial and shared memory parallel processing. Switching from the global structure to the temporary one is based on a cell grouping strategy, following classing cache-blocking principles but handling specifically for this work neighboring data necessary to the efficient treatment of ALE fluxes for cells on the group boundaries. The proposed approach is extensively tested, from the point of views of both the computation time and the access failures into cache memory, confronting the gains obtained within the elementary operations to the potential overhead generated by the data structure switch. Obtained results are very satisfactory, especially

  14. SPINning parallel systems software

    International Nuclear Information System (INIS)

    Matlin, O.S.; Lusk, E.; McCune, W.

    2002-01-01

    We describe our experiences in using Spin to verify parts of the Multi Purpose Daemon (MPD) parallel process management system. MPD is a distributed collection of processes connected by Unix network sockets. MPD is dynamic processes and connections among them are created and destroyed as MPD is initialized, runs user processes, recovers from faults, and terminates. This dynamic nature is easily expressible in the Spin/Promela framework but poses performance and scalability challenges. We present here the results of expressing some of the parallel algorithms of MPD and executing both simulation and verification runs with Spin

  15. Development of an integrated data storage and retrieval system for TEC

    International Nuclear Information System (INIS)

    Kemmerling, G.; Blom, H.; Busch, P.; Kooijman, W.; Korten, M.; Laat, C.T.A.M. de; Lourens, W.; Meer, E. van der; Nideroest, B.; Oomens, A.A.M.; Wijnoltz, F.; Zwoll, K.

    2000-01-01

    The database system for the storage and retrieval of experimental and technical data at TEXTOR-94 has to be revised. A new database has to be developed, which complies with future performance and multiplatform requirements. The concept, to be presented here, is based on the commercial object database Objectivity. Objectivity allows a flexible object oriented data design and is able to cope with the large amount of data, which is expected to be about 1 TByte per year. Furthermore, it offers the possibility of data distribution over several hosts. Thus, parallel data storage from the frontend to the database is possible and can be used to achieve the required storage performance of 200 MByte per min. In order to store configurational and experimental data, an object model is under design. It is aimed at describing the device specific information and the acquired data in a common way such that different aproaches for data access may be applied. There are several methods forseen for remote access. In addition to the C++ and Java interfaces already included in Objectivity/DB, CORBA and socket based C interfaces are currently under development. This could also allow an access by non-supported platforms and enable existing legacy applications an integration of the database for storage and retrieval of data by a minimum of code changes

  16. Hydrogen storage in engineered carbon nanospaces.

    Science.gov (United States)

    Burress, Jacob; Kraus, Michael; Beckner, Matt; Cepel, Raina; Suppes, Galen; Wexler, Carlos; Pfeifer, Peter

    2009-05-20

    It is shown how appropriately engineered nanoporous carbons provide materials for reversible hydrogen storage, based on physisorption, with exceptional storage capacities (approximately 80 g H2/kg carbon, approximately 50 g H2/liter carbon, at 50 bar and 77 K). Nanopores generate high storage capacities (a) by having high surface area to volume ratios, and (b) by hosting deep potential wells through overlapping substrate potentials from opposite pore walls, giving rise to a binding energy nearly twice the binding energy in wide pores. Experimental case studies are presented with surface areas as high as 3100 m(2) g(-1), in which 40% of all surface sites reside in pores of width approximately 0.7 nm and binding energy approximately 9 kJ mol(-1), and 60% of sites in pores of width>1.0 nm and binding energy approximately 5 kJ mol(-1). The findings, including the prevalence of just two distinct binding energies, are in excellent agreement with results from molecular dynamics simulations. It is also shown, from statistical mechanical models, that one can experimentally distinguish between the situation in which molecules do (mobile adsorption) and do not (localized adsorption) move parallel to the surface, how such lateral dynamics affects the hydrogen storage capacity, and how the two situations are controlled by the vibrational frequencies of adsorbed hydrogen molecules parallel and perpendicular to the surface: in the samples presented, adsorption is mobile at 293 K, and localized at 77 K. These findings make a strong case for it being possible to significantly increase hydrogen storage capacities in nanoporous carbons by suitable engineering of the nanopore space.

  17. PRELIMINARY IMPEDANCE BUDGET FOR NSLS-II STORAGE RING.

    Energy Technology Data Exchange (ETDEWEB)

    BLEDNYKH,A.; KRINSKY, S.

    2007-06-25

    The wakefield and impedance produced by the components of the NSLS-II storage ring have been computed for an electron bunch length of 3mm rms. The results are summarized in a table giving for each component, the loss factor ({kappa}{sub {parallel}}), the imaginary part of the longitudinal impedance at low frequency divided by the revolution harmonic (ImZ{sub {parallel}}/n), and the transverse kick factors ({kappa}{sub x}, {kappa}{sub y}).

  18. Distributed-memory matrix computations

    DEFF Research Database (Denmark)

    Balle, Susanne Mølleskov

    1995-01-01

    The main goal of this project is to investigate, develop, and implement algorithms for numerical linear algebra on parallel computers in order to acquire expertise in methods for parallel computations. An important motivation for analyzaing and investigating the potential for parallelism in these......The main goal of this project is to investigate, develop, and implement algorithms for numerical linear algebra on parallel computers in order to acquire expertise in methods for parallel computations. An important motivation for analyzaing and investigating the potential for parallelism...... in these algorithms is that many scientific applications rely heavily on the performance of the involved dense linear algebra building blocks. Even though we consider the distributed-memory as well as the shared-memory programming paradigm, the major part of the thesis is dedicated to distributed-memory architectures....... We emphasize distributed-memory massively parallel computers - such as the Connection Machines model CM-200 and model CM-5/CM-5E - available to us at UNI-C and at Thinking Machines Corporation. The CM-200 was at the time this project started one of the few existing massively parallel computers...

  19. An efficient parallel algorithm: Poststack and prestack Kirchhoff 3D depth migration using flexi-depth iterations

    Science.gov (United States)

    Rastogi, Richa; Srivastava, Abhishek; Khonde, Kiran; Sirasala, Kirannmayi M.; Londhe, Ashutosh; Chavhan, Hitesh

    2015-07-01

    This paper presents an efficient parallel 3D Kirchhoff depth migration algorithm suitable for current class of multicore architecture. The fundamental Kirchhoff depth migration algorithm exhibits inherent parallelism however, when it comes to 3D data migration, as the data size increases the resource requirement of the algorithm also increases. This challenges its practical implementation even on current generation high performance computing systems. Therefore a smart parallelization approach is essential to handle 3D data for migration. The most compute intensive part of Kirchhoff depth migration algorithm is the calculation of traveltime tables due to its resource requirements such as memory/storage and I/O. In the current research work, we target this area and develop a competent parallel algorithm for post and prestack 3D Kirchhoff depth migration, using hybrid MPI+OpenMP programming techniques. We introduce a concept of flexi-depth iterations while depth migrating data in parallel imaging space, using optimized traveltime table computations. This concept provides flexibility to the algorithm by migrating data in a number of depth iterations, which depends upon the available node memory and the size of data to be migrated during runtime. Furthermore, it minimizes the requirements of storage, I/O and inter-node communication, thus making it advantageous over the conventional parallelization approaches. The developed parallel algorithm is demonstrated and analysed on Yuva II, a PARAM series of supercomputers. Optimization, performance and scalability experiment results along with the migration outcome show the effectiveness of the parallel algorithm.

  20. Distributed computing feasibility in a non-dedicated homogeneous distributed system

    Science.gov (United States)

    Leutenegger, Scott T.; Sun, Xian-He

    1993-01-01

    The low cost and availability of clusters of workstations have lead researchers to re-explore distributed computing using independent workstations. This approach may provide better cost/performance than tightly coupled multiprocessors. In practice, this approach often utilizes wasted cycles to run parallel jobs. The feasibility of such a non-dedicated parallel processing environment assuming workstation processes have preemptive priority over parallel tasks is addressed. An analytical model is developed to predict parallel job response times. Our model provides insight into how significantly workstation owner interference degrades parallel program performance. A new term task ratio, which relates the parallel task demand to the mean service demand of nonparallel workstation processes, is introduced. It was proposed that task ratio is a useful metric for determining how large the demand of a parallel applications must be in order to make efficient use of a non-dedicated distributed system.

  1. State-Of-The-Art in Microgrid-Integrated Distributed Energy Storage Sizing

    Directory of Open Access Journals (Sweden)

    Ibrahim Alsaidan

    2017-09-01

    Full Text Available Distributed energy storage (DES plays an important role in microgrid operation and control, as it can potentially improve local reliability and resilience, reduce operation cost, and mitigate challenges caused by high penetration renewable generation. However, to ensure an acceptable economic and technical performance, DES must be optimally sized and placed. This paper reviews the existing DES sizing methods for microgrid applications and presents a generic sizing method that enables microgrid planners to efficiently determine the optimal DES size, technology, and location. The proposed method takes into consideration the impact of DES operation on its lifetime to enhance the obtained results accuracy and practicality. The presented model can be used for both grid-tied (considering both grid-connected and islanded modes and isolated microgrids.

  2. Hierarchical Parallel Matrix Multiplication on Large-Scale Distributed Memory Platforms

    KAUST Repository

    Quintin, Jean-Noel

    2013-10-01

    Matrix multiplication is a very important computation kernel both in its own right as a building block of many scientific applications and as a popular representative for other scientific applications. Cannon\\'s algorithm which dates back to 1969 was the first efficient algorithm for parallel matrix multiplication providing theoretically optimal communication cost. However this algorithm requires a square number of processors. In the mid-1990s, the SUMMA algorithm was introduced. SUMMA overcomes the shortcomings of Cannon\\'s algorithm as it can be used on a nonsquare number of processors as well. Since then the number of processors in HPC platforms has increased by two orders of magnitude making the contribution of communication in the overall execution time more significant. Therefore, the state of the art parallel matrix multiplication algorithms should be revisited to reduce the communication cost further. This paper introduces a new parallel matrix multiplication algorithm, Hierarchical SUMMA (HSUMMA), which is a redesign of SUMMA. Our algorithm reduces the communication cost of SUMMA by introducing a two-level virtual hierarchy into the two-dimensional arrangement of processors. Experiments on an IBM BlueGene/P demonstrate the reduction of communication cost up to 2.08 times on 2048 cores and up to 5.89 times on 16384 cores. © 2013 IEEE.

  3. Hierarchical Parallel Matrix Multiplication on Large-Scale Distributed Memory Platforms

    KAUST Repository

    Quintin, Jean-Noel; Hasanov, Khalid; Lastovetsky, Alexey

    2013-01-01

    Matrix multiplication is a very important computation kernel both in its own right as a building block of many scientific applications and as a popular representative for other scientific applications. Cannon's algorithm which dates back to 1969 was the first efficient algorithm for parallel matrix multiplication providing theoretically optimal communication cost. However this algorithm requires a square number of processors. In the mid-1990s, the SUMMA algorithm was introduced. SUMMA overcomes the shortcomings of Cannon's algorithm as it can be used on a nonsquare number of processors as well. Since then the number of processors in HPC platforms has increased by two orders of magnitude making the contribution of communication in the overall execution time more significant. Therefore, the state of the art parallel matrix multiplication algorithms should be revisited to reduce the communication cost further. This paper introduces a new parallel matrix multiplication algorithm, Hierarchical SUMMA (HSUMMA), which is a redesign of SUMMA. Our algorithm reduces the communication cost of SUMMA by introducing a two-level virtual hierarchy into the two-dimensional arrangement of processors. Experiments on an IBM BlueGene/P demonstrate the reduction of communication cost up to 2.08 times on 2048 cores and up to 5.89 times on 16384 cores. © 2013 IEEE.

  4. Soil Nitrogen Storage, Distribution, and Associated Controlling Factors in the Northeast Tibetan Plateau Shrublands

    Directory of Open Access Journals (Sweden)

    Xiuqing Nie

    2017-11-01

    Full Text Available Although the soils in the Tibetan Plateau shrublands store large amounts of total nitrogen (N, the estimated values remain uncertain because of spatial heterogeneity and a lack of field observations. In this study, we quantified the regional soil N storage, spatial and vertical density distributions, and related climatic controls using 183 soil profiles sampled from 61 sites across the Northeast Tibetan Plateau shrublands during the period of 2011–2013. Our analysis revealed a soil N storage value of 132.40 Tg at a depth of 100 cm, with an average density of 1.21 kg m−2. Soil N density was distributed at greater levels in alpine shrublands, compared with desert shrublands. Spatially, soil N densities decreased from south to north and from east to west, and, vertically, the soil N in the upper 30 and 50 cm accounted for 42% and 64% of the total soil N stocks in the Tibetan Plateau. However, compared with desert shrublands, the surface layers in alpine shrublands exhibited a larger distribution of soil N stocks. Overall, the soil N density in the top 30 cm increased significantly with the mean annual precipitation (MAP and tended to decrease with the mean annual temperature (MAT, although the dominant climatic controls differed among shrubland types. Specifically, MAP in alpine shrublands, and MAT in desert shrubland, had a weak effect on N density. Soil pH can significant affect soil N density in the Tibetan Plateau shrublands. In conclusion, changes in soil N density should be monitored over the long term to provide accurate information about the effects of climatic factors.

  5. OpenMP Issues Arising in the Development of Parallel BLAS and LAPACK Libraries

    Directory of Open Access Journals (Sweden)

    C. Addison

    2003-01-01

    Full Text Available Dense linear algebra libraries need to cope efficiently with a range of input problem sizes and shapes. Inherently this means that parallel implementations have to exploit parallelism wherever it is present. While OpenMP allows relatively fine grain parallelism to be exploited in a shared memory environment it currently lacks features to make it easy to partition computation over multiple array indices or to overlap sequential and parallel computations. The inherent flexible nature of shared memory paradigms such as OpenMP poses other difficulties when it becomes necessary to optimise performance across successive parallel library calls. Notions borrowed from distributed memory paradigms, such as explicit data distributions help address some of these problems, but the focus on data rather than work distribution appears misplaced in an SMP context.

  6. Heat transport and storage

    International Nuclear Information System (INIS)

    Despois, J.

    1977-01-01

    Recalling the close connections existing between heat transport and storage, some general considerations on the problem of heat distribution and transport are presented 'in order to set out the problem' of storage in concrete form. This problem is considered in its overall plane, then studied under the angle of the different technical choices it involves. The two alternatives currently in consideration are described i.e.: storage in a mined cavity and underground storage as captive sheet [fr

  7. Improving hydrogen storage in Ni-doped carbon nanospheres

    Energy Technology Data Exchange (ETDEWEB)

    Zubizarreta, L.; Menendez, J.A.; Pis, J.J.; Arenillas, A. [Instituto Nacional del Carbon, CSIC, Apartado 73, 33080 Oviedo (Spain)

    2009-04-15

    The effect of nickel distribution and content in Ni-doped carbon nanospheres on hydrogen storage capacity under conditions of moderate temperature and pressure was studied. It was found that the nickel distribution, obtained by using different doping techniques and conditions, has a noticeable influence on hydrogen storage capacity. The samples with the most homogeneous nickel distribution, obtained by pre-oxidising the carbon nanospheres, displayed the highest storage capacity. In addition, storage capacity is influenced by the amount of nickel. It was found a higher storage capacity in samples containing 5 wt.% of Ni. This is due to the greater interactions between the nickel and the support that produce a higher activation of the solid through a spillover effect. (author)

  8. Parallel Tensor Compression for Large-Scale Scientific Data.

    Energy Technology Data Exchange (ETDEWEB)

    Kolda, Tamara G. [Sandia National Lab. (SNL-CA), Livermore, CA (United States); Ballard, Grey [Sandia National Lab. (SNL-CA), Livermore, CA (United States); Austin, Woody Nathan [Univ. of Texas, Austin, TX (United States)

    2015-10-01

    As parallel computing trends towards the exascale, scientific data produced by high-fidelity simulations are growing increasingly massive. For instance, a simulation on a three-dimensional spatial grid with 512 points per dimension that tracks 64 variables per grid point for 128 time steps yields 8 TB of data. By viewing the data as a dense five way tensor, we can compute a Tucker decomposition to find inherent low-dimensional multilinear structure, achieving compression ratios of up to 10000 on real-world data sets with negligible loss in accuracy. So that we can operate on such massive data, we present the first-ever distributed memory parallel implementation for the Tucker decomposition, whose key computations correspond to parallel linear algebra operations, albeit with nonstandard data layouts. Our approach specifies a data distribution for tensors that avoids any tensor data redistribution, either locally or in parallel. We provide accompanying analysis of the computation and communication costs of the algorithms. To demonstrate the compression and accuracy of the method, we apply our approach to real-world data sets from combustion science simulations. We also provide detailed performance results, including parallel performance in both weak and strong scaling experiments.

  9. Holographic memory for high-density data storage and high-speed pattern recognition

    Science.gov (United States)

    Gu, Claire

    2002-09-01

    As computers and the internet become faster and faster, more and more information is transmitted, received, and stored everyday. The demand for high density and fast access time data storage is pushing scientists and engineers to explore all possible approaches including magnetic, mechanical, optical, etc. Optical data storage has already demonstrated its potential in the competition against other storage technologies. CD and DVD are showing their advantages in the computer and entertainment market. What motivated the use of optical waves to store and access information is the same as the motivation for optical communication. Light or an optical wave has an enormous capacity (or bandwidth) to carry information because of its short wavelength and parallel nature. In optical storage, there are two types of mechanism, namely localized and holographic memories. What gives the holographic data storage an advantage over localized bit storage is the natural ability to read the stored information in parallel, therefore, meeting the demand for fast access. Another unique feature that makes the holographic data storage attractive is that it is capable of performing associative recall at an incomparable speed. Therefore, volume holographic memory is particularly suitable for high-density data storage and high-speed pattern recognition. In this paper, we review previous works on volume holographic memories and discuss the challenges for this technology to become a reality.

  10. BCYCLIC: A parallel block tridiagonal matrix cyclic solver

    Science.gov (United States)

    Hirshman, S. P.; Perumalla, K. S.; Lynch, V. E.; Sanchez, R.

    2010-09-01

    A block tridiagonal matrix is factored with minimal fill-in using a cyclic reduction algorithm that is easily parallelized. Storage of the factored blocks allows the application of the inverse to multiple right-hand sides which may not be known at factorization time. Scalability with the number of block rows is achieved with cyclic reduction, while scalability with the block size is achieved using multithreaded routines (OpenMP, GotoBLAS) for block matrix manipulation. This dual scalability is a noteworthy feature of this new solver, as well as its ability to efficiently handle arbitrary (non-powers-of-2) block row and processor numbers. Comparison with a state-of-the art parallel sparse solver is presented. It is expected that this new solver will allow many physical applications to optimally use the parallel resources on current supercomputers. Example usage of the solver in magneto-hydrodynamic (MHD), three-dimensional equilibrium solvers for high-temperature fusion plasmas is cited.

  11. Cold chain management in meat storage, distribution and retail: A review

    Science.gov (United States)

    Nastasijević, I.; Lakićević, B.; Petrović, Z.

    2017-09-01

    Meat is a perishable product with a short shelf life and therefore short selling times. Therefore, cold chain management in meat supply is of utmost importance for the maintenance of quality and safety of meat/meat products. Raw meat/meat products are likely to support the growth of pathogenic microorganisms and/or spoilage bacteria, and should be kept at temperatures that do not result in a risk to health. The cold chain should not be interrupted at all times along the meat distribution chain. The complexity of global meat supply chain, with frequently long distribution chains associated with transportation of the product within one country, from one to another country and from one to another continent, makes the solutions for the chilling and freezing regimes, as well as monitoring of time-temperature profiles, very important for the overall success in delivery of product which will be accepted by consumer for its freshness and safety levels. From recently, there are several available options for control and management of the cold chain, such as chilled and frozen storage combinations, superchilling, ionizing radiation, biopreservation, high hydrostatic pressure (HHP), active packaging, wireless sensors, supported with the software-based cold chain database (CCD).

  12. JTpack90: A parallel, object-based, Fortran 90 linear algebra package

    Energy Technology Data Exchange (ETDEWEB)

    Turner, J.A.; Kothe, D.B. [Los Alamos National Lab., NM (United States); Ferrell, R.C. [Cambridge Power Computing Associates, Ltd., Brookline, MA (United States)

    1997-03-01

    The authors have developed an object-based linear algebra package, currently with emphasis on sparse Krylov methods, driven primarily by needs of the Los Alamos National Laboratory parallel unstructured-mesh casting simulation tool Telluride. Support for a number of sparse storage formats, methods, and preconditioners have been implemented, driven primarily by application needs. They describe the object-based Fortran 90 approach, which enhances maintainability, performance, and extensibility, the parallelization approach using a new portable gather/scatter library (PGSLib), current capabilities and future plans, and present preliminary performance results on a variety of platforms.

  13. Time complexity analysis for distributed memory computers: implementation of parallel conjugate gradient method

    NARCIS (Netherlands)

    Hoekstra, A.G.; Sloot, P.M.A.; Haan, M.J.; Hertzberger, L.O.; van Leeuwen, J.

    1991-01-01

    New developments in Computer Science, both hardware and software, offer researchers, such as physicists, unprecedented possibilities to solve their computational intensive problems.However, full exploitation of e.g. new massively parallel computers, parallel languages or runtime environments

  14. Parallel transport of long mean-free-path plasma along open magnetic field lines: Parallel heat flux

    International Nuclear Information System (INIS)

    Guo Zehua; Tang Xianzhu

    2012-01-01

    In a long mean-free-path plasma where temperature anisotropy can be sustained, the parallel heat flux has two components with one associated with the parallel thermal energy and the other the perpendicular thermal energy. Due to the large deviation of the distribution function from local Maxwellian in an open field line plasma with low collisionality, the conventional perturbative calculation of the parallel heat flux closure in its local or non-local form is no longer applicable. Here, a non-perturbative calculation is presented for a collisionless plasma in a two-dimensional flux expander bounded by absorbing walls. Specifically, closures of previously unfamiliar form are obtained for ions and electrons, which relate two distinct components of the species parallel heat flux to the lower order fluid moments such as density, parallel flow, parallel and perpendicular temperatures, and the field quantities such as the magnetic field strength and the electrostatic potential. The plasma source and boundary condition at the absorbing wall enter explicitly in the closure calculation. Although the closure calculation does not take into account wave-particle interactions, the results based on passing orbits from steady-state collisionless drift-kinetic equation show remarkable agreement with fully kinetic-Maxwell simulations. As an example of the physical implications of the theory, the parallel heat flux closures are found to predict a surprising observation in the kinetic-Maxwell simulation of the 2D magnetic flux expander problem, where the parallel heat flux of the parallel thermal energy flows from low to high parallel temperature region.

  15. Parallel Molecular Distributed Detection With Brownian Motion.

    Science.gov (United States)

    Rogers, Uri; Koh, Min-Sung

    2016-12-01

    This paper explores the in vivo distributed detection of an undesired biological agent's (BAs) biomarkers by a group of biological sized nanomachines in an aqueous medium under drift. The term distributed, indicates that the system information relative to the BAs presence is dispersed across the collection of nanomachines, where each nanomachine possesses limited communication, computation, and movement capabilities. Using Brownian motion with drift, a probabilistic detection and optimal data fusion framework, coined molecular distributed detection, will be introduced that combines theory from both molecular communication and distributed detection. Using the optimal data fusion framework as a guide, simulation indicates that a sub-optimal fusion method exists, allowing for a significant reduction in implementation complexity while retaining BA detection accuracy.

  16. 14th ACIS/IEEE International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing

    CERN Document Server

    Studies in Computational Intelligence : Volume 492

    2013-01-01

    This edited book presents scientific results of the 14th ACIS/IEEE International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing (SNPD 2013), held in Honolulu, Hawaii, USA on July 1-3, 2013. The aim of this conference was to bring together scientists, engineers, computer users, and students to share their experiences and exchange new ideas, research results about all aspects (theory, applications and tools) of computer and information science, and to discuss the practical challenges encountered along the way and the solutions adopted to solve them. The conference organizers selected the 17 outstanding papers from those papers accepted for presentation at the conference.  

  17. 15th IEEE/ACIS International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing

    CERN Document Server

    2015-01-01

    This edited book presents scientific results of 15th IEEE/ACIS International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing (SNPD 2014) held on June 30 – July 2, 2014 in Las Vegas Nevada, USA. The aim of this conference was to bring together scientists, engineers, computer users, and students to share their experiences and exchange new ideas, research results about all aspects (theory, applications and tools) of computer and information science, and to discuss the practical challenges encountered along the way and the solutions adopted to solve them. The conference organizers selected the 13 outstanding papers from those papers accepted for presentation at the conference.

  18. Acceleration and parallelization calculation of EFEN-SP_3 method

    International Nuclear Information System (INIS)

    Yang Wen; Zheng Youqi; Wu Hongchun; Cao Liangzhi; Li Yunzhao

    2013-01-01

    Due to the fact that the exponential function expansion nodal-SP_3 (EFEN-SP_3) method needs further improvement in computational efficiency to routinely carry out PWR whole core pin-by-pin calculation, the coarse mesh acceleration and spatial parallelization were investigated in this paper. The coarse mesh acceleration was built by considering discontinuity factor on each coarse mesh interface and preserving neutron balance within each coarse mesh in space, angle and energy. The spatial parallelization based on MPI was implemented by guaranteeing load balancing and minimizing communications cost to fully take advantage of the modern computing and storage abilities. Numerical results based on a commercial nuclear power reactor demonstrate an speedup ratio of about 40 for the coarse mesh acceleration and a parallel efficiency of higher than 60% with 40 CPUs for the spatial parallelization. With these two improvements, the EFEN code can complete a PWR whole core pin-by-pin calculation with 289 × 289 × 218 meshes and 4 energy groups within 100 s by using 48 CPUs (2.40 GHz frequency). (authors)

  19. Environmental consequences of parallel marketing of LPG in India

    International Nuclear Information System (INIS)

    Das, T.K.; Deb, C.; Jash, T.

    2000-01-01

    In 1993, the government of India introduced a parallel marketing system for petroleum products, particularly for liquefied petroleum gas (LPG). LPG, under the parallel marketing system, is distributed at higher price without any subsidy. The government distributors, however, distribute LPG at subsidized rate. The consumers who get LPG at higher price, therefore, opt for a mixed energy use. The present study, based on a field survey, deals with the pattern of this mixed energy use and its environmental consequences. (author)

  20. 21 CFR 1271.260 - Storage.

    Science.gov (United States)

    2010-04-01

    ... TISSUE-BASED PRODUCTS Current Good Tissue Practice § 1271.260 Storage. (a) Control of storage areas. You... distribution. (b) Temperature. You must store HCT/Ps at an appropriate temperature. (c) Expiration date. Where...

  1. Optimal sizing of energy storage system for microgrids

    Indian Academy of Sciences (India)

    Microgrids (MGs) are Low Voltage distribution networks comprising various distributed generators (DG), storage devices and controllable loads that can operate either interconnected or isolated from the main distribution grid as a controlled entity. Energy storage system (ESS) is a vital part of an MG. In this paper, a ...

  2. A multipurpose computing center with distributed resources

    Science.gov (United States)

    Chudoba, J.; Adam, M.; Adamová, D.; Kouba, T.; Mikula, A.; Říkal, V.; Švec, J.; Uhlířová, J.; Vokáč, P.; Svatoš, M.

    2017-10-01

    The Computing Center of the Institute of Physics (CC IoP) of the Czech Academy of Sciences serves a broad spectrum of users with various computing needs. It runs WLCG Tier-2 center for the ALICE and the ATLAS experiments; the same group of services is used by astroparticle physics projects the Pierre Auger Observatory (PAO) and the Cherenkov Telescope Array (CTA). OSG stack is installed for the NOvA experiment. Other groups of users use directly local batch system. Storage capacity is distributed to several locations. DPM servers used by the ATLAS and the PAO are all in the same server room, but several xrootd servers for the ALICE experiment are operated in the Nuclear Physics Institute in Řež, about 10 km away. The storage capacity for the ATLAS and the PAO is extended by resources of the CESNET - the Czech National Grid Initiative representative. Those resources are in Plzen and Jihlava, more than 100 km away from the CC IoP. Both distant sites use a hierarchical storage solution based on disks and tapes. They installed one common dCache instance, which is published in the CC IoP BDII. ATLAS users can use these resources using the standard ATLAS tools in the same way as the local storage without noticing this geographical distribution. Computing clusters LUNA and EXMAG dedicated to users mostly from the Solid State Physics departments offer resources for parallel computing. They are part of the Czech NGI infrastructure MetaCentrum with distributed batch system based on torque with a custom scheduler. Clusters are installed remotely by the MetaCentrum team and a local contact helps only when needed. Users from IoP have exclusive access only to a part of these two clusters and take advantage of higher priorities on the rest (1500 cores in total), which can also be used by any user of the MetaCentrum. IoP researchers can also use distant resources located in several towns of the Czech Republic with a capacity of more than 12000 cores in total.

  3. Two-state ion heating at quasi-parallel shocks

    International Nuclear Information System (INIS)

    Thomsen, M.F.; Gosling, J.T.; Bame, S.J.; Onsager, T.G.; Russell, C.T.

    1990-01-01

    In a previous study of ion heating at quasi-parallel shocks, the authors showed a case in which the ion distributions downstream from the shock alternated between a cooler, denser, core/shoulder type and a hotter, less dense, more Maxwellian type. In this paper they further document the alternating occurrence of two different ion states downstream from several quasi-parallel shocks. Three separate lines of evidence are presented to show that the two states are not related in an evolutionary sense, but rather both are produced alternately at the shock: (1) the asymptotic downstream plasma parameters (density, ion temperature, and flow speed) are intermediate between those characterizing the two different states closer to the shock, suggesting that the asymptotic state is produced by a mixing of the two initial states; (2) examples of apparently interpenetrating (i.e., mixing) distributions can be found during transitions from one state to the other; and (3) examples of both types of distributions can be found at actual crossings of the shock ramp. The alternation between the two different types of ion distribution provides direct observational support for the idea that the dissipative dynamics of at least some quasi-parallel shocks is non-stationary and cyclic in nature, as demonstrated by recent numerical simulations. Typical cycle times between intervals of similar ion heating states are ∼2 upstream ion gyroperiods. Both the simulations and the in situ observations indicate that a process of coherent ion reflection is commonly an important part of the dissipation at quasi-parallel shocks

  4. Storm: A Manager for Storage Resource in Grid

    International Nuclear Information System (INIS)

    Ghiselli, A.; Magnoni, L.; Zappi, R.

    2009-01-01

    Nowadays, data intensive applications demand high-performance and large-storage systems capable of serving up to various Peta bytes of storage space. Therefore, common solutions adopted in data centres include Storage Area Networks (San) and cluster parallel file systems, such as GPFS from IBM and Lustre from Sun Microsystems. In order to make these storage system solutions available in modern Data Grid architectures, standard interfaces are needed. The Grid Storage Resource Manager (SRM) interface is one of these standard interfaces. Grid storage services implementing the SRM standard provide common capabilities and advanced functionality such as dynamic space allocation and file management on shared storage systems. In this paper, we describe Storm (Storage Resource Manager). Storm is a flexible and high-performing implementation of the standard SRM interface version 2.2. The software architecture of Storm allows for an easy integration to different underlying storage systems via a plug-in mechanism. In particular, Storm takes advantage from storage systems based on cluster file systems. Currently, Storm is installed and used in production in various data centres, including the WLCG Italian Tier-1. In addition, Economics and Financial communities, as represented by the EGRID Project, adopt Storm in production as well.

  5. Language constructs for modular parallel programs

    Energy Technology Data Exchange (ETDEWEB)

    Foster, I.

    1996-03-01

    We describe programming language constructs that facilitate the application of modular design techniques in parallel programming. These constructs allow us to isolate resource management and processor scheduling decisions from the specification of individual modules, which can themselves encapsulate design decisions concerned with concurrence, communication, process mapping, and data distribution. This approach permits development of libraries of reusable parallel program components and the reuse of these components in different contexts. In particular, alternative mapping strategies can be explored without modifying other aspects of program logic. We describe how these constructs are incorporated in two practical parallel programming languages, PCN and Fortran M. Compilers have been developed for both languages, allowing experimentation in substantial applications.

  6. Southern company energy storage study :

    Energy Technology Data Exchange (ETDEWEB)

    Ellison, James; Bhatnagar, Dhruv; Black, Clifton; Jenkins, Kip

    2013-03-01

    This study evaluates the business case for additional bulk electric energy storage in the Southern Company service territory for the year 2020. The model was used to examine how system operations are likely to change as additional storage is added. The storage resources were allowed to provide energy time shift, regulation reserve, and spinning reserve services. Several storage facilities, including pumped hydroelectric systems, flywheels, and bulk-scale batteries, were considered. These scenarios were tested against a range of sensitivities: three different natural gas price assumptions, a 15% decrease in coal-fired generation capacity, and a high renewable penetration (10% of total generation from wind energy). Only in the elevated natural gas price sensitivities did some of the additional bulk-scale storage projects appear justifiable on the basis of projected production cost savings. Enabling existing peak shaving hydroelectric plants to provide regulation and spinning reserve, however, is likely to provide savings that justify the project cost even at anticipated natural gas price levels. Transmission and distribution applications of storage were not examined in this study. Allowing new storage facilities to serve both bulk grid and transmission/distribution-level needs may provide for increased benefit streams, and thus make a stronger business case for additional storage.

  7. The Grid Enabled Mass Storage System (GEMSS): the Storage and Data management system used at the INFN Tier1 at CNAF

    International Nuclear Information System (INIS)

    Ricci, Pier Paolo; Cavalli, Alessandro; Dell'Agnello, Luca; Gregori, Daniele; Prosperini, Andrea; Rinaldi, Lorenzo; Sapunenko, Vladimir; Bonacorsi, Daniele; Vagnoni, Vincenzo

    2012-01-01

    The storage system currently used in production at the INFN Tier1 at CNAF is the result of several years of case studies, software development and tests. This solution, called the Grid Enabled Mass Storage System (GEMSS), is based on a custom integration between a fast and reliable parallel filesystem (the IBM General Parallel File System, GPFS), with a complete integrated tape backend based on the Tivoli Storage Manager (TSM), which provides Hierarchical Storage Management (HSM) capabilities, and the Grid Storage Resource Manager (StoRM), providing access to grid users through a standard SRM interface. Since the start of the Large Hadron Collider (LHC) operation, all LHC experiments have been using GEMSS at CNAF for both disk data access and long-term archival on tape media. Moreover, during last year, GEMSS has become the standard solution for all other experiments hosted at CNAF, allowing the definitive consolidation of the data storage layer. Our choice has proved to be very successful during the last two years of production with continuous enhancements, accurate monitoring and effective customizations according to the end-user requests. In this paper a description of the system is reported, addressing recent developments and giving an overview of the administration and monitoring tools. We also discuss the solutions adopted in order to grant the maximum availability of the service and the latest optimization features within the data access process. Finally, we summarize the main results obtained during these last years of activity from the perspective of some of the end-users, showing the reliability and the high performances that can be achieved using GEMSS.

  8. Energy Conversion and Transmission Characteristics Analysis of Ice Storage Air Conditioning System Driven by Distributed Photovoltaic Energy System

    Directory of Open Access Journals (Sweden)

    Yongfeng Xu

    2016-01-01

    Full Text Available In order to reduce the investment and operation cost of distributed PV energy system, ice storage technology was introduced to substitute batteries for solar energy storage. Firstly, the ice storage air conditioning system (ISACS driven by distributed photovoltaic energy system (DPES was proposed and the feasibility studies have been investigated in this paper. And then, the theoretical model has been established and experimental work has been done to analyze the energy coupling and transferring characteristics in light-electricity-cold conversion process. In addition, the structure optimization analysis was investigated. Results revealed that energy losses were high in ice making process of ice slide maker with only 17.38% energy utilization efficiency and the energy efficiency and exergy efficiency of ISACS driven by DPES were 5.44% and 67.30%, respectively. So the immersed evaporator and cointegrated exchanger were adopted for higher energy utilization efficiency and better financial rewards in structure optimization. The COP and exergy efficiency of ice maker can be increased to 1.48 and 81.24%, respectively, after optimization and the energy utilization efficiency of ISACS driven by DPES could be improved 2.88 times. Moreover, ISACS has the out-of-the-box function of ordinary air conditioning system. In conclusion, ISACS driven by DPES will have good application prospects in tropical regions without power grid.

  9. Development of large scale fusion plasma simulation and storage grid on JAERI Origin3800 system

    International Nuclear Information System (INIS)

    Idomura, Yasuhiro; Wang, Xin

    2003-01-01

    Under the Numerical EXperiment of Tokamak (NEXT) research project, various fluid, particle, and hybrid codes have been developed. These codes require a computational environment which consists of high performance processors, high speed storage system, and high speed parallelized visualization system. In this paper, the performance of the JAERI Origin3800 system is examined from a point of view of these requests. In the performance tests, it is shown that the representative particle and fluid codes operate with 15 - 40% of processing efficiency up to 512 processors. A storage area network (SAN) provides high speed parallel data transfer. A parallel visualization system enables order to magnitude faster visualization of a large scale simulation data compared with the previous graphic workstations. Accordingly, an extremely advanced simulation environment is realized on the JAERI Origin3800 system. Recently, development of a storage grid is underway in order to improve a computational environment of remote users. The storage grid is constructed by a combination of SAN and a wavelength division multiplexer (WDM). The preliminary tests show that compared with the existing data transfer methods, it enables dramatically high speed data transfer ∼100 Gbps over a wide area network. (author)

  10. Parallel implementations of 2D explicit Euler solvers

    International Nuclear Information System (INIS)

    Giraud, L.; Manzini, G.

    1996-01-01

    In this work we present a subdomain partitioning strategy applied to an explicit high-resolution Euler solver. We describe the design of a portable parallel multi-domain code suitable for parallel environments. We present several implementations on a representative range of MlMD computers that include shared memory multiprocessors, distributed virtual shared memory computers, as well as networks of workstations. Computational results are given to illustrate the efficiency, the scalability, and the limitations of the different approaches. We discuss also the effect of the communication protocol on the optimal domain partitioning strategy for the distributed memory computers

  11. Parallel and non-parallel laminar mixed convection flow in an inclined tube: The effect of the boundary conditions

    International Nuclear Information System (INIS)

    Barletta, A.

    2008-01-01

    The necessary condition for the onset of parallel flow in the fully developed region of an inclined duct is applied to the case of a circular tube. Parallel flow in inclined ducts is an uncommon regime, since in most cases buoyancy tends to produce the onset of secondary flow. The present study shows how proper thermal boundary conditions may preserve parallel flow regime. Mixed convection flow is studied for a special non-axisymmetric thermal boundary condition that, with a proper choice of a switch parameter, may be compatible with parallel flow. More precisely, a circumferentially variable heat flux distribution is prescribed on the tube wall, expressed as a sinusoidal function of the azimuthal coordinate θ with period 2π. A π/2 rotation in the position of the maximum heat flux, achieved by setting the switch parameter, may allow or not the existence of parallel flow. Two cases are considered corresponding to parallel and non-parallel flow. In the first case, the governing balance equations allow a simple analytical solution. On the contrary, in the second case, the local balance equations are solved numerically by employing a finite element method

  12. Deployment strategy for battery energy storage system in distribution network based on voltage violation regulation

    Science.gov (United States)

    Wu, H.; Zhou, L.; Xu, T.; Fang, W. L.; He, W. G.; Liu, H. M.

    2017-11-01

    In order to improve the situation of voltage violation caused by the grid-connection of photovoltaic (PV) system in a distribution network, a bi-level programming model is proposed for battery energy storage system (BESS) deployment. The objective function of inner level programming is to minimize voltage violation, with the power of PV and BESS as the variables. The objective function of outer level programming is to minimize the comprehensive function originated from inner layer programming and all the BESS operating parameters, with the capacity and rated power of BESS as the variables. The differential evolution (DE) algorithm is applied to solve the model. Based on distribution network operation scenarios with photovoltaic generation under multiple alternative output modes, the simulation results of IEEE 33-bus system prove that the deployment strategy of BESS proposed in this paper is well adapted to voltage violation regulation invariable distribution network operation scenarios. It contributes to regulating voltage violation in distribution network, as well as to improve the utilization of PV systems.

  13. Efficient Parallel Kernel Solvers for Computational Fluid Dynamics Applications

    Science.gov (United States)

    Sun, Xian-He

    1997-01-01

    Distributed-memory parallel computers dominate today's parallel computing arena. These machines, such as Intel Paragon, IBM SP2, and Cray Origin2OO, have successfully delivered high performance computing power for solving some of the so-called "grand-challenge" problems. Despite initial success, parallel machines have not been widely accepted in production engineering environments due to the complexity of parallel programming. On a parallel computing system, a task has to be partitioned and distributed appropriately among processors to reduce communication cost and to attain load balance. More importantly, even with careful partitioning and mapping, the performance of an algorithm may still be unsatisfactory, since conventional sequential algorithms may be serial in nature and may not be implemented efficiently on parallel machines. In many cases, new algorithms have to be introduced to increase parallel performance. In order to achieve optimal performance, in addition to partitioning and mapping, a careful performance study should be conducted for a given application to find a good algorithm-machine combination. This process, however, is usually painful and elusive. The goal of this project is to design and develop efficient parallel algorithms for highly accurate Computational Fluid Dynamics (CFD) simulations and other engineering applications. The work plan is 1) developing highly accurate parallel numerical algorithms, 2) conduct preliminary testing to verify the effectiveness and potential of these algorithms, 3) incorporate newly developed algorithms into actual simulation packages. The work plan has well achieved. Two highly accurate, efficient Poisson solvers have been developed and tested based on two different approaches: (1) Adopting a mathematical geometry which has a better capacity to describe the fluid, (2) Using compact scheme to gain high order accuracy in numerical discretization. The previously developed Parallel Diagonal Dominant (PDD) algorithm

  14. Distributed Coordination Control Based on State-of-Charge for Bidirectional Power Converters in a Hybrid AC/DC Microgrid

    Directory of Open Access Journals (Sweden)

    Zeyan Lv

    2018-04-01

    Full Text Available This paper proposes a distributed coordination control for multiple bidirectional power converters (BPCs in a hybrid AC/DC microgrid with consideration of state-of-charge (SOC of storages. The researched hybrid AC/DC microgrid is composed of both AC and DC subgrids connected by multiple parallel BPCs. In the literature, the storages of a hybrid microgrid are considered to allocate in only the AC subgrid or DC subgrid, which reduces the reliability of the whole system, especially during the islanded mode. Besides, the SOC management has not been considered in BPCs’ operating strategy. This paper considers a hybrid microgrid topology which has energy storages in both AC side and DC side. This ensures the reliability while increasing the complexity of the control strategy at the same time. Further, a distributed coordination control method for multiple BPCs based on SOC was proposed to enhance the reliability of hybrid microgrid. Finally, the performance of the proposed control methods was verified by real-time hardware-in-loop (HIL tests.

  15. Parallelization of pressure equation solver for incompressible N-S equations

    International Nuclear Information System (INIS)

    Ichihara, Kiyoshi; Yokokawa, Mitsuo; Kaburaki, Hideo.

    1996-03-01

    A pressure equation solver in a code for 3-dimensional incompressible flow analysis has been parallelized by using red-black SOR method and PCG method on Fujitsu VPP500, a vector parallel computer with distributed memory. For the comparison of scalability, the solver using the red-black SOR method has been also parallelized on the Intel Paragon, a scalar parallel computer with a distributed memory. The scalability of the red-black SOR method on both VPP500 and Paragon was lost, when number of processor elements was increased. The reason of non-scalability on both systems is increasing communication time between processor elements. In addition, the parallelization by DO-loop division makes the vectorizing efficiency lower on VPP500. For an effective implementation on VPP500, a large scale problem which holds very long vectorized DO-loops in the parallel program should be solved. PCG method with red-black SOR method applied to incomplete LU factorization (red-black PCG) has more iteration steps than normal PCG method with forward and backward substitution, in spite of same number of the floating point operations in a DO-loop of incomplete LU factorization. The parallelized red-black PCG method has less merits than the parallelized red-black SOR method when the computational region has fewer grids, because the low vectorization efficiency is obtained in red-black PCG method. (author)

  16. Hierarchical Control Design for Shipboard Power System with DC Distribution and Energy Storage aboard Future More-Electric Ships

    DEFF Research Database (Denmark)

    Jin, Zheming; Meng, Lexuan; Guerrero, Josep M.

    2018-01-01

    power system (SPS) with DC distribution and energy storage system (ESS) is picked as study case. To meet the requirement of control and management of such a large-scale mobile power system, a hierarchical control design is proposed in this paper. In order to fully exploit the benefit of ESS, as well...

  17. 6th IEEE/ACIS International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing

    CERN Document Server

    2016-01-01

    This edited book presents scientific results of the 16th IEEE/ACIS International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing (SNPD 2015) which was held on June 1 – 3, 2015 in Takamatsu, Japan. The aim of this conference was to bring together researchers and scientists, businessmen and entrepreneurs, teachers, engineers, computer users, and students to discuss the numerous fields of computer science and to share their experiences and exchange new ideas and information in a meaningful way. Research results about all aspects (theory, applications and tools) of computer and information science, and to discuss the practical challenges encountered along the way and the solutions adopted to solve them.

  18. 17th IEEE/ACIS International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing

    CERN Document Server

    SNPD 2016

    2016-01-01

    This edited book presents scientific results of the 17th IEEE/ACIS International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing (SNPD 2016) which was held on May 30 - June 1, 2016 in Shanghai, China. The aim of this conference was to bring together researchers and scientists, businessmen and entrepreneurs, teachers, engineers, computer users, and students to discuss the numerous fields of computer science and to share their experiences and exchange new ideas and information in a meaningful way. Research results about all aspects (theory, applications and tools) of computer and information science, and to discuss the practical challenges encountered along the way and the solutions adopted to solve them.

  19. NDL-v2.0: A new version of the numerical differentiation library for parallel architectures

    Science.gov (United States)

    Hadjidoukas, P. E.; Angelikopoulos, P.; Voglis, C.; Papageorgiou, D. G.; Lagaris, I. E.

    2014-07-01

    We present a new version of the numerical differentiation library (NDL) used for the numerical estimation of first and second order partial derivatives of a function by finite differencing. In this version we have restructured the serial implementation of the code so as to achieve optimal task-based parallelization. The pure shared-memory parallelization of the library has been based on the lightweight OpenMP tasking model allowing for the full extraction of the available parallelism and efficient scheduling of multiple concurrent library calls. On multicore clusters, parallelism is exploited by means of TORC, an MPI-based multi-threaded tasking library. The new MPI implementation of NDL provides optimal performance in terms of function calls and, furthermore, supports asynchronous execution of multiple library calls within legacy MPI programs. In addition, a Python interface has been implemented for all cases, exporting the functionality of our library to sequential Python codes. Catalog identifier: AEDG_v2_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEDG_v2_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 63036 No. of bytes in distributed program, including test data, etc.: 801872 Distribution format: tar.gz Programming language: ANSI Fortran-77, ANSI C, Python. Computer: Distributed systems (clusters), shared memory systems. Operating system: Linux, Unix. Has the code been vectorized or parallelized?: Yes. RAM: The library uses O(N) internal storage, N being the dimension of the problem. It can use up to O(N2) internal storage for Hessian calculations, if a task throttling factor has not been set by the user. Classification: 4.9, 4.14, 6.5. Catalog identifier of previous version: AEDG_v1_0 Journal reference of previous version: Comput. Phys. Comm. 180

  20. Molecular diagnosis of glycogen storage disease and disorders with overlapping clinical symptoms by massive parallel sequencing.

    Science.gov (United States)

    Vega, Ana I; Medrano, Celia; Navarrete, Rosa; Desviat, Lourdes R; Merinero, Begoña; Rodríguez-Pombo, Pilar; Vitoria, Isidro; Ugarte, Magdalena; Pérez-Cerdá, Celia; Pérez, Belen

    2016-10-01

    Glycogen storage disease (GSD) is an umbrella term for a group of genetic disorders that involve the abnormal metabolism of glycogen; to date, 23 types of GSD have been identified. The nonspecific clinical presentation of GSD and the lack of specific biomarkers mean that Sanger sequencing is now widely relied on for making a diagnosis. However, this gene-by-gene sequencing technique is both laborious and costly, which is a consequence of the number of genes to be sequenced and the large size of some genes. This work reports the use of massive parallel sequencing to diagnose patients at our laboratory in Spain using either a customized gene panel (targeted exome sequencing) or the Illumina Clinical-Exome TruSight One Gene Panel (clinical exome sequencing (CES)). Sequence variants were matched against biochemical and clinical hallmarks. Pathogenic mutations were detected in 23 patients. Twenty-two mutations were recognized (mostly loss-of-function mutations), including 11 that were novel in GSD-associated genes. In addition, CES detected five patients with mutations in ALDOB, LIPA, NKX2-5, CPT2, or ANO5. Although these genes are not involved in GSD, they are associated with overlapping phenotypic characteristics such as hepatic, muscular, and cardiac dysfunction. These results show that next-generation sequencing, in combination with the detection of biochemical and clinical hallmarks, provides an accurate, high-throughput means of making genetic diagnoses of GSD and related diseases.Genet Med 18 10, 1037-1043.

  1. A Hybrid Parallel Preconditioning Algorithm For CFD

    Science.gov (United States)

    Barth,Timothy J.; Tang, Wei-Pai; Kwak, Dochan (Technical Monitor)

    1995-01-01

    A new hybrid preconditioning algorithm will be presented which combines the favorable attributes of incomplete lower-upper (ILU) factorization with the favorable attributes of the approximate inverse method recently advocated by numerous researchers. The quality of the preconditioner is adjustable and can be increased at the cost of additional computation while at the same time the storage required is roughly constant and approximately equal to the storage required for the original matrix. In addition, the preconditioning algorithm suggests an efficient and natural parallel implementation with reduced communication. Sample calculations will be presented for the numerical solution of multi-dimensional advection-diffusion equations. The matrix solver has also been embedded into a Newton algorithm for solving the nonlinear Euler and Navier-Stokes equations governing compressible flow. The full paper will show numerous examples in CFD to demonstrate the efficiency and robustness of the method.

  2. Parallel definition of tear film maps on distributed-memory clusters for the support of dry eye diagnosis.

    Science.gov (United States)

    González-Domínguez, Jorge; Remeseiro, Beatriz; Martín, María J

    2017-02-01

    The analysis of the interference patterns on the tear film lipid layer is a useful clinical test to diagnose dry eye syndrome. This task can be automated with a high degree of accuracy by means of the use of tear film maps. However, the time required by the existing applications to generate them prevents a wider acceptance of this method by medical experts. Multithreading has been previously successfully employed by the authors to accelerate the tear film map definition on multicore single-node machines. In this work, we propose a hybrid message-passing and multithreading parallel approach that further accelerates the generation of tear film maps by exploiting the computational capabilities of distributed-memory systems such as multicore clusters and supercomputers. The algorithm for drawing tear film maps is parallelized using Message Passing Interface (MPI) for inter-node communications and the multithreading support available in the C++11 standard for intra-node parallelization. The original algorithm is modified to reduce the communications and increase the scalability. The hybrid method has been tested on 32 nodes of an Intel cluster (with two 12-core Haswell 2680v3 processors per node) using 50 representative images. Results show that maximum runtime is reduced from almost two minutes using the previous only-multithreaded approach to less than ten seconds using the hybrid method. The hybrid MPI/multithreaded implementation can be used by medical experts to obtain tear film maps in only a few seconds, which will significantly accelerate and facilitate the diagnosis of the dry eye syndrome. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.

  3. The Role of Energy Reservoirs in Distributed Computing: Manufacturing, Implementing, and Optimizing Energy Storage in Energy-Autonomous Sensor Nodes

    Science.gov (United States)

    Cowell, Martin Andrew

    The world already hosts more internet connected devices than people, and that ratio is only increasing. These devices seamlessly integrate with peoples lives to collect rich data and give immediate feedback about complex systems from business, health care, transportation, and security. As every aspect of global economies integrate distributed computing into their industrial systems and these systems benefit from rich datasets. Managing the power demands of these distributed computers will be paramount to ensure the continued operation of these networks, and is elegantly addressed by including local energy harvesting and storage on a per-node basis. By replacing non-rechargeable batteries with energy harvesting, wireless sensor nodes will increase their lifetimes by an order of magnitude. This work investigates the coupling of high power energy storage with energy harvesting technologies to power wireless sensor nodes; with sections covering device manufacturing, system integration, and mathematical modeling. First we consider the energy storage mechanism of supercapacitors and batteries, and identify favorable characteristics in both reservoir types. We then discuss experimental methods used to manufacture high power supercapacitors in our labs. We go on to detail the integration of our fabricated devices with collaborating labs to create functional sensor node demonstrations. With the practical knowledge gained through in-lab manufacturing and system integration, we build mathematical models to aid in device and system design. First, we model the mechanism of energy storage in porous graphene supercapacitors to aid in component architecture optimization. We then model the operation of entire sensor nodes for the purpose of optimally sizing the energy harvesting and energy reservoir components. In consideration of deploying these sensor nodes in real-world environments, we model the operation of our energy harvesting and power management systems subject to

  4. Evidence for parallel consolidation of motion direction and orientation into visual short-term memory.

    Science.gov (United States)

    Rideaux, Reuben; Apthorp, Deborah; Edwards, Mark

    2015-02-12

    Recent findings have indicated the capacity to consolidate multiple items into visual short-term memory in parallel varies as a function of the type of information. That is, while color can be consolidated in parallel, evidence suggests that orientation cannot. Here we investigated the capacity to consolidate multiple motion directions in parallel and reexamined this capacity using orientation. This was achieved by determining the shortest exposure duration necessary to consolidate a single item, then examining whether two items, presented simultaneously, could be consolidated in that time. The results show that parallel consolidation of direction and orientation information is possible, and that parallel consolidation of direction appears to be limited to two. Additionally, we demonstrate the importance of adequate separation between feature intervals used to define items when attempting to consolidate in parallel, suggesting that when multiple items are consolidated in parallel, as opposed to serially, the resolution of representations suffer. Finally, we used facilitation of spatial attention to show that the deterioration of item resolution occurs during parallel consolidation, as opposed to storage. © 2015 ARVO.

  5. A distributed parallel genetic algorithm of placement strategy for virtual machines deployment on cloud platform.

    Science.gov (United States)

    Dong, Yu-Shuang; Xu, Gao-Chao; Fu, Xiao-Dong

    2014-01-01

    The cloud platform provides various services to users. More and more cloud centers provide infrastructure as the main way of operating. To improve the utilization rate of the cloud center and to decrease the operating cost, the cloud center provides services according to requirements of users by sharding the resources with virtualization. Considering both QoS for users and cost saving for cloud computing providers, we try to maximize performance and minimize energy cost as well. In this paper, we propose a distributed parallel genetic algorithm (DPGA) of placement strategy for virtual machines deployment on cloud platform. It executes the genetic algorithm parallelly and distributedly on several selected physical hosts in the first stage. Then it continues to execute the genetic algorithm of the second stage with solutions obtained from the first stage as the initial population. The solution calculated by the genetic algorithm of the second stage is the optimal one of the proposed approach. The experimental results show that the proposed placement strategy of VM deployment can ensure QoS for users and it is more effective and more energy efficient than other placement strategies on the cloud platform.

  6. A Distributed Parallel Genetic Algorithm of Placement Strategy for Virtual Machines Deployment on Cloud Platform

    Directory of Open Access Journals (Sweden)

    Yu-Shuang Dong

    2014-01-01

    Full Text Available The cloud platform provides various services to users. More and more cloud centers provide infrastructure as the main way of operating. To improve the utilization rate of the cloud center and to decrease the operating cost, the cloud center provides services according to requirements of users by sharding the resources with virtualization. Considering both QoS for users and cost saving for cloud computing providers, we try to maximize performance and minimize energy cost as well. In this paper, we propose a distributed parallel genetic algorithm (DPGA of placement strategy for virtual machines deployment on cloud platform. It executes the genetic algorithm parallelly and distributedly on several selected physical hosts in the first stage. Then it continues to execute the genetic algorithm of the second stage with solutions obtained from the first stage as the initial population. The solution calculated by the genetic algorithm of the second stage is the optimal one of the proposed approach. The experimental results show that the proposed placement strategy of VM deployment can ensure QoS for users and it is more effective and more energy efficient than other placement strategies on the cloud platform.

  7. Technical and Economic Potential of Distributed Energy Storages for the Integration of Renewable Energy

    DEFF Research Database (Denmark)

    Sveinbjörnsson, Dadi Þorsteinn; Trier, Daniel; Hansen, Kenneth

    Very high penetration of fluctuating renewable energy sources can lead to new challenges in balancing energy supply and demand in future energy systems. This work, carried out as a part of Annex 28 of the IEA ECES programme, addresses this. The aim of the study is to identify which role decentral...... indicate that sector coupling along with an intelligent choice of distributed energy storage technologies can enable the integration of large shares of fluctuating renewable energy in an energy efficient and cost-effective way.......Very high penetration of fluctuating renewable energy sources can lead to new challenges in balancing energy supply and demand in future energy systems. This work, carried out as a part of Annex 28 of the IEA ECES programme, addresses this. The aim of the study is to identify which role...... decentralised energy storages (DES) should play in integrating fluctuating renewable energy sources. The technical and economic potential for DES solutions is quantified using energy system modelling, and it is identified which DES technologies have the largest total (technical and economic) potential. For this...

  8. Parallel and Distributed Data Processing Using Autonomous ...

    African Journals Online (AJOL)

    Looking at the distributed nature of these networks, data is processed by remote login or Remote Procedure Calls (RPC), this causes congestion in the network bandwidth. This paper proposes a framework where software agents are assigned duties to be processing the distributed data concurrently and assembling the ...

  9. Acoustic simulation in architecture with parallel algorithm

    Science.gov (United States)

    Li, Xiaohong; Zhang, Xinrong; Li, Dan

    2004-03-01

    In allusion to complexity of architecture environment and Real-time simulation of architecture acoustics, a parallel radiosity algorithm was developed. The distribution of sound energy in scene is solved with this method. And then the impulse response between sources and receivers at frequency segment, which are calculated with multi-process, are combined into whole frequency response. The numerical experiment shows that parallel arithmetic can improve the acoustic simulating efficiency of complex scene.

  10. Development of a parallel DBMS on the basis of PostgreSQL

    OpenAIRE

    Pan, C.

    2011-01-01

    The paper describes the architecture and the design of PargreSQL parallel database management system (DBMS) for distributed memory multiprocessors. PargreSQL is based upon PostgreSQL open-source DBMS and exploits partitioned parallelism.

  11. HAlign-II: efficient ultra-large multiple sequence alignment and phylogenetic tree reconstruction with distributed and parallel computing.

    Science.gov (United States)

    Wan, Shixiang; Zou, Quan

    2017-01-01

    Multiple sequence alignment (MSA) plays a key role in biological sequence analyses, especially in phylogenetic tree construction. Extreme increase in next-generation sequencing results in shortage of efficient ultra-large biological sequence alignment approaches for coping with different sequence types. Distributed and parallel computing represents a crucial technique for accelerating ultra-large (e.g. files more than 1 GB) sequence analyses. Based on HAlign and Spark distributed computing system, we implement a highly cost-efficient and time-efficient HAlign-II tool to address ultra-large multiple biological sequence alignment and phylogenetic tree construction. The experiments in the DNA and protein large scale data sets, which are more than 1GB files, showed that HAlign II could save time and space. It outperformed the current software tools. HAlign-II can efficiently carry out MSA and construct phylogenetic trees with ultra-large numbers of biological sequences. HAlign-II shows extremely high memory efficiency and scales well with increases in computing resource. THAlign-II provides a user-friendly web server based on our distributed computing infrastructure. HAlign-II with open-source codes and datasets was established at http://lab.malab.cn/soft/halign.

  12. Development of Industrial High-Speed Transfer Parallel Robot

    International Nuclear Information System (INIS)

    Kim, Byung In; Kyung, Jin Ho; Do, Hyun Min; Jo, Sang Hyun

    2013-01-01

    Parallel robots used in industry require high stiffness or high speed because of their structural characteristics. Nowadays, the importance of rapid transportation has increased in the distribution industry. In this light, an industrial parallel robot has been developed for high-speed transfer. The developed parallel robot can handle a maximum payload of 3 kg. For a payload of 0.1 kg, the trajectory cycle time is 0.3 s (come and go), and the maximum velocity is 4.5 m/s (pick amp, place work, adept cycle). In this motion, its maximum acceleration is very high and reaches approximately 13g. In this paper, the design, analysis, and performance test results of the developed parallel robot system are introduced

  13. Temperature distribution by the effect of groundwater flow in an aquifer thermal energy storage system model

    Science.gov (United States)

    Shim, B.

    2005-12-01

    Aquifer thermal energy storage (ATES) can be a cost-effective and renewable energy source, depending on site-specific thermohydraulic conditions. To design an effective ATES system, the understanding of thermohydraulic processes is necessary. The heat transfer phenomena of an aquifer heat storage system are simulated with the scenario of heat pump operation of pumping and waste water reinjection in a two layered confined aquifer model having the effect of groundwater movement. Temperature distribution of the aquifer model is generated, and hydraulic heads and temperature variations are monitored at both wells during simulation days. The average groundwater velocities are determined with two assumed hydraulic gradients set by boundary conditions, and the effect of groundwater flow are shown at the generated thermal distributions at three different depth slices. The generated temperature contour lines at the hydraulic gradient of 0.001 are shaped circular, and the center is moved less than 5 m to the east in 365 days. However at the hydraulic gradient of 0.01, the contour centers of the east well at each depth slice are moved near the east boundary and the movement of temperature distribution is increased at the lower aquifer. By the analysis of thermal interference data between two wells the efficiency of a heat pump operation model is validated, and the variation of heads is monitored at injection, pumping and stabilized state. The thermal efficiency of the ATES system model is represented as highly depended on groundwater flow velocity and direction. Therefore the hydrogeologic condition for the system site should be carefully surveyed.

  14. Evaluation of DEC`s GIGAswitch for distributed parallel computing

    Energy Technology Data Exchange (ETDEWEB)

    Chen, H.; Hutchins, J.; Brandt, J.

    1993-10-01

    One of Sandia`s research efforts is to reduce the end-to-end communication delay in a parallel-distributed computing environment. GIGAswitch is DEC`s implementation of a gigabit local area network based on switched FDDI technology. Using the GIGAswitch, the authors intend to minimize the medium access latency suffered by shared-medium FDDI technology. Experimental results show that the GIGAswitch adds 16.5 microseconds of switching and bridging delay to an end-to-end communication. Although the added latency causes a 1.8% throughput degradation and a 5% line efficiency degradation, the availability of dedicated bandwidth is much more than what is available to a workstation on a shared medium. For example, ten directly connected workstations each would have a dedicated bandwidth of 95 Mbps, but if they were sharing the FDDI bandwidth, each would have 10% of the total bandwidth, i.e., less than 10 Mbps. In addition, they have found that when there is no output port contention, the switch`s aggregate bandwidth will scale up to multiples of its port bandwidth. However, with output port contention, the throughput and latency performance suffered significantly. Their mathematical and simulation models indicate that the GIGAswitch line efficiency could be as low as 63% when there are nine input ports contending for the same output port. The data indicate that the delay introduced by contention at the server workstation is 50 times that introduced by the GIGAswitch. The authors conclude that the GIGAswitch meets the performance requirements of today`s high-end workstations and that the switched FDDI technology provides an alternative that utilizes existing workstation interfaces while increasing the aggregate bandwidth. However, because the speed of workstations is increasing by a factor of 2 every 1.5 years, the switched FDDI technology is only good as an interim solution.

  15. Optical information storage

    International Nuclear Information System (INIS)

    Woike, T.

    1996-01-01

    In order to increase storage capacity and data transfer velocity by about three orders of magnitude compared to CD or magnetic disc it is necessary to work with optical techniques, especially with holography. About 100 TByte can be stored in a waver of an area of 50 cm 2 via holograms which corresponds to a density of 2.10 9 Byte/mm 2 . Every hologram contains data of 1 MByte, so that parallel-processing is possible for read-out. Using high-speed CCD-arrays a read-out velocity of 1 MByte/μsec can be reached. Further, holographic technics are very important in solid state physics. We will discuss the existence of a space charge field in Sr 1-x Ba x Nb 2 O 6 doped with cerium and the physical properties of metastable states, which are suited for information storage. (author) 19 figs., 9 refs

  16. Two-Level Control for Fast Electrical Vehicle Charging Stations with Multi Flywheel Energy Storage System

    DEFF Research Database (Denmark)

    SUN, BO; Dragicevic, Tomislav; Vasquez, Juan Carlos

    2015-01-01

    This paper applies a hierarchical control for a fast charging station (FCS) composed of paralleled PWM rectifier and dedicated paralleled multiple flywheel energy storage systems (FESSs), in order to mitigate peak power shock on grid caused by sudden connection of electrical vehicle (EV) chargers...

  17. Leveraging storage assets to meet winter power demand

    Energy Technology Data Exchange (ETDEWEB)

    Charleson, D. [Enbridge Gas Distribution, Toronto, ON (Canada)

    2004-07-01

    Toronto-based Enbridge Gas Distribution serves 1.7 million customers by distributing 420 billion cubic feet (BCF) of natural gas over more than 31,000 km of pipelines. A map of the franchise area was presented. The utility has one of the lowest operating and maintenance costs in North America. Daily gas requirements were outlined along with the historic role of storage in gas utilities. Storage is used by heat sensitive local distribution companies, marketers, large industrials, and power generators. Storage locations in North America were reviewed with reference to baseload electricity production versus peak load; depleted reservoirs; salt caverns; aquifers; and liquefied natural gas (LNG). Enbridge operates 98 BCF of storage facilities for a maximum deliverability of 1.7 BCF per day. tabs., figs.

  18. Leveraging storage assets to meet winter power demand

    International Nuclear Information System (INIS)

    Charleson, D.

    2004-01-01

    Toronto-based Enbridge Gas Distribution serves 1.7 million customers by distributing 420 billion cubic feet (BCF) of natural gas over more than 31,000 km of pipelines. A map of the franchise area was presented. The utility has one of the lowest operating and maintenance costs in North America. Daily gas requirements were outlined along with the historic role of storage in gas utilities. Storage is used by heat sensitive local distribution companies, marketers, large industrials, and power generators. Storage locations in North America were reviewed with reference to baseload electricity production versus peak load; depleted reservoirs; salt caverns; aquifers; and liquefied natural gas (LNG). Enbridge operates 98 BCF of storage facilities for a maximum deliverability of 1.7 BCF per day. tabs., figs

  19. Parallel processing of Monte Carlo code MCNP for particle transport problem

    Energy Technology Data Exchange (ETDEWEB)

    Higuchi, Kenji; Kawasaki, Takuji

    1996-06-01

    It is possible to vectorize or parallelize Monte Carlo codes (MC code) for photon and neutron transport problem, making use of independency of the calculation for each particle. Applicability of existing MC code to parallel processing is mentioned. As for parallel computer, we have used both vector-parallel processor and scalar-parallel processor in performance evaluation. We have made (i) vector-parallel processing of MCNP code on Monte Carlo machine Monte-4 with four vector processors, (ii) parallel processing on Paragon XP/S with 256 processors. In this report we describe the methodology and results for parallel processing on two types of parallel or distributed memory computers. In addition, we mention the evaluation of parallel programming environments for parallel computers used in the present work as a part of the work developing STA (Seamless Thinking Aid) Basic Software. (author)

  20. Comparison of some parallelization strategies of thermalhydraulic codes on GPUs

    International Nuclear Information System (INIS)

    Jendoubi, T.; Bergeaud, V.; Geay, A.

    2013-01-01

    Modern supercomputers architecture is now often based on hybrid concepts combining parallelism to distributed memory, parallelism to shared memory and also to GPUs (Graphic Process Units). In this work, we propose a new approach to take advantage of these graphic cards in thermohydraulics algorithms. (authors)

  1. The determination of the cesium distribution coefficient of the interim storage soil from Abadia de Goias, Go, Brazil

    International Nuclear Information System (INIS)

    Marumo, J.T.; Suarez, A.A.

    1989-01-01

    In September, 1987, an unauthorized removal of a cesium-therapy unit and its violation caused an accident, where several places of Goiania's city, capital of Goias, Brazil, were contaminated. The removal of the radioactive wastes generated from decontamination process, was made to Abadia de Goias city (near Goiania), where an interim storage was constructed. Soil samples collected from the 57th Street (Goiania) and from the interim storage permitted to determine, through static method, the cesium distribution coefficient for different cesium solution concentrations. Those results allows for some migration/retention evaluations in disposal site selection. Some soils parameters (water content, density, granulometric analysis, etc) as well as clay minerals constituents were also determined. (author) [pt

  2. Metabolomic analysis of platelets during storage

    DEFF Research Database (Denmark)

    Paglia, Giuseppe; Sigurjónsson, Ólafur E; Rolfsson, Óttar

    2015-01-01

    BACKGROUND: Platelet concentrates (PCs) can be prepared using three methods: platelet (PLT)-rich plasma, apheresis, and buffy coat. The aim of this study was to obtain a comprehensive data set that describes metabolism of buffy coat-derived PLTs during storage and to compare it with a previously...... published parallel data set obtained for apheresis-derived PLTs. STUDY DESIGN AND METHODS: During storage we measured more than 150 variables in 8 PLT units, prepared by the buffy coat method. Samples were collected at seven different time points resulting in a data set containing more than 8000...... after their collection. The transition was evident in PLT produced by both production methods. Apheresis-derived PLTs showed a clearer phenotype of PLT activation during early days of storage. The activated phenotype of apheresis PLTs was accompanied by a higher metabolic activity, especially related...

  3. Autonomous Control of Distributed Generation and Storage to Coordinate P/Q Sharing in Islanded Microgrids

    DEFF Research Database (Denmark)

    Wu, Dan; Tang, Fen; Guerrero, Josep M.

    2014-01-01

    In this paper, a decentralized control for coordinate both active and reactive powers is proposed for islanded microgrids. Compared with the conventional droop control strategies, the proposed control realizes decentralized power distribution among renewable energy sources (RES) and energy storage...... systems (ESS) according to the local source conditions. Based on bus-signaling method, the ESS is able to limit charging power by decreasing RES power generation automatically. As well, the reactive power coordinated control makes the RES units able to support reactive power in a decentralized way, which...... allows ESS providing for more active power availability. Moreover, the reactive power is distributed according to the apparent power capacity of each unit. The control strategy principle is simple and easy to implement without extra communication requirements. Real time hardware-in-the-loop results...

  4. On radial flow between parallel disks

    International Nuclear Information System (INIS)

    Wee, A Y L; Gorin, A

    2015-01-01

    Approximate analytical solutions are presented for converging flow in between two parallel non rotating disks. The static pressure distribution and radial component of the velocity are developed by averaging the inertial term across the gap in between parallel disks. The predicted results from the first approximation are favourable to experimental results as well as results presented by other authors. The second approximation shows that as the fluid approaches the center, the velocity at the mid channel slows down which is due to the struggle between the inertial term and the flowrate. (paper)

  5. MulticoreBSP for C : A high-performance library for shared-memory parallel programming

    NARCIS (Netherlands)

    Yzelman, A. N.; Bisseling, R. H.; Roose, D.; Meerbergen, K.

    2014-01-01

    The bulk synchronous parallel (BSP) model, as well as parallel programming interfaces based on BSP, classically target distributed-memory parallel architectures. In earlier work, Yzelman and Bisseling designed a MulticoreBSP for Java library specifically for shared-memory architectures. In the

  6. Soil moisture storage and hillslope stability

    Directory of Open Access Journals (Sweden)

    A. Talebi

    2007-09-01

    Full Text Available Recently, we presented a steady-state analytical hillslope stability model to study rain-induced shallow landslides. This model is based on kinematic wave dynamics of saturated subsurface storage and the infinite slope stability assumption. Here we apply the model to investigate the effect of neglecting the unsaturated storage on the assessment of slope stability in the steady-state hydrology. For that purpose we extend the hydrological model to compute the soil pore pressure distribution over the entire flow domain. We also apply this model for hillslopes with non-constant soil depth to compare the stability of different hillslopes and to find the critical slip surface in hillslopes with different geometric characteristics. In order to do this, we incorporate more complex approaches to compute slope stability (Janbu's non-circular method and Bishop's simplified method in the steady-state analytical hillslope stability model. We compare the safety factor (FS derived from the infinite slope stability method and the more complex approach for two cases: with and without the soil moisture profile in the unsaturated zone. We apply this extended hillslope stability model to nine characteristic hillslope types with three different profile curvatures (concave, straight, convex and three different plan shapes (convergent, parallel, divergent. Overall, we find that unsaturated zone storage does not play a critical role in determining the factor of safety for shallow and deep landslides. As a result, the effect of the unsaturated zone storage on slope stability can be neglected in the steady-state hydrology and one can assume the same bulk specific weight below and above the water table. We find that steep slopes with concave profile and convergent plan shape have the least stability. We also demonstrate that in hillslopes with non-constant soil depth (possible deep landslides, the ones with convex profiles and convergent plan shapes have

  7. Parallel pic plasma simulation through particle decomposition techniques

    International Nuclear Information System (INIS)

    Briguglio, S.; Vlad, G.; Di Martino, B.; Naples, Univ. 'Federico II'

    1998-02-01

    Particle-in-cell (PIC) codes are among the major candidates to yield a satisfactory description of the detail of kinetic effects, such as the resonant wave-particle interaction, relevant in determining the transport mechanism in magnetically confined plasmas. A significant improvement of the simulation performance of such codes con be expected from parallelization, e.g., by distributing the particle population among several parallel processors. Parallelization of a hybrid magnetohydrodynamic-gyrokinetic code has been accomplished within the High Performance Fortran (HPF) framework, and tested on the IBM SP2 parallel system, using a 'particle decomposition' technique. The adopted technique requires a moderate effort in porting the code in parallel form and results in intrinsic load balancing and modest inter processor communication. The performance tests obtained confirm the hypothesis of high effectiveness of the strategy, if targeted towards moderately parallel architectures. Optimal use of resources is also discussed with reference to a specific physics problem [it

  8. Modeling nurses' attitude toward using automated unit-based medication storage and distribution systems: an extension of the technology acceptance model.

    Science.gov (United States)

    Escobar-Rodríguez, Tomás; Romero-Alonso, María Mercedes

    2013-05-01

    This article analyzes the attitude of nurses toward the use of automated unit-based medication storage and distribution systems and identifies influencing factors. Understanding these factors provides an opportunity to explore actions that might be taken to boost adoption by potential users. The theoretical grounding for this research is the Technology Acceptance Model. The Technology Acceptance Model specifies the causal relationships between perceived usefulness, perceived ease of use, attitude toward using, and actual usage behavior. The research model has six constructs, and nine hypotheses were generated from connections between these six constructs. These constructs include perceived risks, experience level, and training. The findings indicate that these three external variables are related to the perceived ease of use and perceived usefulness of automated unit-based medication storage and distribution systems, and therefore, they have a significant influence on attitude toward the use of these systems.

  9. On the edge of change : a growing role for Alberta storage

    International Nuclear Information System (INIS)

    Daniel, R.

    1998-01-01

    The growing role for Alberta regarding storage of natural gas was the focus of this presentation. The history of the development of natural gas storage facilities in the province was outlined. An updated list of storage facilities in the province with their working gas distribution and deliverability distribution was provided. Storage capacity in North America as a whole was assessed on the same basis. Services of storage companies, and the price sensitivity of storage activity were also discussed. figs

  10. Portable programming on parallel/networked computers using the Application Portable Parallel Library (APPL)

    Science.gov (United States)

    Quealy, Angela; Cole, Gary L.; Blech, Richard A.

    1993-01-01

    The Application Portable Parallel Library (APPL) is a subroutine-based library of communication primitives that is callable from applications written in FORTRAN or C. APPL provides a consistent programmer interface to a variety of distributed and shared-memory multiprocessor MIMD machines. The objective of APPL is to minimize the effort required to move parallel applications from one machine to another, or to a network of homogeneous machines. APPL encompasses many of the message-passing primitives that are currently available on commercial multiprocessor systems. This paper describes APPL (version 2.3.1) and its usage, reports the status of the APPL project, and indicates possible directions for the future. Several applications using APPL are discussed, as well as performance and overhead results.

  11. Erasure Coded Storage on a Changing Network

    DEFF Research Database (Denmark)

    Sipos, Marton A.; Venkat, Narayan; Oran, David

    2016-01-01

    As faster storage devices become commercially viable alternatives to disk drives, the network is increasingly becoming the bottleneck in achieving good performance in distributed storage systems. This is especially true for erasure coded storage, where the reconstruction of lost data can signific...

  12. In-cylinder diesel spray combustion simulations using parallel computation: A performance benchmarking study

    International Nuclear Information System (INIS)

    Pang, Kar Mun; Ng, Hoon Kiat; Gan, Suyin

    2012-01-01

    Highlights: ► A performance benchmarking exercise is conducted for diesel combustion simulations. ► The reduced chemical mechanism shows its advantages over base and skeletal models. ► High efficiency and great reduction of CPU runtime are achieved through 4-node solver. ► Increasing ISAT memory from 0.1 to 2 GB reduces the CPU runtime by almost 35%. ► Combustion and soot processes are predicted well with minimal computational cost. - Abstract: In the present study, in-cylinder diesel combustion simulation was performed with parallel processing on an Intel Xeon Quad-Core platform to allow both fluid dynamics and chemical kinetics of the surrogate diesel fuel model to be solved simultaneously on multiple processors. Here, Cartesian Z-Coordinate was selected as the most appropriate partitioning algorithm since it computationally bisects the domain such that the dynamic load associated with fuel particle tracking was evenly distributed during parallel computations. Other variables examined included number of compute nodes, chemistry sizes and in situ adaptive tabulation (ISAT) parameters. Based on the performance benchmarking test conducted, parallel configuration of 4-compute node was found to reduce the computational runtime most efficiently whereby a parallel efficiency of up to 75.4% was achieved. The simulation results also indicated that accuracy level was insensitive to the number of partitions or the partitioning algorithms. The effect of reducing the number of species on computational runtime was observed to be more significant than reducing the number of reactions. Besides, the study showed that an increase in the ISAT maximum storage of up to 2 GB reduced the computational runtime by 50%. Also, the ISAT error tolerance of 10 −3 was chosen to strike a balance between results accuracy and computational runtime. The optimised parameters in parallel processing and ISAT, as well as the use of the in-house reduced chemistry model allowed accurate

  13. Hydrogen storage inside graphene-oxide frameworks

    International Nuclear Information System (INIS)

    Chan Yue; Hill, James M

    2011-01-01

    In this paper, we use applied mathematical modelling to investigate the storage of hydrogen molecules inside graphene-oxide frameworks, which comprise two parallel graphenes rigidly separated by perpendicular ligands. Hydrogen uptake is calculated for graphene-oxide frameworks using the continuous approximation and an equation of state for both the bulk and adsorption gas phases. We first validate our approach by obtaining results for two parallel graphene sheets. This result agrees well with an existing theoretical result, namely 1.85 wt% from our calculations, and 2 wt% arising from an ab initio and grand canonical Monte Carlo calculation. This provides confidence to the determination of the hydrogen uptake for the four graphene-oxide frameworks, GOF-120, GOF-66, GOF-28 and GOF-6, and we obtain 1.68, 2, 6.33 and 0 wt%, respectively. The high value obtained for GOF-28 may be partly explained by the fact that the benzenediboronic acid pillars between graphene sheets not only provide mechanical support and porous spaces for the molecular structure but also provide the higher binding energy to enhance the hydrogen storage inside graphene-oxide frameworks. For the other three structures, this binding energy is not as large in comparison to that of GOF-28 and this effect diminishes as the ligand density decreases. In the absence of conflicting data, the present work indicates GOF-28 as a likely contender for practical hydrogen storage.

  14. HPDC ´12 : proceedings of the 21st ACM symposium on high-performance parallel and distributed computing, June 18-22, 2012, Delft, The Netherlands

    NARCIS (Netherlands)

    Epema, D.H.J.; Kielmann, T.; Ripeanu, M.

    2012-01-01

    Welcome to ACM HPDC 2012! This is the twenty-first year of HPDC and we are pleased to report that our community continues to grow in size, quality and reputation. The program consists of three days packed with presentations on the latest developments in high-performance parallel and distributed

  15. Distributed parallel cooperative coevolutionary multi-objective large-scale immune algorithm for deployment of wireless sensor networks

    DEFF Research Database (Denmark)

    Cao, Bin; Zhao, Jianwei; Yang, Po

    2018-01-01

    -objective evolutionary algorithms the Cooperative Coevolutionary Generalized Differential Evolution 3, the Cooperative Multi-objective Differential Evolution and the Nondominated Sorting Genetic Algorithm III, the proposed algorithm addresses the deployment optimization problem efficiently and effectively.......Using immune algorithms is generally a time-intensive process especially for problems with a large number of variables. In this paper, we propose a distributed parallel cooperative coevolutionary multi-objective large-scale immune algorithm that is implemented using the message passing interface...... (MPI). The proposed algorithm is composed of three layers: objective, group and individual layers. First, for each objective in the multi-objective problem to be addressed, a subpopulation is used for optimization, and an archive population is used to optimize all the objectives. Second, the large...

  16. Applications of Parallel Processing in Mobile Banking

    Directory of Open Access Journals (Sweden)

    2007-01-01

    Full Text Available The future of mobile banking will be represented by such applications that support mobile, Internet banking and EFT (Electronic Funds Transfer transactions in a single user interface. In such a way, the mobile banking will be able to cover all the types of applications demanded at the market level. The parallel processing of credit card bank transactions could be performed with the help of a grid network. Excluding some limitations, the grid processing offers huge opportunities to exploit the parallelism. For this reason, a lot of applications of waiting queues in grid processing were developed in the last years. Grid networks represent a distinctive and very modern field of the parallel and distributed processing.

  17. CO2 storage in Sweden

    International Nuclear Information System (INIS)

    Ekstroem, Clas; Andersson, Annika; Kling, Aasa; Bernstone, Christian; Carlsson, Anders; Liljemark, Stefan; Wall, Caroline; Erstedt, Thomas; Lindroth, Maria; Tengborg, Per; Edstroem, Mikael

    2004-07-01

    This study considers options, that could be feasible for Sweden, to transport and geologically store CO 2 , providing that technology for electricity production with CO 2 capture will be available in the future and also acceptable from cost- and reliability point of view. As a starting point, it is assumed that a new 600-1000 MW power plant, fired with coal or natural gas, will be constructed with CO 2 capture and localised to the Stockholm, Malmoe or Goeteborg areas. Of vital importance for storage of carbon dioxide in a reservoir is the possibility to monitor its distribution, i.e. its migration within the reservoir. It has been shown in the SACS-project that the distribution of carbon dioxide within the reservoir can be monitored successfully, mainly by seismic methods. Suitable geologic conditions and a large storage potential seems to exist mainly in South West Scania, where additional knowledge on geology/hydrogeology has been obtained since the year 2000 in connection to geothermal energy projects, and in the Eastern part of Denmark, bordering on South West Scania. Storage of carbon dioxide from the Stockholm area should not be excluded, but more studies are needed to clarify the storage options within this area. The possibilities to use CO 2 for enhanced oil recovery, EOR, in i.a. the North Sea should be investigated, in order to receive incomes from the CO 2 and shared costs for infrastructure, and by this also make the CO 2 regarded as a trading commodity, and thereby achieving a more favourable position concerning acceptance, legal issues and regulations. The dimensions of CO 2 -pipelines should be similar to those for natural natural gas, although regarding some aspects they have different design and construction prerequisites. To obtain cost efficiency, the transport distances should be kept short, and possibilities for co-ordinated networks with short distribution pipelines connected to common main pipelines, should be searched for. Also, synergies

  18. Leveraging human oversight and intervention in large-scale parallel processing of open-source data

    Science.gov (United States)

    Casini, Enrico; Suri, Niranjan; Bradshaw, Jeffrey M.

    2015-05-01

    The popularity of cloud computing along with the increased availability of cheap storage have led to the necessity of elaboration and transformation of large volumes of open-source data, all in parallel. One way to handle such extensive volumes of information properly is to take advantage of distributed computing frameworks like Map-Reduce. Unfortunately, an entirely automated approach that excludes human intervention is often unpredictable and error prone. Highly accurate data processing and decision-making can be achieved by supporting an automatic process through human collaboration, in a variety of environments such as warfare, cyber security and threat monitoring. Although this mutual participation seems easily exploitable, human-machine collaboration in the field of data analysis presents several challenges. First, due to the asynchronous nature of human intervention, it is necessary to verify that once a correction is made, all the necessary reprocessing is done in chain. Second, it is often needed to minimize the amount of reprocessing in order to optimize the usage of resources due to limited availability. In order to improve on these strict requirements, this paper introduces improvements to an innovative approach for human-machine collaboration in the processing of large amounts of open-source data in parallel.

  19. Fast disk array for image storage

    Science.gov (United States)

    Feng, Dan; Zhu, Zhichun; Jin, Hai; Zhang, Jiangling

    1997-01-01

    A fast disk array is designed for the large continuous image storage. It includes a high speed data architecture and the technology of data striping and organization on the disk array. The high speed data path which is constructed by two dual port RAM and some control circuit is configured to transfer data between a host system and a plurality of disk drives. The bandwidth can be more than 100 MB/s if the data path based on PCI (peripheral component interconnect). The organization of data stored on the disk array is similar to RAID 4. Data are striped on a plurality of disk, and each striping unit is equal to a track. I/O instructions are performed in parallel on the disk drives. An independent disk is used to store the parity information in the fast disk array architecture. By placing the parity generation circuit directly on the SCSI (or SCSI 2) bus, the parity information can be generated on the fly. It will affect little on the data writing in parallel on the other disks. The fast disk array architecture designed in the paper can meet the demands of the image storage.

  20. Building a columnar database on shared main memory-based storage

    OpenAIRE

    Tinnefeld, Christian

    2014-01-01

    In the field of disk-based parallel database management systems exists a great variety of solutions based on a shared-storage or a shared-nothing architecture. In contrast, main memory-based parallel database management systems are dominated solely by the shared-nothing approach as it preserves the in-memory performance advantage by processing data locally on each server. We argue that this unilateral development is going to cease due to the combination of the following three trends: a) Nowad...

  1. Organic Carbon Storage in China's Urban Areas

    Science.gov (United States)

    Zhao, Shuqing; Zhu, Chao; Zhou, Decheng; Huang, Dian; Werner, Jeremy

    2013-01-01

    China has been experiencing rapid urbanization in parallel with its economic boom over the past three decades. To date, the organic carbon storage in China's urban areas has not been quantified. Here, using data compiled from literature review and statistical yearbooks, we estimated that total carbon storage in China's urban areas was 577±60 Tg C (1 Tg  = 1012 g) in 2006. Soil was the largest contributor to total carbon storage (56%), followed by buildings (36%), and vegetation (7%), while carbon storage in humans was relatively small (1%). The carbon density in China's urban areas was 17.1±1.8 kg C m−2, about two times the national average of all lands. The most sensitive variable in estimating urban carbon storage was urban area. Examining urban carbon storages over a wide range of spatial extents in China and in the United States, we found a strong linear relationship between total urban carbon storage and total urban area, with a specific urban carbon storage of 16 Tg C for every 1,000 km2 urban area. This value might be useful for estimating urban carbon storage at regional to global scales. Our results also showed that the fraction of carbon storage in urban green spaces was still much lower in China relative to western countries, suggesting a great potential to mitigate climate change through urban greening and green spaces management in China. PMID:23991014

  2. The Fermilab data storage infrastructure

    International Nuclear Information System (INIS)

    Jon A Bakken et al.

    2003-01-01

    Fermilab, in collaboration with the DESY laboratory in Hamburg, Germany, has created a petabyte scale data storage infrastructure to meet the requirements of experiments to store and access large data sets. The Fermilab data storage infrastructure consists of the following major storage and data transfer components: Enstore mass storage system, DCache distributed data cache, ftp and Grid ftp for primarily external data transfers. This infrastructure provides a data throughput sufficient for transferring data from experiments' data acquisition systems. It also allows access to data in the Grid framework

  3. An Intrinsic Algorithm for Parallel Poisson Disk Sampling on Arbitrary Surfaces.

    Science.gov (United States)

    Ying, Xiang; Xin, Shi-Qing; Sun, Qian; He, Ying

    2013-03-08

    Poisson disk sampling plays an important role in a variety of visual computing, due to its useful statistical property in distribution and the absence of aliasing artifacts. While many effective techniques have been proposed to generate Poisson disk distribution in Euclidean space, relatively few work has been reported to the surface counterpart. This paper presents an intrinsic algorithm for parallel Poisson disk sampling on arbitrary surfaces. We propose a new technique for parallelizing the dart throwing. Rather than the conventional approaches that explicitly partition the spatial domain to generate the samples in parallel, our approach assigns each sample candidate a random and unique priority that is unbiased with regard to the distribution. Hence, multiple threads can process the candidates simultaneously and resolve conflicts by checking the given priority values. It is worth noting that our algorithm is accurate as the generated Poisson disks are uniformly and randomly distributed without bias. Our method is intrinsic in that all the computations are based on the intrinsic metric and are independent of the embedding space. This intrinsic feature allows us to generate Poisson disk distributions on arbitrary surfaces. Furthermore, by manipulating the spatially varying density function, we can obtain adaptive sampling easily.

  4. FPGA-based prototype storage system with phase change memory

    Science.gov (United States)

    Li, Gezi; Chen, Xiaogang; Chen, Bomy; Li, Shunfen; Zhou, Mi; Han, Wenbing; Song, Zhitang

    2016-10-01

    With the ever-increasing amount of data being stored via social media, mobile telephony base stations, and network devices etc. the database systems face severe bandwidth bottlenecks when moving vast amounts of data from storage to the processing nodes. At the same time, Storage Class Memory (SCM) technologies such as Phase Change Memory (PCM) with unique features like fast read access, high density, non-volatility, byte-addressability, positive response to increasing temperature, superior scalability, and zero standby leakage have changed the landscape of modern computing and storage systems. In such a scenario, we present a storage system called FLEET which can off-load partial or whole SQL queries to the storage engine from CPU. FLEET uses an FPGA rather than conventional CPUs to implement the off-load engine due to its highly parallel nature. We have implemented an initial prototype of FLEET with PCM-based storage. The results demonstrate that significant performance and CPU utilization gains can be achieved by pushing selected query processing components inside in PCM-based storage.

  5. Parallel Processing and Applied Mathematics. 10th International Conference, PPAM 2013. Revised Selected Papers

    DEFF Research Database (Denmark)

    The following topics are dealt with: parallel scientific computing; numerical algorithms; parallel nonnumerical algorithms; cloud computing; evolutionary computing; metaheuristics; applied mathematics; GPU computing; multicore systems; hybrid architectures; hierarchical parallelism; HPC systems......; power monitoring; energy monitoring; and distributed computing....

  6. Optical information storage

    Energy Technology Data Exchange (ETDEWEB)

    Woike, T [Koeln Univ., Inst. fuer Kristallography, Koeln (Germany)

    1996-11-01

    In order to increase storage capacity and data transfer velocity by about three orders of magnitude compared to CD or magnetic disc it is necessary to work with optical techniques, especially with holography. About 100 TByte can be stored in a waver of an area of 50 cm{sup 2} via holograms which corresponds to a density of 2.10{sup 9} Byte/mm{sup 2}. Every hologram contains data of 1 MByte, so that parallel-processing is possible for read-out. Using high-speed CCD-arrays a read-out velocity of 1 MByte/{mu}sec can be reached. Further, holographic technics are very important in solid state physics. We will discuss the existence of a space charge field in Sr{sub 1-x}Ba{sub x}Nb{sub 2}O{sub 6} doped with cerium and the physical properties of metastable states, which are suited for information storage. (author) 19 figs., 9 refs.

  7. Magnetic energy storage

    International Nuclear Information System (INIS)

    Rogers, J.D.

    1980-01-01

    The fusion program embraces low loss superconductor strand development with integration into cables capable of carrying 50 kA in pulsed mode at high fields. This evolvement has been paralleled with pulsed energy storage coil development and testing from tens of kJ at low fields to a 20 MJ prototype tokamak induction coil at 7.5 T. Energy transfer times have ranged from 0.7 ms to several seconds. Electric utility magnetic storage for prospective application is for diurnal load leveling with massive systems to store 10 GWh at 1.8 K in a dewar structure supported on bedrock underground. An immediate utility application is a 30 MJ system to be used to damp power oscillations on the Bonneville Power Administration electric transmission lines. An off-shoot of this last work is a new program for electric utility VAR control with the potential for use to suppress subsynchronous resonance. This paper presents work in progress, work planned, and recently completed unusual work

  8. 41 CFR 109-28.000-51 - Storage guidelines.

    Science.gov (United States)

    2010-07-01

    ... 41 Public Contracts and Property Management 3 2010-07-01 2010-07-01 false Storage guidelines. 109-28.000-51 Section 109-28.000-51 Public Contracts and Property Management Federal Property Management...-STORAGE AND DISTRIBUTION § 109-28.000-51 Storage guidelines. (a) Indoor storage areas should be arranged...

  9. On the Minimum Cable Tensions for the Cable-Based Parallel Robots

    Directory of Open Access Journals (Sweden)

    Peng Liu

    2014-01-01

    Full Text Available This paper investigates the minimum cable tension distributions in the workspace for cable-based parallel robots to find out more information on the stability. First, the kinematic model of a cable-based parallel robot is derived based on the wrench matrix. Then, a noniterative polynomial-based optimization algorithm with the proper optimal objective function is presented based on the convex optimization theory, in which the minimum cable tension at any pose is determined. Additionally, three performance indices are proposed to show the distributions of the minimum cable tensions in a specified region of the workspace. An important thing is that the three performance indices can be used to evaluate the stability of the cable-based parallel robots. Furthermore, a new workspace, the Specified Minimum Cable Tension Workspace (SMCTW, is introduced, within which all the minimum tensions exceed a specified value, therefore meeting the specified stability requirement. Finally, a camera robot parallel driven by four cables for aerial panoramic photographing is selected to illustrate the distributions of the minimum cable tensions in the workspace and the relationship between the three performance indices and the stability.

  10. Optimal Active Power Control of A Wind Farm Equipped with Energy Storage System based on Distributed Model Predictive Control

    DEFF Research Database (Denmark)

    Zhao, Haoran; Wu, Qiuwei; Guo, Qinglai

    2016-01-01

    This paper presents the Distributed Model Predictive Control (D-MPC) of a wind farm equipped with fast and short-term Energy Storage System (ESS) for optimal active power control using the fast gradient method via dual decomposition. The primary objective of the D-MPC control of the wind farm...... is power reference tracking from system operators. Besides, by optimal distribution of the power references to individual wind turbines and the ESS unit, the wind turbine mechanical loads are alleviated. With the fast gradient method, the convergence rate of the DMPC is significantly improved which leads...

  11. A tool for simulating parallel branch-and-bound methods

    Science.gov (United States)

    Golubeva, Yana; Orlov, Yury; Posypkin, Mikhail

    2016-01-01

    The Branch-and-Bound method is known as one of the most powerful but very resource consuming global optimization methods. Parallel and distributed computing can efficiently cope with this issue. The major difficulty in parallel B&B method is the need for dynamic load redistribution. Therefore design and study of load balancing algorithms is a separate and very important research topic. This paper presents a tool for simulating parallel Branchand-Bound method. The simulator allows one to run load balancing algorithms with various numbers of processors, sizes of the search tree, the characteristics of the supercomputer's interconnect thereby fostering deep study of load distribution strategies. The process of resolution of the optimization problem by B&B method is replaced by a stochastic branching process. Data exchanges are modeled using the concept of logical time. The user friendly graphical interface to the simulator provides efficient visualization and convenient performance analysis.

  12. A quantitative assessment of the Hadoop framework for analyzing massively parallel DNA sequencing data.

    Science.gov (United States)

    Siretskiy, Alexey; Sundqvist, Tore; Voznesenskiy, Mikhail; Spjuth, Ola

    2015-01-01

    New high-throughput technologies, such as massively parallel sequencing, have transformed the life sciences into a data-intensive field. The most common e-infrastructure for analyzing this data consists of batch systems that are based on high-performance computing resources; however, the bioinformatics software that is built on this platform does not scale well in the general case. Recently, the Hadoop platform has emerged as an interesting option to address the challenges of increasingly large datasets with distributed storage, distributed processing, built-in data locality, fault tolerance, and an appealing programming methodology. In this work we introduce metrics and report on a quantitative comparison between Hadoop and a single node of conventional high-performance computing resources for the tasks of short read mapping and variant calling. We calculate efficiency as a function of data size and observe that the Hadoop platform is more efficient for biologically relevant data sizes in terms of computing hours for both split and un-split data files. We also quantify the advantages of the data locality provided by Hadoop for NGS problems, and show that a classical architecture with network-attached storage will not scale when computing resources increase in numbers. Measurements were performed using ten datasets of different sizes, up to 100 gigabases, using the pipeline implemented in Crossbow. To make a fair comparison, we implemented an improved preprocessor for Hadoop with better performance for splittable data files. For improved usability, we implemented a graphical user interface for Crossbow in a private cloud environment using the CloudGene platform. All of the code and data in this study are freely available as open source in public repositories. From our experiments we can conclude that the improved Hadoop pipeline scales better than the same pipeline on high-performance computing resources, we also conclude that Hadoop is an economically viable

  13. Hydraulic Profiling of a Parallel Channel Type Reactor Core

    International Nuclear Information System (INIS)

    Seo, Kyong-Won; Hwang, Dae-Hyun; Lee, Chung-Chan

    2006-01-01

    An advanced reactor core which consisted of closed multiple parallel channels was optimized to maximize the thermal margin of the core. The closed multiple parallel channel configurations have different characteristics to the open channels of conventional PWRs. The channels, usually assemblies, are isolated hydraulically from each other and there is no cross flow between channels. The distribution of inlet flow rate between channels is a very important design parameter in the core because distribution of inlet flow is directly proportional to a margin for a certain hydraulic parameter. The thermal hydraulic parameter may be the boiling margin, maximum fuel temperature, and critical heat flux. The inlet flow distribution of the core was optimized for the boiling margins by grouping the inlet orifices by several hydraulic regions. The procedure is called a hydraulic profiling

  14. Economics of long-distance transmission, storage, and distribution of heat from nuclear plants with existing and newer techniques

    International Nuclear Information System (INIS)

    Margen, P.H.

    1978-01-01

    Conventional and newer types of hot-water pipes are applied to the bulk transport of reject heat from central nuclear power plants to the district heating network of cities or groups of cities. With conventional pipes, the transport of 300 to 2000 MW of heat over distances of 30 to 100 km can be justified, while with newer pipe types, even longer distances would often be economic. For medium-size district heating schemes, low-temperature heat transport from simple heat-only reactors suitable for closer location to cities is of interest. For daily storage of heat on district heating systems, steel heat accumulators are currently used in Sweden. The development of more advanced cheaper heat accumulators, such as lake storage schemes, could make even seasonal heat storage economic. Newer distribution technology extends the economic field of penetration of district heating even to suburban one-family house districts. With proper design and optimization, nuclear district heating can be competitive in a wide market and achieve very substantial fossil-fuel savings

  15. Massively Parallel and Scalable Implicit Time Integration Algorithms for Structural Dynamics

    Science.gov (United States)

    Farhat, Charbel

    1997-01-01

    Explicit codes are often used to simulate the nonlinear dynamics of large-scale structural systems, even for low frequency response, because the storage and CPU requirements entailed by the repeated factorizations traditionally found in implicit codes rapidly overwhelm the available computing resources. With the advent of parallel processing, this trend is accelerating because of the following additional facts: (a) explicit schemes are easier to parallelize than implicit ones, and (b) explicit schemes induce short range interprocessor communications that are relatively inexpensive, while the factorization methods used in most implicit schemes induce long range interprocessor communications that often ruin the sought-after speed-up. However, the time step restriction imposed by the Courant stability condition on all explicit schemes cannot yet be offset by the speed of the currently available parallel hardware. Therefore, it is essential to develop efficient alternatives to direct methods that are also amenable to massively parallel processing because implicit codes using unconditionally stable time-integration algorithms are computationally more efficient when simulating the low-frequency dynamics of aerospace structures.

  16. Storage facility for highly radioactive solid waste

    International Nuclear Information System (INIS)

    Kitano, Shozo

    1996-01-01

    A heat insulation plate is disposed at an intermediate portion between a ceiling wall of a storage chamber and an upper plate of a storage pit in parallel with them. A large number of highly radioactive solid wastes contained in canisters are contained in the storage pit. Cooling air is introduced from an air suction port, passes a channel on the upper side of the heat insulation plate formed by the ceiling of the storage chamber and the heat insulation plate, and flows from a flow channel on the side of the wall of the storage chamber to the lower portion of the storage pit. Afterheat is removed by the air flown from the lower portion to ventilation tubes at the outer side of container tubes. The air heated to a high temperature through the flow channel on the lower side of the heat insulation plate between the heat insulation plate and the upper plate of the storage pit, and is exhausted to an exhaustion port. Further, a portion of a heat insulation plate as a boundary between the cooling air and a high temperature air formed on the upper portion of the storage pit is formed as a heat transfer plate, so that the heat of the high temperature air is removed by the cooling air flowing the upper flow channel. This can prevent heating of the ceiling wall of the storage chamber. (I.N.)

  17. Parallelism in computations in quantum and statistical mechanics

    International Nuclear Information System (INIS)

    Clementi, E.; Corongiu, G.; Detrich, J.H.

    1985-01-01

    Often very fundamental biochemical and biophysical problems defy simulations because of limitations in today's computers. We present and discuss a distributed system composed of two IBM 4341 s and/or an IBM 4381 as front-end processors and ten FPS-164 attached array processors. This parallel system - called LCAP - has presently a peak performance of about 110 Mflops; extensions to higher performance are discussed. Presently, the system applications use a modified version of VM/SP as the operating system: description of the modifications is given. Three applications programs have been migrated from sequential to parallel: a molecular quantum mechanical, a Metropolis-Monte Carlo and a molecular dynamics program. Descriptions of the parallel codes are briefly outlined. Use of these parallel codes has already opened up new capabilities for our research. The very positive performance comparisons with today's supercomputers allow us to conclude that parallel computers and programming, of the type we have considered, represent a pragmatic answer to many computationally intensive problems. (orig.)

  18. Highly efficient distributed generation and high-capacity energy storage

    DEFF Research Database (Denmark)

    Hemmes, Kas; Guerrero, Josep M.; Zhelev, Toshko

    2012-01-01

    With the growing amount of decentralized power production the design and operation of the grid has to be reconsidered. New problems include the two-way flow of electricity and maintaining the power balance given the increased amount of uncertain and fluctuating renewable energy sources like wind...... and solar that deliver electricity to the grid. Solution directions are the development of smart grids, demand side management, virtual power plants and storage of electricity. These are directions that, rightly so, are already attracting a lot of attention and R&D funding. In this paper critical issues...... and fuel that can also fulfill a storage function....

  19. Progress on H5Part: A Portable High Performance Parallel Data Interface for Electromagnetics Simulations

    International Nuclear Information System (INIS)

    Adelmann, Andreas; Gsell, Achim; Oswald, Benedikt; Schietinger, Thomas; Bethel, Wes; Shalf, John; Siegerist, Cristina; Stockinger, Kurt

    2007-01-01

    Significant problems facing all experimental and computational sciences arise from growing data size and complexity. Common to all these problems is the need to perform efficient data I/O on diverse computer architectures. In our scientific application, the largest parallel particle simulations generate vast quantities of six-dimensional data. Such a simulation run produces data for an aggregate data size up to several TB per run. Motivated by the need to address data I/O and access challenges, we have implemented H5Part, an open source data I/O API that simplifies the use of the Hierarchical Data Format v5 library (HDF5). HDF5 is an industry standard for high performance, cross-platform data storage and retrieval that runs on all contemporary architectures from large parallel supercomputers to laptops. H5Part, which is oriented to the needs of the particle physics and cosmology communities, provides support for parallel storage and retrieval of particles, structured and in the future unstructured meshes. In this paper, we describe recent work focusing on I/O support for particles and structured meshes and provide data showing performance on modern supercomputer architectures like the IBM POWER 5

  20. Fuzzy pool balance: An algorithm to achieve a two dimensional balance in distribute storage systems

    International Nuclear Information System (INIS)

    Wu, Wenjing; Chen, Gang

    2014-01-01

    The limitation of scheduling modules and the gradual addition of disk pools in distributed storage systems often result in imbalances among their disk pools in terms of both disk usage and file count. This can cause various problems to the storage system such as single point of failure, low system throughput and imbalanced resource utilization and system loads. An algorithm named Fuzzy Pool Balance (FPB) is proposed here to solve this problem. The input of FPB is the current file distribution among disk pools and the output is a file migration plan indicating what files are to be migrated to which pools. FPB uses an array to classify the files by their sizes. The file classification array is dynamically calculated with a defined threshold named T max that defines the allowed pool disk usage deviations. File classification is the basis of file migration. FPB also defines the Immigration Pool (IP) and Emigration Pool (EP) according to the pool disk usage and File Quantity Ratio (FQR) that indicates the percentage of each category of files in each disk pool, so files with higher FQR in an EP will be migrated to IP(s) with a lower FQR of this file category. To verify this algorithm, we implemented FPB on an ATLAS Tier2 dCache production system. The results show that FPB can achieve a very good balance in both free space and file counts, and adjusting the threshold value T max and the correction factor to the average FQR can achieve a tradeoff between free space and file count.