Scientific computer simulation review
International Nuclear Information System (INIS)
Kaizer, Joshua S.; Heller, A. Kevin; Oberkampf, William L.
2015-01-01
Before the results of a scientific computer simulation are used for any purpose, it should be determined if those results can be trusted. Answering that question of trust is the domain of scientific computer simulation review. There is limited literature that focuses on simulation review, and most is specific to the review of a particular type of simulation. This work is intended to provide a foundation for a common understanding of simulation review. This is accomplished through three contributions. First, scientific computer simulation review is formally defined. This definition identifies the scope of simulation review and provides the boundaries of the review process. Second, maturity assessment theory is developed. This development clarifies the concepts of maturity criteria, maturity assessment sets, and maturity assessment frameworks, which are essential for performing simulation review. Finally, simulation review is described as the application of a maturity assessment framework. This is illustrated through evaluating a simulation review performed by the U.S. Nuclear Regulatory Commission. In making these contributions, this work provides a means for a more objective assessment of a simulation’s trustworthiness and takes the next step in establishing scientific computer simulation review as its own field. - Highlights: • We define scientific computer simulation review. • We develop maturity assessment theory. • We formally define a maturity assessment framework. • We describe simulation review as the application of a maturity framework. • We provide an example of a simulation review using a maturity framework
International Nuclear Information System (INIS)
Pongratz, M.B.
1981-01-01
The motivation for active experiments in space is considered, taking into account the use of active techniques to obtain a better understanding of the natural space environment, the utilization of the advantages of space as a laboratory to study fundamental plasma physics, and the employment of active techniques to determine the magnitude, degree, and consequences of artificial modification of the space environment. It is pointed out that mass-injection experiments in space plasmas began about twenty years ago with the Project Firefly releases. Attention is given to mass-release techniques and diagnostics, operational aspects of mass release active experiments, the active observation of mass release experiments, active perturbation mass release experiments, simulating an artificial modification of the space environment, and active experiments to study fundamental plasma physics
Speedup predictions on large scientific parallel programs
International Nuclear Information System (INIS)
Williams, E.; Bobrowicz, F.
1985-01-01
How much speedup can we expect for large scientific parallel programs running on supercomputers. For insight into this problem we extend the parallel processing environment currently existing on the Cray X-MP (a shared memory multiprocessor with at most four processors) to a simulated N-processor environment, where N greater than or equal to 1. Several large scientific parallel programs from Los Alamos National Laboratory were run in this simulated environment, and speedups were predicted. A speedup of 14.4 on 16 processors was measured for one of the three most used codes at the Laboratory
Scientific Modeling and simulations
Diaz de la Rubia, Tomás
2009-01-01
Showcases the conceptual advantages of modeling which, coupled with the unprecedented computing power through simulations, allow scientists to tackle the formibable problems of our society, such as the search for hydrocarbons, understanding the structure of a virus, or the intersection between simulations and real data in extreme environments
Software Engineering for Scientific Computer Simulations
Post, Douglass E.; Henderson, Dale B.; Kendall, Richard P.; Whitney, Earl M.
2004-11-01
Computer simulation is becoming a very powerful tool for analyzing and predicting the performance of fusion experiments. Simulation efforts are evolving from including only a few effects to many effects, from small teams with a few people to large teams, and from workstations and small processor count parallel computers to massively parallel platforms. Successfully making this transition requires attention to software engineering issues. We report on the conclusions drawn from a number of case studies of large scale scientific computing projects within DOE, academia and the DoD. The major lessons learned include attention to sound project management including setting reasonable and achievable requirements, building a good code team, enforcing customer focus, carrying out verification and validation and selecting the optimum computational mathematics approaches.
Learning from large scale neural simulations
DEFF Research Database (Denmark)
Serban, Maria
2017-01-01
Large-scale neural simulations have the marks of a distinct methodology which can be fruitfully deployed to advance scientific understanding of the human brain. Computer simulation studies can be used to produce surrogate observational data for better conceptual models and new how...
Computational Simulations and the Scientific Method
Kleb, Bil; Wood, Bill
2005-01-01
As scientific simulation software becomes more complicated, the scientific-software implementor's need for component tests from new model developers becomes more crucial. The community's ability to follow the basic premise of the Scientific Method requires independently repeatable experiments, and model innovators are in the best position to create these test fixtures. Scientific software developers also need to quickly judge the value of the new model, i.e., its cost-to-benefit ratio in terms of gains provided by the new model and implementation risks such as cost, time, and quality. This paper asks two questions. The first is whether other scientific software developers would find published component tests useful, and the second is whether model innovators think publishing test fixtures is a feasible approach.
LARGE BUILDING HVAC SIMULATION
The report discusses the monitoring and collection of data relating to indoor pressures and radon concentrations under several test conditions in a large school building in Bartow, Florida. The Florida Solar Energy Center (FSEC) used an integrated computational software, FSEC 3.0...
Numerical simulation of turbulent combustion: Scientific challenges
Ren, ZhuYin; Lu, Zhen; Hou, LingYun; Lu, LiuYan
2014-08-01
Predictive simulation of engine combustion is key to understanding the underlying complicated physicochemical processes, improving engine performance, and reducing pollutant emissions. Critical issues as turbulence modeling, turbulence-chemistry interaction, and accommodation of detailed chemical kinetics in complex flows remain challenging and essential for high-fidelity combustion simulation. This paper reviews the current status of the state-of-the-art large eddy simulation (LES)/prob-ability density function (PDF)/detailed chemistry approach that can address the three challenging modelling issues. PDF as a subgrid model for LES is formulated and the hybrid mesh-particle method for LES/PDF simulations is described. Then the development need in micro-mixing models for the PDF simulations of turbulent premixed combustion is identified. Finally the different acceleration methods for detailed chemistry are reviewed and a combined strategy is proposed for further development.
Large Eddy Simulation of turbulence
International Nuclear Information System (INIS)
Poullet, P.; Sancandi, M.
1994-12-01
Results of Large Eddy Simulation of 3D isotropic homogeneous turbulent flows are presented. A computer code developed on Connexion Machine (CM5) has allowed to compare two turbulent viscosity models (Smagorinsky and structure function). The numerical scheme influence on the energy density spectrum is also studied [fr
Application of Logic Models in a Large Scientific Research Program
O'Keefe, Christine M.; Head, Richard J.
2011-01-01
It is the purpose of this article to discuss the development and application of a logic model in the context of a large scientific research program within the Commonwealth Scientific and Industrial Research Organisation (CSIRO). CSIRO is Australia's national science agency and is a publicly funded part of Australia's innovation system. It conducts…
Advanced I/O for large-scale scientific applications
International Nuclear Information System (INIS)
Klasky, Scott; Schwan, Karsten; Oldfield, Ron A.; Lofstead, Gerald F. II
2010-01-01
As scientific simulations scale to use petascale machines and beyond, the data volumes generated pose a dual problem. First, with increasing machine sizes, the careful tuning of IO routines becomes more and more important to keep the time spent in IO acceptable. It is not uncommon, for instance, to have 20% of an application's runtime spent performing IO in a 'tuned' system. Careful management of the IO routines can move that to 5% or even less in some cases. Second, the data volumes are so large, on the order of 10s to 100s of TB, that trying to discover the scientifically valid contributions requires assistance at runtime to both organize and annotate the data. Waiting for offline processing is not feasible due both to the impact on the IO system and the time required. To reduce this load and improve the ability of scientists to use the large amounts of data being produced, new techniques for data management are required. First, there is a need for techniques for efficient movement of data from the compute space to storage. These techniques should understand the underlying system infrastructure and adapt to changing system conditions. Technologies include aggregation networks, data staging nodes for a closer parity to the IO subsystem, and autonomic IO routines that can detect system bottlenecks and choose different approaches, such as splitting the output into multiple targets, staggering output processes. Such methods must be end-to-end, meaning that even with properly managed asynchronous techniques, it is still essential to properly manage the later synchronous interaction with the storage system to maintain acceptable performance. Second, for the data being generated, annotations and other metadata must be incorporated to help the scientist understand output data for the simulation run as a whole, to select data and data features without concern for what files or other storage technologies were employed. All of these features should be attained while
Studying Scientific Discovery by Computer Simulation.
1983-03-30
Mendel’s laws of inheritance, the law of Gay- Lussac for gaseous reactions, tile law of Dulong and Petit, the derivation of atomic weights by Avogadro...neceseary mid identify by block number) scientific discovery -ittri sic properties physical laws extensive terms data-driven heuristics intensive...terms theory-driven heuristics conservation laws 20. ABSTRACT (Continue on revere. side It necessary and identify by block number) Scientific discovery
The Large-Scale Structure of Scientific Method
Kosso, Peter
2009-01-01
The standard textbook description of the nature of science describes the proposal, testing, and acceptance of a theoretical idea almost entirely in isolation from other theories. The resulting model of science is a kind of piecemeal empiricism that misses the important network structure of scientific knowledge. Only the large-scale description of…
Technologies for Large Data Management in Scientific Computing
Pace, A
2014-01-01
In recent years, intense usage of computing has been the main strategy of investigations in several scientific research projects. The progress in computing technology has opened unprecedented opportunities for systematic collection of experimental data and the associated analysis that were considered impossible only few years ago. This paper focusses on the strategies in use: it reviews the various components that are necessary for an effective solution that ensures the storage, the long term preservation, and the worldwide distribution of large quantities of data that are necessary in a large scientific research project. The paper also mentions several examples of data management solutions used in High Energy Physics for the CERN Large Hadron Collider (LHC) experiments in Geneva, Switzerland which generate more than 30,000 terabytes of data every year that need to be preserved, analyzed, and made available to a community of several tenth of thousands scientists worldwide.
On-demand Overlay Networks for Large Scientific Data Transfers
Energy Technology Data Exchange (ETDEWEB)
Ramakrishnan, Lavanya [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Guok, Chin [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Jackson, Keith [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Kissel, Ezra [Univ. of Delaware, Newark, DE (United States); Swany, D. Martin [Univ. of Delaware, Newark, DE (United States); Agarwal, Deborah [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States)
2009-10-12
Large scale scientific data transfers are central to scientific processes. Data from large experimental facilities have to be moved to local institutions for analysis or often data needs to be moved between local clusters and large supercomputing centers. In this paper, we propose and evaluate a network overlay architecture to enable highthroughput, on-demand, coordinated data transfers over wide-area networks. Our work leverages Phoebus and On-demand Secure Circuits and AdvanceReservation System (OSCARS) to provide high performance wide-area network connections. OSCARS enables dynamic provisioning of network paths with guaranteed bandwidth and Phoebus enables the coordination and effective utilization of the OSCARS network paths. Our evaluation shows that this approach leads to improved end-to-end data transfer throughput with minimal overheads. The achievedthroughput using our overlay was limited only by the ability of the end hosts to sink the data.
Large-scale numerical simulations of plasmas
International Nuclear Information System (INIS)
Hamaguchi, Satoshi
2004-01-01
The recent trend of large scales simulations of fusion plasma and processing plasmas is briefly summarized. Many advanced simulation techniques have been developed for fusion plasmas and some of these techniques are now applied to analyses of processing plasmas. (author)
Modeling and simulation of large HVDC systems
Energy Technology Data Exchange (ETDEWEB)
Jin, H.; Sood, V.K.
1993-01-01
This paper addresses the complexity and the amount of work in preparing simulation data and in implementing various converter control schemes and the excessive simulation time involved in modelling and simulation of large HVDC systems. The Power Electronic Circuit Analysis program (PECAN) is used to address these problems and a large HVDC system with two dc links is simulated using PECAN. A benchmark HVDC system is studied to compare the simulation results with those from other packages. The simulation time and results are provided in the paper.
Software quality and process improvement in scientific simulation codes
Energy Technology Data Exchange (ETDEWEB)
Ambrosiano, J.; Webster, R. [Los Alamos National Lab., NM (United States)
1997-11-01
This report contains viewgraphs on the quest to develope better simulation code quality through process modeling and improvement. This study is based on the experience of the authors and interviews with ten subjects chosen from simulation code development teams at LANL. This study is descriptive rather than scientific.
Distributed simulation of large computer systems
International Nuclear Information System (INIS)
Marzolla, M.
2001-01-01
Sequential simulation of large complex physical systems is often regarded as a computationally expensive task. In order to speed-up complex discrete-event simulations, the paradigm of Parallel and Distributed Discrete Event Simulation (PDES) has been introduced since the late 70s. The authors analyze the applicability of PDES to the modeling and analysis of large computer system; such systems are increasingly common in the area of High Energy and Nuclear Physics, because many modern experiments make use of large 'compute farms'. Some feasibility tests have been performed on a prototype distributed simulator
Parallel Tensor Compression for Large-Scale Scientific Data.
Energy Technology Data Exchange (ETDEWEB)
Kolda, Tamara G. [Sandia National Lab. (SNL-CA), Livermore, CA (United States); Ballard, Grey [Sandia National Lab. (SNL-CA), Livermore, CA (United States); Austin, Woody Nathan [Univ. of Texas, Austin, TX (United States)
2015-10-01
As parallel computing trends towards the exascale, scientific data produced by high-fidelity simulations are growing increasingly massive. For instance, a simulation on a three-dimensional spatial grid with 512 points per dimension that tracks 64 variables per grid point for 128 time steps yields 8 TB of data. By viewing the data as a dense five way tensor, we can compute a Tucker decomposition to find inherent low-dimensional multilinear structure, achieving compression ratios of up to 10000 on real-world data sets with negligible loss in accuracy. So that we can operate on such massive data, we present the first-ever distributed memory parallel implementation for the Tucker decomposition, whose key computations correspond to parallel linear algebra operations, albeit with nonstandard data layouts. Our approach specifies a data distribution for tensors that avoids any tensor data redistribution, either locally or in parallel. We provide accompanying analysis of the computation and communication costs of the algorithms. To demonstrate the compression and accuracy of the method, we apply our approach to real-world data sets from combustion science simulations. We also provide detailed performance results, including parallel performance in both weak and strong scaling experiments.
Cray XT4: An Early Evaluation for Petascale Scientific Simulation
International Nuclear Information System (INIS)
Alam, Sadaf R.; Barrett, Richard F.; Fahey, Mark R.; Kuehn, Jeffery A.; Sankaran, Ramanan; Worley, Patrick H.; Larkin, Jeffrey M.
2007-01-01
The scientific simulation capabilities of next generation high-end computing technology will depend on striking a balance among memory, processor, I/O, and local and global network performance across the breadth of the scientific simulation space. The Cray XT4 combines commodity AMD dual core Opteron processor technology with the second generation of Cray's custom communication accelerator in a system design whose balance is claimed to be driven by the demands of scientific simulation. This paper presents an evaluation of the Cray XT4 using microbenchmarks to develop a controlled understanding of individual system components, providing the context for analyzing and comprehending the performance of several petascale-ready applications. Results gathered from several strategic application domains are compared with observations on the previous generation Cray XT3 and other high-end computing systems, demonstrating performance improvements across a wide variety of application benchmark problems.
Guiquan, Xi; Lin, Cong; Xuehui, Jin
2018-05-01
As an important platform for scientific and technological development, large -scale scientific facilities are the cornerstone of technological innovation and a guarantee for economic and social development. Researching management of large-scale scientific facilities can play a key role in scientific research, sociology and key national strategy. This paper reviews the characteristics of large-scale scientific facilities, and summarizes development status of China's large-scale scientific facilities. At last, the construction, management, operation and evaluation of large-scale scientific facilities is analyzed from the perspective of sustainable development.
Large Eddy Simulations using oodlesDST
2016-01-01
Research Agency DST-Group-TR-3205 ABSTRACT The oodlesDST code is based on OpenFOAM software and performs Large Eddy Simulations of......maritime platforms using a variety of simulation techniques. He is currently using OpenFOAM software to perform both Reynolds Averaged Navier-Stokes
Large eddy simulation of bundle turbulent flows
International Nuclear Information System (INIS)
Hassan, Y.A.; Barsamian, H.R.
1995-01-01
Large eddy simulation may be defined as simulation of a turbulent flow in which the large scale motions are explicitly resolved while the small scale motions are modeled. This results into a system of equations that require closure models. The closure models relate the effects of the small scale motions onto the large scale motions. There have been several models developed, the most popular is the Smagorinsky eddy viscosity model. A new model has recently been introduced by Lee that modified the Smagorinsky model. Using both of the above mentioned closure models, two different geometric arrangements were used in the simulation of turbulent cross flow within rigid tube bundles. An inlined array simulations was performed for a deep bundle (10,816 nodes) as well as an inlet/outlet simulation (57,600 nodes). Comparisons were made to available experimental data. Flow visualization enabled the distinction of different characteristics within the flow such as jet switching effects in the wake of the bundle flow for the inlet/outlet simulation case, as well as within tube bundles. The results indicate that the large eddy simulation technique is capable of turbulence prediction and may be used as a viable engineering tool with the careful consideration of the subgrid scale model. (author)
Web-based visualization of very large scientific astronomy imagery
Bertin, E.; Pillay, R.; Marmo, C.
2015-04-01
Visualizing and navigating through large astronomy images from a remote location with current astronomy display tools can be a frustrating experience in terms of speed and ergonomics, especially on mobile devices. In this paper, we present a high performance, versatile and robust client-server system for remote visualization and analysis of extremely large scientific images. Applications of this work include survey image quality control, interactive data query and exploration, citizen science, as well as public outreach. The proposed software is entirely open source and is designed to be generic and applicable to a variety of datasets. It provides access to floating point data at terabyte scales, with the ability to precisely adjust image settings in real-time. The proposed clients are light-weight, platform-independent web applications built on standard HTML5 web technologies and compatible with both touch and mouse-based devices. We put the system to the test and assess the performance of the system and show that a single server can comfortably handle more than a hundred simultaneous users accessing full precision 32 bit astronomy data.
Numerical simulation of large deformation polycrystalline plasticity
International Nuclear Information System (INIS)
Inal, K.; Neale, K.W.; Wu, P.D.; MacEwen, S.R.
2000-01-01
A finite element model based on crystal plasticity has been developed to simulate the stress-strain response of sheet metal specimens in uniaxial tension. Each material point in the sheet is considered to be a polycrystalline aggregate of FCC grains. The Taylor theory of crystal plasticity is assumed. The numerical analysis incorporates parallel computing features enabling simulations of realistic models with large number of grains. Simulations have been carried out for the AA3004-H19 aluminium alloy and the results are compared with experimental data. (author)
Computer simulations and the changing face of scientific experimentation
Duran, Juan M
2013-01-01
Computer simulations have become a central tool for scientific practice. Their use has replaced, in many cases, standard experimental procedures. This goes without mentioning cases where the target system is empirical but there are no techniques for direct manipulation of the system, such as astronomical observation. To these cases, computer simulations have proved to be of central importance. The question about their use and implementation, therefore, is not only a technical one but represents a challenge for the humanities as well. In this volume, scientists, historians, and philosophers joi
Regularization modeling for large-eddy simulation
Geurts, Bernardus J.; Holm, D.D.
2003-01-01
A new modeling approach for large-eddy simulation (LES) is obtained by combining a "regularization principle" with an explicit filter and its inversion. This regularization approach allows a systematic derivation of the implied subgrid model, which resolves the closure problem. The central role of
Remote collaboration system based on large scale simulation
International Nuclear Information System (INIS)
Kishimoto, Yasuaki; Sugahara, Akihiro; Li, J.Q.
2008-01-01
Large scale simulation using super-computer, which generally requires long CPU time and produces large amount of data, has been extensively studied as a third pillar in various advanced science fields in parallel to theory and experiment. Such a simulation is expected to lead new scientific discoveries through elucidation of various complex phenomena, which are hardly identified only by conventional theoretical and experimental approaches. In order to assist such large simulation studies for which many collaborators working at geographically different places participate and contribute, we have developed a unique remote collaboration system, referred to as SIMON (simulation monitoring system), which is based on client-server system control introducing an idea of up-date processing, contrary to that of widely used post-processing. As a key ingredient, we have developed a trigger method, which transmits various requests for the up-date processing from the simulation (client) running on a super-computer to a workstation (server). Namely, the simulation running on a super-computer actively controls the timing of up-date processing. The server that has received the requests from the ongoing simulation such as data transfer, data analyses, and visualizations, etc. starts operations according to the requests during the simulation. The server makes the latest results available to web browsers, so that the collaborators can monitor the results at any place and time in the world. By applying the system to a specific simulation project of laser-matter interaction, we have confirmed that the system works well and plays an important role as a collaboration platform on which many collaborators work with one another
nanoHUB.org: Experiences and Challenges in Software Sustainability for a Large Scientific Community
Directory of Open Access Journals (Sweden)
Lynn Zentner
2014-07-01
Full Text Available The science gateway nanoHUB.org, funded by the National Science Foundation (NSF, serves a large scientific community dedicated to research and education in nanotechnology with community-contributed simulation codes as well as a vast repository of other materials such as recorded presentations, teaching materials, and workshops and courses. Nearly 330,000 users annually access over 4400 items of content on nanoHUB, including 343 simulation tools. Arguably the largest nanotechnology facility in the world, nanoHUB has led the way not only in providing open access to scientific code in the nanotechnology community, but also in lowering barriers to the use of that code, by providing a platform where developers are able to easily and quickly deploy code written in a variety of languages with user-friendly graphical user interfaces and where users can run the latest versions of codes transparently on the grid or other powerful resources without ever having to download or update code. Being a leader in open access code deployment provides nanoHUB with opportunities and challenges as it meets the current and future needs of its community. This paper discusses the experiences of nanoHUB in addressing and adapting to the changing landscape of scientific software in ways that best serve its community and meet the needs of the largest portion of its user base.
Unified Access Architecture for Large-Scale Scientific Datasets
Karna, Risav
2014-05-01
Data-intensive sciences have to deploy diverse large scale database technologies for data analytics as scientists have now been dealing with much larger volume than ever before. While array databases have bridged many gaps between the needs of data-intensive research fields and DBMS technologies (Zhang 2011), invocation of other big data tools accompanying these databases is still manual and separate the database management's interface. We identify this as an architectural challenge that will increasingly complicate the user's work flow owing to the growing number of useful but isolated and niche database tools. Such use of data analysis tools in effect leaves the burden on the user's end to synchronize the results from other data manipulation analysis tools with the database management system. To this end, we propose a unified access interface for using big data tools within large scale scientific array database using the database queries themselves to embed foreign routines belonging to the big data tools. Such an invocation of foreign data manipulation routines inside a query into a database can be made possible through a user-defined function (UDF). UDFs that allow such levels of freedom as to call modules from another language and interface back and forth between the query body and the side-loaded functions would be needed for this purpose. For the purpose of this research we attempt coupling of four widely used tools Hadoop (hadoop1), Matlab (matlab1), R (r1) and ScaLAPACK (scalapack1) with UDF feature of rasdaman (Baumann 98), an array-based data manager, for investigating this concept. The native array data model used by an array-based data manager provides compact data storage and high performance operations on ordered data such as spatial data, temporal data, and matrix-based data for linear algebra operations (scidbusr1). Performances issues arising due to coupling of tools with different paradigms, niche functionalities, separate processes and output
Direct and large-eddy simulation IX
Kuerten, Hans; Geurts, Bernard; Armenio, Vincenzo
2015-01-01
This volume reflects the state of the art of numerical simulation of transitional and turbulent flows and provides an active forum for discussion of recent developments in simulation techniques and understanding of flow physics. Following the tradition of earlier DLES workshops, these papers address numerous theoretical and physical aspects of transitional and turbulent flows. At an applied level it contributes to the solution of problems related to energy production, transportation, magneto-hydrodynamics and the environment. A special session is devoted to quality issues of LES. The ninth Workshop on 'Direct and Large-Eddy Simulation' (DLES-9) was held in Dresden, April 3-5, 2013, organized by the Institute of Fluid Mechanics at Technische Universität Dresden. This book is of interest to scientists and engineers, both at an early level in their career and at more senior levels.
Field simulations for large dipole magnets
International Nuclear Information System (INIS)
Lazzaro, A.; Cappuzzello, F.; Cunsolo, A.; Cavallaro, M.; Foti, A.; Khouaja, A.; Orrigo, S.E.A.; Winfield, J.S.
2007-01-01
The problem of the description of magnetic field for large bending magnets is addressed in relation to the requirements of modern techniques of trajectory reconstruction. The crucial question of the interpolation and extrapolation of fields known at a discrete number of points is analysed. For this purpose a realistic field model of the large dipole of the MAGNEX spectrometer, obtained with finite elements three dimensional simulations, is used. The influence of the uncertainties in the measured field to the quality of the trajectory reconstruction is treated in detail. General constraints for field measurements in terms of required resolutions, step sizes and precisions are thus extracted
Large data management and systematization of simulation
International Nuclear Information System (INIS)
Ueshima, Yutaka; Saitho, Kanji; Koga, James; Isogai, Kentaro
2004-01-01
In the advanced photon research large-scale simulations are powerful tools. In the numerical experiments, real-time visualization and steering system are thought as hopeful methods of data analysis. This approach is valid in the stereotype analysis at one time or short-cycle simulation. In the research for an unknown problem, it is necessary that the output data can be analyzed many times because profitable analysis is difficult at the first time. Consequently, output data should be filed to refer and analyze at any time. To support the research, we need the followed automatic functions, transporting data files from data generator to data storage, analyzing data, tracking history of data handling, and so on. The Large Data Management system will be functional Problem Solving Environment distributed system. (author)
Large Eddy Simulation for Compressible Flows
Garnier, E; Sagaut, P
2009-01-01
Large Eddy Simulation (LES) of compressible flows is still a widely unexplored area of research. The authors, whose books are considered the most relevant monographs in this field, provide the reader with a comprehensive state-of-the-art presentation of the available LES theory and application. This book is a sequel to "Large Eddy Simulation for Incompressible Flows", as most of the research on LES for compressible flows is based on variable density extensions of models, methods and paradigms that were developed within the incompressible flow framework. The book addresses both the fundamentals and the practical industrial applications of LES in order to point out gaps in the theoretical framework as well as to bridge the gap between LES research and the growing need to use it in engineering modeling. After introducing the fundamentals on compressible turbulence and the LES governing equations, the mathematical framework for the filtering paradigm of LES for compressible flow equations is established. Instead ...
Large eddy simulation of hydrodynamic cavitation
Bhatt, Mrugank; Mahesh, Krishnan
2017-11-01
Large eddy simulation is used to study sheet to cloud cavitation over a wedge. The mixture of water and water vapor is represented using a homogeneous mixture model. Compressible Navier-Stokes equations for mixture quantities along with transport equation for vapor mass fraction employing finite rate mass transfer between the two phases, are solved using the numerical method of Gnanaskandan and Mahesh. The method is implemented on unstructured grid with parallel MPI capabilities. Flow over a wedge is simulated at Re = 200 , 000 and the performance of the homogeneous mixture model is analyzed in predicting different regimes of sheet to cloud cavitation; namely, incipient, transitory and periodic, as observed in the experimental investigation of Harish et al.. This work is supported by the Office of Naval Research.
Large-eddy simulation of contrails
Energy Technology Data Exchange (ETDEWEB)
Chlond, A [Max-Planck-Inst. fuer Meteorologie, Hamburg (Germany)
1998-12-31
A large eddy simulation (LES) model has been used to investigate the role of various external parameters and physical processes in the life-cycle of contrails. The model is applied to conditions that are typical for those under which contrails could be observed, i.e. in an atmosphere which is supersaturated with respect to ice and at a temperature of approximately 230 K or colder. The sensitivity runs indicate that the contrail evolution is controlled primarily by humidity, temperature and static stability of the ambient air and secondarily by the baroclinicity of the atmosphere. Moreover, it turns out that the initial ice particle concentration and radiative processes are of minor importance in the evolution of contrails at least during the 30 minutes simulation period. (author) 9 refs.
Large eddy simulation of breaking waves
DEFF Research Database (Denmark)
Christensen, Erik Damgaard; Deigaard, Rolf
2001-01-01
A numerical model is used to simulate wave breaking, the large scale water motions and turbulence induced by the breaking process. The model consists of a free surface model using the surface markers method combined with a three-dimensional model that solves the flow equations. The turbulence....... The incoming waves are specified by a flux boundary condition. The waves are approaching in the shore-normal direction and are breaking on a plane, constant slope beach. The first few wave periods are simulated by a two-dimensional model in the vertical plane normal to the beach line. The model describes...... the steepening and the overturning of the wave. At a given instant, the model domain is extended to three dimensions, and the two-dimensional flow field develops spontaneously three-dimensional flow features with turbulent eddies. After a few wave periods, stationary (periodic) conditions are achieved...
Large-eddy simulation of contrails
Energy Technology Data Exchange (ETDEWEB)
Chlond, A. [Max-Planck-Inst. fuer Meteorologie, Hamburg (Germany)
1997-12-31
A large eddy simulation (LES) model has been used to investigate the role of various external parameters and physical processes in the life-cycle of contrails. The model is applied to conditions that are typical for those under which contrails could be observed, i.e. in an atmosphere which is supersaturated with respect to ice and at a temperature of approximately 230 K or colder. The sensitivity runs indicate that the contrail evolution is controlled primarily by humidity, temperature and static stability of the ambient air and secondarily by the baroclinicity of the atmosphere. Moreover, it turns out that the initial ice particle concentration and radiative processes are of minor importance in the evolution of contrails at least during the 30 minutes simulation period. (author) 9 refs.
Scientific and Computational Challenges of the Fusion Simulation Program (FSP)
International Nuclear Information System (INIS)
Tang, William M.
2011-01-01
This paper highlights the scientific and computational challenges facing the Fusion Simulation Program (FSP) a major national initiative in the United States with the primary objective being to enable scientific discovery of important new plasma phenomena with associated understanding that emerges only upon integration. This requires developing a predictive integrated simulation capability for magnetically-confined fusion plasmas that are properly validated against experiments in regimes relevant for producing practical fusion energy. It is expected to provide a suite of advanced modeling tools for reliably predicting fusion device behavior with comprehensive and targeted science-based simulations of nonlinearly-coupled phenomena in the core plasma, edge plasma, and wall region on time and space scales required for fusion energy production. As such, it will strive to embody the most current theoretical and experimental understanding of magnetic fusion plasmas and to provide a living framework for the simulation of such plasmas as the associated physics understanding continues to advance over the next several decades. Substantive progress on answering the outstanding scientific questions in the field will drive the FSP toward its ultimate goal of developing the ability to predict the behavior of plasma discharges in toroidal magnetic fusion devices with high physics fidelity on all relevant time and space scales. From a computational perspective, this will demand computing resources in the petascale range and beyond together with the associated multi-core algorithmic formulation needed to address burning plasma issues relevant to ITER - a multibillion dollar collaborative experiment involving seven international partners representing over half the world's population. Even more powerful exascale platforms will be needed to meet the future challenges of designing a demonstration fusion reactor (DEMO). Analogous to other major applied physics modeling projects (e
Scientific and computational challenges of the fusion simulation program (FSP)
International Nuclear Information System (INIS)
Tang, William M.
2011-01-01
This paper highlights the scientific and computational challenges facing the Fusion Simulation Program (FSP) - a major national initiative in the United States with the primary objective being to enable scientific discovery of important new plasma phenomena with associated understanding that emerges only upon integration. This requires developing a predictive integrated simulation capability for magnetically-confined fusion plasmas that are properly validated against experiments in regimes relevant for producing practical fusion energy. It is expected to provide a suite of advanced modeling tools for reliably predicting fusion device behavior with comprehensive and targeted science-based simulations of nonlinearly-coupled phenomena in the core plasma, edge plasma, and wall region on time and space scales required for fusion energy production. As such, it will strive to embody the most current theoretical and experimental understanding of magnetic fusion plasmas and to provide a living framework for the simulation of such plasmas as the associated physics understanding continues to advance over the next several decades. Substantive progress on answering the outstanding scientific questions in the field will drive the FSP toward its ultimate goal of developing the ability to predict the behavior of plasma discharges in toroidal magnetic fusion devices with high physics fidelity on all relevant time and space scales. From a computational perspective, this will demand computing resources in the petascale range and beyond together with the associated multi-core algorithmic formulation needed to address burning plasma issues relevant to ITER - a multibillion dollar collaborative experiment involving seven international partners representing over half the world's population. Even more powerful exascale platforms will be needed to meet the future challenges of designing a demonstration fusion reactor (DEMO). Analogous to other major applied physics modeling projects (e
Large eddy simulations of compressible magnetohydrodynamic turbulence
International Nuclear Information System (INIS)
Grete, Philipp
2016-01-01
Supersonic, magnetohydrodynamic (MHD) turbulence is thought to play an important role in many processes - especially in astrophysics, where detailed three-dimensional observations are scarce. Simulations can partially fill this gap and help to understand these processes. However, direct simulations with realistic parameters are often not feasible. Consequently, large eddy simulations (LES) have emerged as a viable alternative. In LES the overall complexity is reduced by simulating only large and intermediate scales directly. The smallest scales, usually referred to as subgrid-scales (SGS), are introduced to the simulation by means of an SGS model. Thus, the overall quality of an LES with respect to properly accounting for small-scale physics crucially depends on the quality of the SGS model. While there has been a lot of successful research on SGS models in the hydrodynamic regime for decades, SGS modeling in MHD is a rather recent topic, in particular, in the compressible regime. In this thesis, we derive and validate a new nonlinear MHD SGS model that explicitly takes compressibility effects into account. A filter is used to separate the large and intermediate scales, and it is thought to mimic finite resolution effects. In the derivation, we use a deconvolution approach on the filter kernel. With this approach, we are able to derive nonlinear closures for all SGS terms in MHD: the turbulent Reynolds and Maxwell stresses, and the turbulent electromotive force (EMF). We validate the new closures both a priori and a posteriori. In the a priori tests, we use high-resolution reference data of stationary, homogeneous, isotropic MHD turbulence to compare exact SGS quantities against predictions by the closures. The comparison includes, for example, correlations of turbulent fluxes, the average dissipative behavior, and alignment of SGS vectors such as the EMF. In order to quantify the performance of the new nonlinear closure, this comparison is conducted from the
Large eddy simulations of compressible magnetohydrodynamic turbulence
Grete, Philipp
2017-02-01
Supersonic, magnetohydrodynamic (MHD) turbulence is thought to play an important role in many processes - especially in astrophysics, where detailed three-dimensional observations are scarce. Simulations can partially fill this gap and help to understand these processes. However, direct simulations with realistic parameters are often not feasible. Consequently, large eddy simulations (LES) have emerged as a viable alternative. In LES the overall complexity is reduced by simulating only large and intermediate scales directly. The smallest scales, usually referred to as subgrid-scales (SGS), are introduced to the simulation by means of an SGS model. Thus, the overall quality of an LES with respect to properly accounting for small-scale physics crucially depends on the quality of the SGS model. While there has been a lot of successful research on SGS models in the hydrodynamic regime for decades, SGS modeling in MHD is a rather recent topic, in particular, in the compressible regime. In this thesis, we derive and validate a new nonlinear MHD SGS model that explicitly takes compressibility effects into account. A filter is used to separate the large and intermediate scales, and it is thought to mimic finite resolution effects. In the derivation, we use a deconvolution approach on the filter kernel. With this approach, we are able to derive nonlinear closures for all SGS terms in MHD: the turbulent Reynolds and Maxwell stresses, and the turbulent electromotive force (EMF). We validate the new closures both a priori and a posteriori. In the a priori tests, we use high-resolution reference data of stationary, homogeneous, isotropic MHD turbulence to compare exact SGS quantities against predictions by the closures. The comparison includes, for example, correlations of turbulent fluxes, the average dissipative behavior, and alignment of SGS vectors such as the EMF. In order to quantify the performance of the new nonlinear closure, this comparison is conducted from the
Large eddy simulations of compressible magnetohydrodynamic turbulence
Energy Technology Data Exchange (ETDEWEB)
Grete, Philipp
2016-09-09
Supersonic, magnetohydrodynamic (MHD) turbulence is thought to play an important role in many processes - especially in astrophysics, where detailed three-dimensional observations are scarce. Simulations can partially fill this gap and help to understand these processes. However, direct simulations with realistic parameters are often not feasible. Consequently, large eddy simulations (LES) have emerged as a viable alternative. In LES the overall complexity is reduced by simulating only large and intermediate scales directly. The smallest scales, usually referred to as subgrid-scales (SGS), are introduced to the simulation by means of an SGS model. Thus, the overall quality of an LES with respect to properly accounting for small-scale physics crucially depends on the quality of the SGS model. While there has been a lot of successful research on SGS models in the hydrodynamic regime for decades, SGS modeling in MHD is a rather recent topic, in particular, in the compressible regime. In this thesis, we derive and validate a new nonlinear MHD SGS model that explicitly takes compressibility effects into account. A filter is used to separate the large and intermediate scales, and it is thought to mimic finite resolution effects. In the derivation, we use a deconvolution approach on the filter kernel. With this approach, we are able to derive nonlinear closures for all SGS terms in MHD: the turbulent Reynolds and Maxwell stresses, and the turbulent electromotive force (EMF). We validate the new closures both a priori and a posteriori. In the a priori tests, we use high-resolution reference data of stationary, homogeneous, isotropic MHD turbulence to compare exact SGS quantities against predictions by the closures. The comparison includes, for example, correlations of turbulent fluxes, the average dissipative behavior, and alignment of SGS vectors such as the EMF. In order to quantify the performance of the new nonlinear closure, this comparison is conducted from the
Trends in scientific publishing: Dark clouds loom large.
Vinny, Pulikottil Wilson; Vishnu, Venugopalan Y; Lal, Vivek
2016-04-15
The world wide web has brought about a paradigm shift in the way medical research is published and accessed. The ease with which a new journal can be started/hosted by publishing start-ups is unprecedented. The tremendous capabilities of the world wide web and the open access revolution when combined with a highly profitable business have attracted unscrupulous fraudulent operators to the publishing industry. The intent of these fraudulent publishers is solely driven by profit with utter disregard to scientific content, peer reviews and ethics. This phenomenon has been referred to as "predatory publishing". The "international" tag of such journals often betrays their true origins. The gold open access model of publishing, where the author pays the publisher, when coupled with a non-existent peer review threatens to blur the distinction between science and pseudoscience. The average researcher needs to be made more aware of this clear and present danger to the scientific community. Prevention is better than cure. Copyright © 2016 Elsevier B.V. All rights reserved.
Scientific and computational challenges of the fusion simulation project (FSP)
International Nuclear Information System (INIS)
Tang, W M
2008-01-01
This paper highlights the scientific and computational challenges facing the Fusion Simulation Project (FSP). The primary objective is to develop advanced software designed to use leadership-class computers for carrying out multiscale physics simulations to provide information vital to delivering a realistic integrated fusion simulation model with unprecedented physics fidelity. This multiphysics capability will be unprecedented in that in the current FES applications domain, the largest-scale codes are used to carry out first-principles simulations of mostly individual phenomena in realistic 3D geometry while the integrated models are much smaller-scale, lower-dimensionality codes with significant empirical elements used for modeling and designing experiments. The FSP is expected to be the most up-to-date embodiment of the theoretical and experimental understanding of magnetically confined thermonuclear plasmas and to provide a living framework for the simulation of such plasmas as the associated physics understanding continues to advance over the next several decades. Substantive progress on answering the outstanding scientific questions in the field will drive the FSP toward its ultimate goal of developing a reliable ability to predict the behavior of plasma discharges in toroidal magnetic fusion devices on all relevant time and space scales. From a computational perspective, the fusion energy science application goal to produce high-fidelity, whole-device modeling capabilities will demand computing resources in the petascale range and beyond, together with the associated multicore algorithmic formulation needed to address burning plasma issues relevant to ITER - a multibillion dollar collaborative device involving seven international partners representing over half the world's population. Even more powerful exascale platforms will be needed to meet the future challenges of designing a demonstration fusion reactor (DEMO). Analogous to other major applied physics
Large eddy simulation of cavitating flows
Gnanaskandan, Aswin; Mahesh, Krishnan
2014-11-01
Large eddy simulation on unstructured grids is used to study hydrodynamic cavitation. The multiphase medium is represented using a homogeneous equilibrium model that assumes thermal equilibrium between the liquid and the vapor phase. Surface tension effects are ignored and the governing equations are the compressible Navier Stokes equations for the liquid/vapor mixture along with a transport equation for the vapor mass fraction. A characteristic-based filtering scheme is developed to handle shocks and material discontinuities in non-ideal gases and mixtures. A TVD filter is applied as a corrector step in a predictor-corrector approach with the predictor scheme being non-dissipative and symmetric. The method is validated for canonical one dimensional flows and leading edge cavitation over a hydrofoil, and applied to study sheet to cloud cavitation over a wedge. This work is supported by the Office of Naval Research.
Large-eddy simulations for turbulent flows
International Nuclear Information System (INIS)
Husson, S.
2007-07-01
The aim of this work is to study the impact of thermal gradients on a turbulent channel flow with imposed wall temperatures and friction Reynolds numbers of 180 and 395. In this configuration, temperature variations can be strong and induce significant variations of the fluid properties. We consider the low Mach number equations and carry out large eddy simulations. We first validate our simulations thanks to comparisons of some of our LES results with DNS data. Then, we investigate the influence of the variations of the conductivity and the viscosity and show that we can assume these properties constant only for weak temperature gradients. We also study the thermal sub-grid-scale modelling and find no difference when the sub-grid-scale Prandtl number is taken constant or dynamically calculated. The analysis of the effects of strongly increasing the temperature ratio mainly shows a dissymmetry of the profiles. The physical mechanism responsible of these modifications is explained. Finally, we use semi-local scaling and the Van Driest transformation and we show that they lead to a better correspondence of the low and high temperature ratios profiles. (author)
Dynamic large eddy simulation: Stability via realizability
Mokhtarpoor, Reza; Heinz, Stefan
2017-10-01
The concept of dynamic large eddy simulation (LES) is highly attractive: such methods can dynamically adjust to changing flow conditions, which is known to be highly beneficial. For example, this avoids the use of empirical, case dependent approximations (like damping functions). Ideally, dynamic LES should be local in physical space (without involving artificial clipping parameters), and it should be stable for a wide range of simulation time steps, Reynolds numbers, and numerical schemes. These properties are not trivial, but dynamic LES suffers from such problems over decades. We address these questions by performing dynamic LES of periodic hill flow including separation at a high Reynolds number Re = 37 000. For the case considered, the main result of our studies is that it is possible to design LES that has the desired properties. It requires physical consistency: a PDF-realizable and stress-realizable LES model, which requires the inclusion of the turbulent kinetic energy in the LES calculation. LES models that do not honor such physical consistency can become unstable. We do not find support for the previous assumption that long-term correlations of negative dynamic model parameters are responsible for instability. Instead, we concluded that instability is caused by the stable spatial organization of significant unphysical states, which are represented by wall-type gradient streaks of the standard deviation of the dynamic model parameter. The applicability of our realizability stabilization to other dynamic models (including the dynamic Smagorinsky model) is discussed.
Large-scale Intelligent Transporation Systems simulation
Energy Technology Data Exchange (ETDEWEB)
Ewing, T.; Canfield, T.; Hannebutte, U.; Levine, D.; Tentner, A.
1995-06-01
A prototype computer system has been developed which defines a high-level architecture for a large-scale, comprehensive, scalable simulation of an Intelligent Transportation System (ITS) capable of running on massively parallel computers and distributed (networked) computer systems. The prototype includes the modelling of instrumented ``smart`` vehicles with in-vehicle navigation units capable of optimal route planning and Traffic Management Centers (TMC). The TMC has probe vehicle tracking capabilities (display position and attributes of instrumented vehicles), and can provide 2-way interaction with traffic to provide advisories and link times. Both the in-vehicle navigation module and the TMC feature detailed graphical user interfaces to support human-factors studies. The prototype has been developed on a distributed system of networked UNIX computers but is designed to run on ANL`s IBM SP-X parallel computer system for large scale problems. A novel feature of our design is that vehicles will be represented by autonomus computer processes, each with a behavior model which performs independent route selection and reacts to external traffic events much like real vehicles. With this approach, one will be able to take advantage of emerging massively parallel processor (MPP) systems.
Sensitivity technologies for large scale simulation
International Nuclear Information System (INIS)
Collis, Samuel Scott; Bartlett, Roscoe Ainsworth; Smith, Thomas Michael; Heinkenschloss, Matthias; Wilcox, Lucas C.; Hill, Judith C.; Ghattas, Omar; Berggren, Martin Olof; Akcelik, Volkan; Ober, Curtis Curry; van Bloemen Waanders, Bart Gustaaf; Keiter, Eric Richard
2005-01-01
Sensitivity analysis is critically important to numerous analysis algorithms, including large scale optimization, uncertainty quantification,reduced order modeling, and error estimation. Our research focused on developing tools, algorithms and standard interfaces to facilitate the implementation of sensitivity type analysis into existing code and equally important, the work was focused on ways to increase the visibility of sensitivity analysis. We attempt to accomplish the first objective through the development of hybrid automatic differentiation tools, standard linear algebra interfaces for numerical algorithms, time domain decomposition algorithms and two level Newton methods. We attempt to accomplish the second goal by presenting the results of several case studies in which direct sensitivities and adjoint methods have been effectively applied, in addition to an investigation of h-p adaptivity using adjoint based a posteriori error estimation. A mathematical overview is provided of direct sensitivities and adjoint methods for both steady state and transient simulations. Two case studies are presented to demonstrate the utility of these methods. A direct sensitivity method is implemented to solve a source inversion problem for steady state internal flows subject to convection diffusion. Real time performance is achieved using novel decomposition into offline and online calculations. Adjoint methods are used to reconstruct initial conditions of a contamination event in an external flow. We demonstrate an adjoint based transient solution. In addition, we investigated time domain decomposition algorithms in an attempt to improve the efficiency of transient simulations. Because derivative calculations are at the root of sensitivity calculations, we have developed hybrid automatic differentiation methods and implemented this approach for shape optimization for gas dynamics using the Euler equations. The hybrid automatic differentiation method was applied to a first
TESLA: Large Signal Simulation Code for Klystrons
International Nuclear Information System (INIS)
Vlasov, Alexander N.; Cooke, Simon J.; Chernin, David P.; Antonsen, Thomas M. Jr.; Nguyen, Khanh T.; Levush, Baruch
2003-01-01
TESLA (Telegraphist's Equations Solution for Linear Beam Amplifiers) is a new code designed to simulate linear beam vacuum electronic devices with cavities, such as klystrons, extended interaction klystrons, twistrons, and coupled cavity amplifiers. The model includes a self-consistent, nonlinear solution of the three-dimensional electron equations of motion and the solution of time-dependent field equations. The model differs from the conventional Particle in Cell approach in that the field spectrum is assumed to consist of a carrier frequency and its harmonics with slowly varying envelopes. Also, fields in the external cavities are modeled with circuit like equations and couple to fields in the beam region through boundary conditions on the beam tunnel wall. The model in TESLA is an extension of the model used in gyrotron code MAGY. The TESLA formulation has been extended to be capable to treat the multiple beam case, in which each beam is transported inside its own tunnel. The beams interact with each other as they pass through the gaps in their common cavities. The interaction is treated by modification of the boundary conditions on the wall of each tunnel to include the effect of adjacent beams as well as the fields excited in each cavity. The extended version of TESLA for the multiple beam case, TESLA-MB, has been developed for single processor machines, and can run on UNIX machines and on PC computers with a large memory (above 2GB). The TESLA-MB algorithm is currently being modified to simulate multiple beam klystrons on multiprocessor machines using the MPI (Message Passing Interface) environment. The code TESLA has been verified by comparison with MAGIC for single and multiple beam cases. The TESLA code and the MAGIC code predict the same power within 1% for a simple two cavity klystron design while the computational time for TESLA is orders of magnitude less than for MAGIC 2D. In addition, recently TESLA was used to model the L-6048 klystron, code
Promoting access to and use of seismic data in a large scientific community
Directory of Open Access Journals (Sweden)
Michel Eric
2017-01-01
Full Text Available The growing amount of seismic data available from space missions (SOHO, CoRoT, Kepler, SDO,… but also from ground-based facilities (GONG, BiSON, ground-based large programmes…, stellar modelling and numerical simulations, creates new scientific perspectives such as characterizing stellar populations in our Galaxy or planetary systems by providing model-independent global properties of stars such as mass, radius, and surface gravity within several percent accuracy, as well as constraints on the age. These applications address a broad scientific community beyond the solar and stellar one and require combining indices elaborated with data from different databases (e.g. seismic archives and ground-based spectroscopic surveys. It is thus a basic requirement to develop a simple and effcient access to these various data resources and dedicated tools. In the framework of the European project SpaceInn (FP7, several data sources have been developed or upgraded. The Seismic Plus Portal has been developed, where synthetic descriptions of the most relevant existing data sources can be found, as well as tools allowing to localize existing data for given objects or period and helping the data query. This project has been developed within the Virtual Observatory (VO framework. In this paper, we give a review of the various facilities and tools developed within this programme. The SpaceInn project (Exploitation of Space Data for Innovative Helio- and Asteroseismology has been initiated by the European Helio- and Asteroseismology Network (HELAS.
Development of a common data model for scientific simulations
Energy Technology Data Exchange (ETDEWEB)
Ambrosiano, J. [Los Alamos National Lab., NM (United States); Butler, D.M. [Limit Point Systems, Inc. (United States); Matarazzo, C.; Miller, M. [Lawrence Livermore National Lab., CA (United States); Schoof, L. [Sandia National Lab., Albuquerque, NM (United States)
1999-06-01
The problem of sharing data among scientific simulation models is a difficult and persistent one. Computational scientists employ an enormous variety of discrete approximations in modeling physical processes on computers. Problems occur when models based on different representations are required to exchange data with one another, or with some other software package. Within the DOE`s Accelerated Strategic Computing Initiative (ASCI), a cross-disciplinary group called the Data Models and Formats (DMF) group, has been working to develop a common data model. The current model is comprised of several layers of increasing semantic complexity. One of these layers is an abstract model based on set theory and topology called the fiber bundle kernel (FBK). This layer provides the flexibility needed to describe a wide range of mesh-approximated functions as well as other entities. This paper briefly describes the ASCI common data model, its mathematical basis, and ASCI prototype development. These prototypes include an object-oriented data management library developed at Los Alamos called the Common Data Model Library or CDMlib, the Vector Bundle API from the Lawrence Livermore Laboratory, and the DMF API from Sandia National Laboratory.
Large scale molecular simulations of nanotoxicity.
Jimenez-Cruz, Camilo A; Kang, Seung-gu; Zhou, Ruhong
2014-01-01
The widespread use of nanomaterials in biomedical applications has been accompanied by an increasing interest in understanding their interactions with tissues, cells, and biomolecules, and in particular, on how they might affect the integrity of cell membranes and proteins. In this mini-review, we present a summary of some of the recent studies on this important subject, especially from the point of view of large scale molecular simulations. The carbon-based nanomaterials and noble metal nanoparticles are the main focus, with additional discussions on quantum dots and other nanoparticles as well. The driving forces for adsorption of fullerenes, carbon nanotubes, and graphene nanosheets onto proteins or cell membranes are found to be mainly hydrophobic interactions and the so-called π-π stacking (between aromatic rings), while for the noble metal nanoparticles the long-range electrostatic interactions play a bigger role. More interestingly, there are also growing evidences showing that nanotoxicity can have implications in de novo design of nanomedicine. For example, the endohedral metallofullerenol Gd@C₈₂(OH)₂₂ is shown to inhibit tumor growth and metastasis by inhibiting enzyme MMP-9, and graphene is illustrated to disrupt bacteria cell membranes by insertion/cutting as well as destructive extraction of lipid molecules. These recent findings have provided a better understanding of nanotoxicity at the molecular level and also suggested therapeutic potential by using the cytotoxicity of nanoparticles against cancer or bacteria cells. © 2014 Wiley Periodicals, Inc.
Large-Eddy Simulation of Subsonic Jets
International Nuclear Information System (INIS)
Vuorinen, Ville; Wehrfritz, Armin; Yu Jingzhou; Kaario, Ossi; Larmi, Martti; Boersma, Bendiks Jan
2011-01-01
The present study deals with development and validation of a fully explicit, compressible Runge-Kutta-4 (RK4) Navier-Stokes solver in the opensource CFD programming environment OpenFOAM. The background motivation is to shift towards explicit density based solution strategy and thereby avoid using the pressure based algorithms which are currently proposed in the standard OpenFOAM release for Large-Eddy Simulation (LES). This shift is considered necessary in strongly compressible flows when Ma > 0.5. Our application of interest is related to the pre-mixing stage in direct injection gas engines where high injection pressures are typically utilized. First, the developed flow solver is discussed and validated. Then, the implementation of subsonic inflow conditions using a forcing region in combination with a simplified nozzle geometry is discussed and validated. After this, LES of mixing in compressible, round jets at Ma = 0.3, 0.5 and 0.65 are carried out. Respectively, the Reynolds numbers of the jets correspond to Re = 6000, 10000 and 13000. Results for two meshes are presented. The results imply that the present solver produces turbulent structures, resolves a range of turbulent eddy frequencies and gives also mesh independent results within satisfactory limits for mean flow and turbulence statistics.
Large eddy simulation of stably stratified turbulence
International Nuclear Information System (INIS)
Shen Zhi; Zhang Zhaoshun; Cui Guixiang; Xu Chunxiao
2011-01-01
Stably stratified turbulence is a common phenomenon in atmosphere and ocean. In this paper the large eddy simulation is utilized for investigating homogeneous stably stratified turbulence numerically at Reynolds number Re = uL/v = 10 2 ∼10 3 and Froude number Fr = u/NL = 10 −2 ∼10 0 in which u is root mean square of velocity fluctuations, L is integral scale and N is Brunt-Vaïsälä frequency. Three sets of computation cases are designed with different initial conditions, namely isotropic turbulence, Taylor Green vortex and internal waves, to investigate the statistical properties from different origins. The computed horizontal and vertical energy spectra are consistent with observation in atmosphere and ocean when the composite parameter ReFr 2 is greater than O(1). It has also been found in this paper that the stratification turbulence can be developed under different initial velocity conditions and the internal wave energy is dominated in the developed stably stratified turbulence.
Background simulations for the Large Area Detector onboard LOFT
DEFF Research Database (Denmark)
Campana, Riccardo; Feroci, Marco; Ettore, Del Monte
2013-01-01
and magnetic fields around compact objects and in supranuclear density conditions. Having an effective area of similar to 10 m(2) at 8 keV, LOFT will be able to measure with high sensitivity very fast variability in the X-ray fluxes and spectra. A good knowledge of the in-orbit background environment...... is essential to assess the scientific performance of the mission and optimize the design of its main instrument, the Large Area Detector (LAD). In this paper the results of an extensive Geant-4 simulation of the instrumentwillbe discussed, showing the main contributions to the background and the design...... an anticipated modulation of the background rate as small as 10 % over the orbital timescale. The intrinsic photonic origin of the largest background component also allows for an efficient modelling, supported by an in-flight active monitoring, allowing to predict systematic residuals significantly better than...
Paul Scherrer Institute Scientific and Technical Report 2000. Volume VI: Large Research Facilities
International Nuclear Information System (INIS)
Foroughi, Fereydoun; Bercher, Renate; Buechli, Carmen; Zumkeller, Lotty
2001-01-01
The PSI Department Large Research Facilities (GFA) joins the efforts to provide an excellent research environment to Swiss and foreign research groups on the experimental facilities driven by our high intensity proton accelerator complex. Its divisions care for the running, maintenance and enhancement of the accelerator complex, the primary proton beamlines, the targets and the secondary beams as well as the neutron spallation source SINQ. The division for technical support and coordination provides for technical support to the research facility complementary to the basic logistic available from the department for logistics and marketing. Besides running the facilities, the staff of the department is also involved in theoretical and experimental research projects. Some of them address basic scientific questions mainly concerning the properties of micro- or nanostructured materials: experiments as well as large scale computer simulations of molecular dynamics were performed to investigate nonclassical materials properties. Others are related to improvements or extensions of the capabilities of our facilities. We also report on intriguing results from applications of the neutron capture radiography, the prompt gamma activation method and the isotope production facility at SINQ
Paul Scherrer Institute Scientific and Technical Report 2000. Volume VI: Large Research Facilities
Energy Technology Data Exchange (ETDEWEB)
Foroughi, Fereydoun; Bercher, Renate; Buechli, Carmen; Zumkeller, Lotty [eds.
2001-07-01
The PSI Department Large Research Facilities (GFA) joins the efforts to provide an excellent research environment to Swiss and foreign research groups on the experimental facilities driven by our high intensity proton accelerator complex. Its divisions care for the running, maintenance and enhancement of the accelerator complex, the primary proton beamlines, the targets and the secondary beams as well as the neutron spallation source SINQ. The division for technical support and coordination provides for technical support to the research facility complementary to the basic logistic available from the department for logistics and marketing. Besides running the facilities, the staff of the department is also involved in theoretical and experimental research projects. Some of them address basic scientific questions mainly concerning the properties of micro- or nanostructured materials: experiments as well as large scale computer simulations of molecular dynamics were performed to investigate nonclassical materials properties. Others are related to improvements or extensions of the capabilities of our facilities. We also report on intriguing results from applications of the neutron capture radiography, the prompt gamma activation method and the isotope production facility at SINQ.
Large-Signal Klystron Simulations Using KLSC
Carlsten, B. E.; Ferguson, P.
1997-05-01
We describe a new, 2-1/2 dimensional, klystron-simulation code, KLSC. This code has a sophisticated input cavity model for calculating the klystron gain with arbitrary input cavity matching and tuning, and is capable of modeling coupled output cavities. We will discuss the input and output cavity models, and present simulation results from a high-power, S-band design. We will use these results to explore tuning issues with coupled output cavities.
Large-Scale Assessment, Rationality, and Scientific Management: The Case of No Child Left Behind
Roach, Andrew T.; Frank, Jennifer
2007-01-01
This article examines the ways in which NCLB and the movement towards large-scale assessment systems are based on Weber's concept of formal rationality and tradition of scientific management. Building on these ideas, the authors use Ritzer's McDonaldization thesis to examine some of the core features of large-scale assessment and accountability…
A high performance scientific cloud computing environment for materials simulations
Jorissen, Kevin; Vila, Fernando D.; Rehr, John J.
2011-01-01
We describe the development of a scientific cloud computing (SCC) platform that offers high performance computation capability. The platform consists of a scientific virtual machine prototype containing a UNIX operating system and several materials science codes, together with essential interface tools (an SCC toolset) that offers functionality comparable to local compute clusters. In particular, our SCC toolset provides automatic creation of virtual clusters for parallel computing, including...
Advancements in Large-Scale Data/Metadata Management for Scientific Data.
Guntupally, K.; Devarakonda, R.; Palanisamy, G.; Frame, M. T.
2017-12-01
the future enhancements of these tools which enable users to retrieve fast search results, along with parallelizing the retrieval process from online and High Performance Storage Systems. In addition, these improvements to the tools will support additional metadata formats like the Large-Eddy Simulation (LES) ARM Symbiotic and Observation (LASSO) bundle data.
A high performance scientific cloud computing environment for materials simulations
Jorissen, K.; Vila, F. D.; Rehr, J. J.
2012-09-01
We describe the development of a scientific cloud computing (SCC) platform that offers high performance computation capability. The platform consists of a scientific virtual machine prototype containing a UNIX operating system and several materials science codes, together with essential interface tools (an SCC toolset) that offers functionality comparable to local compute clusters. In particular, our SCC toolset provides automatic creation of virtual clusters for parallel computing, including tools for execution and monitoring performance, as well as efficient I/O utilities that enable seamless connections to and from the cloud. Our SCC platform is optimized for the Amazon Elastic Compute Cloud (EC2). We present benchmarks for prototypical scientific applications and demonstrate performance comparable to local compute clusters. To facilitate code execution and provide user-friendly access, we have also integrated cloud computing capability in a JAVA-based GUI. Our SCC platform may be an alternative to traditional HPC resources for materials science or quantum chemistry applications.
Erdemir, Ahmet
2016-01-01
Virtual representations of the knee joint can provide clinicians, scientists, and engineers the tools to explore mechanical function of the knee and its tissue structures in health and disease. Modeling and simulation approaches such as finite element analysis also provide the possibility to understand the influence of surgical procedures and implants on joint stresses and tissue deformations. A large number of knee joint models are described in the biomechanics literature. However, freely accessible, customizable, and easy-to-use models are scarce. Availability of such models can accelerate clinical translation of simulations, where labor intensive reproduction of model development steps can be avoided. The interested parties can immediately utilize readily available models for scientific discovery and for clinical care. Motivated by this gap, this study aims to describe an open source and freely available finite element representation of the tibiofemoral joint, namely Open Knee, which includes detailed anatomical representation of the joint's major tissue structures, their nonlinear mechanical properties and interactions. Three use cases illustrate customization potential of the model, its predictive capacity, and its scientific and clinical utility: prediction of joint movements during passive flexion, examining the role of meniscectomy on contact mechanics and joint movements, and understanding anterior cruciate ligament mechanics. A summary of scientific and clinically directed studies conducted by other investigators are also provided. The utilization of this open source model by groups other than its developers emphasizes the premise of model sharing as an accelerator of simulation-based medicine. Finally, the imminent need to develop next generation knee models are noted. These are anticipated to incorporate individualized anatomy and tissue properties supported by specimen-specific joint mechanics data for evaluation, all acquired in vitro from varying age
Proceedings of the meeting on large scale computer simulation research
International Nuclear Information System (INIS)
2004-04-01
The meeting to summarize the collaboration activities for FY2003 on the Large Scale Computer Simulation Research was held January 15-16, 2004 at Theory and Computer Simulation Research Center, National Institute for Fusion Science. Recent simulation results, methodologies and other related topics were presented. (author)
Large TileCal magnetic field simulation
International Nuclear Information System (INIS)
Nessi, M.; Bergsma, F.; Vorozhtsov, S.B.; Borisov, O.N.; Lomakina, O.V.; Karamysheva, G.A.; Budagov, Yu.A.
1994-01-01
The ATLAS magnetic field map has been estimated in the presence of the hadron tile calorimeter. This is an important issue in order to quantify the needs for individual PMT shielding, the effect on the scintillator light yield and its implications on the calibration. The field source is based on a central solenoid and 8 superconducting air-core toroidal coils. The maximum induction value in the scintillating tiles does not exceed 6 mT. When an iron plate is used to close the open drawer window the field inside the PMT near to the extended barrel edge does not exceed 0.6 mT. Estimation of ponder motive force distribution, acting on individual units of the system was performed. VF electromagnetic software OPERA-TOSCA and CERN POISCR code were used for the field simulation of the system. 10 refs., 4 figs
Mathematics of large eddy simulation of turbulent flows
Energy Technology Data Exchange (ETDEWEB)
Berselli, L.C. [Pisa Univ. (Italy). Dept. of Applied Mathematics ' ' U. Dini' ' ; Iliescu, T. [Virginia Polytechnic Inst. and State Univ., Blacksburg, VA (United States). Dept. of Mathematics; Layton, W.J. [Pittsburgh Univ., PA (United States). Dept. of Mathematics
2006-07-01
Large eddy simulation (LES) is a method of scientific computation seeking to predict the dynamics of organized structures in turbulent flows by approximating local, spatial averages of the flow. Since its birth in 1970, LES has undergone an explosive development and has matured into a highly-developed computational technology. It uses the tools of turbulence theory and the experience gained from practical computation. This book focuses on the mathematical foundations of LES and its models and provides a connection between the powerful tools of applied mathematics, partial differential equations and LES. Thus, it is concerned with fundamental aspects not treated so deeply in the other books in the field, aspects such as well-posedness of the models, their energy balance and the connection to the Leray theory of weak solutions of the Navier-Stokes equations. The authors give a mathematically informed and detailed treatment of an interesting selection of models, focusing on issues connected with understanding and expanding the correctness and universality of LES. This volume offers a useful entry point into the field for PhD students in applied mathematics, computational mathematics and partial differential equations. Non-mathematicians will appreciate it as a reference that introduces them to current tools and advances in the mathematical theory of LES. (orig.)
Large Scale Simulation Platform for NODES Validation Study
Energy Technology Data Exchange (ETDEWEB)
Sotorrio, P. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Qin, Y. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Min, L. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)
2017-04-27
This report summarizes the Large Scale (LS) simulation platform created for the Eaton NODES project. The simulation environment consists of both wholesale market simulator and distribution simulator and includes the CAISO wholesale market model and a PG&E footprint of 25-75 feeders to validate the scalability under a scenario of 33% RPS in California with additional 17% of DERS coming from distribution and customers. The simulator can generate hourly unit commitment, 5-minute economic dispatch, and 4-second AGC regulation signals. The simulator is also capable of simulating greater than 10k individual controllable devices. Simulated DERs include water heaters, EVs, residential and light commercial HVAC/buildings, and residential-level battery storage. Feeder-level voltage regulators and capacitor banks are also simulated for feeder-level real and reactive power management and Vol/Var control.
Efficient Feature-Driven Visualization of Large-Scale Scientific Data
Energy Technology Data Exchange (ETDEWEB)
Lu, Aidong
2012-12-12
Very large, complex scientific data acquired in many research areas creates critical challenges for scientists to understand, analyze, and organize their data. The objective of this project is to expand the feature extraction and analysis capabilities to develop powerful and accurate visualization tools that can assist domain scientists with their requirements in multiple phases of scientific discovery. We have recently developed several feature-driven visualization methods for extracting different data characteristics of volumetric datasets. Our results verify the hypothesis in the proposal and will be used to develop additional prototype systems.
Scientific codes developed and used at GRS. Nuclear simulation chain
Energy Technology Data Exchange (ETDEWEB)
Schaffrath, Andreas; Sonnenkalb, Martin; Sievers, Juergen; Luther, Wolfgang; Velkov, Kiril [Gesellschaft fuer Anlagen und Reaktorsicherheit (GRS) gGmbH, Garching/Muenchen (Germany). Forschungszentrum
2016-05-15
Over 60 technical experts of the reactor safety research division of the Gesellschaft fuer Anlagen- und Reaktorsicherheit (GRS) gGmbH are developing and validating reliable methods and computer codes - summarized under the term nuclear simulation chain - for the safety-related assessment for all types of nuclear power plants (NPP) and other nuclear facilities considering the current state of science and technology. This nuclear simulation chain has to be able to simulate and assess all relevant physical processes and phenomena for all operating states and (severe) accidents. In the present contribution, the nuclear simulation chain developed and applied by GRS as well as selected examples of its application are presented. The latter demonstrate impressively the width of its scope and its performance. The GRS codes can be passed on request to other (national as well as international) organizations. This contributes to a worldwide increase of the nuclear safety standards. The code transfer is especially important for developing and emerging countries lacking the financial means and/or the necessary know-how for this purpose. At the end of this contribution, the respective course of action is described.
Visualization of the Flux Rope Generation Process Using Large Quantities of MHD Simulation Data
Directory of Open Access Journals (Sweden)
Y Kubota
2013-03-01
Full Text Available We present a new concept of analysis using visualization of large quantities of simulation data. The time development of 3D objects with high temporal resolution provides the opportunity for scientific discovery. We visualize large quantities of simulation data using the visualization application 'Virtual Aurora' based on AVS (Advanced Visual Systems and the parallel distributed processing at "Space Weather Cloud" in NICT based on Gfarm technology. We introduce two results of high temporal resolution visualization: the magnetic flux rope generation process and dayside reconnection using a system of magnetic field line tracing.
International Nuclear Information System (INIS)
Veleiro Blanco, A. M.
2011-01-01
The cryogenic plants and their control in Scientific-Nuclear Facilities is complicated by the large number of variables and the wide range of variation during operation. Initially the design and control of these systems in CERN was based on stationary calculations which non yielded the expected results. Due to its complexity, the dynamic simulation is the only way to get adequate results during operational transients.
A Combined Ethical and Scientific Analysis of Large-scale Tests of Solar Climate Engineering
Ackerman, T. P.
2017-12-01
Our research group recently published an analysis of the combined ethical and scientific issues surrounding large-scale testing of stratospheric aerosol injection (SAI; Lenferna et al., 2017, Earth's Future). We are expanding this study in two directions. The first is extending this same analysis to other geoengineering techniques, particularly marine cloud brightening (MCB). MCB has substantial differences to SAI in this context because MCB can be tested over significantly smaller areas of the planet and, following injection, has a much shorter lifetime of weeks as opposed to years for SAI. We examine issues such as the role of intent, the lesser of two evils, and the nature of consent. In addition, several groups are currently considering climate engineering governance tools such as a code of ethics and a registry. We examine how these tools might influence climate engineering research programs and, specifically, large-scale testing. The second direction of expansion is asking whether ethical and scientific issues associated with large-scale testing are so significant that they effectively preclude moving ahead with climate engineering research and testing. Some previous authors have suggested that no research should take place until these issues are resolved. We think this position is too draconian and consider a more nuanced version of this argument. We note, however, that there are serious questions regarding the ability of the scientific research community to move to the point of carrying out large-scale tests.
Utilization of Large Cohesive Interface Elements for Delamination Simulation
DEFF Research Database (Denmark)
Bak, Brian Lau Verndal; Lund, Erik
2012-01-01
This paper describes the difficulties of utilizing large interface elements in delamination simulation. Solutions to increase the size of applicable interface elements are described and cover numerical integration of the element and modifications of the cohesive law....
Large eddy simulation of premixed and non-premixed combustion
Malalasekera, W; Ibrahim, SS; Masri, AR; Sadasivuni, SK; Gubba, SR
2010-01-01
This paper summarises the authors experience in using the Large Eddy Simulation (LES) technique for the modelling of premixed and non-premixed combustion. The paper describes the application of LES based combustion modelling technique to two well defined experimental configurations where high quality data is available for validation. The large eddy simulation technique for the modelling flow and turbulence is based on the solution of governing equations for continuity and momentum in a struct...
Energy Technology Data Exchange (ETDEWEB)
Ahrens, James P [ORNL; Heitmann, Katrin [ORNL; Petersen, Mark R [ORNL; Woodring, Jonathan [Los Alamos National Laboratory (LANL); Williams, Sean [Los Alamos National Laboratory (LANL); Fasel, Patricia [Los Alamos National Laboratory (LANL); Ahrens, Christine [Los Alamos National Laboratory (LANL); Hsu, Chung-Hsing [ORNL; Geveci, Berk [ORNL
2010-11-01
This article presents a visualization-assisted process that verifies scientific-simulation codes. Code verification is necessary because scientists require accurate predictions to interpret data confidently. This verification process integrates iterative hypothesis verification with comparative, feature, and quantitative visualization. Following this process can help identify differences in cosmological and oceanographic simulations.
On asymptotically efficient simulation of large deviation probabilities.
Dieker, A.B.; Mandjes, M.R.H.
2005-01-01
ABSTRACT: Consider a family of probabilities for which the decay is governed by a large deviation principle. To find an estimate for a fixed member of this family, one is often forced to use simulation techniques. Direct Monte Carlo simulation, however, is often impractical, particularly if the
Design of General-purpose Industrial signal acquisition system in a large scientific device
Ren, Bin; Yang, Lei
2018-02-01
In order to measure the industrial signal of a large scientific device experiment, a set of industrial data general-purpose acquisition system has been designed. It can collect 4~20mA current signal and 0~10V voltage signal. Through the practical experiments, it shows that the system is flexible, reliable, convenient and economical, and the system has characters of high definition and strong anti-interference ability. Thus, the system fully meets the design requirements..
Research of Impact Load in Large Electrohydraulic Load Simulator
Directory of Open Access Journals (Sweden)
Yongguang Liu
2014-01-01
Full Text Available The stronger impact load will appear in the initial phase when the large electric cylinder is tested in the hardware-in-loop simulation. In this paper, the mathematical model is built based on AMESim, and then the reason of the impact load is investigated through analyzing the changing tendency of parameters in the simulation results. The inhibition methods of impact load are presented according to the structural invariability principle and applied to the actual system. The final experimental result indicates that the impact load is inhibited, which provides a good experimental condition for the electric cylinder and promotes the study of large load simulator.
Quality and Reliability of Large-Eddy Simulations
Meyers, Johan; Sagaut, Pierre
2008-01-01
Computational resources have developed to the level that, for the first time, it is becoming possible to apply large-eddy simulation (LES) to turbulent flow problems of realistic complexity. Many examples can be found in technology and in a variety of natural flows. This puts issues related to assessing, assuring, and predicting the quality of LES into the spotlight. Several LES studies have been published in the past, demonstrating a high level of accuracy with which turbulent flow predictions can be attained, without having to resort to the excessive requirements on computational resources imposed by direct numerical simulations. However, the setup and use of turbulent flow simulations requires a profound knowledge of fluid mechanics, numerical techniques, and the application under consideration. The susceptibility of large-eddy simulations to errors in modelling, in numerics, and in the treatment of boundary conditions, can be quite large due to nonlinear accumulation of different contributions over time, ...
SIMON: Remote collaboration system based on large scale simulation
International Nuclear Information System (INIS)
Sugawara, Akihiro; Kishimoto, Yasuaki
2003-01-01
Development of SIMON (SImulation MONitoring) system is described. SIMON aims to investigate many physical phenomena of tokamak type nuclear fusion plasma by simulation and to exchange information and to carry out joint researches with scientists in the world using internet. The characteristics of SIMON are followings; 1) decrease load of simulation by trigger sending method, 2) visualization of simulation results and hierarchical structure of analysis, 3) decrease of number of license by using command line when software is used, 4) improvement of support for using network of simulation data output by use of HTML (Hyper Text Markup Language), 5) avoidance of complex built-in work in client part and 6) small-sized and portable software. The visualization method of large scale simulation, remote collaboration system by HTML, trigger sending method, hierarchical analytical method, introduction into three-dimensional electromagnetic transportation code and technologies of SIMON system are explained. (S.Y.)
McPhaden, Michael; Leinen, Margaret; McEntee, Christine; Townsend, Randy; Williams, Billy
2016-04-01
The American Geophysical Union, a scientific society of 62,000 members worldwide, has established a set of scientific integrity and professional ethics guidelines for the actions of its members, for the governance of the union in its internal activities, and for the operations and participation in its publications and scientific meetings. This presentation will provide an overview of the Ethics program at AGU, highlighting the reasons for its establishment, the process of dealing ethical breaches, the number and types of cases considered, how AGU helps educate its members on Ethics issues, and the rapidly evolving efforts at AGU to address issues related to the emerging field of GeoEthics. The presentation will also cover the most recent AGU Ethics program focus on the role for AGU and other scientific societies in addressing sexual harassment, and AGU's work to provide additional program strength in this area.
Believability in simplifications of large scale physically based simulation
Han, Donghui; Hsu, Shu-wei; McNamara, Ann; Keyser, John
2013-01-01
We verify two hypotheses which are assumed to be true only intuitively in many rigid body simulations. I: In large scale rigid body simulation, viewers may not be able to perceive distortion incurred by an approximated simulation method. II: Fixing objects under a pile of objects does not affect the visual plausibility. Visual plausibility of scenarios simulated with these hypotheses assumed true are measured using subjective rating from viewers. As expected, analysis of results supports the truthfulness of the hypotheses under certain simulation environments. However, our analysis discovered four factors which may affect the authenticity of these hypotheses: number of collisions simulated simultaneously, homogeneity of colliding object pairs, distance from scene under simulation to camera position, and simulation method used. We also try to find an objective metric of visual plausibility from eye-tracking data collected from viewers. Analysis of these results indicates that eye-tracking does not present a suitable proxy for measuring plausibility or distinguishing between types of simulations. © 2013 ACM.
Large-scale computing techniques for complex system simulations
Dubitzky, Werner; Schott, Bernard
2012-01-01
Complex systems modeling and simulation approaches are being adopted in a growing number of sectors, including finance, economics, biology, astronomy, and many more. Technologies ranging from distributed computing to specialized hardware are explored and developed to address the computational requirements arising in complex systems simulations. The aim of this book is to present a representative overview of contemporary large-scale computing technologies in the context of complex systems simulations applications. The intention is to identify new research directions in this field and
Time simulation of flutter with large stiffness changes
Karpel, Mordechay; Wieseman, Carol D.
1992-01-01
Time simulation of flutter, involving large local structural changes, is formulated with a state-space model that is based on a relatively small number of generalized coordinates. Free-free vibration modes are first calculated for a nominal finite-element model with relatively large fictitious masses located at the area of structural changes. A low-frequency subset of these modes is then transformed into a set of structural modal coordinates with which the entire simulation is performed. These generalized coordinates and the associated oscillatory aerodynamic force coefficient matrices are used to construct an efficient time-domain, state-space model for a basic aeroelastic case. The time simulation can then be performed by simply changing the mass, stiffness, and damping coupling terms when structural changes occur. It is shown that the size of the aeroelastic model required for time simulation with large structural changes at a few apriori known locations is similar to that required for direct analysis of a single structural case. The method is applied to the simulation of an aeroelastic wind-tunnel model. The diverging oscillations are followed by the activation of a tip-ballast decoupling mechanism that stabilizes the system but may cause significant transient overshoots.
A Polar Rover for Large-Scale Scientific Surveys: Design, Implementation and Field Test Results
Directory of Open Access Journals (Sweden)
Yuqing He
2015-10-01
Full Text Available Exploration of polar regions is of great importance to scientific research. Unfortunately, due to the harsh environment, most of the regions on the Antarctic continent are still unreachable for humankind. Therefore, in 2011, the Chinese National Antarctic Research Expedition (CHINARE launched a project to design a rover to conduct large-scale scientific surveys on the Antarctic. The main challenges for the rover are twofold: one is the mobility, i.e., how to make a rover that could survive the harsh environment and safely move on the uneven, icy and snowy terrain; the other is the autonomy, in that the robot should be able to move at a relatively high speed with little or no human intervention so that it can explore a large region in a limit time interval under the communication constraints. In this paper, the corresponding techniques, especially the polar rover's design and autonomous navigation algorithms, are introduced in detail. Subsequently, an experimental report of the fields tests on the Antarctic is given to show some preliminary evaluation of the rover. Finally, experiences and existing challenging problems are summarized.
Wu, Xingfu; Taylor, Valerie
2013-01-01
In this paper, we present a performance modeling framework based on memory bandwidth contention time and a parameterized communication model to predict the performance of OpenMP, MPI and hybrid applications with weak scaling on three large-scale multicore supercomputers: IBM POWER4, POWER5+ and BlueGene/P, and analyze the performance of these MPI, OpenMP and hybrid applications. We use STREAM memory benchmarks and Intel's MPI benchmarks to provide initial performance analysis and model validation of MPI and OpenMP applications on these multicore supercomputers because the measured sustained memory bandwidth can provide insight into the memory bandwidth that a system should sustain on scientific applications with the same amount of workload per core. In addition to using these benchmarks, we also use a weak-scaling hybrid MPI/OpenMP large-scale scientific application: Gyrokinetic Toroidal Code (GTC) in magnetic fusion to validate our performance model of the hybrid application on these multicore supercomputers. The validation results for our performance modeling method show less than 7.77% error rate in predicting the performance of hybrid MPI/OpenMP GTC on up to 512 cores on these multicore supercomputers. © 2013 Elsevier Inc.
Wu, Xingfu
2013-12-01
In this paper, we present a performance modeling framework based on memory bandwidth contention time and a parameterized communication model to predict the performance of OpenMP, MPI and hybrid applications with weak scaling on three large-scale multicore supercomputers: IBM POWER4, POWER5+ and BlueGene/P, and analyze the performance of these MPI, OpenMP and hybrid applications. We use STREAM memory benchmarks and Intel\\'s MPI benchmarks to provide initial performance analysis and model validation of MPI and OpenMP applications on these multicore supercomputers because the measured sustained memory bandwidth can provide insight into the memory bandwidth that a system should sustain on scientific applications with the same amount of workload per core. In addition to using these benchmarks, we also use a weak-scaling hybrid MPI/OpenMP large-scale scientific application: Gyrokinetic Toroidal Code (GTC) in magnetic fusion to validate our performance model of the hybrid application on these multicore supercomputers. The validation results for our performance modeling method show less than 7.77% error rate in predicting the performance of hybrid MPI/OpenMP GTC on up to 512 cores on these multicore supercomputers. © 2013 Elsevier Inc.
Large Atmospheric Computation on the Earth Simulator: The LACES Project
Directory of Open Access Journals (Sweden)
Michel Desgagné
2006-01-01
Full Text Available The Large Atmospheric Computation on the Earth Simulator (LACES project is a joint initiative between Canadian and Japanese meteorological services and academic institutions that focuses on the high resolution simulation of Hurricane Earl (1998. The unique aspect of this effort is the extent of the computational domain, which covers all of North America and Europe with a grid spacing of 1 km. The Canadian Mesoscale Compressible Community (MC2 model is shown to parallelize effectively on the Japanese Earth Simulator (ES supercomputer; however, even using the extensive computing resources of the ES Center (ESC, the full simulation for the majority of Hurricane Earl's lifecycle takes over eight days to perform and produces over 5.2 TB of raw data. Preliminary diagnostics show that the results of the LACES simulation for the tropical stage of Hurricane Earl's lifecycle compare well with available observations for the storm. Further studies involving advanced diagnostics have commenced, taking advantage of the uniquely large spatial extent of the high resolution LACES simulation to investigate multiscale interactions in the hurricane and its environment. It is hoped that these studies will enhance our understanding of processes occurring within the hurricane and between the hurricane and its planetary-scale environment.
Paul Scherrer Institute Scientific and Technical Report 1999. Volume VI: Large Research Facilities
Energy Technology Data Exchange (ETDEWEB)
Foroughi, Fereydoun; Bercher, Renate; Buechli, Carmen; Meyer, Rosa [eds.
2000-07-01
The department GFA (Grossforschungsanlagen, Large Research Facilities) has been established in October 1998. Its main duty is operation, maintenance and development of the PSI accelerators, the spallation neutron source and the beam transport systems for pions and muons. A large effort of this group concerns the planning and co-ordination of new projects like e.g. the assembly of the synchrotron light source (SLS), design studies of a new proton therapy facility, the ultracold neutron source and a new intensive secondary beam line for low energy muons. A large fraction of this report is devoted to research especially in the field of materials Science. The studies include large scale molecular dynamics computer simulations on the elastic and plastic behavior of nanostructured metals, complemented by experimental mechanical testing using micro-indentation and miniaturized tensile testing, as well as microstructural characterisation and strain field mapping of metallic coatings and thin ceramic layers, the latter done with synchrotron radiation.
Paul Scherrer Institute Scientific and Technical Report 1999. Volume VI: Large Research Facilities
International Nuclear Information System (INIS)
Foroughi, Fereydoun; Bercher, Renate; Buechli, Carmen; Meyer, Rosa
2000-01-01
The department GFA (Grossforschungsanlagen, Large Research Facilities) has been established in October 1998. Its main duty is operation, maintenance and development of the PSI accelerators, the spallation neutron source and the beam transport systems for pions and muons. A large effort of this group concerns the planning and co-ordination of new projects like e.g. the assembly of the synchrotron light source (SLS), design studies of a new proton therapy facility, the ultracold neutron source and a new intensive secondary beam line for low energy muons. A large fraction of this report is devoted to research especially in the field of materials Science. The studies include large scale molecular dynamics computer simulations on the elastic and plastic behavior of nanostructured metals, complemented by experimental mechanical testing using micro-indentation and miniaturized tensile testing, as well as microstructural characterisation and strain field mapping of metallic coatings and thin ceramic layers, the latter done with synchrotron radiation
Simulating experiments using a Comsol application for teaching scientific research methods
Schijndel, van A.W.M.
2015-01-01
For universities it is important to teach the principles of scientific methods as soon as possible. However, in case of performing experiments, students need to have some knowledge and skills before start doing measurements. In this case, Comsol can be helpfully by simulating the experiments before
Large Scale Simulations of the Euler Equations on GPU Clusters
Liebmann, Manfred; Douglas, Craig C.; Haase, Gundolf; Horvá th, Zoltá n
2010-01-01
The paper investigates the scalability of a parallel Euler solver, using the Vijayasundaram method, on a GPU cluster with 32 Nvidia Geforce GTX 295 boards. The aim of this research is to enable large scale fluid dynamics simulations with up to one
Large Eddy Simulation of Sydney Swirl Non-Reaction Jets
DEFF Research Database (Denmark)
Yang, Yang; Kær, Søren Knudsen; Yin, Chungen
The Sydney swirl burner non-reaction case was studied using large eddy simulation. The two-point correlation method was introduced and used to estimate grid resolution. Energy spectra and instantaneous pressure and velocity plots were used to identify features in flow field. By using these method......, vortex breakdown and precessing vortex core are identified and different flow zones are shown....
Large interface simulation in an averaged two-fluid code
International Nuclear Information System (INIS)
Henriques, A.
2006-01-01
Different ranges of size of interfaces and eddies are involved in multiphase flow phenomena. Classical formalisms focus on a specific range of size. This study presents a Large Interface Simulation (LIS) two-fluid compressible formalism taking into account different sizes of interfaces. As in the single-phase Large Eddy Simulation, a filtering process is used to point out Large Interface (LI) simulation and Small interface (SI) modelization. The LI surface tension force is modelled adapting the well-known CSF method. The modelling of SI transfer terms is done calling for classical closure laws of the averaged approach. To simulate accurately LI transfer terms, we develop a LI recognition algorithm based on a dimensionless criterion. The LIS model is applied in a classical averaged two-fluid code. The LI transfer terms modelling and the LI recognition are validated on analytical and experimental tests. A square base basin excited by a horizontal periodic movement is studied with the LIS model. The capability of the model is also shown on the case of the break-up of a bubble in a turbulent liquid flow. The break-up of a large bubble at a grid impact performed regime transition between two different scales of interface from LI to SI and from PI to LI. (author) [fr
Large-eddy simulation of highly underexpanded transient gas jets
Vuorinen, V.; Yu, J.; Tirunagari, S.; Kaario, O.; Larmi, M.; Duwig, C.; Boersma, B.J.
2013-01-01
Large-eddy simulations (LES) based on scale-selective implicit filtering are carried out in order to study the effect of nozzle pressure ratios on the characteristics of highly underexpanded jets. Pressure ratios ranging from 4.5 to 8.5 with Reynolds numbers of the order 75?000–140?000 are
Large signal simulation of photonic crystal Fano laser
DEFF Research Database (Denmark)
Zali, Aref Rasoulzadeh; Yu, Yi; Moravvej-Farshi, Mohammad Kazem
2017-01-01
be modulated at frequencies exceeding 1 THz which is much higher than its corresponding relaxation oscillation frequency. Large signal simulation of the Fano laser is also investigated based on pseudorandom bit sequence at 0.5 Tbit/s. It shows eye patterns are open at such high modulation frequency, verifying...
Large eddy simulations of an airfoil in turbulent inflow
DEFF Research Database (Denmark)
Gilling, Lasse; Sørensen, Niels N.
2008-01-01
Wind turbines operate in the turbulent boundary layer of the atmosphere and due to the rotational sampling effect the blades experience a high level of turbulence [1]. In this project the effect of turbulence is investigated by large eddy simulations of the turbulent flow past a NACA 0015 airfoil...
Large scale particle simulations in a virtual memory computer
International Nuclear Information System (INIS)
Gray, P.C.; Million, R.; Wagner, J.S.; Tajima, T.
1983-01-01
Virtual memory computers are capable of executing large-scale particle simulations even when the memory requirements exceeds the computer core size. The required address space is automatically mapped onto slow disc memory the the operating system. When the simulation size is very large, frequent random accesses to slow memory occur during the charge accumulation and particle pushing processes. Assesses to slow memory significantly reduce the excecution rate of the simulation. We demonstrate in this paper that with the proper choice of sorting algorithm, a nominal amount of sorting to keep physically adjacent particles near particles with neighboring array indices can reduce random access to slow memory, increase the efficiency of the I/O system, and hence, reduce the required computing time. (orig.)
Real-time simulation of large-scale floods
Liu, Q.; Qin, Y.; Li, G. D.; Liu, Z.; Cheng, D. J.; Zhao, Y. H.
2016-08-01
According to the complex real-time water situation, the real-time simulation of large-scale floods is very important for flood prevention practice. Model robustness and running efficiency are two critical factors in successful real-time flood simulation. This paper proposed a robust, two-dimensional, shallow water model based on the unstructured Godunov- type finite volume method. A robust wet/dry front method is used to enhance the numerical stability. An adaptive method is proposed to improve the running efficiency. The proposed model is used for large-scale flood simulation on real topography. Results compared to those of MIKE21 show the strong performance of the proposed model.
How to simulate global cosmic strings with large string tension
Energy Technology Data Exchange (ETDEWEB)
Klaer, Vincent B.; Moore, Guy D., E-mail: vklaer@theorie.ikp.physik.tu-darmstadt.de, E-mail: guy.moore@physik.tu-darmstadt.de [Institut für Kernphysik, Technische Universität Darmstadt, Schlossgartenstraße 2, Darmstadt, D-64289 Germany (Germany)
2017-10-01
Global string networks may be relevant in axion production in the early Universe, as well as other cosmological scenarios. Such networks contain a large hierarchy of scales between the string core scale and the Hubble scale, ln( f {sub a} / H ) ∼ 70, which influences the network dynamics by giving the strings large tensions T ≅ π f {sub a} {sup 2} ln( f {sub a} / H ). We present a new numerical approach to simulate such global string networks, capturing the tension without an exponentially large lattice.
Large Eddy Simulation of Turbulent Flows in Wind Energy
DEFF Research Database (Denmark)
Chivaee, Hamid Sarlak
This research is devoted to the Large Eddy Simulation (LES), and to lesser extent, wind tunnel measurements of turbulent flows in wind energy. It starts with an introduction to the LES technique associated with the solution of the incompressible Navier-Stokes equations, discretized using a finite......, should the mesh resolution, numerical discretization scheme, time averaging period, and domain size be chosen wisely. A thorough investigation of the wind turbine wake interactions is also conducted and the simulations are validated against available experimental data from external sources. The effect...... Reynolds numbers, and thereafter, the fully-developed infinite wind farm boundary later simulations are performed. Sources of inaccuracy in the simulations are investigated and it is found that high Reynolds number flows are more sensitive to the choice of the SGS model than their low Reynolds number...
Large-eddy simulation of atmospheric flow over complex terrain
DEFF Research Database (Denmark)
Bechmann, Andreas
2007-01-01
The present report describes the development and validation of a turbulence model designed for atmospheric flows based on the concept of Large-Eddy Simulation (LES). The background for the work is the high Reynolds number k - #epsilon# model, which has been implemented on a finite-volume code...... turbulence model is able to handle both engineering and atmospheric flows and can be run in both RANS or LES mode. For LES simulations a time-dependent wind field that accurately represents the turbulent structures of a wind environment must be prescribed at the computational inlet. A method is implemented...... where the turbulent wind field from a separate LES simulation can be used as inflow. To avoid numerical dissipation of turbulence special care is paid to the numerical method, e.g. the turbulence model is calibrated with the specific numerical scheme used. This is done by simulating decaying isotropic...
DoSSiER: Database of Scientific Simulation and Experimental Results
Wenzel, Hans; Genser, Krzysztof; Elvira, Daniel; Pokorski, Witold; Carminati, Federico; Konstantinov, Dmitri; Ribon, Alberto; Folger, Gunter; Dotti, Andrea
2017-01-01
The Geant4, GeantV and GENIE collaborations regularly perform validation and regression tests for simulation results. DoSSiER (Database of Scientific Simulation and Experimental Results) is being developed as a central repository to store the simulation results as well as the experimental data used for validation. DoSSiER can be easily accessed via a web application. In addition, a web service allows for programmatic access to the repository to extract records in json or xml exchange formats. In this article, we describe the functionality and the current status of various components of DoSSiER as well as the technology choices we made.
Large-Eddy Simulations of Flows in Complex Terrain
Kosovic, B.; Lundquist, K. A.
2011-12-01
Large-eddy simulation as a methodology for numerical simulation of turbulent flows was first developed to study turbulent flows in atmospheric by Lilly (1967). The first LES were carried by Deardorff (1970) who used these simulations to study atmospheric boundary layers. Ever since, LES has been extensively used to study canonical atmospheric boundary layers, in most cases flat plate boundary layers under the assumption of horizontal homogeneity. Carefully designed LES of canonical convective and neutrally stratified and more recently stably stratified atmospheric boundary layers have contributed significantly to development of better understanding of these flows and their parameterizations in large scale models. These simulations were often carried out using codes specifically designed and developed for large-eddy simulations of horizontally homogeneous flows with periodic lateral boundary conditions. Recent developments in multi-scale numerical simulations of atmospheric flows enable numerical weather prediction (NWP) codes such as ARPS (Chow and Street, 2009), COAMPS (Golaz et al., 2009) and Weather Research and Forecasting model, to be used nearly seamlessly across a wide range of atmospheric scales from synoptic down to turbulent scales in atmospheric boundary layers. Before we can with confidence carry out multi-scale simulations of atmospheric flows, NWP codes must be validated for accurate performance in simulating flows over complex or inhomogeneous terrain. We therefore carry out validation of WRF-LES for simulations of flows over complex terrain using data from Askervein Hill (Taylor and Teunissen, 1985, 1987) and METCRAX (Whiteman et al., 2008) field experiments. WRF's nesting capability is employed with a one-way nested inner domain that includes complex terrain representation while the coarser outer nest is used to spin up fully developed atmospheric boundary layer turbulence and thus represent accurately inflow to the inner domain. LES of a
International Nuclear Information System (INIS)
Sekimura, Naoto; Okita, Taira
2006-01-01
Scientific computational methods have advanced remarkably with the progress of nuclear development. They have played the role of weft connecting each realm of nuclear engineering and then an introductory course of advanced scientific computational methods and their applications to nuclear technologies were prepared in serial form. This is the fourth issue showing the overview of scientific computational methods with the introduction of continuum simulation methods and their applications. Simulation methods on physical radiation effects on materials are reviewed based on the process such as binary collision approximation, molecular dynamics, kinematic Monte Carlo method, reaction rate method and dislocation dynamics. (T. Tanaka)
Shishkova, Evgenia; Kwiecien, Nicholas W.; Hebert, Alexander S.; Westphall, Michael S.; Prenni, Jessica E.; Coon, Joshua J.
2017-12-01
Speaking engagements, serving as session chairs, and receiving awards at national meetings are essential stepping stones towards professional success for scientific researchers. Studies of gender parity in meetings of national scientific societies repeatedly uncover bias in speaker selection, engendering underrepresentation of women among featured presenters. To continue this dialogue, we analyzed membership data and annual conference programs of a large scientific society (>7000 members annually) in a male-rich ( 70% males), technology-oriented STEM subfield. We detected a pronounced skew towards males among invited keynote lecturers, plenary speakers, and recipients of the society's Senior Investigator award (15%, 13%, and 8% females, respectively). However, the proportion of females among Mid-Career and Young Investigator award recipients and oral session chairs resembled the current gender distribution of the general membership. Female members were more likely to present at the conferences and equally likely to apply and be accepted for oral presentations as their male counterparts. The gender of a session chair had no effect on the gender distribution of selected applicants. Interestingly, we identified several research subareas that were naturally enriched (i.e., not influenced by unequal selection of presenters) for either female or male participants, illustrating within a single subfield the gender divide along biology-technology line typical of all STEM disciplines. Two female-enriched topics experienced a rapid growth in popularity within the examined period, more than doubling the number of associated researchers. Collectively, these findings contribute to the contemporary discourse on gender in science and hopefully will propel positive changes within this and other societies. [Figure not available: see fulltext.
Large-eddy simulation of sand dune morphodynamics
Khosronejad, Ali; Sotiropoulos, Fotis; St. Anthony Falls Laboratory, University of Minnesota Team
2015-11-01
Sand dunes are natural features that form under complex interaction between turbulent flow and bed morphodynamics. We employ a fully-coupled 3D numerical model (Khosronejad and Sotiropoulos, 2014, Journal of Fluid Mechanics, 753:150-216) to perform high-resolution large-eddy simulations of turbulence and bed morphodynamics in a laboratory scale mobile-bed channel to investigate initiation, evolution and quasi-equilibrium of sand dunes (Venditti and Church, 2005, J. Geophysical Research, 110:F01009). We employ a curvilinear immersed boundary method along with convection-diffusion and bed-morphodynamics modules to simulate the suspended sediment and the bed-load transports respectively. The coupled simulation were carried out on a grid with more than 100 million grid nodes and simulated about 3 hours of physical time of dune evolution. The simulations provide the first complete description of sand dune formation and long-term evolution. The geometric characteristics of the simulated dunes are shown to be in excellent agreement with observed data obtained across a broad range of scales. This work was supported by NSF Grants EAR-0120914 (as part of the National Center for Earth-Surface Dynamics). Computational resources were provided by the University of Minnesota Supercomputing Institute.
Large-scale computation at PSI scientific achievements and future requirements
International Nuclear Information System (INIS)
Adelmann, A.; Markushin, V.
2008-11-01
Computational modelling and simulation are among the disciplines that have seen the most dramatic growth in capabilities in the 2Oth Century. Within the past two decades, scientific computing has become an important contributor to all scientific research programs. Computational modelling and simulation are particularly indispensable for solving research problems that are unsolvable by traditional theoretical and experimental approaches, hazardous to study, or time consuming or expensive to solve by traditional means. Many such research areas are found in PSI's research portfolio. Advances in computing technologies (including hardware and software) during the past decade have set the stage for a major step forward in modelling and simulation. We have now arrived at a situation where we have a number of otherwise unsolvable problems, where simulations are as complex as the systems under study. In 2008 the High-Performance Computing (HPC) community entered the petascale area with the heterogeneous Opteron/Cell machine, called Road Runner built by IBM for the Los Alamos National Laboratory. We are on the brink of a time where the availability of many hundreds of thousands of cores will open up new challenging possibilities in physics, algorithms (numerical mathematics) and computer science. However, to deliver on this promise, it is not enough to provide 'peak' performance in terms of peta-flops, the maximum theoretical speed a computer can attain. Most important, this must be translated into corresponding increase in the capabilities of scientific codes. This is a daunting problem that can only be solved by increasing investment in hardware, in the accompanying system software that enables the reliable use of high-end computers, in scientific competence i.e. the mathematical (parallel) algorithms that are the basis of the codes, and education. In the case of Switzerland, the white paper 'Swiss National Strategic Plan for High Performance Computing and Networking
Large-scale computation at PSI scientific achievements and future requirements
Energy Technology Data Exchange (ETDEWEB)
Adelmann, A.; Markushin, V
2008-11-15
Computational modelling and simulation are among the disciplines that have seen the most dramatic growth in capabilities in the 2Oth Century. Within the past two decades, scientific computing has become an important contributor to all scientific research programs. Computational modelling and simulation are particularly indispensable for solving research problems that are unsolvable by traditional theoretical and experimental approaches, hazardous to study, or time consuming or expensive to solve by traditional means. Many such research areas are found in PSI's research portfolio. Advances in computing technologies (including hardware and software) during the past decade have set the stage for a major step forward in modelling and simulation. We have now arrived at a situation where we have a number of otherwise unsolvable problems, where simulations are as complex as the systems under study. In 2008 the High-Performance Computing (HPC) community entered the petascale area with the heterogeneous Opteron/Cell machine, called Road Runner built by IBM for the Los Alamos National Laboratory. We are on the brink of a time where the availability of many hundreds of thousands of cores will open up new challenging possibilities in physics, algorithms (numerical mathematics) and computer science. However, to deliver on this promise, it is not enough to provide 'peak' performance in terms of peta-flops, the maximum theoretical speed a computer can attain. Most important, this must be translated into corresponding increase in the capabilities of scientific codes. This is a daunting problem that can only be solved by increasing investment in hardware, in the accompanying system software that enables the reliable use of high-end computers, in scientific competence i.e. the mathematical (parallel) algorithms that are the basis of the codes, and education. In the case of Switzerland, the white paper 'Swiss National Strategic Plan for High Performance Computing
Large eddy simulation of a wing-body junction flow
Ryu, Sungmin; Emory, Michael; Campos, Alejandro; Duraisamy, Karthik; Iaccarino, Gianluca
2014-11-01
We present numerical simulations of the wing-body junction flow experimentally investigated by Devenport & Simpson (1990). Wall-junction flows are common in engineering applications but relevant flow physics close to the corner region is not well understood. Moreover, performance of turbulence models for the body-junction case is not well characterized. Motivated by the insufficient investigations, we have numerically investigated the case with Reynolds-averaged Naiver-Stokes equation (RANS) and Large Eddy Simulation (LES) approaches. The Vreman model applied for the LES and SST k- ω model for the RANS simulation are validated focusing on the ability to predict turbulence statistics near the junction region. Moreover, a sensitivity study of the form of the Vreman model will also be presented. This work is funded under NASA Cooperative Agreement NNX11AI41A (Technical Monitor Dr. Stephen Woodruff)
Lightweight computational steering of very large scale molecular dynamics simulations
International Nuclear Information System (INIS)
Beazley, D.M.
1996-01-01
We present a computational steering approach for controlling, analyzing, and visualizing very large scale molecular dynamics simulations involving tens to hundreds of millions of atoms. Our approach relies on extensible scripting languages and an easy to use tool for building extensions and modules. The system is extremely easy to modify, works with existing C code, is memory efficient, and can be used from inexpensive workstations and networks. We demonstrate how we have used this system to manipulate data from production MD simulations involving as many as 104 million atoms running on the CM-5 and Cray T3D. We also show how this approach can be used to build systems that integrate common scripting languages (including Tcl/Tk, Perl, and Python), simulation code, user extensions, and commercial data analysis packages
Large Eddy Simulations of Severe Convection Induced Turbulence
Ahmad, Nash'at; Proctor, Fred
2011-01-01
Convective storms can pose a serious risk to aviation operations since they are often accompanied by turbulence, heavy rain, hail, icing, lightning, strong winds, and poor visibility. They can cause major delays in air traffic due to the re-routing of flights, and by disrupting operations at the airports in the vicinity of the storm system. In this study, the Terminal Area Simulation System is used to simulate five different convective events ranging from a mesoscale convective complex to isolated storms. The occurrence of convection induced turbulence is analyzed from these simulations. The validation of model results with the radar data and other observations is reported and an aircraft-centric turbulence hazard metric calculated for each case is discussed. The turbulence analysis showed that large pockets of significant turbulence hazard can be found in regions of low radar reflectivity. Moderate and severe turbulence was often found in building cumulus turrets and overshooting tops.
Large Eddy Simulation of Cryogenic Injection Processes at Supercritical Pressure
Oefelein, Joseph C.
2002-01-01
This paper highlights results from the first of a series of hierarchical simulations aimed at assessing the modeling requirements for application of the large eddy simulation technique to cryogenic injection and combustion processes in liquid rocket engines. The focus is on liquid-oxygen-hydrogen coaxial injectors at a condition where the liquid-oxygen is injected at a subcritical temperature into a supercritical environment. For this situation a diffusion dominated mode of combustion occurs in the presence of exceedingly large thermophysical property gradients. Though continuous, these gradients approach the behavior of a contact discontinuity. Significant real gas effects and transport anomalies coexist locally in colder regions of the flow, with ideal gas and transport characteristics occurring within the flame zone. The current focal point is on the interfacial region between the liquid-oxygen core and the coaxial hydrogen jet where the flame anchors itself.
Experimental simulation of microinteractions in large scale explosions
Energy Technology Data Exchange (ETDEWEB)
Chen, X.; Luo, R.; Yuen, W.W.; Theofanous, T.G. [California Univ., Santa Barbara, CA (United States). Center for Risk Studies and Safety
1998-01-01
This paper presents data and analysis of recent experiments conducted in the SIGMA-2000 facility to simulate microinteractions in large scale explosions. Specifically, the fragmentation behavior of a high temperature molten steel drop under high pressure (beyond critical) conditions are investigated. The current data demonstrate, for the first time, the effect of high pressure in suppressing the thermal effect of fragmentation under supercritical conditions. The results support the microinteractions idea, and the ESPROSE.m prediction of fragmentation rate. (author)
Simulation requirements for the Large Deployable Reflector (LDR)
Soosaar, K.
1984-01-01
Simulation tools for the large deployable reflector (LDR) are discussed. These tools are often the transfer function variety equations. However, transfer functions are inadequate to represent time-varying systems for multiple control systems with overlapping bandwidths characterized by multi-input, multi-output features. Frequency domain approaches are the useful design tools, but a full-up simulation is needed. Because of the need for a dedicated computer for high frequency multi degree of freedom components encountered, non-real time smulation is preferred. Large numerical analysis software programs are useful only to receive inputs and provide output to the next block, and should be kept out of the direct loop of simulation. The following blocks make up the simulation. The thermal model block is a classical heat transfer program. It is a non-steady state program. The quasistatic block deals with problems associated with rigid body control of reflector segments. The steady state block assembles data into equations of motion and dynamics. A differential raytrace is obtained to establish a change in wave aberrations. The observation scene is described. The focal plane module converts the photon intensity impinging on it into electron streams or into permanent film records.
Exploring the large-scale structure of Taylor–Couette turbulence through Large-Eddy Simulations
Ostilla-Mónico, Rodolfo; Zhu, Xiaojue; Verzicco, Roberto
2018-04-01
Large eddy simulations (LES) of Taylor-Couette (TC) flow, the flow between two co-axial and independently rotating cylinders are performed in an attempt to explore the large-scale axially-pinned structures seen in experiments and simulations. Both static and dynamic LES models are used. The Reynolds number is kept fixed at Re = 3.4 · 104, and the radius ratio η = ri /ro is set to η = 0.909, limiting the effects of curvature and resulting in frictional Reynolds numbers of around Re τ ≈ 500. Four rotation ratios from Rot = ‑0.0909 to Rot = 0.3 are simulated. First, the LES of TC is benchmarked for different rotation ratios. Both the Smagorinsky model with a constant of cs = 0.1 and the dynamic model are found to produce reasonable results for no mean rotation and cyclonic rotation, but deviations increase for increasing rotation. This is attributed to the increasing anisotropic character of the fluctuations. Second, “over-damped” LES, i.e. LES with a large Smagorinsky constant is performed and is shown to reproduce some features of the large-scale structures, even when the near-wall region is not adequately modeled. This shows the potential for using over-damped LES for fast explorations of the parameter space where large-scale structures are found.
Subgrid-scale models for large-eddy simulation of rotating turbulent channel flows
Silvis, Maurits H.; Bae, Hyunji Jane; Trias, F. Xavier; Abkar, Mahdi; Moin, Parviz; Verstappen, Roel
2017-11-01
We aim to design subgrid-scale models for large-eddy simulation of rotating turbulent flows. Rotating turbulent flows form a challenging test case for large-eddy simulation due to the presence of the Coriolis force. The Coriolis force conserves the total kinetic energy while transporting it from small to large scales of motion, leading to the formation of large-scale anisotropic flow structures. The Coriolis force may also cause partial flow laminarization and the occurrence of turbulent bursts. Many subgrid-scale models for large-eddy simulation are, however, primarily designed to parametrize the dissipative nature of turbulent flows, ignoring the specific characteristics of transport processes. We, therefore, propose a new subgrid-scale model that, in addition to the usual dissipative eddy viscosity term, contains a nondissipative nonlinear model term designed to capture transport processes, such as those due to rotation. We show that the addition of this nonlinear model term leads to improved predictions of the energy spectra of rotating homogeneous isotropic turbulence as well as of the Reynolds stress anisotropy in spanwise-rotating plane-channel flows. This work is financed by the Netherlands Organisation for Scientific Research (NWO) under Project Number 613.001.212.
Schüller, Anton; Schweitzer, Marc
2017-01-01
The contributions gathered here provide an overview of current research projects and selected software products of the Fraunhofer Institute for Algorithms and Scientific Computing SCAI. They show the wide range of challenges that scientific computing currently faces, the solutions it offers, and its important role in developing applications for industry. Given the exciting field of applied collaborative research and development it discusses, the book will appeal to scientists, practitioners, and students alike. The Fraunhofer Institute for Algorithms and Scientific Computing SCAI combines excellent research and application-oriented development to provide added value for our partners. SCAI develops numerical techniques, parallel algorithms and specialized software tools to support and optimize industrial simulations. Moreover, it implements custom software solutions for production and logistics, and offers calculations on high-performance computers. Its services and products are based on state-of-the-art metho...
Towards Large Eddy Simulation of gas turbine compressors
McMullan, W. A.; Page, G. J.
2012-07-01
With increasing computing power, Large Eddy Simulation could be a useful simulation tool for gas turbine axial compressor design. This paper outlines a series of simulations performed on compressor geometries, ranging from a Controlled Diffusion Cascade stator blade to the periodic sector of a stage in a 3.5 stage axial compressor. The simulation results show that LES may offer advantages over traditional RANS methods when off-design conditions are considered - flow regimes where RANS models often fail to converge. The time-dependent nature of LES permits the resolution of transient flow structures, and can elucidate new mechanisms of vorticity generation on blade surfaces. It is shown that accurate LES is heavily reliant on both the near-wall mesh fidelity and the ability of the imposed inflow condition to recreate the conditions found in the reference experiment. For components embedded in a compressor this requires the generation of turbulence fluctuations at the inlet plane. A recycling method is developed that improves the quality of the flow in a single stage calculation of an axial compressor, and indicates that future developments in both the recycling technique and computing power will bring simulations of axial compressors within reach of industry in the coming years.
Do large-scale assessments measure students' ability to integrate scientific knowledge?
Lee, Hee-Sun
2010-03-01
Large-scale assessments are used as means to diagnose the current status of student achievement in science and compare students across schools, states, and countries. For efficiency, multiple-choice items and dichotomously-scored open-ended items are pervasively used in large-scale assessments such as Trends in International Math and Science Study (TIMSS). This study investigated how well these items measure secondary school students' ability to integrate scientific knowledge. This study collected responses of 8400 students to 116 multiple-choice and 84 open-ended items and applied an Item Response Theory analysis based on the Rasch Partial Credit Model. Results indicate that most multiple-choice items and dichotomously-scored open-ended items can be used to determine whether students have normative ideas about science topics, but cannot measure whether students integrate multiple pieces of relevant science ideas. Only when the scoring rubric is redesigned to capture subtle nuances of student open-ended responses, open-ended items become a valid and reliable tool to assess students' knowledge integration ability.
Improving the trust in results of numerical simulations and scientific data analytics
Energy Technology Data Exchange (ETDEWEB)
Cappello, Franck [Argonne National Lab. (ANL), Argonne, IL (United States); Constantinescu, Emil [Argonne National Lab. (ANL), Argonne, IL (United States); Hovland, Paul [Argonne National Lab. (ANL), Argonne, IL (United States); Peterka, Tom [Argonne National Lab. (ANL), Argonne, IL (United States); Phillips, Carolyn [Argonne National Lab. (ANL), Argonne, IL (United States); Snir, Marc [Argonne National Lab. (ANL), Argonne, IL (United States); Wild, Stefan [Argonne National Lab. (ANL), Argonne, IL (United States)
2015-04-30
This white paper investigates several key aspects of the trust that a user can give to the results of numerical simulations and scientific data analytics. In this document, the notion of trust is related to the integrity of numerical simulations and data analytics applications. This white paper complements the DOE ASCR report on Cybersecurity for Scientific Computing Integrity by (1) exploring the sources of trust loss; (2) reviewing the definitions of trust in several areas; (3) providing numerous cases of result alteration, some of them leading to catastrophic failures; (4) examining the current notion of trust in numerical simulation and scientific data analytics; (5) providing a gap analysis; and (6) suggesting two important research directions and their respective research topics. To simplify the presentation without loss of generality, we consider that trust in results can be lost (or the results’ integrity impaired) because of any form of corruption happening during the execution of the numerical simulation or the data analytics application. In general, the sources of such corruption are threefold: errors, bugs, and attacks. Current applications are already using techniques to deal with different types of corruption. However, not all potential corruptions are covered by these techniques. We firmly believe that the current level of trust that a user has in the results is at least partially founded on ignorance of this issue or the hope that no undetected corruptions will occur during the execution. This white paper explores the notion of trust and suggests recommendations for developing a more scientifically grounded notion of trust in numerical simulation and scientific data analytics. We first formulate the problem and show that it goes beyond previous questions regarding the quality of results such as V&V, uncertainly quantification, and data assimilation. We then explore the complexity of this difficult problem, and we sketch complementary general
Parallel continuous simulated tempering and its applications in large-scale molecular simulations
Energy Technology Data Exchange (ETDEWEB)
Zang, Tianwu; Yu, Linglin; Zhang, Chong [Applied Physics Program and Department of Bioengineering, Rice University, Houston, Texas 77005 (United States); Ma, Jianpeng, E-mail: jpma@bcm.tmc.edu [Applied Physics Program and Department of Bioengineering, Rice University, Houston, Texas 77005 (United States); Verna and Marrs McLean Department of Biochemistry and Molecular Biology, Baylor College of Medicine, One Baylor Plaza, BCM-125, Houston, Texas 77030 (United States)
2014-07-28
In this paper, we introduce a parallel continuous simulated tempering (PCST) method for enhanced sampling in studying large complex systems. It mainly inherits the continuous simulated tempering (CST) method in our previous studies [C. Zhang and J. Ma, J. Chem. Phys. 130, 194112 (2009); C. Zhang and J. Ma, J. Chem. Phys. 132, 244101 (2010)], while adopts the spirit of parallel tempering (PT), or replica exchange method, by employing multiple copies with different temperature distributions. Differing from conventional PT methods, despite the large stride of total temperature range, the PCST method requires very few copies of simulations, typically 2–3 copies, yet it is still capable of maintaining a high rate of exchange between neighboring copies. Furthermore, in PCST method, the size of the system does not dramatically affect the number of copy needed because the exchange rate is independent of total potential energy, thus providing an enormous advantage over conventional PT methods in studying very large systems. The sampling efficiency of PCST was tested in two-dimensional Ising model, Lennard-Jones liquid and all-atom folding simulation of a small globular protein trp-cage in explicit solvent. The results demonstrate that the PCST method significantly improves sampling efficiency compared with other methods and it is particularly effective in simulating systems with long relaxation time or correlation time. We expect the PCST method to be a good alternative to parallel tempering methods in simulating large systems such as phase transition and dynamics of macromolecules in explicit solvent.
Large-eddy simulation of atmospheric flow over complex terrain
Energy Technology Data Exchange (ETDEWEB)
Bechmann, A.
2006-11-15
The present report describes the development and validation of a turbulence model designed for atmospheric flows based on the concept of Large-Eddy Simulation (LES). The background for the work is the high Reynolds number k - epsilon model, which has been implemented on a finite-volume code of the incompressible Reynolds-averaged Navier-Stokes equations (RANS). The k - epsilon model is traditionally used for RANS computations, but is here developed to also enable LES. LES is able to provide detailed descriptions of a wide range of engineering flows at low Reynolds numbers. For atmospheric flows, however, the high Reynolds numbers and the rough surface of the earth provide difficulties normally not compatible with LES. Since these issues are most severe near the surface they are addressed by handling the near surface region with RANS and only use LES above this region. Using this method, the developed turbulence model is able to handle both engineering and atmospheric flows and can be run in both RANS or LES mode. For LES simulations a time-dependent wind field that accurately represents the turbulent structures of a wind environment must be prescribed at the computational inlet. A method is implemented where the turbulent wind field from a separate LES simulation can be used as inflow. To avoid numerical dissipation of turbulence special care is paid to the numerical method, e.g. the turbulence model is calibrated with the specific numerical scheme used. This is done by simulating decaying isotropic and homogeneous turbulence. Three atmospheric test cases are investigated in order to validate the behavior of the presented turbulence model. Simulation of the neutral atmospheric boundary layer, illustrates the turbulence model ability to generate and maintain the turbulent structures responsible for boundary layer transport processes. Velocity and turbulence profiles are in good agreement with measurements. Simulation of the flow over the Askervein hill is also
Tool Support for Parametric Analysis of Large Software Simulation Systems
Schumann, Johann; Gundy-Burlet, Karen; Pasareanu, Corina; Menzies, Tim; Barrett, Tony
2008-01-01
The analysis of large and complex parameterized software systems, e.g., systems simulation in aerospace, is very complicated and time-consuming due to the large parameter space, and the complex, highly coupled nonlinear nature of the different system components. Thus, such systems are generally validated only in regions local to anticipated operating points rather than through characterization of the entire feasible operational envelope of the system. We have addressed the factors deterring such an analysis with a tool to support envelope assessment: we utilize a combination of advanced Monte Carlo generation with n-factor combinatorial parameter variations to limit the number of cases, but still explore important interactions in the parameter space in a systematic fashion. Additional test-cases, automatically generated from models (e.g., UML, Simulink, Stateflow) improve the coverage. The distributed test runs of the software system produce vast amounts of data, making manual analysis impossible. Our tool automatically analyzes the generated data through a combination of unsupervised Bayesian clustering techniques (AutoBayes) and supervised learning of critical parameter ranges using the treatment learner TAR3. The tool has been developed around the Trick simulation environment, which is widely used within NASA. We will present this tool with a GN&C (Guidance, Navigation and Control) simulation of a small satellite system.
Large Eddy Simulation of High-Speed, Premixed Ethylene Combustion
Ramesh, Kiran; Edwards, Jack R.; Chelliah, Harsha; Goyne, Christopher; McDaniel, James; Rockwell, Robert; Kirik, Justin; Cutler, Andrew; Danehy, Paul
2015-01-01
A large-eddy simulation / Reynolds-averaged Navier-Stokes (LES/RANS) methodology is used to simulate premixed ethylene-air combustion in a model scramjet designed for dual mode operation and equipped with a cavity for flameholding. A 22-species reduced mechanism for ethylene-air combustion is employed, and the calculations are performed on a mesh containing 93 million cells. Fuel plumes injected at the isolator entrance are processed by the isolator shock train, yielding a premixed fuel-air mixture at an equivalence ratio of 0.42 at the cavity entrance plane. A premixed flame is anchored within the cavity and propagates toward the opposite wall. Near complete combustion of ethylene is obtained. The combustor is highly dynamic, exhibiting a large-scale oscillation in global heat release and mass flow rate with a period of about 2.8 ms. Maximum heat release occurs when the flame front reaches its most downstream extent, as the flame surface area is larger. Minimum heat release is associated with flame propagation toward the cavity and occurs through a reduction in core flow velocity that is correlated with an upstream movement of the shock train. Reasonable agreement between simulation results and available wall pressure, particle image velocimetry, and OH-PLIF data is obtained, but it is not yet clear whether the system-level oscillations seen in the calculations are actually present in the experiment.
Simulations of Large-Area Electron Beam Diodes
Swanekamp, S. B.; Friedman, M.; Ludeking, L.; Smithe, D.; Obenschain, S. P.
1999-11-01
Large area electron beam diodes are typically used to pump the amplifiers of KrF lasers. Simulations of large-area electron beam diodes using the particle-in-cell code MAGIC3D have shown the electron flow in the diode to be unstable. Since this instability can potentially produce a non-uniform current and energy distribution in the hibachi structure and lasing medium it can be detrimental to laser efficiency. These results are similar to simulations performed using the ISIS code.(M.E. Jones and V.A. Thomas, Proceedings of the 8^th) International Conference on High-Power Particle Beams, 665 (1990). We have identified the instability as the so called ``transit-time" instability(C.K. Birdsall and W.B. Bridges, Electrodynamics of Diode Regions), (Academic Press, New York, 1966).^,(T.M. Antonsen, W.H. Miner, E. Ott, and A.T. Drobot, Phys. Fluids 27), 1257 (1984). and have investigated the role of the applied magnetic field and diode geometry. Experiments are underway to characterize the instability on the Nike KrF laser system and will be compared to simulation. Also some possible ways to mitigate the instability will be presented.
International Nuclear Information System (INIS)
Oka, Yoshiaki; Okuda, Hiroshi
2006-01-01
Scientific computational methods have advanced remarkably with the progress of nuclear development. They have played the role of weft connecting each realm of nuclear engineering and then an introductory course of advanced scientific computational methods and their applications to nuclear technologies were prepared in serial form. This is the first issue showing their overview and introduction of continuum simulation methods. Finite element method as their applications is also reviewed. (T. Tanaka)
Large breast compressions: Observations and evaluation of simulations
Energy Technology Data Exchange (ETDEWEB)
Tanner, Christine; White, Mark; Guarino, Salvatore; Hall-Craggs, Margaret A.; Douek, Michael; Hawkes, David J. [Centre of Medical Image Computing, UCL, London WC1E 6BT, United Kingdom and Computer Vision Laboratory, ETH Zuerich, 8092 Zuerich (Switzerland); Centre of Medical Image Computing, UCL, London WC1E 6BT (United Kingdom); Department of Surgery, UCL, London W1P 7LD (United Kingdom); Department of Imaging, UCL Hospital, London NW1 2BU (United Kingdom); Department of Surgery, UCL, London W1P 7LD (United Kingdom); Centre of Medical Image Computing, UCL, London WC1E 6BT (United Kingdom)
2011-02-15
Purpose: Several methods have been proposed to simulate large breast compressions such as those occurring during x-ray mammography. However, the evaluation of these methods against real data is rare. The aim of this study is to learn more about the deformation behavior of breasts and to assess a simulation method. Methods: Magnetic resonance (MR) images of 11 breasts before and after applying a relatively large in vivo compression in the medial direction were acquired. Nonrigid registration was employed to study the deformation behavior. Optimal material properties for finite element modeling were determined and their prediction performance was assessed. The realism of simulated compressions was evaluated by comparing the breast shapes on simulated and real mammograms. Results: Following image registration, 19 breast compressions from 8 women were studied. An anisotropic deformation behavior, with a reduced elongation in the anterior-posterior direction and an increased stretch in the inferior-superior direction was observed. Using finite element simulations, the performance of isotropic and transverse isotropic material models to predict the displacement of internal landmarks was compared. Isotropic materials reduced the mean displacement error of the landmarks from 23.3 to 4.7 mm, on average, after optimizing material properties with respect to breast surface alignment and image similarity. Statistically significantly smaller errors were achieved with transverse isotropic materials (4.1 mm, P=0.0045). Homogeneous material models performed substantially worse (transverse isotropic: 5.5 mm; isotropic: 6.7 mm). Of the parameters varied, the amount of anisotropy had the greatest influence on the results. Optimal material properties varied less when grouped by patient rather than by compression magnitude (mean: 0.72 vs 1.44). Employing these optimal materials for simulating mammograms from ten MR breast images of a different cohort resulted in more realistic breast
Large breast compressions: observations and evaluation of simulations.
Tanner, Christine; White, Mark; Guarino, Salvatore; Hall-Craggs, Margaret A; Douek, Michael; Hawkes, David J
2011-02-01
Several methods have been proposed to simulate large breast compressions such as those occurring during x-ray mammography. However, the evaluation of these methods against real data is rare. The aim of this study is to learn more about the deformation behavior of breasts and to assess a simulation method. Magnetic resonance (MR) images of 11 breasts before and after applying a relatively large in vivo compression in the medial direction were acquired. Nonrigid registration was employed to study the deformation behavior. Optimal material properties for finite element modeling were determined and their prediction performance was assessed. The realism of simulated compressions was evaluated by comparing the breast shapes on simulated and real mammograms. Following image registration, 19 breast compressions from 8 women were studied. An anisotropic deformation behavior, with a reduced elongation in the anterior-posterior direction and an increased stretch in the inferior-superior direction was observed. Using finite element simulations, the performance of isotropic and transverse isotropic material models to predict the displacement of internal landmarks was compared. Isotropic materials reduced the mean displacement error of the landmarks from 23.3 to 4.7 mm, on average, after optimizing material properties with respect to breast surface alignment and image similarity. Statistically significantly smaller errors were achieved with transverse isotropic materials (4.1 mm, P=0.0045). Homogeneous material models performed substantially worse (transverse isotropic: 5.5 mm; isotropic: 6.7 mm). Of the parameters varied, the amount of anisotropy had the greatest influence on the results. Optimal material properties varied less when grouped by patient rather than by compression magnitude (mean: 0.72 vs. 1.44). Employing these optimal materials for simulating mammograms from ten MR breast images of a different cohort resulted in more realistic breast shapes than when using
Large breast compressions: Observations and evaluation of simulations
International Nuclear Information System (INIS)
Tanner, Christine; White, Mark; Guarino, Salvatore; Hall-Craggs, Margaret A.; Douek, Michael; Hawkes, David J.
2011-01-01
Purpose: Several methods have been proposed to simulate large breast compressions such as those occurring during x-ray mammography. However, the evaluation of these methods against real data is rare. The aim of this study is to learn more about the deformation behavior of breasts and to assess a simulation method. Methods: Magnetic resonance (MR) images of 11 breasts before and after applying a relatively large in vivo compression in the medial direction were acquired. Nonrigid registration was employed to study the deformation behavior. Optimal material properties for finite element modeling were determined and their prediction performance was assessed. The realism of simulated compressions was evaluated by comparing the breast shapes on simulated and real mammograms. Results: Following image registration, 19 breast compressions from 8 women were studied. An anisotropic deformation behavior, with a reduced elongation in the anterior-posterior direction and an increased stretch in the inferior-superior direction was observed. Using finite element simulations, the performance of isotropic and transverse isotropic material models to predict the displacement of internal landmarks was compared. Isotropic materials reduced the mean displacement error of the landmarks from 23.3 to 4.7 mm, on average, after optimizing material properties with respect to breast surface alignment and image similarity. Statistically significantly smaller errors were achieved with transverse isotropic materials (4.1 mm, P=0.0045). Homogeneous material models performed substantially worse (transverse isotropic: 5.5 mm; isotropic: 6.7 mm). Of the parameters varied, the amount of anisotropy had the greatest influence on the results. Optimal material properties varied less when grouped by patient rather than by compression magnitude (mean: 0.72 vs 1.44). Employing these optimal materials for simulating mammograms from ten MR breast images of a different cohort resulted in more realistic breast
Large Eddy Simulation of the spray formation in confinements
International Nuclear Information System (INIS)
Lampa, A.; Fritsching, U.
2013-01-01
Highlights: • Process stability of confined spray processes is affected by the geometric design of the spray confinement. • LES simulations of confined spray flow have been performed successfully. • Clustering processes of droplets is predicted in simulations and validated with experiments. • Criteria for specific coherent gas flow patterns and droplet clustering behaviour are found. -- Abstract: The particle and powder properties produced within spray drying processes are influenced by various unsteady transport phenomena in the dispersed multiphase spray flow in a confined spray chamber. In this context differently scaled spray structures in a confined spray environment have been analyzed in experiments and numerical simulations. The experimental investigations have been carried out with Particle-Image-Velocimetry to determine the velocity of the gas and the discrete phase. Large-Eddy-Simulations have been set up to predict the transient behaviour of the spray process and have given more insight into the sensitivity of the spray flow structures in dependency from the spray chamber design
Inviscid Wall-Modeled Large Eddy Simulations for Improved Efficiency
Aikens, Kurt; Craft, Kyle; Redman, Andrew
2015-11-01
The accuracy of an inviscid flow assumption for wall-modeled large eddy simulations (LES) is examined because of its ability to reduce simulation costs. This assumption is not generally applicable for wall-bounded flows due to the high velocity gradients found near walls. In wall-modeled LES, however, neither the viscous near-wall region or the viscous length scales in the outer flow are resolved. Therefore, the viscous terms in the Navier-Stokes equations have little impact on the resolved flowfield. Zero pressure gradient flat plate boundary layer results are presented for both viscous and inviscid simulations using a wall model developed previously. The results are very similar and compare favorably to those from another wall model methodology and experimental data. Furthermore, the inviscid assumption reduces simulation costs by about 25% and 39% for supersonic and subsonic flows, respectively. Future research directions are discussed as are preliminary efforts to extend the wall model to include the effects of unresolved wall roughness. This work used the Extreme Science and Engineering Discovery Environment (XSEDE), which is supported by National Science Foundation grant number ACI-1053575. Computational resources on TACC Stampede were provided under XSEDE allocation ENG150001.
Large eddy simulation of a fuel rod subchannel
International Nuclear Information System (INIS)
Mayer, Gusztav
2007-01-01
In a VVER-440 reactor the measured outlet temperature is related to fuel limit parameters and the power upgrading plans of VVER-440 reactors motivated us to obtain more information on the mixing process of the fuel assemblies. In a VVER-440 rod bundle the fuel rods are arranged in triangular array. Measurement shows (Krauss and Meyer, 1998) that the classical engineering approach, which tries to trace the characterization of such systems back to equivalent (hydraulic diameter) pipe flows, does not give reasonable results. Due to the different turbulence characteristics, the mixing is more intensive in rod bundles than it would be expected based on equivalent pipe flow correlations. As a possible explanation of the high mixing, secondary flow was deduced from measurements by several experimentalists (Trupp and Azad, 1975). Another candidate to explain the high mixing is the so-called flow pulsation phenomenon (Krauss and Meyer, 1998). In this paper we present subchannel simulations (Mayer et al. 2007) using large eddy simulation (LES) methodology and the lattice Boltzmann method (LBM) without the spacers at Reynolds number 21000. The simulation results are compared with the measurements of Trupp and Azad (1975). The mean axial velocity profile shows good agreement with the measurement data. Secondary flow has been observed directly in the simulation results. Reasonable agreement has been achieved for most Reynolds stresses. Nevertheless, the calculated normal stresses show small, but systematic deviation from the measurement data. (author)
Accelerating large-scale phase-field simulations with GPU
Directory of Open Access Journals (Sweden)
Xiaoming Shi
2017-10-01
Full Text Available A new package for accelerating large-scale phase-field simulations was developed by using GPU based on the semi-implicit Fourier method. The package can solve a variety of equilibrium equations with different inhomogeneity including long-range elastic, magnetostatic, and electrostatic interactions. Through using specific algorithm in Compute Unified Device Architecture (CUDA, Fourier spectral iterative perturbation method was integrated in GPU package. The Allen-Cahn equation, Cahn-Hilliard equation, and phase-field model with long-range interaction were solved based on the algorithm running on GPU respectively to test the performance of the package. From the comparison of the calculation results between the solver executed in single CPU and the one on GPU, it was found that the speed on GPU is enormously elevated to 50 times faster. The present study therefore contributes to the acceleration of large-scale phase-field simulations and provides guidance for experiments to design large-scale functional devices.
Quality and Reliability of Large-Eddy Simulations II
Salvetti, Maria Vittoria; Meyers, Johan; Sagaut, Pierre
2011-01-01
The second Workshop on "Quality and Reliability of Large-Eddy Simulations", QLES2009, was held at the University of Pisa from September 9 to September 11, 2009. Its predecessor, QLES2007, was organized in 2007 in Leuven (Belgium). The focus of QLES2009 was on issues related to predicting, assessing and assuring the quality of LES. The main goal of QLES2009 was to enhance the knowledge on error sources and on their interaction in LES and to devise criteria for the prediction and optimization of simulation quality, by bringing together mathematicians, physicists and engineers and providing a platform specifically addressing these aspects for LES. Contributions were made by leading experts in the field. The present book contains the written contributions to QLES2009 and is divided into three parts, which reflect the main topics addressed at the workshop: (i) SGS modeling and discretization errors; (ii) Assessment and reduction of computational errors; (iii) Mathematical analysis and foundation for SGS modeling.
Large Eddy Simulation (LES for IC Engine Flows
Directory of Open Access Journals (Sweden)
Kuo Tang-Wei
2013-10-01
Full Text Available Numerical computations are carried out using an engineering-level Large Eddy Simulation (LES model that is provided by a commercial CFD code CONVERGE. The analytical framework and experimental setup consist of a single cylinder engine with Transparent Combustion Chamber (TCC under motored conditions. A rigorous working procedure for comparing and analyzing the results from simulation and high speed Particle Image Velocimetry (PIV experiments is documented in this work. The following aspects of LES are analyzed using this procedure: number of cycles required for convergence with adequate accuracy; effect of mesh size, time step, sub-grid-scale (SGS turbulence models and boundary condition treatments; application of the proper orthogonal decomposition (POD technique.
Large Eddy Simulation for Incompressible Flows An Introduction
Sagaut, P
2005-01-01
The first and most exhaustive work of its kind devoted entirely to the subject, Large Eddy Simulation presents a comprehensive account and a unified view of this young but very rich discipline. LES is the only efficient technique for approaching high Reynolds numbers when simulating industrial, natural or experimental configurations. The author concentrates on incompressible fluids and chooses his topics in treating with care both the mathematical ideas and their applications. The book addresses researchers as well as graduate students and engineers. The second edition was a greatly enriched version motivated both by the increasing theoretical interest in LES and the increasing number of applications. Two entirely new chapters were devoted to the coupling of LES with multiresolution multidomain techniques and to the new hybrid approaches that relate the LES procedures to the classical statistical methods based on the Reynolds-Averaged Navier-Stokes equations. This 3rd edition adds various sections to the text...
Aero-Acoustic Modelling using Large Eddy Simulation
International Nuclear Information System (INIS)
Shen, W Z; Soerensen, J N
2007-01-01
The splitting technique for aero-acoustic computations is extended to simulate three-dimensional flow and acoustic waves from airfoils. The aero-acoustic model is coupled to a sub-grid-scale turbulence model for Large-Eddy Simulations. In the first test case, the model is applied to compute laminar flow past a NACA 0015 airfoil at a Reynolds number of 800, a Mach number of 0.2 and an angle of attack of 20 deg. The model is then applied to compute turbulent flow past a NACA 0015 airfoil at a Reynolds number of 100 000, a Mach number of 0.2 and an angle of attack of 20 deg. The predicted noise spectrum is compared to experimental data
Very large eddy simulation of the Red Sea overflow
Ilıcak, Mehmet; Özgökmen, Tamay M.; Peters, Hartmut; Baumert, Helmut Z.; Iskandarani, Mohamed
Mixing between overflows and ambient water masses is a critical problem of deep-water mass formation in the downwelling branch of the meridional overturning circulation of the ocean. Modeling approaches that have been tested so far rely either on algebraic parameterizations in hydrostatic ocean circulation models, or on large eddy simulations that resolve most of the mixing using nonhydrostatic models. In this study, we examine the performance of a set of turbulence closures, that have not been tested in comparison to observational data for overflows before. We employ the so-called very large eddy simulation (VLES) technique, which allows the use of k-ɛ models in nonhydrostatic models. This is done by applying a dynamic spatial filtering to the k-ɛ equations. To our knowledge, this is the first time that the VLES approach is adopted for an ocean modeling problem. The performance of k-ɛ and VLES models are evaluated by conducting numerical simulations of the Red Sea overflow and comparing them to observations from the Red Sea Outflow Experiment (REDSOX). The computations are constrained to one of the main channels transporting the overflow, which is narrow enough to permit the use of a two-dimensional (and nonhydrostatic) model. A large set of experiments are conducted using different closure models, Reynolds numbers and spatial resolutions. It is found that, when no turbulence closure is used, the basic structure of the overflow, consisting of a well-mixed bottom layer (BL) and entraining interfacial layer (IL), cannot be reproduced. The k-ɛ model leads to unrealistic thicknesses for both BL and IL, while VLES results in the most realistic reproduction of the REDSOX observations.
Large-eddy simulation of swirling pulverized-coal combustion
Energy Technology Data Exchange (ETDEWEB)
Hu, L.Y.; Luo, Y.H. [Shanghai Jiaotong Univ. (China). School of Mechanical Engineering; Zhou, L.X.; Xu, C.S. [Tsinghua Univ., Beijing (China). Dept. of Engineering Mechanics
2013-07-01
A Eulerian-Lagrangian large-eddy simulation (LES) with a Smagorinsky-Lilly sub-grid scale stress model, presumed-PDF fast chemistry and EBU gas combustion models, particle devolatilization and particle combustion models are used to study the turbulence and flame structures of swirling pulverized-coal combustion. The LES statistical results are validated by the measurement results. The instantaneous LES results show that the coherent structures for pulverized coal combustion is stronger than that for swirling gas combustion. The particles are concentrated in the periphery of the coherent structures. The flame is located at the high vorticity and high particle concentration zone.
Large Scale Simulations of the Euler Equations on GPU Clusters
Liebmann, Manfred
2010-08-01
The paper investigates the scalability of a parallel Euler solver, using the Vijayasundaram method, on a GPU cluster with 32 Nvidia Geforce GTX 295 boards. The aim of this research is to enable large scale fluid dynamics simulations with up to one billion elements. We investigate communication protocols for the GPU cluster to compensate for the slow Gigabit Ethernet network between the GPU compute nodes and to maintain overall efficiency. A diesel engine intake-port and a nozzle, meshed in different resolutions, give good real world examples for the scalability tests on the GPU cluster. © 2010 IEEE.
Large Eddy Simulation of the ventilated wave boundary layer
DEFF Research Database (Denmark)
Lohmann, Iris P.; Fredsøe, Jørgen; Sumer, B. Mutlu
2006-01-01
A Large Eddy Simulation (LES) of (1) a fully developed turbulent wave boundary layer and (2) case 1 subject to ventilation (i.e., suction and injection varying alternately in phase) has been performed, using the Smagorinsky subgrid-scale model to express the subgrid viscosity. The model was found...... slows down the flow in the full vertical extent of the boundary layer, destabilizes the flow and decreases the mean bed shear stress significantly; whereas suction generally speeds up the flow in the full vertical extent of the boundary layer, stabilizes the flow and increases the mean bed shear stress...
Roundtree, Aimee Kendall
2013-01-01
Computer simulations help advance climatology, astrophysics, and other scientific disciplines. They are also at the crux of several high-profile cases of science in the news. How do simulation scientists, with little or no direct observations, make decisions about what to represent? What is the nature of simulated evidence, and how do we evaluate its strength? Aimee Kendall Roundtree suggests answers in Computer Simulation, Rhetoric, and the Scientific Imagination. She interprets simulations in the sciences by uncovering the argumentative strategies that underpin the production and disseminati
Multiscale Data Assimilation for Large-Eddy Simulations
Li, Z.; Cheng, X.; Gustafson, W. I., Jr.; Xiao, H.; Vogelmann, A. M.; Endo, S.; Toto, T.
2017-12-01
Large-eddy simulation (LES) is a powerful tool for understanding atmospheric turbulence, boundary layer physics and cloud development, and there is a great need for developing data assimilation methodologies that can constrain LES models. The U.S. Department of Energy Atmospheric Radiation Measurement (ARM) User Facility has been developing the capability to routinely generate ensembles of LES. The LES ARM Symbiotic Simulation and Observation (LASSO) project (https://www.arm.gov/capabilities/modeling/lasso) is generating simulations for shallow convection days at the ARM Southern Great Plains site in Oklahoma. One of major objectives of LASSO is to develop the capability to observationally constrain LES using a hierarchy of ARM observations. We have implemented a multiscale data assimilation (MSDA) scheme, which allows data assimilation to be implemented separately for distinct spatial scales, so that the localized observations can be effectively assimilated to constrain the mesoscale fields in the LES area of about 15 km in width. The MSDA analysis is used to produce forcing data that drive LES. With such LES workflow we have examined 13 days with shallow convection selected from the period May-August 2016. We will describe the implementation of MSDA, present LES results, and address challenges and opportunities for applying data assimilation to LES studies.
Large eddy simulation of turbulent and stably-stratified flows
International Nuclear Information System (INIS)
Fallon, Benoit
1994-01-01
The unsteady turbulent flow over a backward-facing step is studied by mean of Large Eddy Simulations with structure function sub grid model, both in isothermal and stably-stratified configurations. Without stratification, the flow develops highly-distorted Kelvin-Helmholtz billows, undergoing to helical pairing, with A-shaped vortices shed downstream. We show that forcing injected by recirculation fluctuations governs this oblique mode instabilities development. The statistical results show good agreements with the experimental measurements. For stably-stratified configurations, the flow remains more bi-dimensional. We show with increasing stratification, how the shear layer growth is frozen by inhibition of pairing process then of Kelvin-Helmholtz instabilities, and the development of gravity waves or stable density interfaces. Eddy structures of the flow present striking analogies with the stratified mixing layer. Additional computations show the development of secondary Kelvin-Helmholtz instabilities on the vorticity layers between two primary structures. This important mechanism based on baroclinic effects (horizontal density gradients) constitutes an additional part of the turbulent mixing process. Finally, the feasibility of Large Eddy Simulation is demonstrated for industrial flows, by studying a complex stratified cavity. Temperature fluctuations are compared to experimental measurements. We also develop three-dimensional un-stationary animations, in order to understand and visualize turbulent interactions. (author) [fr
Large Eddy Simulation of Film-Cooling Jets
Iourokina, Ioulia
2005-11-01
Large Eddy Simulation of inclined jets issuing into a turbulent boundary layer crossflow has been performed. The simulation models film-cooling experiments of Pietrzyk et al. (J. of. Turb., 1989), consisting of a large plenum feeding an array of jets inclined at 35ÃÂ° to the flat surface with a pitch 3D and L/D=3.5. The blowing ratio is 0.5 with unity density ratio. The numerical method used is a hybrid combining external compressible solver with a low-Mach number code for the plenum and film holes. Vorticity dynamics pertinent to jet-in-crossflow interactions is analyzed and three-dimensional vortical structures are revealed. Turbulence statistics are compared to the experimental data. The turbulence production due to shearing in the crossflow is compared to that within the jet hole. The influence of three-dimensional coherent structures on the wall heat transfer is investigated and strategies to increase film- cooling performance are discussed.
Large Eddy Simulation of Supercritical CO2 Through Bend Pipes
He, Xiaoliang; Apte, Sourabh; Dogan, Omer
2017-11-01
Supercritical Carbon Dioxide (sCO2) is investigated as working fluid for power generation in thermal solar, fossil energy and nuclear power plants at high pressures. Severe erosion has been observed in the sCO2 test loops, particularly in nozzles, turbine blades and pipe bends. It is hypothesized that complex flow features such as flow separation and property variations may lead to large oscillations in the wall shear stresses and result in material erosion. In this work, large eddy simulations are conducted at different Reynolds numbers (5000, 27,000 and 50,000) to investigate the effect of heat transfer in a 90 degree bend pipe with unit radius of curvature in order to identify the potential causes of the erosion. The simulation is first performed without heat transfer to validate the flow solver against available experimental and computational studies. Mean flow statistics, turbulent kinetic energy, shear stresses and wall force spectra are computed and compared with available experimental data. Formation of counter-rotating vortices, named Dean vortices, are observed. Secondary flow pattern and swirling-switching flow motions are identified and visualized. Effects of heat transfer on these flow phenomena are then investigated by applying a constant heat flux at the wall. DOE Fossil Energy Crosscutting Technology Research Program.
Cleaning of scientific references in large patent databases using rule-based scoring and clustering
Caron, Emiel
2017-01-01
Patent databases contain patent related data, organized in a relational data model, and are used to produce various patent statistics. These databases store raw data about scientific references cited by patents. For example, Patstat holds references to tens of millions of scientific journal
Numerical techniques for large cosmological N-body simulations
International Nuclear Information System (INIS)
Efstathiou, G.; Davis, M.; Frenk, C.S.; White, S.D.M.
1985-01-01
We describe and compare techniques for carrying out large N-body simulations of the gravitational evolution of clustering in the fundamental cube of an infinite periodic universe. In particular, we consider both particle mesh (PM) codes and P 3 M codes in which a higher resolution force is obtained by direct summation of contributions from neighboring particles. We discuss the mesh-induced anisotropies in the forces calculated by these schemes, and the extent to which they can model the desired 1/r 2 particle-particle interaction. We also consider how transformation of the time variable can improve the efficiency with which the equations of motion are integrated. We present tests of the accuracy with which the resulting schemes conserve energy and are able to follow individual particle trajectories. We have implemented an algorithm which allows initial conditions to be set up to model any desired spectrum of linear growing mode density fluctuations. A number of tests demonstrate the power of this algorithm and delineate the conditions under which it is effective. We carry out several test simulations using a variety of techniques in order to show how the results are affected by dynamic range limitations in the force calculations, by boundary effects, by residual artificialities in the initial conditions, and by the number of particles employed. For most purposes cosmological simulations are limited by the resolution of their force calculation rather than by the number of particles they can employ. For this reason, while PM codes are quite adequate to study the evolution of structure on large scale, P 3 M methods are to be preferred, in spite of their greater cost and complexity, whenever the evolution of small-scale structure is important
Large Eddy Simulation for an inherent boron dilution transient
International Nuclear Information System (INIS)
Jayaraju, S.T.; Sathiah, P.; Komen, E.M.J.; Baglietto, E.
2013-01-01
Highlights: • Large Eddy Simulation is performed for a transient boron dilution scenario in the scaled experimental facility of ROCOM. • Fully conformal polyhedral grid of 14 million is created to capture all details of the domain. • Systematic multi-step validation methodology is followed to assess the accuracy of LES model. • For the presently simulated BDT scenario, the LES results lend support to its reliability in consistently predicting the slug transport in the RPV. -- Abstract: The present paper focuses on the validation and applicability of large eddy simulation (LES) to analyze the transport and mixing in the reactor pressure vessel (RPV) during an inherent boron dilution transient (BDT) scenario. Extensive validation data comes from relevant integral tests performed in the scaled ROCOM experimental facility. The modeling of sub-grid-scales is based on the WALE model. A fully conformal polyhedral grid of about 15 million cells is constructed to capture all details in the domain, including the complex structures of the lower-plenum. Detailed qualitative and quantitative validations are performed by following a systematic multi-step validation methodology. Qualitative comparisons to the experimental data in the cold legs, downcomer and the core inlet showed good predictions by the LES model. Minor deviations seen in the quantitative comparisons are rigorously quantified. A key parameter which is affecting the core neutron kinetics response is the value of highest deborated slug concentration that occurs at the core inlet during the transient. Detailed analyses are made at the core inlet to evaluate not only the value of the maximum slug concentration, but also the location and the time at which it occurs during the transient. The relative differences between the ensemble averaged experimental data and CFD predictions were within the range of relative differences seen within 10 different experimental realizations. For the studied scenario, the
Large-eddy simulation of maritime deep tropical convection
Directory of Open Access Journals (Sweden)
Peter A Bogenschutz
2009-12-01
Full Text Available This study represents an attempt to apply Large-Eddy Simulation (LES resolution to simulate deep tropical convection in near equilibrium for 24 hours over an area of about 205 x 205 km2, which is comparable to that of a typical horizontal grid cell in a global climate model. The simulation is driven by large-scale thermodynamic tendencies derived from mean conditions during the GATE Phase III field experiment. The LES uses 2048 x 2048 x 256 grid points with horizontal grid spacing of 100 m and vertical grid spacing ranging from 50 m in the boundary layer to 100 m in the free troposphere. The simulation reaches a near equilibrium deep convection regime in 12 hours. The simulated vertical cloud distribution exhibits a trimodal vertical distribution of deep, middle and shallow clouds similar to that often observed in Tropics. A sensitivity experiment in which cold pools are suppressed by switching off the evaporation of precipitation results in much lower amounts of shallow and congestus clouds. Unlike the benchmark LES where the new deep clouds tend to appear along the edges of spreading cold pools, the deep clouds in the no-cold-pool experiment tend to reappear at the sites of the previous deep clouds and tend to be surrounded by extensive areas of sporadic shallow clouds. The vertical velocity statistics of updraft and downdraft cores below 6 km height are compared to aircraft observations made during GATE. The comparison shows generally good agreement, and strongly suggests that the LES simulation can be used as a benchmark to represent the dynamics of tropical deep convection on scales ranging from large turbulent eddies to mesoscale convective systems. The effect of horizontal grid resolution is examined by running the same case with progressively larger grid sizes of 200, 400, 800, and 1600 m. These runs show a reasonable agreement with the benchmark LES in statistics such as convective available potential energy, convective inhibition
Nesting Large-Eddy Simulations Within Mesoscale Simulations for Wind Energy Applications
Lundquist, J. K.; Mirocha, J. D.; Chow, F. K.; Kosovic, B.; Lundquist, K. A.
2008-12-01
With increasing demand for more accurate atmospheric simulations for wind turbine micrositing, for operational wind power forecasting, and for more reliable turbine design, simulations of atmospheric flow with resolution of tens of meters or higher are required. These time-dependent large-eddy simulations (LES) account for complex terrain and resolve individual atmospheric eddies on length scales smaller than turbine blades. These small-domain high-resolution simulations are possible with a range of commercial and open- source software, including the Weather Research and Forecasting (WRF) model. In addition to "local" sources of turbulence within an LES domain, changing weather conditions outside the domain can also affect flow, suggesting that a mesoscale model provide boundary conditions to the large-eddy simulations. Nesting a large-eddy simulation within a mesoscale model requires nuanced representations of turbulence. Our group has improved the Weather and Research Forecating model's (WRF) LES capability by implementing the Nonlinear Backscatter and Anisotropy (NBA) subfilter stress model following Kosoviæ (1997) and an explicit filtering and reconstruction technique to compute the Resolvable Subfilter-Scale (RSFS) stresses (following Chow et al, 2005). We have also implemented an immersed boundary method (IBM) in WRF to accommodate complex terrain. These new models improve WRF's LES capabilities over complex terrain and in stable atmospheric conditions. We demonstrate approaches to nesting LES within a mesoscale simulation for farms of wind turbines in hilly regions. Results are sensitive to the nesting method, indicating that care must be taken to provide appropriate boundary conditions, and to allow adequate spin-up of turbulence in the LES domain. This work is performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
Large eddy simulation of soot evolution in an aircraft combustor
Mueller, Michael E.; Pitsch, Heinz
2013-11-01
An integrated kinetics-based Large Eddy Simulation (LES) approach for soot evolution in turbulent reacting flows is applied to the simulation of a Pratt & Whitney aircraft gas turbine combustor, and the results are analyzed to provide insights into the complex interactions of the hydrodynamics, mixing, chemistry, and soot. The integrated approach includes detailed models for soot, combustion, and the unresolved interactions between soot, chemistry, and turbulence. The soot model is based on the Hybrid Method of Moments and detailed descriptions of soot aggregates and the various physical and chemical processes governing their evolution. The detailed kinetics of jet fuel oxidation and soot precursor formation is described with the Radiation Flamelet/Progress Variable model, which has been modified to account for the removal of soot precursors from the gas-phase. The unclosed filtered quantities in the soot and combustion models, such as source terms, are closed with a novel presumed subfilter PDF approach that accounts for the high subfilter spatial intermittency of soot. For the combustor simulation, the integrated approach is combined with a Lagrangian parcel method for the liquid spray and state-of-the-art unstructured LES technology for complex geometries. Two overall fuel-to-air ratios are simulated to evaluate the ability of the model to make not only absolute predictions but also quantitative predictions of trends. The Pratt & Whitney combustor is a Rich-Quench-Lean combustor in which combustion first occurs in a fuel-rich primary zone characterized by a large recirculation zone. Dilution air is then added downstream of the recirculation zone, and combustion continues in a fuel-lean secondary zone. The simulations show that large quantities of soot are formed in the fuel-rich recirculation zone, and, furthermore, the overall fuel-to-air ratio dictates both the dominant soot growth process and the location of maximum soot volume fraction. At the higher fuel
Robust large-scale parallel nonlinear solvers for simulations.
Energy Technology Data Exchange (ETDEWEB)
Bader, Brett William; Pawlowski, Roger Patrick; Kolda, Tamara Gibson (Sandia National Laboratories, Livermore, CA)
2005-11-01
This report documents research to develop robust and efficient solution techniques for solving large-scale systems of nonlinear equations. The most widely used method for solving systems of nonlinear equations is Newton's method. While much research has been devoted to augmenting Newton-based solvers (usually with globalization techniques), little has been devoted to exploring the application of different models. Our research has been directed at evaluating techniques using different models than Newton's method: a lower order model, Broyden's method, and a higher order model, the tensor method. We have developed large-scale versions of each of these models and have demonstrated their use in important applications at Sandia. Broyden's method replaces the Jacobian with an approximation, allowing codes that cannot evaluate a Jacobian or have an inaccurate Jacobian to converge to a solution. Limited-memory methods, which have been successful in optimization, allow us to extend this approach to large-scale problems. We compare the robustness and efficiency of Newton's method, modified Newton's method, Jacobian-free Newton-Krylov method, and our limited-memory Broyden method. Comparisons are carried out for large-scale applications of fluid flow simulations and electronic circuit simulations. Results show that, in cases where the Jacobian was inaccurate or could not be computed, Broyden's method converged in some cases where Newton's method failed to converge. We identify conditions where Broyden's method can be more efficient than Newton's method. We also present modifications to a large-scale tensor method, originally proposed by Bouaricha, for greater efficiency, better robustness, and wider applicability. Tensor methods are an alternative to Newton-based methods and are based on computing a step based on a local quadratic model rather than a linear model. The advantage of Bouaricha's method is that it can use any
Using an Agent-Based Modeling Simulation and Game to Teach Socio-Scientific Topics
Directory of Open Access Journals (Sweden)
Lori L. Scarlatos
2014-02-01
Full Text Available In our modern world, where science, technology and society are tightly interwoven, it is essential that all students be able to evaluate scientific evidence and make informed decisions. Energy Choices, an agent-based simulation with a multiplayer game interface, was developed as a learning tool that models the interdependencies between the energy choices that are made, growth in local economies, and climate change on a global scale. This paper presents the results of pilot testing Energy Choices in two different settings, using two different modes of delivery.
Hulshof, Casper D.; de Jong, Ton
2006-01-01
Students encounter many obstacles during scientific discovery learning with computer-based simulations. It is hypothesized that an effective type of support, that does not interfere with the scientific discovery learning process, should be delivered on a "just-in-time" base. This study explores the effect of facilitating access to…
Large-scale ground motion simulation using GPGPU
Aoi, S.; Maeda, T.; Nishizawa, N.; Aoki, T.
2012-12-01
Huge computation resources are required to perform large-scale ground motion simulations using 3-D finite difference method (FDM) for realistic and complex models with high accuracy. Furthermore, thousands of various simulations are necessary to evaluate the variability of the assessment caused by uncertainty of the assumptions of the source models for future earthquakes. To conquer the problem of restricted computational resources, we introduced the use of GPGPU (General purpose computing on graphics processing units) which is the technique of using a GPU as an accelerator of the computation which has been traditionally conducted by the CPU. We employed the CPU version of GMS (Ground motion Simulator; Aoi et al., 2004) as the original code and implemented the function for GPU calculation using CUDA (Compute Unified Device Architecture). GMS is a total system for seismic wave propagation simulation based on 3-D FDM scheme using discontinuous grids (Aoi&Fujiwara, 1999), which includes the solver as well as the preprocessor tools (parameter generation tool) and postprocessor tools (filter tool, visualization tool, and so on). The computational model is decomposed in two horizontal directions and each decomposed model is allocated to a different GPU. We evaluated the performance of our newly developed GPU version of GMS on the TSUBAME2.0 which is one of the Japanese fastest supercomputer operated by the Tokyo Institute of Technology. First we have performed a strong scaling test using the model with about 22 million grids and achieved 3.2 and 7.3 times of the speed-up by using 4 and 16 GPUs. Next, we have examined a weak scaling test where the model sizes (number of grids) are increased in proportion to the degree of parallelism (number of GPUs). The result showed almost perfect linearity up to the simulation with 22 billion grids using 1024 GPUs where the calculation speed reached to 79.7 TFlops and about 34 times faster than the CPU calculation using the same number
Large Eddy Simulations of turbulent flows at supercritical pressure
Energy Technology Data Exchange (ETDEWEB)
Kunik, C.; Otic, I.; Schulenberg, T., E-mail: claus.kunik@kit.edu, E-mail: ivan.otic@kit.edu, E-mail: thomas.schulenberg@kit.edu [Karlsruhe Inst. of Tech. (KIT), Karlsruhe (Germany)
2011-07-01
A Large Eddy Simulation (LES) method is used to investigate turbulent heat transfer to CO{sub 2} at supercritical pressure for upward flows. At those pressure conditions the fluid undergoes strong variations of fluid properties in a certain temperature range, which can lead to a deterioration of heat transfer (DHT). In this analysis, the LES method is applied on turbulent forced convection conditions to investigate the influence of several subgrid scale models (SGS-model). At first, only velocity profiles of the so-called inflow generator are considered, whereas in the second part temperature profiles of the heated section are investigated in detail. The results are statistically analyzed and compared with DNS data from the literature. (author)
Generating wind fluctuations for Large Eddy Simulation inflow boundary condition
International Nuclear Information System (INIS)
Bekele, S.A.; Hangan, H.
2004-01-01
Large Eddy Simulation (LES) studies of flows over bluff bodies immersed in a boundary layer wind environment require instantaneous wind characteristics. The influences of the wind environment on the building pressure distribution are a well-established fact in the experimental study of wind engineering. Measured wind data of full or model scale are available only at a limited number of points. A method of obtaining instantaneous wind data at all mesh points of the inlet boundary for LES computation is necessary. Herein previous and new wind inflow generation techniques are presented. The generated wind data is then applied to a LES computation of a channel flow. The characteristics of the generated wind fluctuations in comparison to the measured data and the properties of the flow field computed from these two wind data are discussed. (author)
Large eddy simulation of vortex breakdown behind a delta wing
International Nuclear Information System (INIS)
Mary, I.
2003-01-01
A large eddy simulation (LES) of a turbulent flow past a 70 deg. sweep angle delta wing is performed and compared with wind tunnel experiments. The angle of attack and the Reynolds number based on the root chord are equal to 27 deg. and 1.6x10 6 , respectively. Due to the high value of the Reynolds number and the three-dimensional geometry, the mesh resolution usually required by LES cannot be reached. Therefore a local mesh refinement technique based on semi-structured grids is proposed, whereas different wall functions are assessed in this paper. The goal is to evaluate if these techniques are sufficient to provide an accurate solution of such flow on available supercomputers. An implicit Miles model is retained for the subgrid scale (SGS) modelling because the resolution is too coarse to take advantage of more sophisticated SGS models. The solution sensitivity to grid refinement in the streamwise and wall normal direction is investigated
Langevin dynamics simulations of large frustrated Josephson junction arrays
International Nuclear Information System (INIS)
Groenbech-Jensen, N.; Bishop, A.R.; Lomdahl, P.S.
1991-01-01
Long-time Langevin dynamics simulations of large (N x N,N = 128) 2-dimensional arrays of Josephson junctions in a uniformly frustrating external magnetic field are reported. The results demonstrate: (1) Relaxation from an initially random flux configuration as a universal fit to a glassy stretched-exponential type of relaxation for the intermediate temperatures T(0.3 T c approx-lt T approx-lt 0.7 T c ), and an activated dynamic behavior for T ∼ T c ; (2) a glassy (multi-time, multi-length scale) voltage response to an applied current. Intrinsic dynamical symmetry breaking induced by boundaries as nucleation sites for flux lattice defects gives rise to transverse and noisy voltage response
Langevin dynamics simulations of large frustrated Josephson junction arrays
International Nuclear Information System (INIS)
Gronbech-Jensen, N.; Bishop, A.R.; Lomdahl, P.S.
1991-01-01
Long-time Langevin dynamics simulations of large (N x N, N = 128) 2-dimensional arrays of Josephson junctions in a uniformly frustrating external magnetic field are reported. The results demonstrate: Relaxation from an initially random flux configuration as a ''universal'' fit to a ''glassy'' stretched-exponential type of relaxation for the intermediate temperatures T (0.3 T c approx-lt T approx-lt 0.7 T c ), and an ''activated dynamic'' behavior for T ∼ T c A glassy (multi-time, multi-length scale) voltage response to an applied current. Intrinsic dynamical symmetry breaking induced by boundaries as nucleation sites for flux lattice defects gives rise to transverse and noisy voltage response
Large-Eddy-Simulation of turbulent magnetohydrodynamic flows
Directory of Open Access Journals (Sweden)
Woelck Johannes
2017-01-01
Full Text Available A magnetohydrodynamic turbulent channel flow under the influence of a wallnormal magnetic field is investigated using the Large-Eddy-Simulation technique and k-equation subgrid-scale-model. Therefore, the new solver MHDpisoFoam is implemented in the OpenFOAM CFD-Code. The temporal decay of an initial turbulent field for different magnetic parameters is investigated. The rms values of the averaged velocity fluctuations show a similar, trend for each coordinate direction. 80% of the fluctuations are damped out in the range between 0 < Ha < < 75 at Re = 6675. The trend can be approximated via an exponential of the form exp(−a·Ha, where a is a scaling parameter. At higher Hartmann numbers the fluctuations decrease in an almost linear way. Therefore, the results of this study show that it may be possible to construct a general law for the turbulence damping due to action of magnetic fields.
Large eddy simulation of the flow through a swirl generator
Energy Technology Data Exchange (ETDEWEB)
Conway, Stephen
1998-12-01
The advances made in computer technology over recent years have led to a great increase in the engineering problems that can be studied using CFD. The computation of flows over and through complex geometries at relatively high Reynolds numbers is becoming more common using the Large Eddy Simulation (LES) technique. Direct numerical simulations of such flows is still beyond the capacity of todays fastest supercomputers, requiring excessive computational times and memory. In addition, traditional Reynolds Averaged Navier Stokes (RANS) methods are known to have limited applicability in a wide range of engineering flow situations. In this thesis LES has been used to simulate the flow through a cascade of guidance vanes, more commonly known as a swirl generator, positioned at the inlet to a gas turbine combustion chamber. This flow case is of interest because of the complex flow phenomena which occur within the swirl generator, which include compressibility effects, different types of flow instabilities, transition, laminar and turbulent separation and near wall turbulence. It is also of interest because it fits very well into the range of engineering applications that can be studied using LES. Two computational grids with different resolutions and two subgrid scale stress models were used in the study. The effects of separation and transition are investigated. A vortex shedding frequency from the guidance vanes is determined which is seen to be dependent on the angle of incident air flow. Interaction between the movement of the separation region and the shedding frequency is also noted. Such vortex shedding phenomena can directly affect the quality of fuel and air mixing within the combustion chamber and can in some cases induce vibrations in the gas turbine structure. Comparisons between the results obtained using different grid resolutions with an implicit and a dynamic divergence (DDM) subgrid scale stress models are also made 32 refs, 35 figs, 2 tabs
Large Eddy Simulation of Vertical Axis Wind Turbine Wakes
Directory of Open Access Journals (Sweden)
Sina Shamsoddin
2014-02-01
Full Text Available In this study, large eddy simulation (LES is combined with a turbine model to investigate the wake behind a vertical-axis wind turbine (VAWT in a three-dimensional turbulent flow. Two methods are used to model the subgrid-scale (SGS stresses: (a the Smagorinsky model; and (b the modulated gradient model. To parameterize the effects of the VAWT on the flow, two VAWT models are developed: (a the actuator swept-surface model (ASSM, in which the time-averaged turbine-induced forces are distributed on a surface swept by the turbine blades, i.e., the actuator swept surface; and (b the actuator line model (ALM, in which the instantaneous blade forces are only spatially distributed on lines representing the blades, i.e., the actuator lines. This is the first time that LES has been applied and validated for the simulation of VAWT wakes by using either the ASSM or the ALM techniques. In both models, blade-element theory is used to calculate the lift and drag forces on the blades. The results are compared with flow measurements in the wake of a model straight-bladed VAWT, carried out in the Institute de Méchanique et Statistique de la Turbulence (IMST water channel. Different combinations of SGS models with VAWT models are studied, and a fairly good overall agreement between simulation results and measurement data is observed. In general, the ALM is found to better capture the unsteady-periodic nature of the wake and shows a better agreement with the experimental data compared with the ASSM. The modulated gradient model is also found to be a more reliable SGS stress modeling technique, compared with the Smagorinsky model, and it yields reasonable predictions of the mean flow and turbulence characteristics of a VAWT wake using its theoretically-determined model coefficient.
Simulation of ODE/PDE models with MATLAB, OCTAVE and SCILAB scientific and engineering applications
Vande Wouwer, Alain; Vilas, Carlos
2014-01-01
Simulation of ODE/PDE Models with MATLAB®, OCTAVE and SCILAB shows the reader how to exploit a fuller array of numerical methods for the analysis of complex scientific and engineering systems than is conventionally employed. The book is dedicated to numerical simulation of distributed parameter systems described by mixed systems of algebraic equations, ordinary differential equations (ODEs) and partial differential equations (PDEs). Special attention is paid to the numerical method of lines (MOL), a popular approach to the solution of time-dependent PDEs, which proceeds in two basic steps: spatial discretization and time integration. Besides conventional finite-difference and -element techniques, more advanced spatial-approximation methods are examined in some detail, including nonoscillatory schemes and adaptive-grid approaches. A MOL toolbox has been developed within MATLAB®/OCTAVE/SCILAB. In addition to a set of spatial approximations and time integrators, this toolbox includes a collection of applicatio...
Study of Hydrokinetic Turbine Arrays with Large Eddy Simulation
Sale, Danny; Aliseda, Alberto
2014-11-01
Marine renewable energy is advancing towards commercialization, including electrical power generation from ocean, river, and tidal currents. The focus of this work is to develop numerical simulations capable of predicting the power generation potential of hydrokinetic turbine arrays-this includes analysis of unsteady and averaged flow fields, turbulence statistics, and unsteady loadings on turbine rotors and support structures due to interaction with rotor wakes and ambient turbulence. The governing equations of large-eddy-simulation (LES) are solved using a finite-volume method, and the presence of turbine blades are approximated by the actuator-line method in which hydrodynamic forces are projected to the flow field as a body force. The actuator-line approach captures helical wake formation including vortex shedding from individual blades, and the effects of drag and vorticity generation from the rough seabed surface are accounted for by wall-models. This LES framework was used to replicate a previous flume experiment consisting of three hydrokinetic turbines tested under various operating conditions and array layouts. Predictions of the power generation, velocity deficit and turbulence statistics in the wakes are compared between the LES and experimental datasets.
Huang, Jingfeng; Hsu, N. Christina; Tsay, Si-Chee; Zhang, Chidong; Jeong, Myeong Jae; Gautam, Ritesh; Bettenhausen, Corey; Sayer, Andrew M.; Hansell, Richard A.; Liu, Xiaohong;
2012-01-01
One of the seven scientific areas of interests of the 7-SEAS field campaign is to evaluate the impact of aerosol on cloud and precipitation (http://7-seas.gsfc.nasa.gov). However, large-scale covariability between aerosol, cloud and precipitation is complicated not only by ambient environment and a variety of aerosol effects, but also by effects from rain washout and climate factors. This study characterizes large-scale aerosol-cloud-precipitation covariability through synergy of long-term multi ]sensor satellite observations with model simulations over the 7-SEAS region [10S-30N, 95E-130E]. Results show that climate factors such as ENSO significantly modulate aerosol and precipitation over the region simultaneously. After removal of climate factor effects, aerosol and precipitation are significantly anti-correlated over the southern part of the region, where high aerosols loading is associated with overall reduced total precipitation with intensified rain rates and decreased rain frequency, decreased tropospheric latent heating, suppressed cloud top height and increased outgoing longwave radiation, enhanced clear-sky shortwave TOA flux but reduced all-sky shortwave TOA flux in deep convective regimes; but such covariability becomes less notable over the northern counterpart of the region where low ]level stratus are found. Using CO as a proxy of biomass burning aerosols to minimize the washout effect, large-scale covariability between CO and precipitation was also investigated and similar large-scale covariability observed. Model simulations with NCAR CAM5 were found to show similar effects to observations in the spatio-temporal patterns. Results from both observations and simulations are valuable for improving our understanding of this region's meteorological system and the roles of aerosol within it. Key words: aerosol; precipitation; large-scale covariability; aerosol effects; washout; climate factors; 7- SEAS; CO; CAM5
The Roles of Sparse Direct Methods in Large-scale Simulations
International Nuclear Information System (INIS)
Li, Xiaoye S.; Gao, Weiguo; Husbands, Parry J.R.; Yang, Chao; Ng, Esmond G.
2005-01-01
Sparse systems of linear equations and eigen-equations arise at the heart of many large-scale, vital simulations in DOE. Examples include the Accelerator Science and Technology SciDAC (Omega3P code, electromagnetic problem), the Center for Extended Magnetohydrodynamic Modeling SciDAC(NIMROD and M3D-C1 codes, fusion plasma simulation). The Terascale Optimal PDE Simulations (TOPS)is providing high-performance sparse direct solvers, which have had significant impacts on these applications. Over the past several years, we have been working closely with the other SciDAC teams to solve their large, sparse matrix problems arising from discretization of the partial differential equations. Most of these systems are very ill-conditioned, resulting in extremely poor convergence deployed our direct methods techniques in these applications, which achieved significant scientific results as well as performance gains. These successes were made possible through the SciDAC model of computer scientists and application scientists working together to take full advantage of terascale computing systems and new algorithms research
The Roles of Sparse Direct Methods in Large-scale Simulations
Energy Technology Data Exchange (ETDEWEB)
Li, Xiaoye S.; Gao, Weiguo; Husbands, Parry J.R.; Yang, Chao; Ng, Esmond G.
2005-06-27
Sparse systems of linear equations and eigen-equations arise at the heart of many large-scale, vital simulations in DOE. Examples include the Accelerator Science and Technology SciDAC (Omega3P code, electromagnetic problem), the Center for Extended Magnetohydrodynamic Modeling SciDAC(NIMROD and M3D-C1 codes, fusion plasma simulation). The Terascale Optimal PDE Simulations (TOPS)is providing high-performance sparse direct solvers, which have had significant impacts on these applications. Over the past several years, we have been working closely with the other SciDAC teams to solve their large, sparse matrix problems arising from discretization of the partial differential equations. Most of these systems are very ill-conditioned, resulting in extremely poor convergence deployed our direct methods techniques in these applications, which achieved significant scientific results as well as performance gains. These successes were made possible through the SciDAC model of computer scientists and application scientists working together to take full advantage of terascale computing systems and new algorithms research.
Large-Eddy Simulation of turbulent vortex shedding
International Nuclear Information System (INIS)
Archambeau, F.
1995-06-01
This thesis documents the development and application of a computational algorithm for Large-Eddy Simulation. Unusually, the method adopts a fully collocated variable storage arrangement and is applicable to complex, non-rectilinear geometries. A Reynolds-averaged Navier-Stokes algorithm has formed the starting point of the development, but has been modified substantially: the spatial approximation of convection is effected by an energy-conserving central-differencing scheme; a second-order time-marching Adams-Bashforth scheme has been introduced; the pressure field is determined by solving the pressure-Poisson equation; this equation is solved either by use of preconditioned Conjugate-Gradient methods or with the Generalised Minimum Residual method; two types of sub-grid scale models have been introduced and examined. The algorithm has been validated by reference to a hierarchy of unsteady flows of increasing complexity starting with unsteady lid-driven cavity flows and ending with 3-D turbulent vortex shedding behind a square prism. In the latter case, for which extensive experimental data are available, special emphasis has been put on examining the dependence of the results on mesh density, near-wall treatment and the nature of the sub-grid-scale model, one of which is an advanced dynamic model. The LES scheme is shown to return time-average and phase-averaged results which agree well with experimental data and which support the view that LES is a promising approach for unsteady flows dominated by large periodic structures. (author)
Density-functional theory simulation of large quantum dots
Jiang, Hong; Baranger, Harold U.; Yang, Weitao
2003-10-01
Kohn-Sham spin-density functional theory provides an efficient and accurate model to study electron-electron interaction effects in quantum dots, but its application to large systems is a challenge. Here an efficient method for the simulation of quantum dots using density-function theory is developed; it includes the particle-in-the-box representation of the Kohn-Sham orbitals, an efficient conjugate-gradient method to directly minimize the total energy, a Fourier convolution approach for the calculation of the Hartree potential, and a simplified multigrid technique to accelerate the convergence. We test the methodology in a two-dimensional model system and show that numerical studies of large quantum dots with several hundred electrons become computationally affordable. In the noninteracting limit, the classical dynamics of the system we study can be continuously varied from integrable to fully chaotic. The qualitative difference in the noninteracting classical dynamics has an effect on the quantum properties of the interacting system: integrable classical dynamics leads to higher-spin states and a broader distribution of spacing between Coulomb blockade peaks.
Large Eddy Simulation Study for Fluid Disintegration and Mixing
Bellan, Josette; Taskinoglu, Ezgi
2011-01-01
A new modeling approach is based on the concept of large eddy simulation (LES) within which the large scales are computed and the small scales are modeled. The new approach is expected to retain the fidelity of the physics while also being computationally efficient. Typically, only models for the small-scale fluxes of momentum, species, and enthalpy are used to reintroduce in the simulation the physics lost because the computation only resolves the large scales. These models are called subgrid (SGS) models because they operate at a scale smaller than the LES grid. In a previous study of thermodynamically supercritical fluid disintegration and mixing, additional small-scale terms, one in the momentum and one in the energy conservation equations, were identified as requiring modeling. These additional terms were due to the tight coupling between dynamics and real-gas thermodynamics. It was inferred that if these terms would not be modeled, the high density-gradient magnitude regions, experimentally identified as a characteristic feature of these flows, would not be accurately predicted without the additional term in the momentum equation; these high density-gradient magnitude regions were experimentally shown to redistribute turbulence in the flow. And it was also inferred that without the additional term in the energy equation, the heat flux magnitude could not be accurately predicted; the heat flux to the wall of combustion devices is a crucial quantity that determined necessary wall material properties. The present work involves situations where only the term in the momentum equation is important. Without this additional term in the momentum equation, neither the SGS-flux constant-coefficient Smagorinsky model nor the SGS-flux constant-coefficient Gradient model could reproduce in LES the pressure field or the high density-gradient magnitude regions; the SGS-flux constant- coefficient Scale-Similarity model was the most successful in this endeavor although not
Large Eddy Simulation of Heat Entrainment Under Arctic Sea Ice
Ramudu, Eshwan; Gelderloos, Renske; Yang, Di; Meneveau, Charles; Gnanadesikan, Anand
2018-01-01
Arctic sea ice has declined rapidly in recent decades. The faster than projected retreat suggests that free-running large-scale climate models may not be accurately representing some key processes. The small-scale turbulent entrainment of heat from the mixed layer could be one such process. To better understand this mechanism, we model the Arctic Ocean's Canada Basin, which is characterized by a perennial anomalously warm Pacific Summer Water (PSW) layer residing at the base of the mixed layer and a summertime Near-Surface Temperature Maximum (NSTM) within the mixed layer trapping heat from solar radiation. We use large eddy simulation (LES) to investigate heat entrainment for different ice-drift velocities and different initial temperature profiles. The value of LES is that the resolved turbulent fluxes are greater than the subgrid-scale fluxes for most of our parameter space. The results show that the presence of the NSTM enhances heat entrainment from the mixed layer. Additionally there is no PSW heat entrained under the parameter space considered. We propose a scaling law for the ocean-to-ice heat flux which depends on the initial temperature anomaly in the NSTM layer and the ice-drift velocity. A case study of "The Great Arctic Cyclone of 2012" gives a turbulent heat flux from the mixed layer that is approximately 70% of the total ocean-to-ice heat flux estimated from the PIOMAS model often used for short-term predictions. Present results highlight the need for large-scale climate models to account for the NSTM layer.
Wu, Xingfu; Taylor, Valerie
2013-01-01
In this paper, we investigate the performance characteristics of five hybrid MPI/OpenMP scientific applications (two NAS Parallel benchmarks Multi-Zone SP-MZ and BT-MZ, an earthquake simulation PEQdyna, an aerospace application PMLB and a 3D particle-in-cell application GTC) on a large-scale multithreaded Blue Gene/Q supercomputer at Argonne National laboratory, and quantify the performance gap resulting from using different number of threads per node. We use performance tools and MPI profile and trace libraries available on the supercomputer to analyze and compare the performance of these hybrid scientific applications with increasing the number OpenMP threads per node, and find that increasing the number of threads to some extent saturates or worsens performance of these hybrid applications. For the strong-scaling hybrid scientific applications such as SP-MZ, BT-MZ, PEQdyna and PLMB, using 32 threads per node results in much better application efficiency than using 64 threads per node, and as increasing the number of threads per node, the FPU (Floating Point Unit) percentage decreases, and the MPI percentage (except PMLB) and IPC (Instructions per cycle) per core (except BT-MZ) increase. For the weak-scaling hybrid scientific application such as GTC, the performance trend (relative speedup) is very similar with increasing number of threads per node no matter how many nodes (32, 128, 512) are used. © 2013 IEEE.
Wu, Xingfu
2013-07-01
In this paper, we investigate the performance characteristics of five hybrid MPI/OpenMP scientific applications (two NAS Parallel benchmarks Multi-Zone SP-MZ and BT-MZ, an earthquake simulation PEQdyna, an aerospace application PMLB and a 3D particle-in-cell application GTC) on a large-scale multithreaded Blue Gene/Q supercomputer at Argonne National laboratory, and quantify the performance gap resulting from using different number of threads per node. We use performance tools and MPI profile and trace libraries available on the supercomputer to analyze and compare the performance of these hybrid scientific applications with increasing the number OpenMP threads per node, and find that increasing the number of threads to some extent saturates or worsens performance of these hybrid applications. For the strong-scaling hybrid scientific applications such as SP-MZ, BT-MZ, PEQdyna and PLMB, using 32 threads per node results in much better application efficiency than using 64 threads per node, and as increasing the number of threads per node, the FPU (Floating Point Unit) percentage decreases, and the MPI percentage (except PMLB) and IPC (Instructions per cycle) per core (except BT-MZ) increase. For the weak-scaling hybrid scientific application such as GTC, the performance trend (relative speedup) is very similar with increasing number of threads per node no matter how many nodes (32, 128, 512) are used. © 2013 IEEE.
Large-Eddy Simulation of turbulent vortex shedding
Energy Technology Data Exchange (ETDEWEB)
Archambeau, F
1995-06-01
This thesis documents the development and application of a computational algorithm for Large-Eddy Simulation. Unusually, the method adopts a fully collocated variable storage arrangement and is applicable to complex, non-rectilinear geometries. A Reynolds-averaged Navier-Stokes algorithm has formed the starting point of the development, but has been modified substantially: the spatial approximation of convection is effected by an energy-conserving central-differencing scheme; a second-order time-marching Adams-Bashforth scheme has been introduced; the pressure field is determined by solving the pressure-Poisson equation; this equation is solved either by use of preconditioned Conjugate-Gradient methods or with the Generalised Minimum Residual method; two types of sub-grid scale models have been introduced and examined. The algorithm has been validated by reference to a hierarchy of unsteady flows of increasing complexity starting with unsteady lid-driven cavity flows and ending with 3-D turbulent vortex shedding behind a square prism. In the latter case, for which extensive experimental data are available, special emphasis has been put on examining the dependence of the results on mesh density, near-wall treatment and the nature of the sub-grid-scale model, one of which is an advanced dynamic model. The LES scheme is shown to return time-average and phase-averaged results which agree well with experimental data and which support the view that LES is a promising approach for unsteady flows dominated by large periodic structures. (author) 87 refs.
Energy Technology Data Exchange (ETDEWEB)
Brown, D L
2009-05-01
Many complex systems of importance to the U.S. Department of Energy consist of networks of discrete components. Examples are cyber networks, such as the internet and local area networks over which nearly all DOE scientific, technical and administrative data must travel, the electric power grid, social networks whose behavior can drive energy demand, and biological networks such as genetic regulatory networks and metabolic networks. In spite of the importance of these complex networked systems to all aspects of DOE's operations, the scientific basis for understanding these systems lags seriously behind the strong foundations that exist for the 'physically-based' systems usually associated with DOE research programs that focus on such areas as climate modeling, fusion energy, high-energy and nuclear physics, nano-science, combustion, and astrophysics. DOE has a clear opportunity to develop a similarly strong scientific basis for understanding the structure and dynamics of networked systems by supporting a strong basic research program in this area. Such knowledge will provide a broad basis for, e.g., understanding and quantifying the efficacy of new security approaches for computer networks, improving the design of computer or communication networks to be more robust against failures or attacks, detecting potential catastrophic failure on the power grid and preventing or mitigating its effects, understanding how populations will respond to the availability of new energy sources or changes in energy policy, and detecting subtle vulnerabilities in large software systems to intentional attack. This white paper outlines plans for an aggressive new research program designed to accelerate the advancement of the scientific basis for complex networked systems of importance to the DOE. It will focus principally on four research areas: (1) understanding network structure, (2) understanding network dynamics, (3) predictive modeling and simulation for complex
International Nuclear Information System (INIS)
Brown, D.L.
2009-01-01
Many complex systems of importance to the U.S. Department of Energy consist of networks of discrete components. Examples are cyber networks, such as the internet and local area networks over which nearly all DOE scientific, technical and administrative data must travel, the electric power grid, social networks whose behavior can drive energy demand, and biological networks such as genetic regulatory networks and metabolic networks. In spite of the importance of these complex networked systems to all aspects of DOE's operations, the scientific basis for understanding these systems lags seriously behind the strong foundations that exist for the 'physically-based' systems usually associated with DOE research programs that focus on such areas as climate modeling, fusion energy, high-energy and nuclear physics, nano-science, combustion, and astrophysics. DOE has a clear opportunity to develop a similarly strong scientific basis for understanding the structure and dynamics of networked systems by supporting a strong basic research program in this area. Such knowledge will provide a broad basis for, e.g., understanding and quantifying the efficacy of new security approaches for computer networks, improving the design of computer or communication networks to be more robust against failures or attacks, detecting potential catastrophic failure on the power grid and preventing or mitigating its effects, understanding how populations will respond to the availability of new energy sources or changes in energy policy, and detecting subtle vulnerabilities in large software systems to intentional attack. This white paper outlines plans for an aggressive new research program designed to accelerate the advancement of the scientific basis for complex networked systems of importance to the DOE. It will focus principally on four research areas: (1) understanding network structure, (2) understanding network dynamics, (3) predictive modeling and simulation for complex networked systems
Simulation of fatigue crack growth under large scale yielding conditions
Schweizer, Christoph; Seifert, Thomas; Riedel, Hermann
2010-07-01
A simple mechanism based model for fatigue crack growth assumes a linear correlation between the cyclic crack-tip opening displacement (ΔCTOD) and the crack growth increment (da/dN). The objective of this work is to compare analytical estimates of ΔCTOD with results of numerical calculations under large scale yielding conditions and to verify the physical basis of the model by comparing the predicted and the measured evolution of the crack length in a 10%-chromium-steel. The material is described by a rate independent cyclic plasticity model with power-law hardening and Masing behavior. During the tension-going part of the cycle, nodes at the crack-tip are released such that the crack growth increment corresponds approximately to the crack-tip opening. The finite element analysis performed in ABAQUS is continued for so many cycles until a stabilized value of ΔCTOD is reached. The analytical model contains an interpolation formula for the J-integral, which is generalized to account for cyclic loading and crack closure. Both simulated and estimated ΔCTOD are reasonably consistent. The predicted crack length evolution is found to be in good agreement with the behavior of microcracks observed in a 10%-chromium steel.
Contextual Compression of Large-Scale Wind Turbine Array Simulations
Energy Technology Data Exchange (ETDEWEB)
Gruchalla, Kenny M [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Brunhart-Lupo, Nicholas J [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Potter, Kristin C [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Clyne, John [National Center for Atmospheric Research (NCAR)
2017-12-04
Data sizes are becoming a critical issue particularly for HPC applications. We have developed a user-driven lossy wavelet-based storage model to facilitate the analysis and visualization of large-scale wind turbine array simulations. The model stores data as heterogeneous blocks of wavelet coefficients, providing high-fidelity access to user-defined data regions believed the most salient, while providing lower-fidelity access to less salient regions on a block-by-block basis. In practice, by retaining the wavelet coefficients as a function of feature saliency, we have seen data reductions in excess of 94 percent, while retaining lossless information in the turbine-wake regions most critical to analysis and providing enough (low-fidelity) contextual information in the upper atmosphere to track incoming coherent turbulent structures. Our contextual wavelet compression approach has allowed us to deliver interative visual analysis while providing the user control over where data loss, and thus reduction in accuracy, in the analysis occurs. We argue this reduced but contextualized representation is a valid approach and encourages contextual data management.
Large-eddy simulations of unidirectional water flow over dunes
Grigoriadis, D. G. E.; Balaras, E.; Dimas, A. A.
2009-06-01
The unidirectional, subcritical flow over fixed dunes is studied numerically using large-eddy simulation, while the immersed boundary method is implemented to incorporate the bed geometry. Results are presented for a typical dune shape and two Reynolds numbers, Re = 17,500 and Re = 93,500, on the basis of bulk velocity and water depth. The numerical predictions of velocity statistics at the low Reynolds number are in very good agreement with available experimental data. A primary recirculation region develops downstream of the dune crest at both Reynolds numbers, while a secondary region develops at the toe of the dune crest only for the low Reynolds number. Downstream of the reattachment point, on the dune stoss, the turbulence intensity in the developing boundary layer is weaker than in comparable equilibrium boundary layers. Coherent vortical structures are identified using the fluctuating pressure field and the second invariant of the velocity gradient tensor. Vorticity is primarily generated at the dune crest in the form of spanwise "roller" structures. Roller structures dominate the flow dynamics near the crest, and are responsible for perturbing the boundary layer downstream of the reattachment point, which leads to the formation of "horseshoe" structures. Horseshoe structures dominate the near-wall dynamics after the reattachment point, do not rise to the free surface, and are distorted by the shear layer of the next crest. The occasional interaction between roller and horseshoe structures generates tube-like "kolk" structures, which rise to the free surface and persist for a long time before attenuating.
Enabling parallel simulation of large-scale HPC network systems
International Nuclear Information System (INIS)
Mubarak, Misbah; Carothers, Christopher D.; Ross, Robert B.; Carns, Philip
2016-01-01
Here, with the increasing complexity of today’s high-performance computing (HPC) architectures, simulation has become an indispensable tool for exploring the design space of HPC systems—in particular, networks. In order to make effective design decisions, simulations of these systems must possess the following properties: (1) have high accuracy and fidelity, (2) produce results in a timely manner, and (3) be able to analyze a broad range of network workloads. Most state-of-the-art HPC network simulation frameworks, however, are constrained in one or more of these areas. In this work, we present a simulation framework for modeling two important classes of networks used in today’s IBM and Cray supercomputers: torus and dragonfly networks. We use the Co-Design of Multi-layer Exascale Storage Architecture (CODES) simulation framework to simulate these network topologies at a flit-level detail using the Rensselaer Optimistic Simulation System (ROSS) for parallel discrete-event simulation. Our simulation framework meets all the requirements of a practical network simulation and can assist network designers in design space exploration. First, it uses validated and detailed flit-level network models to provide an accurate and high-fidelity network simulation. Second, instead of relying on serial time-stepped or traditional conservative discrete-event simulations that limit simulation scalability and efficiency, we use the optimistic event-scheduling capability of ROSS to achieve efficient and scalable HPC network simulations on today’s high-performance cluster systems. Third, our models give network designers a choice in simulating a broad range of network workloads, including HPC application workloads using detailed network traces, an ability that is rarely offered in parallel with high-fidelity network simulations
Simulating large-scale spiking neuronal networks with NEST
Schücker, Jannis; Eppler, Jochen Martin
2014-01-01
The Neural Simulation Tool NEST [1, www.nest-simulator.org] is the simulator for spiking neural networkmodels of the HBP that focuses on the dynamics, size and structure of neural systems rather than on theexact morphology of individual neurons. Its simulation kernel is written in C++ and it runs on computinghardware ranging from simple laptops to clusters and supercomputers with thousands of processor cores.The development of NEST is coordinated by the NEST Initiative [www.nest-initiative.or...
Energy Technology Data Exchange (ETDEWEB)
Murakami, Y.; Shi, B. [Geological Survey of Japan, Tsukuba (Japan); Matsushima, J. [The University of Tokyo, Tokyo (Japan). Faculty of Engineering
1997-05-27
Large deformation of the crust is generated by relatively large displacement of the mediums on both sides along a fault. In the conventional finite element method, faults are dealt with by special elements which are called joint elements, but joint elements, elements microscopic in width, generate numerical instability if large shear displacement is given. Therefore, by introducing the master slave (MO) method used for contact analysis in the metal processing field, developed was a large deformation simulator for analyzing diastrophism including large displacement along the fault. Analysis examples were shown in case the upper basement and lower basement were relatively dislocated with the fault as a boundary. The bottom surface and right end boundary of the lower basement are fixed boundaries. The left end boundary of the lower basement is fixed, and to the left end boundary of the upper basement, the horizontal speed, 3{times}10{sup -7}m/s, was given. In accordance with the horizontal movement of the upper basement, the boundary surface largely deformed. Stress is almost at right angles at the boundary surface. As to the analysis of faults by the MO method, it has been used for a single simple fault, but should be spread to lots of faults in the future. 13 refs., 2 figs.
Nuclear EMP simulation for large-scale urban environments. FDTD for electrically large problems.
Energy Technology Data Exchange (ETDEWEB)
Smith, William S. [Los Alamos National Laboratory; Bull, Jeffrey S. [Los Alamos National Laboratory; Wilcox, Trevor [Los Alamos National Laboratory; Bos, Randall J. [Los Alamos National Laboratory; Shao, Xuan-Min [Los Alamos National Laboratory; Goorley, John T. [Los Alamos National Laboratory; Costigan, Keeley R. [Los Alamos National Laboratory
2012-08-13
In case of a terrorist nuclear attack in a metropolitan area, EMP measurement could provide: (1) a prompt confirmation of the nature of the explosion (chemical or nuclear) for emergency response; and (2) and characterization parameters of the device (reaction history, yield) for technical forensics. However, urban environment could affect the fidelity of the prompt EMP measurement (as well as all other types of prompt measurement): (1) Nuclear EMP wavefront would no longer be coherent, due to incoherent production, attenuation, and propagation of gamma and electrons; and (2) EMP propagation from source region outward would undergo complicated transmission, reflection, and diffraction processes. EMP simulation for electrically-large urban environment: (1) Coupled MCNP/FDTD (Finite-difference time domain Maxwell solver) approach; and (2) FDTD tends to be limited to problems that are not 'too' large compared to the wavelengths of interest because of numerical dispersion and anisotropy. We use a higher-order low-dispersion, isotropic FDTD algorithm for EMP propagation.
A scientific operations plan for the large space telescope. [ground support system design
West, D. K.
1977-01-01
The paper describes an LST ground system which is compatible with the operational requirements of the LST. The goal of the approach is to minimize the cost of post launch operations without seriously compromising the quality and total throughput of LST science. Attention is given to cost constraints and guidelines, the telemetry operations processing systems (TELOPS), the image processing facility, ground system planning and data flow, and scientific interfaces.
The TeraShake Computational Platform for Large-Scale Earthquake Simulations
Cui, Yifeng; Olsen, Kim; Chourasia, Amit; Moore, Reagan; Maechling, Philip; Jordan, Thomas
Geoscientific and computer science researchers with the Southern California Earthquake Center (SCEC) are conducting a large-scale, physics-based, computationally demanding earthquake system science research program with the goal of developing predictive models of earthquake processes. The computational demands of this program continue to increase rapidly as these researchers seek to perform physics-based numerical simulations of earthquake processes for larger meet the needs of this research program, a multiple-institution team coordinated by SCEC has integrated several scientific codes into a numerical modeling-based research tool we call the TeraShake computational platform (TSCP). A central component in the TSCP is a highly scalable earthquake wave propagation simulation program called the TeraShake anelastic wave propagation (TS-AWP) code. In this chapter, we describe how we extended an existing, stand-alone, wellvalidated, finite-difference, anelastic wave propagation modeling code into the highly scalable and widely used TS-AWP and then integrated this code into the TeraShake computational platform that provides end-to-end (initialization to analysis) research capabilities. We also describe the techniques used to enhance the TS-AWP parallel performance on TeraGrid supercomputers, as well as the TeraShake simulations phases including input preparation, run time, data archive management, and visualization. As a result of our efforts to improve its parallel efficiency, the TS-AWP has now shown highly efficient strong scaling on over 40K processors on IBM’s BlueGene/L Watson computer. In addition, the TSCP has developed into a computational system that is useful to many members of the SCEC community for performing large-scale earthquake simulations.
Characteristics of Tornado-Like Vortices Simulated in a Large-Scale Ward-Type Simulator
Tang, Zhuo; Feng, Changda; Wu, Liang; Zuo, Delong; James, Darryl L.
2018-02-01
Tornado-like vortices are simulated in a large-scale Ward-type simulator to further advance the understanding of such flows, and to facilitate future studies of tornado wind loading on structures. Measurements of the velocity fields near the simulator floor and the resulting floor surface pressures are interpreted to reveal the mean and fluctuating characteristics of the flow as well as the characteristics of the static-pressure deficit. We focus on the manner in which the swirl ratio and the radial Reynolds number affect these characteristics. The transition of the tornado-like flow from a single-celled vortex to a dual-celled vortex with increasing swirl ratio and the impact of this transition on the flow field and the surface-pressure deficit are closely examined. The mean characteristics of the surface-pressure deficit caused by tornado-like vortices simulated at a number of swirl ratios compare well with the corresponding characteristics recorded during full-scale tornadoes.
Large-eddy simulation of unidirectional turbulent flow over dunes
Omidyeganeh, Mohammad
We performed large eddy simulation of the flow over a series of two- and three-dimensional dune geometries at laboratory scale using the Lagrangian dynamic eddy-viscosity subgrid-scale model. First, we studied the flow over a standard 2D transverse dune geometry, then bedform three-dimensionality was imposed. Finally, we investigated the turbulent flow over barchan dunes. The results are validated by comparison with simulations and experiments for the 2D dune case, while the results of the 3D dunes are validated qualitatively against experiments. The flow over transverse dunes separates at the dune crest, generating a shear layer that plays a crucial role in the transport of momentum and energy, as well as the generation of coherent structures. Spanwise vortices are generated in the separated shear; as they are advected, they undergo lateral instabilities and develop into horseshoe-like structures and finally reach the surface. The ejection that occurs between the legs of the vortex creates the upwelling and downdrafting events on the free surface known as "boils". The three-dimensional separation of flow at the crestline alters the distribution of wall pressure, which may cause secondary flow across the stream. The mean flow is characterized by a pair of counter-rotating streamwise vortices, with core radii of the order of the flow depth. Staggering the crestlines alters the secondary motion; two pairs of streamwise vortices appear (a strong one, centred about the lobe, and a weaker one, coming from the previous dune, centred around the saddle). The flow over barchan dunes presents significant differences to that over transverse dunes. The flow near the bed, upstream of the dune, diverges from the centerline plane; the flow close to the centerline plane separates at the crest and reattaches on the bed. Away from the centerline plane and along the horns, flow separation occurs intermittently. The flow in the separation bubble is routed towards the horns and leaves
Energy Technology Data Exchange (ETDEWEB)
Lartigue, G.
2004-11-15
The new european laws on pollutants emission impose more and more constraints to motorists. This is particularly true for gas turbines manufacturers, that must design motors operating with very fuel-lean mixtures. Doing so, pollutants formation is significantly reduced but the problem of combustion stability arises. Actually, combustion regimes that have a large excess of air are naturally more sensitive to combustion instabilities. Numerical predictions of these instabilities is thus a key issue for many industrial involved in energy production. This thesis work tries to show that recent numerical tools are now able to predict these combustion instabilities. Particularly, the Large Eddy Simulation method, when implemented in a compressible CFD code, is able to take into account the main processes involved in combustion instabilities, such as acoustics and flame/vortex interaction. This work describes a new formulation of a Large Eddy Simulation numerical code that enables to take into account very precisely thermodynamics and chemistry, that are essential in combustion phenomena. A validation of this work will be presented in a complex geometry (the PRECCINSTA burner). Our numerical results will be successfully compared with experimental data gathered at DLR Stuttgart (Germany). Moreover, a detailed analysis of the acoustics in this configuration will be presented, as well as its interaction with the combustion. For this acoustics analysis, another CERFACS code has been extensively used, the Helmholtz solver AVSP. (author)
A survey of modelling methods for high-fidelity wind farm simulations using large eddy simulation
DEFF Research Database (Denmark)
Breton, Simon-Philippe; Sumner, J.; Sørensen, Jens Nørkær
2017-01-01
surveys the most common schemes available to model the rotor, atmospheric conditions and terrain effects within current state-of-the-art LES codes, of which an overview is provided. A summary of the experimental research data available for validation of LES codes within the context of single and multiple......Large eddy simulations (LES) of wind farms have the capability to provide valuable and detailed information about the dynamics of wind turbine wakes. For this reason, their use within the wind energy research community is on the rise, spurring the development of new models and methods. This review...
Scientific Visualization and Simulation for Multi-dimensional Marine Environment Data
Su, T.; Liu, H.; Wang, W.; Song, Z.; Jia, Z.
2017-12-01
As higher attention on the ocean and rapid development of marine detection, there are increasingly demands for realistic simulation and interactive visualization of marine environment in real time. Based on advanced technology such as GPU rendering, CUDA parallel computing and rapid grid oriented strategy, a series of efficient and high-quality visualization methods, which can deal with large-scale and multi-dimensional marine data in different environmental circumstances, has been proposed in this paper. Firstly, a high-quality seawater simulation is realized by FFT algorithm, bump mapping and texture animation technology. Secondly, large-scale multi-dimensional marine hydrological environmental data is virtualized by 3d interactive technologies and volume rendering techniques. Thirdly, seabed terrain data is simulated with improved Delaunay algorithm, surface reconstruction algorithm, dynamic LOD algorithm and GPU programming techniques. Fourthly, seamless modelling in real time for both ocean and land based on digital globe is achieved by the WebGL technique to meet the requirement of web-based application. The experiments suggest that these methods can not only have a satisfying marine environment simulation effect, but also meet the rendering requirements of global multi-dimension marine data. Additionally, a simulation system for underwater oil spill is established by OSG 3D-rendering engine. It is integrated with the marine visualization method mentioned above, which shows movement processes, physical parameters, current velocity and direction for different types of deep water oil spill particle (oil spill particles, hydrates particles, gas particles, etc.) dynamically and simultaneously in multi-dimension. With such application, valuable reference and decision-making information can be provided for understanding the progress of oil spill in deep water, which is helpful for ocean disaster forecasting, warning and emergency response.
International Nuclear Information System (INIS)
Satake, Shin-ichi; Kunugi, Tomoaki
2006-01-01
Scientific computational methods have advanced remarkably with the progress of nuclear development. They have played the role of weft connecting each realm of nuclear engineering and then an introductory course of advanced scientific computational methods and their applications to nuclear technologies were prepared in serial form. This is the third issue showing the introduction of continuum simulation methods and their applications. Spectral methods and multi-interface calculation methods in fluid dynamics are reviewed. (T. Tanaka)
Paul Scherrer Institute Scientific Report 1998. Volume VI: Large Research Facilities
International Nuclear Information System (INIS)
Bauer, Guenter; Bercher, Renate; Buechli, Carmen; Foroughi, Fereydoun; Meyer, Rosa
1999-01-01
The department GFA (Grossforschungsanlagen, Large Research Facilities) has been established in October 1998 and its main duty is operation, maintenance and development of the PSI accelerators, the spallation neutron source and the beam transport systems for pions and muons. A large effort of this group concerns the planning and co-ordination of the assembly of the Swiss Light Source (SLS). (author)
Taylor, K.A.; Short, A.
2009-01-01
Integrating science into resource management activities is a goal of the CALFED Bay-Delta Program, a multi-agency effort to address water supply reliability, ecological condition, drinking water quality, and levees in the Sacramento-San Joaquin Delta of northern California. Under CALFED, many different strategies were used to integrate science, including interaction between the research and management communities, public dialogues about scientific work, and peer review. This paper explores ways science was (and was not) integrated into CALFED's management actions and decision systems through three narratives describing different patterns of scientific integration and application in CALFED. Though a collaborative process and certain organizational conditions may be necessary for developing new understandings of the system of interest, we find that those factors are not sufficient for translating that knowledge into management actions and decision systems. We suggest that the application of knowledge may be facilitated or hindered by (1) differences in the objectives, approaches, and cultures of scientists operating in the research community and those operating in the management community and (2) other factors external to the collaborative process and organization.
1st ERCOFTAC Workshop on Direct and Large-Eddy Simulation
Kleiser, Leonhard; Chollet, Jean-Pierre
1994-01-01
It is a truism that turbulence is an unsolved problem, whether in scientific, engin eering or geophysical terms. It is strange that this remains largely the case even though we now know how to solve directly, with the help of sufficiently large and powerful computers, accurate approximations to the equations that govern tur bulent flows. The problem lies not with our numerical approximations but with the size of the computational task and the complexity of the solutions we gen erate, which match the complexity of real turbulence precisely in so far as the computations mimic the real flows. The fact that we can now solve some turbu lence in this limited sense is nevertheless an enormous step towards the goal of full understanding. Direct and large-eddy simulations are these numerical solutions of turbulence. They reproduce with remarkable fidelity the statistical, structural and dynamical properties of physical turbulent and transitional flows, though since the simula tions are necessarily time-depen...
Paul Scherrer Institute Scientific Report 1998. Volume VI: Large Research Facilities
Energy Technology Data Exchange (ETDEWEB)
Bauer, Guenter; Bercher, Renate; Buechli, Carmen; Foroughi, Fereydoun; Meyer, Rosa [eds.
1999-09-01
The department GFA (Grossforschungsanlagen, Large Research Facilities) has been established in October 1998 and its main duty is operation, maintenance and development of the PSI accelerators, the spallation neutron source and the beam transport systems for pions and muons. A large effort of this group concerns the planning and co-ordination of the assembly of the Swiss Light Source (SLS). (author) figs., tabs., refs.
Gleam: the GLAST Large Area Telescope Simulation Framework
Boinee, P; De Angelis, Alessandro; Favretto, Dario; Frailis, Marco; Giannitrapani, Riccardo; Milotti, Edoardo; Longo, Francesco; Brigida, Monica; Gargano, Fabio; Giglietto, Nicola; Loparco, Francesco; Mazziotta, Mario Nicola; Cecchi, Claudia; Lubrano, Pasquale; Pepe, Monica; Baldini, Luca; Cohen-Tanugi, Johann; Kuss, Michael; Latronico, Luca; Omodei, Nicola; Spandre, Gloria; Bogart, Joanne R.; Dubois, Richard; Kamae, Tune; Rochester, Leon; Usher, Tracy; Burnett, Thompson H.; Robinson, Sean M.; Bastieri, Denis; Rando, Riccardo
2003-01-01
This paper presents the simulation of the GLAST high energy gamma-ray telescope. The simulation package, written in C++, is based on the Geant4 toolkit, and it is integrated into a general framework used to process events. A detailed simulation of the electronic signals inside Silicon detectors has been provided and it is used for the particle tracking, which is handled by a dedicated software. A unique repository for the geometrical description of the detector has been realized using the XML language and a C++ library to access this information has been designed and implemented.
Directory of Open Access Journals (Sweden)
Nikhil Kalathil
2018-01-01
Full Text Available When assessing the importance of materials (or other components to a given set of applications, machine analysis of a very large corpus of scientific abstracts can provide an analyst a base of insights to develop further. The use of text analytics reduces the time required to conduct an evaluation, while allowing analysts to experiment with a multitude of different hypotheses. Because the scope and quantity of metadata analyzed can, and should, be large, any divergence from what a human analyst determines and what the text analysis shows provides a prompt for the human analyst to reassess any preliminary findings. In this work, we have successfully extracted material–application pairs and ranked them on their importance. This method provides a novel way to map scientific advances in a particular material to the application for which it is used. Approximately 438,000 titles and abstracts of scientific papers published from 1992 to 2011 were used to examine 16 materials. This analysis used coclustering text analysis to associate individual materials with specific clean energy applications, evaluate the importance of materials to specific applications, and assess their importance to clean energy overall. Our analysis reproduced the judgments of experts in assigning material importance to applications. The validated methods were then used to map the replacement of one material with another material in a specific application (batteries.
Parallel Earthquake Simulations on Large-Scale Multicore Supercomputers
Wu, Xingfu; Duan, Benchun; Taylor, Valerie
2011-01-01
, such as California and Japan, scientists have been using numerical simulations to study earthquake rupture propagation along faults and seismic wave propagation in the surrounding media on ever-advancing modern computers over past several decades. In particular
Energy Technology Data Exchange (ETDEWEB)
Ripeanu, Matei [University of British Columbia, Vancouver; Al-Kiswany, Samer [University of British Columbia, Vancouver; Iamnitchi, Adriana [University of South Florida, Tampa; Vazhkudai, Sudharshan S [ORNL
2009-03-01
The avalanche of data from scientific instruments and the ensuing interest from geographically distributed users to analyze and interpret it accentuates the need for efficient data dissemination. A suitable data distribution scheme will find the delicate balance between conflicting requirements of minimizing transfer times, minimizing the impact on the network, and uniformly distributing load among participants. We identify several data distribution techniques, some successfully employed by today's peer-to-peer networks: staging, data partitioning, orthogonal bandwidth exploitation, and combinations of the above. We use simulations to explore the performance of these techniques in contexts similar to those used by today's data-centric scientific collaborations and derive several recommendations for efficient data dissemination. Our experimental results show that the peer-to-peer solutions that offer load balancing and good fault tolerance properties and have embedded participation incentives lead to unjustified costs in today's scientific data collaborations deployed on over-provisioned network cores. However, as user communities grow and these deployments scale, peer-to-peer data delivery mechanisms will likely outperform other techniques.
Real time simulation of large systems on mini-computer
International Nuclear Information System (INIS)
Nakhle, Michel; Roux, Pierre.
1979-01-01
Most simulation languages will only accept an explicit formulation of differential equations, and logical variables hold no special status therein. The pace of the suggested methods of integration is limited by the smallest time constant of the model submitted. The NEPTUNIX 2 simulation software has a language that will take implicit equations and an integration method of which the variable pace is not limited by the time constants of the model. This, together with high time and memory ressources optimization of the code generated, makes NEPTUNIX 2 a basic tool for simulation on mini-computers. Since the logical variables are specific entities under centralized control, correct processing of discontinuities and synchronization with a real process are feasible. The NEPTUNIX 2 is the industrial version of NEPTUNIX 1 [fr
Manufacturing Process Simulation of Large-Scale Cryotanks
Babai, Majid; Phillips, Steven; Griffin, Brian
2003-01-01
NASA's Space Launch Initiative (SLI) is an effort to research and develop the technologies needed to build a second-generation reusable launch vehicle. It is required that this new launch vehicle be 100 times safer and 10 times cheaper to operate than current launch vehicles. Part of the SLI includes the development of reusable composite and metallic cryotanks. The size of these reusable tanks is far greater than anything ever developed and exceeds the design limits of current manufacturing tools. Several design and manufacturing approaches have been formulated, but many factors must be weighed during the selection process. Among these factors are tooling reachability, cycle times, feasibility, and facility impacts. The manufacturing process simulation capabilities available at NASA.s Marshall Space Flight Center have played a key role in down selecting between the various manufacturing approaches. By creating 3-D manufacturing process simulations, the varying approaches can be analyzed in a virtual world before any hardware or infrastructure is built. This analysis can detect and eliminate costly flaws in the various manufacturing approaches. The simulations check for collisions between devices, verify that design limits on joints are not exceeded, and provide cycle times which aide in the development of an optimized process flow. In addition, new ideas and concerns are often raised after seeing the visual representation of a manufacturing process flow. The output of the manufacturing process simulations allows for cost and safety comparisons to be performed between the various manufacturing approaches. This output helps determine which manufacturing process options reach the safety and cost goals of the SLI. As part of the SLI, The Boeing Company was awarded a basic period contract to research and propose options for both a metallic and a composite cryotank. Boeing then entered into a task agreement with the Marshall Space Flight Center to provide manufacturing
Autiero, D; Badertscher, A; Bezrukov, L; Bouchez, J; Bueno, A; Busto, J; Campagne, J -E; Cavata, C; De Bellefon, A; Dumarchez, J; Ebert, J; Enqvist, T; Ereditato, A; Von Feilitzsch, F; Perez, P Fileviez; Goger-Neff, M; Gninenko, S; Gruber, W; Hagner, C; Hess, M; Hochmuth, K A; Kisiel, J; Knecht, L; Kreslo, I; Kudryavtsev, V A; Kuusiniemi, P; Lachenmaier, T; Laffranchi, M; Lefièvre, B; Lightfoot, P K; Lindner, M; Maalampi, J; Maltoni, M; Marchionni, A; Undagoitia, T Marrodan; Meregaglia, A; Messina, M; Mezzetto, M; Mirizzi, A; Mosca, L; Moser, U; Müller, A; Natterer, G; Oberauer, L; Otiougova, P; Patzak, T; Peltoniemi, J; Potzel, W; Pistillo, C; Raffelt, G G; Rondio, E; Roos, M; Rossi, B; Rubbia, André; Savvinov, N; Schwetz, T; Sobczyk, J; Spooner, N J C; Stefan, D; Tonazzo, A; Trzaska, W; Ulbricht, J; Volpe, C; Winter, J; Wurm, M; Zalewska-Bak, A; Zimmermann, R
2007-01-01
This document reports on a series of experimental and theoretical studies conducted to assess the astro-particle physics potential of three future large-scale particle detectors proposed in Europe as next generation underground observatories. The proposed apparatus employ three different and, to some extent, complementary detection techniques: GLACIER (liquid Argon TPC), LENA (liquid scintillator) and MEMPHYS (\\WC), based on the use of large mass of liquids as active detection media. The results of these studies are presented along with a critical discussion of the performance attainable by the three proposed approaches coupled to existing or planned underground laboratories, in relation to open and outstanding physics issues such as the search for matter instability, the detection of astrophysical- and geo-neutrinos and to the possible use of these detectors in future high-intensity neutrino beams.
Performance of large-scale scientific applications on the IBM ASCI Blue-Pacific system
International Nuclear Information System (INIS)
Mirin, A.
1998-01-01
The IBM ASCI Blue-Pacific System is a scalable, distributed/shared memory architecture designed to reach multi-teraflop performance. The IBM SP pieces together a large number of nodes, each having a modest number of processors. The system is designed to accommodate a mixed programming model as well as a pure message-passing paradigm. We examine a number of applications on this architecture and evaluate their performance and scalability
Paul Scherrer Institute Scientific and Technical Report 2001. Volume VI: Large Research Facilities
Energy Technology Data Exchange (ETDEWEB)
Bercher, R.; Buechli, C.; Zumkeller, L. (eds.)
2002-03-01
While the main effort in the past ten years was directed towards increasing the beam current from 100 to 2000 {mu}A and installation of additional user facilities like SINQ to satisfy new needs, we are now concentrating on stable operation at these high beam intensities. Unfortunately, 'stable operation' is not clearly defined. A few years ago, the accelerator physicists considered 80% beam on time excellent but the users complained about poor performance. Today we achieve a yearly mean beam on time of almost 90% at 1.7 mA and we have achieved 95% to 98% of the scheduled beam time for periods of weeks. These numbers seem to be satisfactory for the users. Despite this achievement, we try hard to further reduce the number of serious and long breakdowns, which are the main cause of the reduced yearly mean availability. Furthermore, breakdowns that necessitate long repair times are extremely detrimental for many experiments, which have only been allocated a few days of beam time. As a result of our discussions, we launched a number of activities, which include design and construction of improved power supplies, intensifying preventive maintenance, procuring vital spare parts, and reducing repair times through careful preparation. In addition, we were given permission to strengthen the accelerator staff with highly qualified physicists in order to study and solve several pending problems. We are aware that the planned measures will by no means be fast and will require considerable financial and personnel support. A long-standing issue concerning the improvement of the machine performance is the replacement of the aluminum cavities in the main ring accelerator by new high power copper cavities. The studies and tests on a model cavity are finished and we have ordered a prototype cavity, which will arrive in fall 2002 and be installed in the ring after a rigorous test phase in 2004. A list of scientific publications in 2000 is also provided.
Paul Scherrer Institute Scientific and Technical Report 2001. Volume VI: Large Research Facilities
International Nuclear Information System (INIS)
Bercher, R.; Buechli, C.; Zumkeller, L.
2002-03-01
While the main effort in the past ten years was directed towards increasing the beam current from 100 to 2000 μA and installation of additional user facilities like SINQ to satisfy new needs, we are now concentrating on stable operation at these high beam intensities. Unfortunately, 'stable operation' is not clearly defined. A few years ago, the accelerator physicists considered 80% beam on time excellent but the users complained about poor performance. Today we achieve a yearly mean beam on time of almost 90% at 1.7 mA and we have achieved 95% to 98% of the scheduled beam time for periods of weeks. These numbers seem to be satisfactory for the users. Despite this achievement, we try hard to further reduce the number of serious and long breakdowns, which are the main cause of the reduced yearly mean availability. Furthermore, breakdowns that necessitate long repair times are extremely detrimental for many experiments, which have only been allocated a few days of beam time. As a result of our discussions, we launched a number of activities, which include design and construction of improved power supplies, intensifying preventive maintenance, procuring vital spare parts, and reducing repair times through careful preparation. In addition, we were given permission to strengthen the accelerator staff with highly qualified physicists in order to study and solve several pending problems. We are aware that the planned measures will by no means be fast and will require considerable financial and personnel support. A long-standing issue concerning the improvement of the machine performance is the replacement of the aluminum cavities in the main ring accelerator by new high power copper cavities. The studies and tests on a model cavity are finished and we have ordered a prototype cavity, which will arrive in fall 2002 and be installed in the ring after a rigorous test phase in 2004. A list of scientific publications in 2000 is also provided
Large eddy simulation of turbulent mixing in a T-junction
International Nuclear Information System (INIS)
Kim, Jung Woo
2010-12-01
In this report, large eddy simulation was performed in order to further improve our understanding the physics of turbulent mixing in a T-junction, which is recently regarded as one of the most important problems in nuclear thermal-hydraulics safety. Large eddy simulation technique and the other numerical methods used in this study were presented in Sec. 2, and the numerical results obtained from large eddy simulation were described in Sec. 3. Finally, the summary was written in Sec. 4
Chang, Hsin-Yi; Hsu, Ying-Shao; Wu, Hsin-Kai
2016-01-01
We investigated the impact of an augmented reality (AR) versus interactive simulation (IS) activity incorporated in a computer learning environment to facilitate students' learning of a socio-scientific issue (SSI) on nuclear power plants and radiation pollution. We employed a quasi-experimental research design. Two classes (a total of 45…
Energy Technology Data Exchange (ETDEWEB)
Corren, Dean [Verdant Power, Inc.; Colby, Jonathan [Verdant Power, Inc.; Adonizio, Mary Ann [Verdant Power, Inc.
2013-01-29
Verdant Power, Inc, working in partnership with the National Renewable Energy Laboratory (NREL), Sandia National Laboratories (SNL), and the University of Minnesota St. Anthony Falls Laboratory (SAFL), among other partners, used evolving Computational Fluid Dynamics (CFD) and Finite Element Analysis (FEA) models and techniques to improve the structure and fabrication of large, high-power composite Kinetic Hydropower System (KHPS) rotor blades. The objectives of the project were to: design; analyze; develop for manufacture and fabricate; and thoroughly test, in the lab and at full scale in the water, the improved KHPS rotor blade.
Modeling and Simulation Techniques for Large-Scale Communications Modeling
National Research Council Canada - National Science Library
Webb, Steve
1997-01-01
.... Tests of random number generators were also developed and applied to CECOM models. It was found that synchronization of random number strings in simulations is easy to implement and can provide significant savings for making comparative studies. If synchronization is in place, then statistical experiment design can be used to provide information on the sensitivity of the output to input parameters. The report concludes with recommendations and an implementation plan.
Entropic Lattice Boltzmann: an implicit Large-Eddy Simulation?
Tauzin, Guillaume; Biferale, Luca; Sbragaglia, Mauro; Gupta, Abhineet; Toschi, Federico; Ehrhardt, Matthias; Bartel, Andreas
2017-11-01
We study the modeling of turbulence implied by the unconditionally stable Entropic Lattice Boltzmann Method (ELBM). We first focus on 2D homogeneous turbulence, for which we conduct numerical simulations for a wide range of relaxation times τ. For these simulations, we analyze the effective viscosity obtained by numerically differentiating the kinetic energy and enstrophy balance equations averaged over sub-domains of the computational grid. We aim at understanding the behavior of the implied sub-grid scale model and verify a formulation previously derived using Chapman-Enskog expansion. These ELBM benchmark simulations are thus useful to understand the range of validity of ELBM as a turbulence model. Finally, we will discuss an extension of the previously obtained results to the 3D case. Supported by the European Unions Framework Programme for Research and Innovation Horizon 2020 (2014-2020) under the Marie Sklodowska-Curie Grant Agreement No. 642069 and by the European Research Council under the ERC Grant Agreement No. 339032.
Parallel Earthquake Simulations on Large-Scale Multicore Supercomputers
Wu, Xingfu
2011-01-01
Earthquakes are one of the most destructive natural hazards on our planet Earth. Hugh earthquakes striking offshore may cause devastating tsunamis, as evidenced by the 11 March 2011 Japan (moment magnitude Mw9.0) and the 26 December 2004 Sumatra (Mw9.1) earthquakes. Earthquake prediction (in terms of the precise time, place, and magnitude of a coming earthquake) is arguably unfeasible in the foreseeable future. To mitigate seismic hazards from future earthquakes in earthquake-prone areas, such as California and Japan, scientists have been using numerical simulations to study earthquake rupture propagation along faults and seismic wave propagation in the surrounding media on ever-advancing modern computers over past several decades. In particular, ground motion simulations for past and future (possible) significant earthquakes have been performed to understand factors that affect ground shaking in populated areas, and to provide ground shaking characteristics and synthetic seismograms for emergency preparation and design of earthquake-resistant structures. These simulation results can guide the development of more rational seismic provisions for leading to safer, more efficient, and economical50pt]Please provide V. Taylor author e-mail ID. structures in earthquake-prone regions.
Statistics of LES simulations of large wind farms
DEFF Research Database (Denmark)
Andersen, Søren Juhl; Sørensen, Jens Nørkær; Mikkelsen, Robert Flemming
2016-01-01
. The statistical moments appear to collapse and hence the turbulence inside large wind farms can potentially be scaled accordingly. The thrust coefficient is estimated by two different reference velocities and the generic CT expression by Frandsen. A reference velocity derived from the power production is shown...... to give very good agreement and furthermore enables the very good estimation of the thrust force using only the steady CT-curve, even for very short time samples. Finally, the effective turbulence inside large wind farms and the equivalent loads are examined....
Wu, Xingfu
2011-08-01
In this paper, we present a performance modeling framework based on memory bandwidth contention time and a parameterized communication model to predict the performance of OpenMP, MPI and hybrid applications with weak scaling on three large-scale multicore clusters: IBM POWER4, POWER5+ and Blue Gene/P, and analyze the performance of these MPI, OpenMP and hybrid applications. We use STREAM memory benchmarks to provide initial performance analysis and model validation of MPI and OpenMP applications on these multicore clusters because the measured sustained memory bandwidth can provide insight into the memory bandwidth that a system should sustain on scientific applications with the same amount of workload per core. In addition to using these benchmarks, we also use a weak-scaling hybrid MPI/OpenMP large-scale scientific application: Gyro kinetic Toroidal Code in magnetic fusion to validate our performance model of the hybrid application on these multicore clusters. The validation results for our performance modeling method show less than 7.77% error rate in predicting the performance of hybrid MPI/OpenMP GTC on up to 512 cores on these multicore clusters. © 2011 IEEE.
Wu, Xingfu; Taylor, Valerie
2011-01-01
In this paper, we present a performance modeling framework based on memory bandwidth contention time and a parameterized communication model to predict the performance of OpenMP, MPI and hybrid applications with weak scaling on three large-scale multicore clusters: IBM POWER4, POWER5+ and Blue Gene/P, and analyze the performance of these MPI, OpenMP and hybrid applications. We use STREAM memory benchmarks to provide initial performance analysis and model validation of MPI and OpenMP applications on these multicore clusters because the measured sustained memory bandwidth can provide insight into the memory bandwidth that a system should sustain on scientific applications with the same amount of workload per core. In addition to using these benchmarks, we also use a weak-scaling hybrid MPI/OpenMP large-scale scientific application: Gyro kinetic Toroidal Code in magnetic fusion to validate our performance model of the hybrid application on these multicore clusters. The validation results for our performance modeling method show less than 7.77% error rate in predicting the performance of hybrid MPI/OpenMP GTC on up to 512 cores on these multicore clusters. © 2011 IEEE.
Charlotte: Scientific Modeling and Simulation Under the Software as a Service Paradigm, Phase I
National Aeronautics and Space Administration — NASA spends considerable effort supporting the efforts of collaborating researchers. These researchers are interested in interacting with scientific models provided...
Page, Roderic D M
2011-05-23
The Biodiversity Heritage Library (BHL) is a large digital archive of legacy biological literature, comprising over 31 million pages scanned from books, monographs, and journals. During the digitisation process basic metadata about the scanned items is recorded, but not article-level metadata. Given that the article is the standard unit of citation, this makes it difficult to locate cited literature in BHL. Adding the ability to easily find articles in BHL would greatly enhance the value of the archive. A service was developed to locate articles in BHL based on matching article metadata to BHL metadata using approximate string matching, regular expressions, and string alignment. This article locating service is exposed as a standard OpenURL resolver on the BioStor web site http://biostor.org/openurl/. This resolver can be used on the web, or called by bibliographic tools that support OpenURL. BioStor provides tools for extracting, annotating, and visualising articles from the Biodiversity Heritage Library. BioStor is available from http://biostor.org/.
Ovcharenko, Aleksandr
2012-03-01
This paper introduces a general-purpose communication package built on top of MPI which is aimed at improving inter-processor communications independently of the supercomputer architecture being considered. The package is developed to support parallel applications that rely on computation characterized by large number of messages of various sizes, often small, that are focused within processor neighborhoods. In some cases, such as solvers having static mesh partitions, the number and size of messages are known a priori. However, in other cases such as mesh adaptation, the messages evolve and vary in number and size and include the dynamic movement of partition objects. The current package provides a utility for dynamic applications based on two key attributes that are: (i) explicit consideration of the neighborhood communication pattern to avoid many-to-many calls and also to reduce the number of collective calls to a minimum, and (ii) use of non-blocking MPI functions along with message packing to manage message flow control and reduce the number and time of communication calls. The test application demonstrated is parallel unstructured mesh adaptation. Results on IBM Blue Gene/P and Cray XE6 computers show that the use of neighborhood-based communication control leads to scalable results when executing generally imbalanced mesh adaptation runs. © 2011 Elsevier B.V. All rights reserved.
Large Blast and Thermal Simulator Reflected Wave Eliminator Study
1990-03-01
it delays the passage of this wave through the test section until after the test is complete. The required length of extra duct depends on the strength...tube axis, which acts like an additional contraction effect since Se = Sj/[Cqsin(aj)]. Tii extra area is illustrated best by plotting (Se-Ae)/Ac versus...34Simulation de Choc et de Soaffie. Comimpensateur d’Ondes de Detente de Bouche pour tube a Choc de 2400 mm de diametre de Veine. Description, Compte- Renda
Large-Eddy Simulations of Reacting Liquid Spray
Lederlin, Thomas; Sanjose, Marlene; Gicquel, Laurent; Cuenot, Benedicte; Pitsch, Heinz; Poinsot, Thierry
2008-11-01
Numerical simulation, which is commonly used in many stages of aero-engine design, still has to demonstrate its predictive capability for two-phase reacting flows. This study is a collaboration between Stanford University and CERFACS to perform LES of a realistic spray combustor installed at ONERA, Toulouse. The experimental configuration is computed on the same unstructured mesh with two different solvers: Stanford's CDP code and CERFACS's AVBP code. CDP uses a low-Mach, variable-density solver with implicit time advancement. Droplets are tracked in a Lagrangian point-particle framework. The combustion model uses a flamelet approach, based on two transported scalars, mixture fraction and reaction progress variable. AVBP is a fully compressible solver with explicit time advancement. The liquid phase is described with an Eulerian method. The flame-turbulence interaction is modeled using a dynamically-thickened flame. Results are compared with experimental data for three regimes: purely gaseous non-reacting flow, non-reacting flow with evaporating droplets, reacting flow with droplets. Both simulations show a good agreement with experimental data and also stress the difference and relative advantages of the numerical methods.
Energy Technology Data Exchange (ETDEWEB)
Priere, C
2005-01-15
Nowadays, environmental and economic constraints require considerable research efforts from the gas turbine industry. Objectives aim at lowering pollutants emissions and fuel consumption. These efforts take a primary importance to satisfy a continue growth of energy production and to obey to stringent environmental legislations. Recorded progresses are linked to mixing enhancement in combustors running at lean premixed operating point. Indeed, industry shows itself to be attentive in the mixing enhancement and during the last years, efforts are concentrated on fresh and burned gas dilution. The Jet In Cross Flow (JICF), which constitutes a representative case to further the research effort. It has been to be widely studied both in experimentally and numerically, and is particularly well suited for the evaluation of Large Eddy Simulations (LES). This approach, where large scale phenomena are naturally taken into account in the governing equation while the small scales are modelled, offers the means to well-predict such flows. The main objective of this work is to gauge and to enhance the quality of the LES predictions in JICF configurations by means of numerical tools developed in the compressible AVBP code. Physical and numerical parameters considered in the JICF modelization are taken into account and strategies that are able to enhance quality of LES results are proposed. Configurations studied in this work are the following: - Influences of the boundary conditions and jet injection system on a free JICF - Study of static mixing device in an industrial gas turbine chamber. - Study of a JICF configuration represented a dilution zone in low emissions combustors. (author)
DEFF Research Database (Denmark)
Yang, Yang; Kær, Søren Knudsen
2012-01-01
The flow structure of one isothermal swirling case in the Sydney swirl flame database was studied using two numerical methods. Results from the Reynolds-averaged Navier-Stokes (RANS) approach and large eddy simulation (LES) were compared with experimental measurements. The simulations were applied...
Large Eddy Simulation of stratified flows over structures
Brechler J.; Fuka V.
2013-01-01
We tested the ability of the LES model CLMM (Charles University Large-Eddy Microscale Model) to model the stratified flow around three dimensional hills. We compared the quantities, as the height of the dividing streamline, recirculation zone length or length of the lee waves with experiments by Hunt and Snyder[3] and numerical computations by Ding, Calhoun and Street[5]. The results mostly agreed with the references, but some important differences are present.
Large Eddy Simulation of stratified flows over structures
Directory of Open Access Journals (Sweden)
Brechler J.
2013-04-01
Full Text Available We tested the ability of the LES model CLMM (Charles University Large-Eddy Microscale Model to model the stratified flow around three dimensional hills. We compared the quantities, as the height of the dividing streamline, recirculation zone length or length of the lee waves with experiments by Hunt and Snyder[3] and numerical computations by Ding, Calhoun and Street[5]. The results mostly agreed with the references, but some important differences are present.
Large Eddy Simulation of stratified flows over structures
Fuka, V.; Brechler, J.
2013-04-01
We tested the ability of the LES model CLMM (Charles University Large-Eddy Microscale Model) to model the stratified flow around three dimensional hills. We compared the quantities, as the height of the dividing streamline, recirculation zone length or length of the lee waves with experiments by Hunt and Snyder[3] and numerical computations by Ding, Calhoun and Street[5]. The results mostly agreed with the references, but some important differences are present.
Large-eddy simulation of the temporal mixing layer using the Clark model
Vreman, A.W.; Geurts, B.J.; Kuerten, J.G.M.
1996-01-01
The Clark model for the turbulent stress tensor in large-eddy simulation is investigated from a theoretical and computational point of view. In order to be applicable to compressible turbulent flows, the Clark model has been reformulated. Actual large-eddy simulation of a weakly compressible,
LaWen Hollingsworth; James Menakis
2010-01-01
This project mapped wildland fire potential (WFP) for the conterminous United States by using the large fire simulation system developed for Fire Program Analysis (FPA) System. The large fire simulation system, referred to here as LFSim, consists of modules for weather generation, fire occurrence, fire suppression, and fire growth modeling. Weather was generated with...
Fast Simulation of Large-Scale Floods Based on GPU Parallel Computing
Qiang Liu; Yi Qin; Guodong Li
2018-01-01
Computing speed is a significant issue of large-scale flood simulations for real-time response to disaster prevention and mitigation. Even today, most of the large-scale flood simulations are generally run on supercomputers due to the massive amounts of data and computations necessary. In this work, a two-dimensional shallow water model based on an unstructured Godunov-type finite volume scheme was proposed for flood simulation. To realize a fast simulation of large-scale floods on a personal...
Mud pressure simulation on large horizontal directional drilling
Energy Technology Data Exchange (ETDEWEB)
Placido, Rafael R.; Avesani Neto, Jose O.; Martins, Pedro R.R.; Rocha, Ronaldo [Instituto de Pesquisas Tecnologicas do Estado de Sao Paulo (IPT), Sao Paulo, SP (Brazil)
2009-07-01
Horizontal Directional Drilling (HDD) is being extensively used in Brazil for installation of oil and gas pipelines. This trenchless technology is currently used in crossings of water bodies, environmental sensitive areas, densely populated areas, areas prone to mass movement and anywhere the traditional technology is not suitable because of the risks. One of the unwanted effects of HDD is collapsing of the soil surrounding the bore-hole, leading to loss of fluid. This can result in problems such as reducing the drilling efficiency, ground heave, structures damage, fluid infiltration and other environmental problems. This paper presents four simulations of down-hole fluid pressures which represents two different geometrical characteristics of the drilling and two different soils. The results showed that greater depths are needed in longer drillings to avoid ground rupture. Thus the end section of the drilling often represents the critical stage. (author)
Methodology for analysis and simulation of large multidisciplinary problems
Russell, William C.; Ikeda, Paul J.; Vos, Robert G.
1989-01-01
The Integrated Structural Modeling (ISM) program is being developed for the Air Force Weapons Laboratory and will be available for Air Force work. Its goal is to provide a design, analysis, and simulation tool intended primarily for directed energy weapons (DEW), kinetic energy weapons (KEW), and surveillance applications. The code is designed to run on DEC (VMS and UNIX), IRIS, Alliant, and Cray hosts. Several technical disciplines are included in ISM, namely structures, controls, optics, thermal, and dynamics. Four topics from the broad ISM goal are discussed. The first is project configuration management and includes two major areas: the software and database arrangement and the system model control. The second is interdisciplinary data transfer and refers to exchange of data between various disciplines such as structures and thermal. Third is a discussion of the integration of component models into one system model, i.e., multiple discipline model synthesis. Last is a presentation of work on a distributed processing computing environment.
Cyclic loading of simulated fault gouge to large strains
Jones, Lucile M.
1980-04-01
As part of a study of the mechanics of simulated fault gouge, deformation of Kayenta Sandstone (24% initial porosity) was observed in triaxial stress tests through several stress cycles. Between 50- and 300-MPa effective pressure the specimens deformed stably without stress drops and with deformation occurring throughout the sample. At 400-MPa effective pressure the specimens underwent strain softening with the deformation occurring along one plane. However, the difference in behavior seems to be due to the density variation at different pressures rather than to the difference in pressure. After peak stress was reached in each cycle, the samples dilated such that the volumetric strain and the linear strain maintained a constant ratio (approximately 0.1) at all pressures. The behavior was independent of the number of stress cycles to linear strains up to 90% and was in general agreement with laws of soil behavior derived from experiments conducted at low pressure (below 5 MPa).
Altitude simulation facility for testing large space motors
Katz, U.; Lustig, J.; Cohen, Y.; Malkin, I.
1993-02-01
This work describes the design of an altitude simulation facility for testing the AKM motor installed in the 'Ofeq' satellite launcher. The facility, which is controlled by a computer, consists of a diffuser and a single-stage ejector fed with preheated air. The calculations of performance and dimensions of the gas extraction system were conducted according to a one-dimensional analysis. Tests were carried out on a small-scale model of the facility in order to examine the design concept, then the full-scale facility was constructed and operated. There was good agreement among the results obtained from the small-scale facility, from the full-scale facility, and from calculations.
Numerical simulations of a large scale oxy-coal burner
Energy Technology Data Exchange (ETDEWEB)
Chae, Taeyoung [Korea Institute of Industrial Technology, Cheonan (Korea, Republic of). Energy System R and D Group; Sungkyunkwan Univ., Suwon (Korea, Republic of). School of Mechanical Engineering; Park, Sanghyun; Ryu, Changkook [Sungkyunkwan Univ., Suwon (Korea, Republic of). School of Mechanical Engineering; Yang, Won [Korea Institute of Industrial Technology, Cheonan (Korea, Republic of). Energy System R and D Group
2013-07-01
Oxy-coal combustion is one of promising carbon dioxide capture and storage (CCS) technologies that uses oxygen and recirculated CO{sub 2} as an oxidizer instead of air. Due to difference in physical properties between CO{sub 2} and N{sub 2}, the oxy-coal combustion requires development of burner and boiler based on fundamental understanding of the flame shape, temperature, radiation and heat flux. For design of a new oxy-coal combustion system, computational fluid dynamics (CFD) is an essential tool to evaluate detailed combustion characteristics and supplement experimental results. In this study, CFD analysis was performed to understand the combustion characteristics inside a tangential vane swirl type 30 MW coal burner for air-mode and oxy-mode operations. In oxy-mode operations, various compositions of primary and secondary oxidizers were assessed which depended on the recirculation ratio of flue gas. For the simulations, devolatilization of coal and char burnout by O{sub 2}, CO{sub 2} and H{sub 2}O were predicted with a Lagrangian particle tracking method considering size distribution of pulverized coal and turbulent dispersion. The radiative heat transfer was solved by employing the discrete ordinate method with the weighted sum of gray gases model (WSGGM) optimized for oxy-coal combustion. In the simulation results for oxy-model operation, the reduced swirl strength of secondary oxidizer increased the flame length due to lower specific volume of CO{sub 2} than N{sub 2}. The flame length was also sensitive to the flow rate of primary oxidizer. The oxidizer without N{sub 2} that reduces thermal NO{sub x} formation makes the NO{sub x} lower in oxy-mode than air-mode. The predicted results showed similar trends with measured temperature profiles for various oxidizer compositions. Further numerical investigations are required to improve the burner design combined with more detailed experimental results.
International Nuclear Information System (INIS)
Frolov, D.; Perevertailo, V.; Frolov, O.; Kononenko, Yu.; Pugatch, V.; Rozenfeld, A.
1995-01-01
breakdown of the junction were detected and examined. Examination of gate-controlled junctions allows one to obtain a useful information on generation currents, breakdown effects, protective rings operation in detectors. Results of this research were in design of a series of the following Si detectors: detectors with active areas 0.1 to 20 cm 2 and a wide range of depletion depth; photodetectors capable of working together with scintillator; large area strip detector with 16 and 128 strips as well as an annular strip detector with 40 strips and a max strip diameter of 13 cm. (author)
Large-Eddy Simulation Using Projection onto Local Basis Functions
Pope, S. B.
In the traditional approach to LES for inhomogeneous flows, the resolved fields are obtained by a filtering operation (with filter width Delta). The equations governing the resolved fields are then partial differential equations, which are solved numerically (on a grid of spacing h). For an LES computation of a given magnitude (i.e., given h), there are conflicting considerations in the choice of Delta: to resolve a large range of turbulent motions, Delta should be small; to solve the equations with numerical accuracy, Delta should be large. In the alternative approach advanced here, this conflict is avoided. The resolved fields are defined by projection onto local basis functions, so that the governing equations are ordinary differential equations for the evolution of the basis-function coefficients. There is no issue of numerical spatial discretization errors. A general methodology for modelling the effects of the residual motions is developed. The model is based directly on the basis-function coefficients, and its effect is to smooth the fields where their rates of change are not well resolved by the basis functions. Demonstration calculations are performed for Burgers' equation.
Electrothermal Simulation of Large-Area Semiconductor Devices
Directory of Open Access Journals (Sweden)
C Kirsch
2017-06-01
Full Text Available The lateral charge transport in thin-film semiconductor devices is affected by the sheet resistance of the various layers. This may lead to a non-uniform current distribution across a large-area device resulting in inhomogeneous luminance, for example, as observed in organic light-emitting diodes (Neyts et al., 2006. The resistive loss in electrical energy is converted into thermal energy via Joule heating, which results in a temperature increase inside the device. On the other hand, the charge transport properties of the device materials are also temperature-dependent, such that we are facing a two-way coupled electrothermal problem. It has been demonstrated that adding thermal effects to an electrical model significantly changes the results (Slawinski et al., 2011. We present a mathematical model for the steady-state distribution of the electric potential and of the temperature across one electrode of a large-area semiconductor device, as well as numerical solutions obtained using the finite element method.
Wind Data Analysis and Wind Flow Simulation Over Large Areas
Directory of Open Access Journals (Sweden)
Terziev Angel
2014-03-01
Full Text Available Increasing the share of renewable energy sources is one of the core policies of the European Union. This is because of the fact that this energy is essential in reducing the greenhouse gas emissions and securing energy supplies. Currently, the share of wind energy from all renewable energy sources is relatively low. The choice of location for a certain wind farm installation strongly depends on the wind potential. Therefore the accurate assessment of wind potential is extremely important. In the present paper an analysis is made on the impact of significant possible parameters on the determination of wind energy potential for relatively large areas. In the analysis the type of measurements (short- and long-term on-site measurements, the type of instrumentation and the terrain roughness factor are considered. The study on the impact of turbulence on the wind flow distribution over complex terrain is presented, and it is based on the real on-site data collected by the meteorological tall towers installed in the northern part of Bulgaria. By means of CFD based software a wind map is developed for relatively large areas. Different turbulent models in numerical calculations were tested and recommendations for the usage of the specific models in flows modeling over complex terrains are presented. The role of each parameter in wind map development is made. Different approaches for determination of wind energy potential based on the preliminary developed wind map are presented.
Large eddy simulation and direct numerical simulation of high speed turbulent reacting flows
Adumitroaie, V.; Frankel, S. H.; Madnia, C. K.; Givi, P.
The objective of this research is to make use of Large Eddy Simulation (LES) and Direct Numerical Simulation (DNS) for the computational analyses of high speed reacting flows. Our efforts in the first phase of this research conducted within the past three years have been directed in several issues pertaining to intricate physics of turbulent reacting flows. In our previous 5 semi-annual reports submitted to NASA LaRC, as well as several technical papers in archival journals, the results of our investigations have been fully described. In this progress report which is different in format as compared to our previous documents, we focus only on the issue of LES. The reason for doing so is that LES is the primary issue of interest to our Technical Monitor and that our other findings were needed to support the activities conducted under this prime issue. The outcomes of our related investigations, nevertheless, are included in the appendices accompanying this report. The relevance of the materials in these appendices are, therefore, discussed only briefly within the body of the report. Here, results are presented of a priori and a posterior analyses for validity assessments of assumed Probability Density Function (PDF) methods as potential subgrid scale (SGS) closures for LES of turbulent reacting flows. Simple non-premixed reacting systems involving an isothermal reaction of the type A + B yields Products under both chemical equilibrium and non-equilibrium conditions are considered. A priori analyses are conducted of a homogeneous box flow, and a spatially developing planar mixing layer to investigate the performance of the Pearson Family of PDF's as SGS models. A posteriori analyses are conducted of the mixing layer using a hybrid one-equation Smagorinsky/PDF SGS closure. The Smagorinsky closure augmented by the solution of the subgrid turbulent kinetic energy (TKE) equation is employed to account for hydrodynamic fluctuations, and the PDF is employed for modeling the
Wu, Jinlu
2013-01-01
Laboratory education can play a vital role in developing a learner's autonomy and scientific inquiry skills. In an innovative, mutation-based learning (MBL) approach, students were instructed to redesign a teacher-designed standard experimental protocol by a "mutation" method in a molecular genetics laboratory course. Students could choose to delete, add, reverse, or replace certain steps of the standard protocol to explore questions of interest to them in a given experimental scenario. They wrote experimental proposals to address their rationales and hypotheses for the "mutations"; conducted experiments in parallel, according to both standard and mutated protocols; and then compared and analyzed results to write individual lab reports. Various autonomy-supportive measures were provided in the entire experimental process. Analyses of student work and feedback suggest that students using the MBL approach 1) spend more time discussing experiments, 2) use more scientific inquiry skills, and 3) find the increased autonomy afforded by MBL more enjoyable than do students following regimented instructions in a conventional "cookbook"-style laboratory. Furthermore, the MBL approach does not incur an obvious increase in labor and financial costs, which makes it feasible for easy adaptation and implementation in a large class.
Energy Technology Data Exchange (ETDEWEB)
Burgwinkel, Paul; Vreydal, Daniel; Eltaliawi, Gamil; Vijayakumar, Nandhakumar [RWTH Aachen (DE). Inst. fuer Maschinentechnik der Rohstoffindustrie (IMR)
2010-09-15
For the first time the Co-simulation method was successfully used for full representation of a large belt conveyor for an open cast mine in a simulation model at the Institute for Mechanical Engineering in the Raw Materials Industry at Rhineland-Westphalia Technological University in Aachen. The aim of this project was the development of an electro-mechanical simulation model, which represents all components of a large belt conveyor from the drive motor to the conveyor belt in one simulation model and thus makes the interactions between the individual assemblies verifiable by calculations. With the aid of the developed model it was possible to determine critical operating speeds of the represented large belt conveyor and derive suitable measures to combat undesirable resonance states in the drive assembly. Furthermore it was possible to clarify the advantage of the full numerical representation of an electromechanical drive system. (orig.)
Understanding Large-scale Structure in the SSA22 Protocluster Region Using Cosmological Simulations
Topping, Michael W.; Shapley, Alice E.; Steidel, Charles C.; Naoz, Smadar; Primack, Joel R.
2018-01-01
We investigate the nature and evolution of large-scale structure within the SSA22 protocluster region at z = 3.09 using cosmological simulations. A redshift histogram constructed from current spectroscopic observations of the SSA22 protocluster reveals two separate peaks at z = 3.065 (blue) and z = 3.095 (red). Based on these data, we report updated overdensity and mass calculations for the SSA22 protocluster. We find {δ }b,{gal}=4.8+/- 1.8 and {δ }r,{gal}=9.5+/- 2.0 for the blue and red peaks, respectively, and {δ }t,{gal}=7.6+/- 1.4 for the entire region. These overdensities correspond to masses of {M}b=(0.76+/- 0.17)× {10}15{h}-1 {M}ȯ , {M}r=(2.15+/- 0.32)× {10}15{h}-1 {M}ȯ , and {M}t=(3.19+/- 0.40)× {10}15{h}-1 {M}ȯ for the red, blue, and total peaks, respectively. We use the Small MultiDark Planck (SMDPL) simulation to identify comparably massive z∼ 3 protoclusters, and uncover the underlying structure and ultimate fate of the SSA22 protocluster. For this analysis, we construct mock redshift histograms for each simulated z∼ 3 protocluster, quantitatively comparing them with the observed SSA22 data. We find that the observed double-peaked structure in the SSA22 redshift histogram corresponds not to a single coalescing cluster, but rather the proximity of a ∼ {10}15{h}-1 {M}ȯ protocluster and at least one > {10}14{h}-1 {M}ȯ cluster progenitor. Such associations in the SMDPL simulation are easily understood within the framework of hierarchical clustering of dark matter halos. We finally find that the opportunity to observe such a phenomenon is incredibly rare, with an occurrence rate of 7.4{h}3 {{{Gpc}}}-3. Based on data obtained at the W.M. Keck Observatory, which is operated as a scientific partnership among the California Institute of Technology, the University of California, and the National Aeronautics and Space Administration, and was made possible by the generous financial support of the W.M. Keck Foundation.
Modeling a Large Data Acquisition Network in a Simulation Framework
AUTHOR|(INSPIRE)INSPIRE-00337030; The ATLAS collaboration; Froening, Holger; Garcia, Pedro Javier; Vandelli, Wainer
2015-01-01
The ATLAS detector at CERN records particle collision “events” delivered by the Large Hadron Collider. Its data-acquisition system is a distributed software system that identifies, selects, and stores interesting events in near real-time, with an aggregate throughput of several 10 GB/s. It is a distributed software system executed on a farm of roughly 2000 commodity worker nodes communicating via TCP/IP on an Ethernet network. Event data fragments are received from the many detector readout channels and are buffered, collected together, analyzed and either stored permanently or discarded. This system, and data-acquisition systems in general, are sensitive to the latency of the data transfer from the readout buffers to the worker nodes. Challenges affecting this transfer include the many-to-one communication pattern and the inherently bursty nature of the traffic. In this paper we introduce the main performance issues brought about by this workload, focusing in particular on the so-called TCP incast pathol...
Milde, Anja; Volkwein, Stefan
2018-01-01
This edited monograph collects research contributions and addresses the advancement of efficient numerical procedures in the area of model order reduction (MOR) for simulation, optimization and control. The topical scope includes, but is not limited to, new out-of-the-box algorithmic solutions for scientific computing, e.g. reduced basis methods for industrial problems and MOR approaches for electrochemical processes. The target audience comprises research experts and practitioners in the field of simulation, optimization and control, but the book may also be beneficial for graduate students alike. .
Visual Data-Analytics of Large-Scale Parallel Discrete-Event Simulations
Energy Technology Data Exchange (ETDEWEB)
Ross, Caitlin; Carothers, Christopher D.; Mubarak, Misbah; Carns, Philip; Ross, Robert; Li, Jianping Kelvin; Ma, Kwan-Liu
2016-11-13
Parallel discrete-event simulation (PDES) is an important tool in the codesign of extreme-scale systems because PDES provides a cost-effective way to evaluate designs of highperformance computing systems. Optimistic synchronization algorithms for PDES, such as Time Warp, allow events to be processed without global synchronization among the processing elements. A rollback mechanism is provided when events are processed out of timestamp order. Although optimistic synchronization protocols enable the scalability of large-scale PDES, the performance of the simulations must be tuned to reduce the number of rollbacks and provide an improved simulation runtime. To enable efficient large-scale optimistic simulations, one has to gain insight into the factors that affect the rollback behavior and simulation performance. We developed a tool for ROSS model developers that gives them detailed metrics on the performance of their large-scale optimistic simulations at varying levels of simulation granularity. Model developers can use this information for parameter tuning of optimistic simulations in order to achieve better runtime and fewer rollbacks. In this work, we instrument the ROSS optimistic PDES framework to gather detailed statistics about the simulation engine. We have also developed an interactive visualization interface that uses the data collected by the ROSS instrumentation to understand the underlying behavior of the simulation engine. The interface connects real time to virtual time in the simulation and provides the ability to view simulation data at different granularities. We demonstrate the usefulness of our framework by performing a visual analysis of the dragonfly network topology model provided by the CODES simulation framework built on top of ROSS. The instrumentation needs to minimize overhead in order to accurately collect data about the simulation performance. To ensure that the instrumentation does not introduce unnecessary overhead, we perform a
Large-scale numerical simulations of star formation put to the test
DEFF Research Database (Denmark)
Frimann, Søren; Jørgensen, Jes Kristian; Haugbølle, Troels
2016-01-01
(SEDs), calculated from large-scalenumerical simulations, to observational studies, thereby aiding in boththe interpretation of the observations and in testing the fidelity ofthe simulations. Methods: The adaptive mesh refinement code,RAMSES, is used to simulate the evolution of a 5 pc × 5 pc ×5 pc...... to calculate evolutionary tracers Tbol andLsmm/Lbol. It is shown that, while the observeddistributions of the tracers are well matched by the simulation, theygenerally do a poor job of tracking the protostellar ages. Disks formearly in the simulation, with 40% of the Class 0 protostars beingencircled by one...
GPU-Accelerated Sparse Matrix Solvers for Large-Scale Simulations, Phase I
National Aeronautics and Space Administration — Many large-scale numerical simulations can be broken down into common mathematical routines. While the applications may differ, the need to perform functions such as...
Energy Technology Data Exchange (ETDEWEB)
NONE
2003-03-01
Joint meeting of the 6th Simulation Science Symposium and the NIFS Collaboration Research 'Large Scale Computer Simulation' was held on December 12-13, 2002 at National Institute for Fusion Science, with the aim of promoting interdisciplinary collaborations in various fields of computer simulations. The present meeting attended by more than 40 people consists of the 11 invited and 22 contributed papers, of which topics were extended not only to fusion science but also to related fields such as astrophysics, earth science, fluid dynamics, molecular dynamics, computer science etc. (author)
Hybrid Reynolds-Averaged/Large Eddy Simulation of the Flow in a Model SCRamjet Cavity Flameholder
Baurle, R. A.
2016-01-01
Steady-state and scale-resolving simulations have been performed for flow in and around a model scramjet combustor flameholder. Experimental data available for this configuration include velocity statistics obtained from particle image velocimetry. Several turbulence models were used for the steady-state Reynolds-averaged simulations which included both linear and non-linear eddy viscosity models. The scale-resolving simulations used a hybrid Reynolds-averaged/large eddy simulation strategy that is designed to be a large eddy simulation everywhere except in the inner portion (log layer and below) of the boundary layer. Hence, this formulation can be regarded as a wall-modeled large eddy simulation. This e ort was undertaken to not only assess the performance of the hybrid Reynolds-averaged / large eddy simulation modeling approach in a flowfield of interest to the scramjet research community, but to also begin to understand how this capability can best be used to augment standard Reynolds-averaged simulations. The numerical errors were quantified for the steady-state simulations, and at least qualitatively assessed for the scale-resolving simulations prior to making any claims of predictive accuracy relative to the measurements. The steady-state Reynolds-averaged results displayed a high degree of variability when comparing the flameholder fuel distributions obtained from each turbulence model. This prompted the consideration of applying the higher-fidelity scale-resolving simulations as a surrogate "truth" model to calibrate the Reynolds-averaged closures in a non-reacting setting prior to their use for the combusting simulations. In general, the Reynolds-averaged velocity profile predictions at the lowest fueling level matched the particle imaging measurements almost as well as was observed for the non-reacting condition. However, the velocity field predictions proved to be more sensitive to the flameholder fueling rate than was indicated in the measurements.
Implementation of a Large Eddy Simulation Method Applied to Recirculating Flow in a Ventilated Room
DEFF Research Database (Denmark)
Davidson, Lars
In the present work Large Eddy Simulations are presented. The flow in a ventilated enclosure is studied. We use an explicit, two-steps time-advancement scheme where the pressure is solved from a Poisson equation.......In the present work Large Eddy Simulations are presented. The flow in a ventilated enclosure is studied. We use an explicit, two-steps time-advancement scheme where the pressure is solved from a Poisson equation....
Khosronejad, Ali; Sotiropoulos, Fotis; Stony Brook University Team
2016-11-01
We present a coupled flow and morphodynamic simulations of extreme flooding in 3 km long and 300 m wide reach of the Mississippi River in Minnesota, which includes three islands and hydraulic structures. We employ the large-eddy simulation (LES) and bed-morphodynamic modules of the VFS-Geophysics model to investigate the flow and bed evolution of the river during a 500 year flood. The coupling of the two modules is carried out via a fluid-structure interaction approach using a nested domain approach to enhance the resolution of bridge scour predictions. The geometrical data of the river, islands and structures are obtained from LiDAR, sub-aqueous sonar and in-situ surveying to construct a digital map of the river bathymetry. Our simulation results for the bed evolution of the river reveal complex sediment dynamics near the hydraulic structures. The numerically captured scour depth near some of the structures reach a maximum of about 10 m. The data-driven simulation strategy we present in this work exemplifies a practical simulation-based-engineering-approach to investigate the resilience of infrastructures to extreme flood events in intricate field-scale riverine systems. This work was funded by a Grant from Minnesota Dept. of Transportation.
Large-scale micromagnetics simulations with dipolar interaction using all-to-all communications
Directory of Open Access Journals (Sweden)
Hiroshi Tsukahara
2016-05-01
Full Text Available We implement on our micromagnetics simulator low-complexity parallel fast-Fourier-transform algorithms, which reduces the frequency of all-to-all communications from six to two times. Almost all the computation time of micromagnetics simulation is taken up by the calculation of the magnetostatic field which can be calculated using the fast Fourier transform method. The results show that the simulation time is decreased with good scalability, even if the micromagentics simulation is performed using 8192 physical cores. This high parallelization effect enables large-scale micromagentics simulation using over one billion to be performed. Because massively parallel computing is needed to simulate the magnetization dynamics of real permanent magnets composed of many micron-sized grains, it is expected that our simulator reveals how magnetization dynamics influences the coercivity of the permanent magnet.
arXiv Stochastic locality and master-field simulations of very large lattices
Lüscher, Martin
2018-01-01
In lattice QCD and other field theories with a mass gap, the field variables in distant regions of a physically large lattice are only weakly correlated. Accurate stochastic estimates of the expectation values of local observables may therefore be obtained from a single representative field. Such master-field simulations potentially allow very large lattices to be simulated, but require various conceptual and technical issues to be addressed. In this talk, an introduction to the subject is provided and some encouraging results of master-field simulations of the SU(3) gauge theory are reported.
Development of the simulation package 'ELSES' for extra-large-scale electronic structure calculation
International Nuclear Information System (INIS)
Hoshi, T; Fujiwara, T
2009-01-01
An early-stage version of the simulation package 'ELSES' (extra-large-scale electronic structure calculation) is developed for simulating the electronic structure and dynamics of large systems, particularly nanometer-scale and ten-nanometer-scale systems (see www.elses.jp). Input and output files are written in the extensible markup language (XML) style for general users. Related pre-/post-simulation tools are also available. A practical workflow and an example are described. A test calculation for the GaAs bulk system is shown, to demonstrate that the present code can handle systems with more than one atom species. Several future aspects are also discussed.
Baurle, R. A.
2015-01-01
Steady-state and scale-resolving simulations have been performed for flow in and around a model scramjet combustor flameholder. The cases simulated corresponded to those used to examine this flowfield experimentally using particle image velocimetry. A variety of turbulence models were used for the steady-state Reynolds-averaged simulations which included both linear and non-linear eddy viscosity models. The scale-resolving simulations used a hybrid Reynolds-averaged / large eddy simulation strategy that is designed to be a large eddy simulation everywhere except in the inner portion (log layer and below) of the boundary layer. Hence, this formulation can be regarded as a wall-modeled large eddy simulation. This effort was undertaken to formally assess the performance of the hybrid Reynolds-averaged / large eddy simulation modeling approach in a flowfield of interest to the scramjet research community. The numerical errors were quantified for both the steady-state and scale-resolving simulations prior to making any claims of predictive accuracy relative to the measurements. The steady-state Reynolds-averaged results showed a high degree of variability when comparing the predictions obtained from each turbulence model, with the non-linear eddy viscosity model (an explicit algebraic stress model) providing the most accurate prediction of the measured values. The hybrid Reynolds-averaged/large eddy simulation results were carefully scrutinized to ensure that even the coarsest grid had an acceptable level of resolution for large eddy simulation, and that the time-averaged statistics were acceptably accurate. The autocorrelation and its Fourier transform were the primary tools used for this assessment. The statistics extracted from the hybrid simulation strategy proved to be more accurate than the Reynolds-averaged results obtained using the linear eddy viscosity models. However, there was no predictive improvement noted over the results obtained from the explicit
Realizability conditions for the turbulent stress tensor in large-eddy simulation
Vreman, A.W.; Geurts, Bernardus J.; Kuerten, Johannes G.M.
1994-01-01
The turbulent stress tensor in large-eddy simulation is examined from a theoretical point of view. Realizability conditions for the components of this tensor are derived, which hold if and only if the filter function is positive. The spectral cut-off, one of the filters frequently used in large-eddy
Sensitivity of the scale partition for variational multiscale large-eddy simulation of channel flow
Holmen, J.; Hughes, T.J.R.; Oberai, A.A.; Wells, G.N.
2004-01-01
The variational multiscale method has been shown to perform well for large-eddy simulation (LES) of turbulent flows. The method relies upon a partition of the resolved velocity field into large- and small-scale components. The subgrid model then acts only on the small scales of motion, unlike
Directory of Open Access Journals (Sweden)
Guangtao Zhang
2015-01-01
Full Text Available In the field of hydropower station transient process simulation (HSTPS, characteristic graph-based iterative hydroturbine model (CGIHM has been widely used when large disturbance hydroturbine modeling is involved. However, by this model, iteration should be used to calculate speed and pressure, and slow convergence or no convergence problems may be encountered for some reasons like special characteristic graph profile, inappropriate iterative algorithm, or inappropriate interpolation algorithm, and so forth. Also, other conventional large disturbance hydroturbine models are of some disadvantages and difficult to be used widely in HSTPS. Therefore, to obtain an accurate simulation result, a simple method for hydroturbine modeling is proposed. By this method, both the initial operating point and the transfer coefficients of linear hydroturbine model keep changing during simulation. Hence, it can reflect the nonlinearity of the hydroturbine and be used for Francis turbine simulation under large disturbance condition. To validate the proposed method, both large disturbance and small disturbance simulations of a single hydrounit supplying a resistive, isolated load were conducted. It was shown that the simulation result is consistent with that of field test. Consequently, the proposed method is an attractive option for HSTPS involving Francis turbine modeling under large disturbance condition.
International Nuclear Information System (INIS)
Norman, A.; Boyd, J.; Davies, G.; Flumerfelt, E.; Herner, K.; Mayer, N.; Mhashilhar, P.; Tamsett, M.; Timm, S.
2015-01-01
Modern long baseline neutrino experiments like the NOvA experiment at Fermilab, require large scale, compute intensive simulations of their neutrino beam fluxes and backgrounds induced by cosmic rays. The amount of simulation required to keep the systematic uncertainties in the simulation from dominating the final physics results is often 10x to 100x that of the actual detector exposure. For the first physics results from NOvA this has meant the simulation of more than 2 billion cosmic ray events in the far detector and more than 200 million NuMI beam spill simulations. Performing these high statistics levels of simulation have been made possible for NOvA through the use of the Open Science Grid and through large scale runs on commercial clouds like Amazon EC2. We details the challenges in performing large scale simulation in these environments and how the computing infrastructure for the NOvA experiment has been adapted to seamlessly support the running of different simulation and data processing tasks on these resources. (paper)
Prasad, K.
2017-12-01
Atmospheric transport is usually performed with weather models, e.g., the Weather Research and Forecasting (WRF) model that employs a parameterized turbulence model and does not resolve the fine scale dynamics generated by the flow around buildings and features comprising a large city. The NIST Fire Dynamics Simulator (FDS) is a computational fluid dynamics model that utilizes large eddy simulation methods to model flow around buildings at length scales much smaller than is practical with models like WRF. FDS has the potential to evaluate the impact of complex topography on near-field dispersion and mixing that is difficult to simulate with a mesoscale atmospheric model. A methodology has been developed to couple the FDS model with WRF mesoscale transport models. The coupling is based on nudging the FDS flow field towards that computed by WRF, and is currently limited to one way coupling performed in an off-line mode. This approach allows the FDS model to operate as a sub-grid scale model with in a WRF simulation. To test and validate the coupled FDS - WRF model, the methane leak from the Aliso Canyon underground storage facility was simulated. Large eddy simulations were performed over the complex topography of various natural gas storage facilities including Aliso Canyon, Honor Rancho and MacDonald Island at 10 m horizontal and vertical resolution. The goal of these simulations included improving and validating transport models as well as testing leak hypotheses. Forward simulation results were compared with aircraft and tower based in-situ measurements as well as methane plumes observed using the NASA Airborne Visible InfraRed Imaging Spectrometer (AVIRIS) and the next generation instrument AVIRIS-NG. Comparison of simulation results with measurement data demonstrate the capability of the coupled FDS-WRF models to accurately simulate the transport and dispersion of methane plumes over urban domains. Simulated integrated methane enhancements will be presented and
ROSA-IV Large Scale Test Facility (LSTF) system description for second simulated fuel assembly
International Nuclear Information System (INIS)
1990-10-01
The ROSA-IV Program's Large Scale Test Facility (LSTF) is a test facility for integral simulation of thermal-hydraulic response of a pressurized water reactor (PWR) during small break loss-of-coolant accidents (LOCAs) and transients. In this facility, the PWR core nuclear fuel rods are simulated using electric heater rods. The simulated fuel assembly which was installed during the facility construction was replaced with a new one in 1988. The first test with this second simulated fuel assembly was conducted in December 1988. This report describes the facility configuration and characteristics as of this date (December 1988) including the new simulated fuel assembly design and the facility changes which were made during the testing with the first assembly as well as during the renewal of the simulated fuel assembly. (author)
Initial condition effects on large scale structure in numerical simulations of plane mixing layers
McMullan, W. A.; Garrett, S. J.
2016-01-01
In this paper, Large Eddy Simulations are performed on the spatially developing plane turbulent mixing layer. The simulated mixing layers originate from initially laminar conditions. The focus of this research is on the effect of the nature of the imposed fluctuations on the large-scale spanwise and streamwise structures in the flow. Two simulations are performed; one with low-level three-dimensional inflow fluctuations obtained from pseudo-random numbers, the other with physically correlated fluctuations of the same magnitude obtained from an inflow generation technique. Where white-noise fluctuations provide the inflow disturbances, no spatially stationary streamwise vortex structure is observed, and the large-scale spanwise turbulent vortical structures grow continuously and linearly. These structures are observed to have a three-dimensional internal geometry with branches and dislocations. Where physically correlated provide the inflow disturbances a "streaky" streamwise structure that is spatially stationary is observed, with the large-scale turbulent vortical structures growing with the square-root of time. These large-scale structures are quasi-two-dimensional, on top of which the secondary structure rides. The simulation results are discussed in the context of the varying interpretations of mixing layer growth that have been postulated. Recommendations are made concerning the data required from experiments in order to produce accurate numerical simulation recreations of real flows.
On the rejection-based algorithm for simulation and analysis of large-scale reaction networks
Energy Technology Data Exchange (ETDEWEB)
Thanh, Vo Hong, E-mail: vo@cosbi.eu [The Microsoft Research-University of Trento Centre for Computational and Systems Biology, Piazza Manifattura 1, Rovereto 38068 (Italy); Zunino, Roberto, E-mail: roberto.zunino@unitn.it [Department of Mathematics, University of Trento, Trento (Italy); Priami, Corrado, E-mail: priami@cosbi.eu [The Microsoft Research-University of Trento Centre for Computational and Systems Biology, Piazza Manifattura 1, Rovereto 38068 (Italy); Department of Mathematics, University of Trento, Trento (Italy)
2015-06-28
Stochastic simulation for in silico studies of large biochemical networks requires a great amount of computational time. We recently proposed a new exact simulation algorithm, called the rejection-based stochastic simulation algorithm (RSSA) [Thanh et al., J. Chem. Phys. 141(13), 134116 (2014)], to improve simulation performance by postponing and collapsing as much as possible the propensity updates. In this paper, we analyze the performance of this algorithm in detail, and improve it for simulating large-scale biochemical reaction networks. We also present a new algorithm, called simultaneous RSSA (SRSSA), which generates many independent trajectories simultaneously for the analysis of the biochemical behavior. SRSSA improves simulation performance by utilizing a single data structure across simulations to select reaction firings and forming trajectories. The memory requirement for building and storing the data structure is thus independent of the number of trajectories. The updating of the data structure when needed is performed collectively in a single operation across the simulations. The trajectories generated by SRSSA are exact and independent of each other by exploiting the rejection-based mechanism. We test our new improvement on real biological systems with a wide range of reaction networks to demonstrate its applicability and efficiency.
Cyber-Enabled Scientific Discovery
International Nuclear Information System (INIS)
Chan, Tony; Jameson, Leland
2007-01-01
It is often said that numerical simulation is third in the group of three ways to explore modern science: theory, experiment and simulation. Carefully executed modern numerical simulations can, however, be considered at least as relevant as experiment and theory. In comparison to physical experimentation, with numerical simulation one has the numerically simulated values of every field variable at every grid point in space and time. In comparison to theory, with numerical simulation one can explore sets of very complex non-linear equations such as the Einstein equations that are very difficult to investigate theoretically. Cyber-enabled scientific discovery is not just about numerical simulation but about every possible issue related to scientific discovery by utilizing cyberinfrastructure such as the analysis and storage of large data sets, the creation of tools that can be used by broad classes of researchers and, above all, the education and training of a cyber-literate workforce
Large-signal, dynamic simulation of the slowpoke-3 nuclear heating reactor
International Nuclear Information System (INIS)
Tseng, C.M.; Lepp, R.M.
1983-07-01
A 2 MWt nuclear reactor, called SLOWPOKE-3, is being developed at the Chalk River Nuclear Laboratories (CRNL). This reactor, which is cooled by natural circulation, is designed to produce hot water for commercial space heating and perhaps generate some electricity in remote locations where the costs of alternate forms of energy are high. A large-signal, dynamic simulation of this reactor, without closed-loop control, was developed and implemented on a hybrid computer, using the basic equations of conservation of mass, energy and momentum. The natural circulation of downcomer flow in the pool was simulated using a special filter, capable of modelling various flow conditions. The simulation was then used to study the intermediate and long-term transient response of SLOWPOKE-3 to large disturbances, such as loss of heat sink, loss of regulation, daily load following, and overcooling of the reactor coolant. Results of the simulation show that none of these disturbances produce hazardous transients
Heavy-Ion Collimation at the Large Hadron Collider Simulations and Measurements
AUTHOR|(CDS)2083002; Wessels, Johannes Peter; Bruce, Roderik; Wessels, Johannes Peter; Bruce, Roderik
The CERN Large Hadron Collider (LHC) stores and collides proton and $^{208}$Pb$^{82+}$ beams of unprecedented energy and intensity. Thousands of superconducting magnets, operated at 1.9 K, guide the very intense and energetic particle beams, which have a large potential for destruction. This implies the demand for a multi-stage collimation system to provide protection from beam-induced quenches or even hardware damage. In heavy-ion operation, ion fragments with significant rigidity offsets can still scatter out of the collimation system. When they irradiate the superconducting LHC magnets, the latter risk to quench (lose their superconducting property). These secondary collimation losses can potentially impose a limitation for the stored heavy-ion beam energy. Therefore, their distribution in the LHC needs to be understood by sophisticated simulations. Such simulation tools must accurately simulate the particle motion of many different nuclides in the magnetic LHC lattice and simulate their interaction with t...
Chatterjee, Tanmoy; Peet, Yulia T.
2018-03-01
Length scales of eddies involved in the power generation of infinite wind farms are studied by analyzing the spectra of the turbulent flux of mean kinetic energy (MKE) from large eddy simulations (LES). Large-scale structures with an order of magnitude bigger than the turbine rotor diameter (D ) are shown to have substantial contribution to wind power. Varying dynamics in the intermediate scales (D -10 D ) are also observed from a parametric study involving interturbine distances and hub height of the turbines. Further insight about the eddies responsible for the power generation have been provided from the scaling analysis of two-dimensional premultiplied spectra of MKE flux. The LES code is developed in a high Reynolds number near-wall modeling framework, using an open-source spectral element code Nek5000, and the wind turbines have been modelled using a state-of-the-art actuator line model. The LES of infinite wind farms have been validated against the statistical results from the previous literature. The study is expected to improve our understanding of the complex multiscale dynamics in the domain of large wind farms and identify the length scales that contribute to the power. This information can be useful for design of wind farm layout and turbine placement that take advantage of the large-scale structures contributing to wind turbine power.
Energy Technology Data Exchange (ETDEWEB)
Veleiro Blanco, A. M.
2011-07-01
The cryogenic plants and their control in Scientific-Nuclear Facilities is complicated by the large number of variables and the wide range of variation during operation. Initially the design and control of these systems in CERN was based on stationary calculations which non yielded the expected results. Due to its complexity, the dynamic simulation is the only way to get adequate results during operational transients.
Piomelli, Ugo; Zang, Thomas A.; Speziale, Charles G.; Lund, Thomas S.
1990-01-01
An eddy viscosity model based on the renormalization group theory of Yakhot and Orszag (1986) is applied to the large-eddy simulation of transition in a flat-plate boundary layer. The simulation predicts with satisfactory accuracy the mean velocity and Reynolds stress profiles, as well as the development of the important scales of motion. The evolution of the structures characteristic of the nonlinear stages of transition is also predicted reasonably well.
Establishment of DNS database in a turbulent channel flow by large-scale simulations
Abe, Hiroyuki; Kawamura, Hiroshi; 阿部 浩幸; 河村 洋
2008-01-01
In the present study, we establish statistical DNS (Direct Numerical Simulation) database in a turbulent channel flow with passive scalar transport at high Reynolds numbers and make the data available at our web site (http://murasun.me.noda.tus.ac.jp/turbulence/). The established database is reported together with the implementation of large-scale simulations, representative DNS results and results on turbulence model testing using the DNS data.
VerHulst, Claire; Meneveau, Charles
2014-02-01
In this study, we address the question of how kinetic energy is entrained into large wind turbine arrays and, in particular, how large-scale flow structures contribute to such entrainment. Previous research has shown this entrainment to be an important limiting factor in the performance of very large arrays where the flow becomes fully developed and there is a balance between the forcing of the atmospheric boundary layer and the resistance of the wind turbines. Given the high Reynolds numbers and domain sizes on the order of kilometers, we rely on wall-modeled large eddy simulation (LES) to simulate turbulent flow within the wind farm. Three-dimensional proper orthogonal decomposition (POD) analysis is then used to identify the most energetic flow structures present in the LES data. We quantify the contribution of each POD mode to the kinetic energy entrainment and its dependence on the layout of the wind turbine array. The primary large-scale structures are found to be streamwise, counter-rotating vortices located above the height of the wind turbines. While the flow is periodic, the geometry is not invariant to all horizontal translations due to the presence of the wind turbines and thus POD modes need not be Fourier modes. Differences of the obtained modes with Fourier modes are documented. Some of the modes are responsible for a large fraction of the kinetic energy flux to the wind turbine region. Surprisingly, more flow structures (POD modes) are needed to capture at least 40% of the turbulent kinetic energy, for which the POD analysis is optimal, than are needed to capture at least 40% of the kinetic energy flux to the turbines. For comparison, we consider the cases of aligned and staggered wind turbine arrays in a neutral atmospheric boundary layer as well as a reference case without wind turbines. While the general characteristics of the flow structures are robust, the net kinetic energy entrainment to the turbines depends on the presence and relative
Large-Scale Brain Simulation and Disorders of Consciousness. Mapping Technical and Conceptual Issues
Directory of Open Access Journals (Sweden)
Michele Farisco
2018-04-01
Full Text Available Modeling and simulations have gained a leading position in contemporary attempts to describe, explain, and quantitatively predict the human brain’s operations. Computer models are highly sophisticated tools developed to achieve an integrated knowledge of the brain with the aim of overcoming the actual fragmentation resulting from different neuroscientific approaches. In this paper we investigate the plausibility of simulation technologies for emulation of consciousness and the potential clinical impact of large-scale brain simulation on the assessment and care of disorders of consciousness (DOCs, e.g., Coma, Vegetative State/Unresponsive Wakefulness Syndrome, Minimally Conscious State. Notwithstanding their technical limitations, we suggest that simulation technologies may offer new solutions to old practical problems, particularly in clinical contexts. We take DOCs as an illustrative case, arguing that the simulation of neural correlates of consciousness is potentially useful for improving treatments of patients with DOCs.
Large-scale simulations with distributed computing: Asymptotic scaling of ballistic deposition
International Nuclear Information System (INIS)
Farnudi, Bahman; Vvedensky, Dimitri D
2011-01-01
Extensive kinetic Monte Carlo simulations are reported for ballistic deposition (BD) in (1 + 1) dimensions. The large system sizes L observed for the onset of asymptotic scaling (L ≅ 2 12 ) explains the widespread discrepancies in previous reports for exponents of BD in one and likely in higher dimensions. The exponents obtained directly from our simulations, α = 0.499 ± 0.004 and β = 0.336 ± 0.004, capture the exact values α = 1/2 and β = 1/3 for the one-dimensional Kardar-Parisi-Zhang equation. An analysis of our simulations suggests a criterion for identifying the onset of true asymptotic scaling, which enables a more informed evaluation of exponents for BD in higher dimensions. These simulations were made possible by the Simulation through Social Networking project at the Institute for Advanced Studies in Basic Sciences in 2007, which was re-launched in November 2010.
Dynamics Modeling and Simulation of Large Transport Airplanes in Upset Conditions
Foster, John V.; Cunningham, Kevin; Fremaux, Charles M.; Shah, Gautam H.; Stewart, Eric C.; Rivers, Robert A.; Wilborn, James E.; Gato, William
2005-01-01
As part of NASA's Aviation Safety and Security Program, research has been in progress to develop aerodynamic modeling methods for simulations that accurately predict the flight dynamics characteristics of large transport airplanes in upset conditions. The motivation for this research stems from the recognition that simulation is a vital tool for addressing loss-of-control accidents, including applications to pilot training, accident reconstruction, and advanced control system analysis. The ultimate goal of this effort is to contribute to the reduction of the fatal accident rate due to loss-of-control. Research activities have involved accident analyses, wind tunnel testing, and piloted simulation. Results have shown that significant improvements in simulation fidelity for upset conditions, compared to current training simulations, can be achieved using state-of-the-art wind tunnel testing and aerodynamic modeling methods. This paper provides a summary of research completed to date and includes discussion on key technical results, lessons learned, and future research needs.
International Nuclear Information System (INIS)
Jamali, J.; Aghajafari, R.; Moini, R.; Sadeghi, H.
2002-01-01
A time-domain approach is presented to calculate electromagnetic fields inside a large Electromagnetic Pulse (EMP) simulator. This type of EMP simulator is used for studying the effect of electromagnetic pulses on electrical apparatus in various structures such as vehicles, a reoplanes, etc. The simulator consists of three planar transmission lines. To solve the problem, we first model the metallic structure of the simulator as a grid of conducting wires. The numerical solution of the governing electric field integral equation is then obtained using the method of moments in time domain. To demonstrate the accuracy of the model, we consider a typical EMP simulator. The comparison of our results with those obtained experimentally in the literature validates the model introduced in this paper
Energy Technology Data Exchange (ETDEWEB)
Seeliger, Andreas; Vreydal, Daniel; Eltaliawi, Gamil; Vijayakumar, Nandhakumar [Technische Hochschule Aachen (Germany). Lehrstuhl und Inst. fuer Bergwerks- und Huettenmaschinenkunde
2009-04-28
The aim of the GrobaDyn research project is the complete modelling of a large conveyor system. With the aid of the model possible conversion of the previous drives with a constant speed to variable-speed drives will be simulated in advance of the planning phase of this conversion and any resonance phenomena within the operating speed range analysed and if necessary counter-measures taken. (orig.)
Pierre, M.; Pacaud, F.; Adami, C.; Alis, S.; Altieri, B.; Baran, N.; Benoist, C.; Birkinshaw, M.; Bongiorno, A.; Bremer, M. N.; Brusa, M.; Butler, A.; Ciliegi, P.; Chiappetti, L.; Clerc, N.; Corasaniti, P. S.; Coupon, J.; De Breuck, C.; Democles, J.; Desai, S.; Delhaize, J.; Devriendt, J.; Dubois, Y.; Eckert, D.; Elyiv, A.; Ettori, S.; Evrard, A.; Faccioli, L.; Farahi, A.; Ferrari, C.; Finet, F.; Fotopoulou, S.; Fourmanoit, N.; Gandhi, P.; Gastaldello, F.; Gastaud, R.; Georgantopoulos, I.; Giles, P.; Guennou, L.; Guglielmo, V.; Horellou, C.; Husband, K.; Huynh, M.; Iovino, A.; Kilbinger, M.; Koulouridis, E.; Lavoie, S.; Le Brun, A. M. C.; Le Fevre, J. P.; Lidman, C.; Lieu, M.; Lin, C. A.; Mantz, A.; Maughan, B. J.; Maurogordato, S.; McCarthy, I. G.; McGee, S.; Melin, J. B.; Melnyk, O.; Menanteau, F.; Novak, M.; Paltani, S.; Plionis, M.; Poggianti, B. M.; Pomarede, D.; Pompei, E.; Ponman, T. J.; Ramos-Ceja, M. E.; Ranalli, P.; Rapetti, D.; Raychaudury, S.; Reiprich, T. H.; Rottgering, H.; Rozo, E.; Rykoff, E.; Sadibekova, T.; Santos, J.; Sauvageot, J. L.; Schimd, C.; Sereno, M.; Smith, G. P.; Smolčić, V.; Snowden, S.; Spergel, D.; Stanford, S.; Surdej, J.; Valageas, P.; Valotti, A.; Valtchanov, I.; Vignali, C.; Willis, J.; Ziparo, F.
2016-06-01
Context. The quest for the cosmological parameters that describe our universe continues to motivate the scientific community to undertake very large survey initiatives across the electromagnetic spectrum. Over the past two decades, the Chandra and XMM-Newton observatories have supported numerous studies of X-ray-selected clusters of galaxies, active galactic nuclei (AGNs), and the X-ray background. The present paper is the first in a series reporting results of the XXL-XMM survey; it comes at a time when the Planck mission results are being finalised. Aims: We present the XXL Survey, the largest XMM programme totaling some 6.9 Ms to date and involving an international consortium of roughly 100 members. The XXL Survey covers two extragalactic areas of 25 deg2 each at a point-source sensitivity of ~5 × 10-15 erg s-1 cm-2 in the [0.5-2] keV band (completeness limit). The survey's main goals are to provide constraints on the dark energy equation of state from the space-time distribution of clusters of galaxies and to serve as a pathfinder for future, wide-area X-ray missions. We review science objectives, including cluster studies, AGN evolution, and large-scale structure, that are being conducted with the support of approximately 30 follow-up programmes. Methods: We describe the 542 XMM observations along with the associated multi-λ and numerical simulation programmes. We give a detailed account of the X-ray processing steps and describe innovative tools being developed for the cosmological analysis. Results: The paper provides a thorough evaluation of the X-ray data, including quality controls, photon statistics, exposure and background maps, and sky coverage. Source catalogue construction and multi-λ associations are briefly described. This material will be the basis for the calculation of the cluster and AGN selection functions, critical elements of the cosmological and science analyses. Conclusions: The XXL multi-λ data set will have a unique lasting legacy
Design and simulation of betavoltaic battery using large-grain polysilicon
International Nuclear Information System (INIS)
Yao, Shulin; Song, Zijun; Wang, Xiang; San, Haisheng; Yu, Yuxi
2012-01-01
In this paper, we present the design and simulation of a p–n junction betavoltaic battery based on large-grain polysilicon. By the Monte Carlo simulation, the average penetration depth were obtained, according to which the optimal depletion region width was designed. The carriers transport model of large-grain polysilicon is used to determine the diffusion length of minority carrier. By optimizing the doping concentration, the maximum power conversion efficiency can be achieved to be 0.90% with a 10 mCi/cm 2 Ni-63 source radiation. - Highlights: ► Ni 63 is employed as the pure beta radioisotope source. ► The planar p–n junction betavoltaic battery is based on large-grain polysilicon. ► The carriers transport model of large-grain polysilicon is used to determine the diffusion length of minority carrier. ► The average penetration depth was obtained by using the Monte Carlo Method.
Large-scale simulations of plastic neural networks on neuromorphic hardware
Directory of Open Access Journals (Sweden)
James Courtney Knight
2016-04-01
Full Text Available SpiNNaker is a digital, neuromorphic architecture designed for simulating large-scale spiking neural networks at speeds close to biological real-time. Rather than using bespoke analog or digital hardware, the basic computational unit of a SpiNNaker system is a general-purpose ARM processor, allowing it to be programmed to simulate a wide variety of neuron and synapse models. This flexibility is particularly valuable in the study of biological plasticity phenomena. A recently proposed learning rule based on the Bayesian Confidence Propagation Neural Network (BCPNN paradigm offers a generic framework for modeling the interaction of different plasticity mechanisms using spiking neurons. However, it can be computationally expensive to simulate large networks with BCPNN learning since it requires multiple state variables for each synapse, each of which needs to be updated every simulation time-step. We discuss the trade-offs in efficiency and accuracy involved in developing an event-based BCPNN implementation for SpiNNaker based on an analytical solution to the BCPNN equations, and detail the steps taken to fit this within the limited computational and memory resources of the SpiNNaker architecture. We demonstrate this learning rule by learning temporal sequences of neural activity within a recurrent attractor network which we simulate at scales of up to 20000 neurons and 51200000 plastic synapses: the largest plastic neural network ever to be simulated on neuromorphic hardware. We also run a comparable simulation on a Cray XC-30 supercomputer system and find that, if it is to match the run-time of our SpiNNaker simulation, the super computer system uses approximately more power. This suggests that cheaper, more power efficient neuromorphic systems are becoming useful discovery tools in the study of plasticity in large-scale brain models.
Energy Technology Data Exchange (ETDEWEB)
Yuan, Haomin; Solberg, Jerome; Merzari, Elia; Kraus, Adam; Grindeanu, Iulian
2017-10-01
This paper describes a numerical study of flow-induced vibration in a helical coil steam generator experiment conducted at Argonne National Laboratory in the 1980s. In the experiment, a half-scale sector model of a steam generator helical coil tube bank was subjected to still and flowing air and water, and the vibrational characteristics were recorded. The research detailed in this document utilizes the multi-physics simulation toolkit SHARP developed at Argonne National Laboratory, in cooperation with Lawrence Livermore National Laboratory, to simulate the experiment. SHARP uses the spectral element code Nek5000 for fluid dynamics analysis and the finite element code DIABLO for structural analysis. The flow around the coil tubes is modeled in Nek5000 by using a large eddy simulation turbulence model. Transient pressure data on the tube surfaces is sampled and transferred to DIABLO for the structural simulation. The structural response is simulated in DIABLO via an implicit time-marching algorithm and a combination of continuum elements and structural shells. Tube vibration data (acceleration and frequency) are sampled and compared with the experimental data. Currently, only one-way coupling is used, which means that pressure loads from the fluid simulation are transferred to the structural simulation but the resulting structural displacements are not fed back to the fluid simulation
A Novel CPU/GPU Simulation Environment for Large-Scale Biologically-Realistic Neural Modeling
Directory of Open Access Journals (Sweden)
Roger V Hoang
2013-10-01
Full Text Available Computational Neuroscience is an emerging field that provides unique opportunities to studycomplex brain structures through realistic neural simulations. However, as biological details are added tomodels, the execution time for the simulation becomes longer. Graphics Processing Units (GPUs are now being utilized to accelerate simulations due to their ability to perform computations in parallel. As such, they haveshown significant improvement in execution time compared to Central Processing Units (CPUs. Most neural simulators utilize either multiple CPUs or a single GPU for better performance, but still show limitations in execution time when biological details are not sacrificed. Therefore, we present a novel CPU/GPU simulation environment for large-scale biological networks,the NeoCortical Simulator version 6 (NCS6. NCS6 is a free, open-source, parallelizable, and scalable simula-tor, designed to run on clusters of multiple machines, potentially with high performance computing devicesin each of them. It has built-in leaky-integrate-and-fire (LIF and Izhikevich (IZH neuron models, but usersalso have the capability to design their own plug-in interface for different neuron types as desired. NCS6is currently able to simulate one million cells and 100 million synapses in quasi real time by distributing dataacross these heterogeneous clusters of CPUs and GPUs.
Krumhansl, R. A.; Foster, J.; Peach, C. L.; Busey, A.; Baker, I.
2012-12-01
The practice of science and engineering is being revolutionized by the development of cyberinfrastructure for accessing near real-time and archived observatory data. Large cyberinfrastructure projects have the potential to transform the way science is taught in high school classrooms, making enormous quantities of scientific data available, giving students opportunities to analyze and draw conclusions from many kinds of complex data, and providing students with experiences using state-of-the-art resources and techniques for scientific investigations. However, online interfaces to scientific data are built by scientists for scientists, and their design can significantly impede broad use by novices. Knowledge relevant to the design of student interfaces to complex scientific databases is broadly dispersed among disciplines ranging from cognitive science to computer science and cartography and is not easily accessible to designers of educational interfaces. To inform efforts at bridging scientific cyberinfrastructure to the high school classroom, Education Development Center, Inc. and the Scripps Institution of Oceanography conducted an NSF-funded 2-year interdisciplinary review of literature and expert opinion pertinent to making interfaces to large scientific databases accessible to and usable by precollege learners and their teachers. Project findings are grounded in the fundamentals of Cognitive Load Theory, Visual Perception, Schemata formation and Universal Design for Learning. The Knowledge Status Report (KSR) presents cross-cutting and visualization-specific guidelines that highlight how interface design features can address/ ameliorate challenges novice high school students face as they navigate complex databases to find data, and construct and look for patterns in maps, graphs, animations and other data visualizations. The guidelines present ways to make scientific databases more broadly accessible by: 1) adjusting the cognitive load imposed by the user
International Nuclear Information System (INIS)
Battaglia, N.; Trac, H.; Cen, R.; Loeb, A.
2013-01-01
We present a new method for modeling inhomogeneous cosmic reionization on large scales. Utilizing high-resolution radiation-hydrodynamic simulations with 2048 3 dark matter particles, 2048 3 gas cells, and 17 billion adaptive rays in a L = 100 Mpc h –1 box, we show that the density and reionization redshift fields are highly correlated on large scales (∼> 1 Mpc h –1 ). This correlation can be statistically represented by a scale-dependent linear bias. We construct a parametric function for the bias, which is then used to filter any large-scale density field to derive the corresponding spatially varying reionization redshift field. The parametric model has three free parameters that can be reduced to one free parameter when we fit the two bias parameters to simulation results. We can differentiate degenerate combinations of the bias parameters by combining results for the global ionization histories and correlation length between ionized regions. Unlike previous semi-analytic models, the evolution of the reionization redshift field in our model is directly compared cell by cell against simulations and performs well in all tests. Our model maps the high-resolution, intermediate-volume radiation-hydrodynamic simulations onto lower-resolution, larger-volume N-body simulations (∼> 2 Gpc h –1 ) in order to make mock observations and theoretical predictions
Creech, Angus; Früh, Wolf-Gerrit; Maguire, A. Eoghan
2015-05-01
We present here a computational fluid dynamics (CFD) simulation of Lillgrund offshore wind farm, which is located in the Øresund Strait between Sweden and Denmark. The simulation combines a dynamic representation of wind turbines embedded within a large-eddy simulation CFD solver and uses hr-adaptive meshing to increase or decrease mesh resolution where required. This allows the resolution of both large-scale flow structures around the wind farm, and the local flow conditions at individual turbines; consequently, the response of each turbine to local conditions can be modelled, as well as the resulting evolution of the turbine wakes. This paper provides a detailed description of the turbine model which simulates the interaction between the wind, the turbine rotors, and the turbine generators by calculating the forces on the rotor, the body forces on the air, and instantaneous power output. This model was used to investigate a selection of key wind speeds and directions, investigating cases where a row of turbines would be fully aligned with the wind or at specific angles to the wind. Results shown here include presentations of the spin-up of turbines, the observation of eddies moving through the turbine array, meandering turbine wakes, and an extensive wind farm wake several kilometres in length. The key measurement available for cross-validation with operational wind farm data is the power output from the individual turbines, where the effect of unsteady turbine wakes on the performance of downstream turbines was a main point of interest. The results from the simulations were compared to the performance measurements from the real wind farm to provide a firm quantitative validation of this methodology. Having achieved good agreement between the model results and actual wind farm measurements, the potential of the methodology to provide a tool for further investigations of engineering and atmospheric science problems is outlined.
Large-scale agent-based social simulation : A study on epidemic prediction and control
Zhang, M.
2016-01-01
Large-scale agent-based social simulation is gradually proving to be a versatile methodological approach for studying human societies, which could make contributions from policy making in social science, to distributed artificial intelligence and agent technology in computer science, and to theory
Simulations of muon-induced neutron flux at large depths underground
International Nuclear Information System (INIS)
Kudryavtsev, V.A.; Spooner, N.J.C.; McMillan, J.E.
2003-01-01
The production of neutrons by cosmic-ray muons at large depths underground is discussed. The most recent versions of the muon propagation code MUSIC, and particle transport code FLUKA are used to evaluate muon and neutron fluxes. The results of simulations are compared with experimental data
Comparison of Large Eddy Simulations of a convective boundary layer with wind LIDAR measurements
DEFF Research Database (Denmark)
Pedersen, Jesper Grønnegaard; Kelly, Mark C.; Gryning, Sven-Erik
2012-01-01
Vertical profiles of the horizontal wind speed and of the standard deviation of vertical wind speed from Large Eddy Simulations of a convective atmospheric boundary layer are compared to wind LIDAR measurements up to 1400 m. Fair agreement regarding both types of profiles is observed only when...
Synthetic atmospheric turbulence and wind shear in large eddy simulations of wind turbine wakes
DEFF Research Database (Denmark)
Keck, Rolf-Erik; Mikkelsen, Robert Flemming; Troldborg, Niels
2014-01-01
, superimposed on top of a mean deterministic shear layer consistent with that used in the IEC standard for wind turbine load calculations. First, the method is evaluated by running a series of large-eddy simulations in an empty domain, where the imposed turbulence and wind shear is allowed to reach a fully...
Large-scale tropospheric transport in the Chemistry-Climate Model Initiative (CCMI) simulations
Orbe, Clara; Yang, Huang; Waugh, Darryn W.; Zeng, Guang; Morgenstern, Olaf; Kinnison, Douglas E.; Lamarque, Jean-Francois; Tilmes, Simone; Plummer, David A.; Scinocca, John F.; Josse, Beatrice; Marecal, Virginie; Jöckel, Patrick; Oman, Luke D.; Strahan, Susan E.; Deushi, Makoto; Tanaka, Taichu Y.; Yoshida, Kohei; Akiyoshi, Hideharu; Yamashita, Yousuke; Stenke, Andreas; Revell, Laura; Sukhodolov, Timofei; Rozanov, Eugene; Pitari, Giovanni; Visioni, Daniele; Stone, Kane A.; Schofield, Robyn; Banerjee, Antara
2018-05-01
Understanding and modeling the large-scale transport of trace gases and aerosols is important for interpreting past (and projecting future) changes in atmospheric composition. Here we show that there are large differences in the global-scale atmospheric transport properties among the models participating in the IGAC SPARC Chemistry-Climate Model Initiative (CCMI). Specifically, we find up to 40 % differences in the transport timescales connecting the Northern Hemisphere (NH) midlatitude surface to the Arctic and to Southern Hemisphere high latitudes, where the mean age ranges between 1.7 and 2.6 years. We show that these differences are related to large differences in vertical transport among the simulations, in particular to differences in parameterized convection over the oceans. While stronger convection over NH midlatitudes is associated with slower transport to the Arctic, stronger convection in the tropics and subtropics is associated with faster interhemispheric transport. We also show that the differences among simulations constrained with fields derived from the same reanalysis products are as large as (and in some cases larger than) the differences among free-running simulations, most likely due to larger differences in parameterized convection. Our results indicate that care must be taken when using simulations constrained with analyzed winds to interpret the influence of meteorology on tropospheric composition.
Large-scale tropospheric transport in the Chemistry–Climate Model Initiative (CCMI simulations
Directory of Open Access Journals (Sweden)
C. Orbe
2018-05-01
Full Text Available Understanding and modeling the large-scale transport of trace gases and aerosols is important for interpreting past (and projecting future changes in atmospheric composition. Here we show that there are large differences in the global-scale atmospheric transport properties among the models participating in the IGAC SPARC Chemistry–Climate Model Initiative (CCMI. Specifically, we find up to 40 % differences in the transport timescales connecting the Northern Hemisphere (NH midlatitude surface to the Arctic and to Southern Hemisphere high latitudes, where the mean age ranges between 1.7 and 2.6 years. We show that these differences are related to large differences in vertical transport among the simulations, in particular to differences in parameterized convection over the oceans. While stronger convection over NH midlatitudes is associated with slower transport to the Arctic, stronger convection in the tropics and subtropics is associated with faster interhemispheric transport. We also show that the differences among simulations constrained with fields derived from the same reanalysis products are as large as (and in some cases larger than the differences among free-running simulations, most likely due to larger differences in parameterized convection. Our results indicate that care must be taken when using simulations constrained with analyzed winds to interpret the influence of meteorology on tropospheric composition.
Wind Energy-Related Atmospheric Boundary Layer Large-Eddy Simulation Using OpenFOAM: Preprint
Energy Technology Data Exchange (ETDEWEB)
Churchfield, M.J.; Vijayakumar, G.; Brasseur, J.G.; Moriarty, P.J.
2010-08-01
This paper develops and evaluates the performance of a large-eddy simulation (LES) solver in computing the atmospheric boundary layer (ABL) over flat terrain under a variety of stability conditions, ranging from shear driven (neutral stratification) to moderately convective (unstable stratification).
Modeling and analysis of large-eddy simulations of particle-laden turbulent boundary layer flows
Rahman, Mustafa M.
2017-01-05
We describe a framework for the large-eddy simulation of solid particles suspended and transported within an incompressible turbulent boundary layer (TBL). For the fluid phase, the large-eddy simulation (LES) of incompressible turbulent boundary layer employs stretched spiral vortex subgrid-scale model and a virtual wall model similar to the work of Cheng, Pullin & Samtaney (J. Fluid Mech., 2015). This LES model is virtually parameter free and involves no active filtering of the computed velocity field. Furthermore, a recycling method to generate turbulent inflow is implemented. For the particle phase, the direct quadrature method of moments (DQMOM) is chosen in which the weights and abscissas of the quadrature approximation are tracked directly rather than the moments themselves. The numerical method in this framework is based on a fractional-step method with an energy-conservative fourth-order finite difference scheme on a staggered mesh. This code is parallelized based on standard message passing interface (MPI) protocol and is designed for distributed-memory machines. It is proposed to utilize this framework to examine transport of particles in very large-scale simulations. The solver is validated using the well know result of Taylor-Green vortex case. A large-scale sandstorm case is simulated and the altitude variations of number density along with its fluctuations are quantified.
Hernandez Perez, F.E.
2011-01-01
Hydrogen (H2) enrichment of hydrocarbon fuels in lean premixed systems is desirable since it can lead to a progressive reduction in greenhouse-gas emissions, while paving the way towards pure hydrogen combustion. In recent decades, large-eddy simulation (LES) has emerged as a promising tool to
Vreman, A.W.; Oijen, van J.A.; Goey, de L.P.H.; Bastiaans, R.J.M.
2009-01-01
Large-eddy simulation (LES) of turbulent combustion with premixed flamelets is investigated in this paper. The approach solves the filtered Navier-Stokes equations supplemented with two transport equations, one for the mixture fraction and another for a progress variable. The LES premixed flamelet
Large-eddy simulation with accurate implicit subgrid-scale diffusion
B. Koren (Barry); C. Beets
1996-01-01
textabstractA method for large-eddy simulation is presented that does not use an explicit subgrid-scale diffusion term. Subgrid-scale effects are modelled implicitly through an appropriate monotone (in the sense of Spekreijse 1987) discretization method for the advective terms. Special attention is
Investigation of wake interaction using full-scale lidar measurements and large eddy simulation
DEFF Research Database (Denmark)
Machefaux, Ewan; Larsen, Gunner Chr.; Troldborg, Niels
2016-01-01
dynamics flow solver, using large eddy simulation and fully turbulent inflow. The rotors are modelled using the actuator disc technique. A mutual validation of the computational fluid dynamics model with the measurements is conducted for a selected dataset, where wake interaction occurs. This validation...
Large shear deformation of particle gels studied by Brownian Dynamics simulations
Rzepiela, A.A.; Opheusden, van J.H.J.; Vliet, van T.
2004-01-01
Brownian Dynamics (BD) simulations have been performed to study structure and rheology of particle gels under large shear deformation. The model incorporates soft spherical particles, and reversible flexible bond formation. Two different methods of shear deformation are discussed, namely affine and
DEFF Research Database (Denmark)
Gebhardt, Cristian; Veluri, Badrinath; Preidikman, Sergio
2010-01-01
In this work an aeroelastic model that describes the interaction between aerodynamics and drivetrain dynamics of a large horizontal–axis wind turbine is presented. Traditional designs for wind turbines are based on the output of specific aeroelastic simulation codes. The output of these codes giv...
A large-signal dynamic simulation for the series resonant converter
King, R. J.; Stuart, T. A.
1983-01-01
A simple nonlinear discrete-time dynamic model for the series resonant dc-dc converter is derived using approximations appropriate to most power converters. This model is useful for the dynamic simulation of a series resonant converter using only a desktop calculator. The model is compared with a laboratory converter for a large transient event.
Largenet2: an object-oriented programming library for simulating large adaptive networks.
Zschaler, Gerd; Gross, Thilo
2013-01-15
The largenet2 C++ library provides an infrastructure for the simulation of large dynamic and adaptive networks with discrete node and link states. The library is released as free software. It is available at http://biond.github.com/largenet2. Largenet2 is licensed under the Creative Commons Attribution-NonCommercial 3.0 Unported License. gerd@biond.org
Fast Simulation of Large-Scale Floods Based on GPU Parallel Computing
Directory of Open Access Journals (Sweden)
Qiang Liu
2018-05-01
Full Text Available Computing speed is a significant issue of large-scale flood simulations for real-time response to disaster prevention and mitigation. Even today, most of the large-scale flood simulations are generally run on supercomputers due to the massive amounts of data and computations necessary. In this work, a two-dimensional shallow water model based on an unstructured Godunov-type finite volume scheme was proposed for flood simulation. To realize a fast simulation of large-scale floods on a personal computer, a Graphics Processing Unit (GPU-based, high-performance computing method using the OpenACC application was adopted to parallelize the shallow water model. An unstructured data management method was presented to control the data transportation between the GPU and CPU (Central Processing Unit with minimum overhead, and then both computation and data were offloaded from the CPU to the GPU, which exploited the computational capability of the GPU as much as possible. The parallel model was validated using various benchmarks and real-world case studies. The results demonstrate that speed-ups of up to one order of magnitude can be achieved in comparison with the serial model. The proposed parallel model provides a fast and reliable tool with which to quickly assess flood hazards in large-scale areas and, thus, has a bright application prospect for dynamic inundation risk identification and disaster assessment.
Large-eddy simulation of ethanol spray combustion using a finite-rate combustion model
Energy Technology Data Exchange (ETDEWEB)
Li, K.; Zhou, L.X. [Tsinghua Univ., Beijing (China). Dept. of Engineering Mechanics; Chan, C.K. [Hong Kong Polytechnic Univ. (China). Dept. of Applied Mathematics
2013-07-01
Large-eddy simulation of spray combustion is under its rapid development, but the combustion models are less validated by detailed experimental data. In this paper, large-eddy simulation of ethanol-air spray combustion was made using an Eulerian-Lagrangian approach, a subgrid-scale kinetic energy stress model, and a finite-rate combustion model. The simulation results are validated in detail by experiments. The LES obtained statistically averaged temperature is in agreement with the experimental results in most regions. The instantaneous LES results show the coherent structures of the shear region near the high-temperature flame zone and the fuel vapor concentration map, indicating the droplets are concentrated in this shear region. The droplet sizes are found to be in the range of 20-100{mu}m. The instantaneous temperature map shows the close interaction between the coherent structures and the combustion reaction.
Coupled large-eddy simulation of thermal mixing in a T-junction
International Nuclear Information System (INIS)
Kloeren, D.; Laurien, E.
2011-01-01
Analyzing thermal fatigue due to thermal mixing in T-junctions is part of the safety assessment of nuclear power plants. Results of two large-eddy simulations of mixing flow in a T-junction with coupled and adiabatic boundary condition are presented and compared. The temperature difference is set to 100 K, which leads to strong stratification of the flow. The main and the branch pipe intersect horizontally in this simulation. The flow is characterized by steady wavy pattern of stratification and temperature distribution. The coupled solution approach shows highly reduced temperature fluctuations in the near wall region due to thermal inertia of the wall. A conjugate heat transfer approach is necessary in order to simulate unsteady heat transfer accurately for large inlet temperature differences. (author)
Large-eddy simulation of a turbulent piloted methane/air diffusion flame (Sandia flame D)
International Nuclear Information System (INIS)
Pitsch, H.; Steiner, H.
2000-01-01
The Lagrangian Flamelet Model is formulated as a combustion model for large-eddy simulations of turbulent jet diffusion flames. The model is applied in a large-eddy simulation of a piloted partially premixed methane/air diffusion flame (Sandia flame D). The results of the simulation are compared to experimental data of the mean and RMS of the axial velocity and the mixture fraction and the unconditional and conditional averages of temperature and various species mass fractions, including CO and NO. All quantities are in good agreement with the experiments. The results indicate in accordance with experimental findings that regions of high strain appear in layer like structures, which are directed inwards and tend to align with the reaction zone, where the turbulence is fully developed. The analysis of the conditional temperature and mass fractions reveals a strong influence of the partial premixing of the fuel. (c) 2000 American Institute of Physics
Large-scale particle simulations in a virtual-memory computer
International Nuclear Information System (INIS)
Gray, P.C.; Wagner, J.S.; Tajima, T.; Million, R.
1982-08-01
Virtual memory computers are capable of executing large-scale particle simulations even when the memory requirements exceed the computer core size. The required address space is automatically mapped onto slow disc memory by the operating system. When the simulation size is very large, frequent random accesses to slow memory occur during the charge accumulation and particle pushing processes. Accesses to slow memory significantly reduce the execution rate of the simulation. We demonstrate in this paper that with the proper choice of sorting algorithm, a nominal amount of sorting to keep physically adjacent particles near particles with neighboring array indices can reduce random access to slow memory, increase the efficiency of the I/O system, and hence, reduce the required computing time
Tang, G.; Bartlein, P. J.
2012-01-01
Water balance models of simple structure are easier to grasp and more clearly connect cause and effect than models of complex structure. Such models are essential for studying large spatial scale land surface water balance in the context of climate and land cover change, both natural and anthropogenic. This study aims to (i) develop a large spatial scale water balance model by modifying a dynamic global vegetation model (DGVM), and (ii) test the model's performance in simulating actual evapotranspiration (ET), soil moisture and surface runoff for the coterminous United States (US). Toward these ends, we first introduced development of the "LPJ-Hydrology" (LH) model by incorporating satellite-based land covers into the Lund-Potsdam-Jena (LPJ) DGVM instead of dynamically simulating them. We then ran LH using historical (1982-2006) climate data and satellite-based land covers at 2.5 arc-min grid cells. The simulated ET, soil moisture and surface runoff were compared to existing sets of observed or simulated data for the US. The results indicated that LH captures well the variation of monthly actual ET (R2 = 0.61, p 0.46, p 0.52) with observed values over the years 1982-2006, respectively. The modeled spatial patterns of annual ET and surface runoff are in accordance with previously published data. Compared to its predecessor, LH simulates better monthly stream flow in winter and early spring by incorporating effects of solar radiation on snowmelt. Overall, this study proves the feasibility of incorporating satellite-based land-covers into a DGVM for simulating large spatial scale land surface water balance. LH developed in this study should be a useful tool for studying effects of climate and land cover change on land surface hydrology at large spatial scales.
Parallel Motion Simulation of Large-Scale Real-Time Crowd in a Hierarchical Environmental Model
Directory of Open Access Journals (Sweden)
Xin Wang
2012-01-01
Full Text Available This paper presents a parallel real-time crowd simulation method based on a hierarchical environmental model. A dynamical model of the complex environment should be constructed to simulate the state transition and propagation of individual motions. By modeling of a virtual environment where virtual crowds reside, we employ different parallel methods on a topological layer, a path layer and a perceptual layer. We propose a parallel motion path matching method based on the path layer and a parallel crowd simulation method based on the perceptual layer. The large-scale real-time crowd simulation becomes possible with these methods. Numerical experiments are carried out to demonstrate the methods and results.
International Nuclear Information System (INIS)
Jacques, D.; Perko, J.; Seetharam, S.; Mallants, D.
2012-01-01
This paper presents a methodology to assess the spatial-temporal evolution of chemical degradation fronts in real-size concrete structures typical of a near-surface radioactive waste disposal facility. The methodology consists of the abstraction of a so-called full (complicated) model accounting for the multicomponent - multi-scale nature of concrete to an abstracted (simplified) model which simulates chemical concrete degradation based on a single component in the aqueous and solid phase. The abstracted model is verified against chemical degradation fronts simulated with the full model under both diffusive and advective transport conditions. Implementation in the multi-physics simulation tool COMSOL allows simulation of the spatial-temporal evolution of chemical degradation fronts in large-scale concrete structures. (authors)
A generative model for scientific concept hierarchies.
Datta, Srayan; Adar, Eytan
2018-01-01
In many scientific disciplines, each new 'product' of research (method, finding, artifact, etc.) is often built upon previous findings-leading to extension and branching of scientific concepts over time. We aim to understand the evolution of scientific concepts by placing them in phylogenetic hierarchies where scientific keyphrases from a large, longitudinal academic corpora are used as a proxy of scientific concepts. These hierarchies exhibit various important properties, including power-law degree distribution, power-law component size distribution, existence of a giant component and less probability of extending an older concept. We present a generative model based on preferential attachment to simulate the graphical and temporal properties of these hierarchies which helps us understand the underlying process behind scientific concept evolution and may be useful in simulating and predicting scientific evolution.
A generative model for scientific concept hierarchies
Adar, Eytan
2018-01-01
In many scientific disciplines, each new ‘product’ of research (method, finding, artifact, etc.) is often built upon previous findings–leading to extension and branching of scientific concepts over time. We aim to understand the evolution of scientific concepts by placing them in phylogenetic hierarchies where scientific keyphrases from a large, longitudinal academic corpora are used as a proxy of scientific concepts. These hierarchies exhibit various important properties, including power-law degree distribution, power-law component size distribution, existence of a giant component and less probability of extending an older concept. We present a generative model based on preferential attachment to simulate the graphical and temporal properties of these hierarchies which helps us understand the underlying process behind scientific concept evolution and may be useful in simulating and predicting scientific evolution. PMID:29474409
Energy Technology Data Exchange (ETDEWEB)
Wu Hong [National Key Laboratory of Science and Technology on Aero-Engine Aero-Thermodynamics, Beihang University, Beijing 100191 (China); Wang Jiao, E-mail: wangjiao@sjp.buaa.edu.cn [National Key Laboratory of Science and Technology on Aero-Engine Aero-Thermodynamics, Beihang University, Beijing 100191 (China); Tao Zhi [National Key Laboratory of Science and Technology on Aero-Engine Aero-Thermodynamics, Beihang University, Beijing 100191 (China)
2011-12-15
Highlights: Black-Right-Pointing-Pointer A double MRT-LBM is used to study heat transfer in turbulent channel flow. Black-Right-Pointing-Pointer Turbulent Pr is modeled by dynamic subgrid scale model. Black-Right-Pointing-Pointer Temperature gradients are calculated by the non-equilibrium temperature distribution moments. - Abstract: In this paper, a large eddy simulation based on the lattice Boltzmann framework is carried out to simulate the heat transfer in a turbulent channel flow, in which the temperature can be regarded as a passive scalar. A double multiple relaxation time (DMRT) thermal lattice Boltzmann model is employed. While applying DMRT, a multiple relaxation time D3Q19 model is used to simulate the flow field, and a multiple relaxation time D3Q7 model is used to simulate the temperature field. The dynamic subgrid stress model, in which the turbulent eddy viscosity and the turbulent Prandtl number are dynamically computed, is integrated to describe the subgrid effect. Not only the strain rate but also the temperature gradient is calculated locally by the non-equilibrium moments. The Reynolds number based on the shear velocity and channel half height is 180. The molecular Prandtl numbers are set to be 0.025 and 0.71. Statistical quantities, such as the average velocity, average temperature, Reynolds stress, root mean square (RMS) velocity fluctuations, RMS temperature and turbulent heat flux are obtained and compared with the available data. The results demonstrate great reliability of DMRT-LES in studying turbulence.
Heavy-ion collimation at the Large Hadron Collider. Simulations and measurements
International Nuclear Information System (INIS)
Hermes, Pascal Dominik
2016-01-01
The CERN Large Hadron Collider (LHC) stores and collides proton and 208 Pb 82+ beams of unprecedented energy and intensity. Thousands of superconducting magnets, operated at 1.9 K, guide the very intense and energetic particle beams, which have a large potential for destruction. This implies the demand for a multi-stage collimation system to provide protection from beam-induced quenches or even hardware damage. In heavy-ion operation, ion fragments with significant rigidity offsets can still scatter out of the collimation system. When they irradiate the superconducting LHC magnets, the latter risk to quench (lose their superconducting property). These secondary collimation losses can potentially impose a limitation for the stored heavy-ion beam energy. Therefore, their distribution in the LHC needs to be understood by sophisticated simulations. Such simulation tools must accurately simulate the particle motion of many different nuclides in the magnetic LHC lattice and simulate their interaction with the collimators. Previous simulation tools used simplified models for the simulation of particle-matter interaction and showed discrepancies compared to the measured loss patterns. This thesis describes the development and application of improved heavy-ion collimation simulation tools. Two different approaches are presented to provide these functionalities. In the first presented tool, called STIER, fragmentation at the primary collimator is simulated with the Monte-Carlo event generator FLUKA. The ion fragments scattered out of the primary collimator are subsequently tracked as protons with ion-equivalent rigidities in the existing proton tracking tool SixTrack. This approach was used to prepare the collimator settings for the 2015 LHC heavy-ion run and its predictions allowed reducing undesired losses. More accurate simulation results are obtained with the second presented simulation tool, in which SixTrack is extended to track arbitrary heavy ions. This new tracking
Heavy-ion collimation at the Large Hadron Collider. Simulations and measurements
Energy Technology Data Exchange (ETDEWEB)
Hermes, Pascal Dominik
2016-12-19
The CERN Large Hadron Collider (LHC) stores and collides proton and {sup 208}Pb{sup 82+} beams of unprecedented energy and intensity. Thousands of superconducting magnets, operated at 1.9 K, guide the very intense and energetic particle beams, which have a large potential for destruction. This implies the demand for a multi-stage collimation system to provide protection from beam-induced quenches or even hardware damage. In heavy-ion operation, ion fragments with significant rigidity offsets can still scatter out of the collimation system. When they irradiate the superconducting LHC magnets, the latter risk to quench (lose their superconducting property). These secondary collimation losses can potentially impose a limitation for the stored heavy-ion beam energy. Therefore, their distribution in the LHC needs to be understood by sophisticated simulations. Such simulation tools must accurately simulate the particle motion of many different nuclides in the magnetic LHC lattice and simulate their interaction with the collimators. Previous simulation tools used simplified models for the simulation of particle-matter interaction and showed discrepancies compared to the measured loss patterns. This thesis describes the development and application of improved heavy-ion collimation simulation tools. Two different approaches are presented to provide these functionalities. In the first presented tool, called STIER, fragmentation at the primary collimator is simulated with the Monte-Carlo event generator FLUKA. The ion fragments scattered out of the primary collimator are subsequently tracked as protons with ion-equivalent rigidities in the existing proton tracking tool SixTrack. This approach was used to prepare the collimator settings for the 2015 LHC heavy-ion run and its predictions allowed reducing undesired losses. More accurate simulation results are obtained with the second presented simulation tool, in which SixTrack is extended to track arbitrary heavy ions. This new
Developments and validation of large eddy simulation of turbulent flows in an industrial code
International Nuclear Information System (INIS)
Ackermann, C.
2000-01-01
Large Eddy Simulation, where large scales of the flow are resolved and sub-grid scales are modelled, is well adapted to the study of turbulent flow, in which geometry and/or heat transfer effects lead to unsteady phenomena. To obtain an improved numerical tool, simulations of elementary test cases, Homogeneous Isotropic Turbulence and Turbulent Plane Channel, were clone on both structured and unstructured grids, before moving to more complex geometries. This allowed the influence of the different physical and numerical parameters to be studied separately. On structured grids, the different properties of the numerical methods corresponding to our problem were identified, a new sub-grid model was elaborated and several laws of the wall tested: for this discretization, our numerical tool is yet validated. On unstructured grids, the construction of numerical methods with the same properties as on the structured grids is harder, especially for the convection scheme: several numerical schemes were tested, and sub-grid models and laws of the wall were adapted to unstructured grids. Simulations of the same elementary tests were clone: the results are relatively satisfactorily, even if they are not so good as the one obtained in structured grids, most probably because the numerical methods chosen cannot perfectly isolate the effects between the convection scheme, physical modelling and the mesh chosen. This work is the first stage towards the development of a practical Large Eddy Simulation tool for unstructured grid. (author) [fr
Cerebral methodology based computing to estimate real phenomena from large-scale nuclear simulation
International Nuclear Information System (INIS)
Suzuki, Yoshio
2011-01-01
Our final goal is to estimate real phenomena from large-scale nuclear simulations by using computing processes. Large-scale simulations mean that they include scale variety and physical complexity so that corresponding experiments and/or theories do not exist. In nuclear field, it is indispensable to estimate real phenomena from simulations in order to improve the safety and security of nuclear power plants. Here, the analysis of uncertainty included in simulations is needed to reveal sensitivity of uncertainty due to randomness, to reduce the uncertainty due to lack of knowledge and to lead a degree of certainty by verification and validation (V and V) and uncertainty quantification (UQ) processes. To realize this, we propose 'Cerebral Methodology based Computing (CMC)' as computing processes with deductive and inductive approaches by referring human reasoning processes. Our idea is to execute deductive and inductive simulations contrasted with deductive and inductive approaches. We have established its prototype system and applied it to a thermal displacement analysis of a nuclear power plant. The result shows that our idea is effective to reduce the uncertainty and to get the degree of certainty. (author)
Thermal large Eddy simulations and experiments in the framework of non-isothermal blowing
International Nuclear Information System (INIS)
Brillant, G.
2004-06-01
The aim of this work is to study thermal large-eddy simulations and to determine the nonisothermal blowing impact on a turbulent boundary layer. An experimental study is also carried out in order to complete and validate simulation results. In a first time, we developed a turbulent inlet condition for the velocity and the temperature, which is necessary for the blowing simulations.We studied the asymptotic behavior of the velocity, the temperature and the thermal turbulent fluxes in a large-eddy simulation point of view. We then considered dynamics models for the eddy-diffusivity and we simulated a turbulent channel flow with imposed temperature, imposed flux and adiabatic walls. The numerical and experimental study of blowing permitted to obtain to the modifications of a thermal turbulent boundary layer with the blowing rate. We observed the consequences of the blowing on mean and rms profiles of velocity and temperature but also on velocity-velocity and velocity-temperature correlations. Moreover, we noticed an increase of the turbulent structures in the boundary layer with blowing. (author)
Three-dimensional two-fluid Braginskii simulations of the large plasma device
Energy Technology Data Exchange (ETDEWEB)
Fisher, Dustin M., E-mail: dustin.m.fisher.gr@dartmouth.edu; Rogers, Barrett N., E-mail: barrett.rogers@dartmouth.edu [Department of Physics and Astronomy, Dartmouth College, Hanover, New Hampshire 03755 (United States); Rossi, Giovanni D.; Guice, Daniel S.; Carter, Troy A. [Department of Physics and Astronomy, University of California, Los Angeles, California 90095-1547 (United States)
2015-09-15
The Large Plasma Device (LAPD) is modeled using the 3D Global Braginskii Solver code. Comparisons to experimental measurements are made in the low-bias regime in which there is an intrinsic E × B rotation of the plasma. In the simulations, this rotation is caused primarily by sheath effects and may be a likely mechanism for the intrinsic rotation seen in LAPD. Simulations show strong qualitative agreement with the data, particularly the radial dependence of the density fluctuations, cross-correlation lengths, radial flux dependence outside of the cathode edge, and camera imagery. Kelvin Helmholtz (KH) turbulence at relatively large scales is the dominant driver of cross-field transport in these simulations with smaller-scale drift waves and sheath modes playing a secondary role. Plasma holes and blobs arising from KH vortices in the simulations are consistent with the scale sizes and overall appearance of those in LAPD camera images. The addition of ion-neutral collisions in the simulations at previously theorized values reduces the radial particle flux by about a factor of two, from values that are somewhat larger than the experimentally measured flux to values that are somewhat lower than the measurements. This reduction is due to a modest stabilizing contribution of the collisions on the KH-modes driving the turbulent transport.
Directory of Open Access Journals (Sweden)
Lorenzo L. Pesce
2013-01-01
Full Text Available Our limited understanding of the relationship between the behavior of individual neurons and large neuronal networks is an important limitation in current epilepsy research and may be one of the main causes of our inadequate ability to treat it. Addressing this problem directly via experiments is impossibly complex; thus, we have been developing and studying medium-large-scale simulations of detailed neuronal networks to guide us. Flexibility in the connection schemas and a complete description of the cortical tissue seem necessary for this purpose. In this paper we examine some of the basic issues encountered in these multiscale simulations. We have determined the detailed behavior of two such simulators on parallel computer systems. The observed memory and computation-time scaling behavior for a distributed memory implementation were very good over the range studied, both in terms of network sizes (2,000 to 400,000 neurons and processor pool sizes (1 to 256 processors. Our simulations required between a few megabytes and about 150 gigabytes of RAM and lasted between a few minutes and about a week, well within the capability of most multinode clusters. Therefore, simulations of epileptic seizures on networks with millions of cells should be feasible on current supercomputers.
Pesce, Lorenzo L; Lee, Hyong C; Hereld, Mark; Visser, Sid; Stevens, Rick L; Wildeman, Albert; van Drongelen, Wim
2013-01-01
Our limited understanding of the relationship between the behavior of individual neurons and large neuronal networks is an important limitation in current epilepsy research and may be one of the main causes of our inadequate ability to treat it. Addressing this problem directly via experiments is impossibly complex; thus, we have been developing and studying medium-large-scale simulations of detailed neuronal networks to guide us. Flexibility in the connection schemas and a complete description of the cortical tissue seem necessary for this purpose. In this paper we examine some of the basic issues encountered in these multiscale simulations. We have determined the detailed behavior of two such simulators on parallel computer systems. The observed memory and computation-time scaling behavior for a distributed memory implementation were very good over the range studied, both in terms of network sizes (2,000 to 400,000 neurons) and processor pool sizes (1 to 256 processors). Our simulations required between a few megabytes and about 150 gigabytes of RAM and lasted between a few minutes and about a week, well within the capability of most multinode clusters. Therefore, simulations of epileptic seizures on networks with millions of cells should be feasible on current supercomputers.
Anatomically detailed and large-scale simulations studying synapse loss and synchrony using NeuroBox
Directory of Open Access Journals (Sweden)
Markus eBreit
2016-02-01
Full Text Available The morphology of neurons and networks plays an important role in processing electrical and biochemical signals. Based on neuronal reconstructions, which are becoming abundantly available through databases such as NeuroMorpho.org, numerical simulations of Hodgkin-Huxley-type equations, coupled to biochemical models, can be performed in order to systematically investigate the influence of cellular morphology and the connectivity pattern in networks on the underlying function. Development in the area of synthetic neural network generation and morphology reconstruction from microscopy data has brought forth the software tool NeuGen. Coupling this morphology data (either from databases, synthetic or reconstruction to the simulation platform UG 4 (which harbors a neuroscientific portfolio and VRL-Studio, has brought forth the extendible toolbox NeuroBox. NeuroBox allows users to perform numerical simulations on hybrid-dimensional morphology representations. The code basis is designed in a modular way, such that e.g. new channel or synapse types can be added to the library. Workflows can be specified through scripts or through the VRL-Studio graphical workflow representation. Third-party tools, such as ImageJ, can be added to NeuroBox workflows. In this paper, NeuroBox is used to study the electrical and biochemical effects of synapse loss vs. synchrony in neurons, to investigate large morphology data sets within detailed biophysical simulations, and used to demonstrate the capability of utilizing high-performance computing infrastructure for large scale network simulations. Using new synapse distribution methods and Finite Volume based numerical solvers for compartment-type models, our results demonstrate how an increase in synaptic synchronization can compensate synapse loss at the electrical and calcium level, and how detailed neuronal morphology can be integrated in large-scale network simulations.
Hybrid Reynolds-Averaged/Large-Eddy Simulations of a Coaxial Supersonic Free-Jet Experiment
Baurle, Robert A.; Edwards, Jack R.
2010-01-01
Reynolds-averaged and hybrid Reynolds-averaged/large-eddy simulations have been applied to a supersonic coaxial jet flow experiment. The experiment was designed to study compressible mixing flow phenomenon under conditions that are representative of those encountered in scramjet combustors. The experiment utilized either helium or argon as the inner jet nozzle fluid, and the outer jet nozzle fluid consisted of laboratory air. The inner and outer nozzles were designed and operated to produce nearly pressure-matched Mach 1.8 flow conditions at the jet exit. The purpose of the computational effort was to assess the state-of-the-art for each modeling approach, and to use the hybrid Reynolds-averaged/large-eddy simulations to gather insight into the deficiencies of the Reynolds-averaged closure models. The Reynolds-averaged simulations displayed a strong sensitivity to choice of turbulent Schmidt number. The initial value chosen for this parameter resulted in an over-prediction of the mixing layer spreading rate for the helium case, but the opposite trend was observed when argon was used as the injectant. A larger turbulent Schmidt number greatly improved the comparison of the results with measurements for the helium simulations, but variations in the Schmidt number did not improve the argon comparisons. The hybrid Reynolds-averaged/large-eddy simulations also over-predicted the mixing layer spreading rate for the helium case, while under-predicting the rate of mixing when argon was used as the injectant. The primary reason conjectured for the discrepancy between the hybrid simulation results and the measurements centered around issues related to the transition from a Reynolds-averaged state to one with resolved turbulent content. Improvements to the inflow conditions were suggested as a remedy to this dilemma. Second-order turbulence statistics were also compared to their modeled Reynolds-averaged counterparts to evaluate the effectiveness of common turbulence closure
A regularized vortex-particle mesh method for large eddy simulation
DEFF Research Database (Denmark)
Spietz, Henrik Juul; Walther, Jens Honore; Hejlesen, Mads Mølholm
We present recent developments of the remeshed vortex particle-mesh method for simulating incompressible ﬂuid ﬂow. The presented method relies on a parallel higher-order FFT based solver for the Poisson equation. Arbitrary high order is achieved through regularization of singular Green’s function...... solutions to the Poisson equation and recently we have derived novel high order solutions for a mixture of open and periodic domains. With this approach the simulated variables may formally be viewed as the approximate solution to the ﬁltered Navier Stokes equations, hence we use the method for Large Eddy...
A simple atmospheric boundary layer model applied to large eddy simulations of wind turbine wakes
DEFF Research Database (Denmark)
Troldborg, Niels; Sørensen, Jens Nørkær; Mikkelsen, Robert Flemming
2014-01-01
A simple model for including the influence of the atmospheric boundary layer in connection with large eddy simulations of wind turbine wakes is presented and validated by comparing computed results with measurements as well as with direct numerical simulations. The model is based on an immersed...... boundary type technique where volume forces are used to introduce wind shear and atmospheric turbulence. The application of the model for wake studies is demonstrated by combining it with the actuator line method, and predictions are compared with field measurements. Copyright © 2013 John Wiley & Sons, Ltd....
Crystallisation of a Lennard-Jones fluid by large scale molecular dynamics simulation
International Nuclear Information System (INIS)
Snook, I.
1998-01-01
Full text: The evolution of the structure of a large system of atoms interacting via a Lennard-Jones pair potential was simulated by the use of the Molecular Dynamics computer simulation technique. The system was initially equilibrated in the one phase region of the phase diagram at a temperature above critical then a temperature quench was performed which placed the system in a region were the single fluid phase was unstable. Quenches to below the triple point temperature gave rise to crystallisation The mechanism and final morphology is shown to depend strongly on the starting conditions e.g. the starting density
Large Eddy Simulation of Flows Associated with Offshore Oil and Gas Pipeline
Directory of Open Access Journals (Sweden)
Nizamani Z.
2017-01-01
Full Text Available Fluid structure interaction (FSI applications are of wide range from offshore fixed and floating structures to offshore pipelines. Reynolds Averaged Navier Stoke (RANS solution has limitation for unsteady and turbulent flow modelling. A possible approach is Large Eddy Simulation (LES and it is applied to flows past a circular cylinder located far above, near and on a flat seabed. The Reynolds number considered is based on the real situation off Malaysia Coast and is sub-critical around 105. Hydrodynamic quantities in terms of mean pressure are predicted and vortex shedding mechanism is evaluated. The results are validated by comparing the simulation and experimental previous studies.
Simulation test of PIUS-type reactor with large scale experimental apparatus
International Nuclear Information System (INIS)
Tamaki, M.; Tsuji, Y.; Ito, T.; Tasaka, K.; Kukita, Yutaka
1995-01-01
A large scale experimental apparatus for simulating the PIUS-type reactor has been constructed keeping the volumetric scaling ratio to the realistic reactor model. Fundamental experiments such as a steady state operation and a pump trip simulation were performed. Experimental results were compared with those obtained by the small scale apparatus in JAERI. We have already reported the effectiveness of the feedback control for the primary loop pump speed (PI control) for the stable operation. In this paper this feedback system is modified and the PID control is introduced. This new system worked well for the operation of the PIUS-type reactor even in a rapid transient condition. (author)
Mizell, Carolyn; Malone, Linda
2007-01-01
It is very difficult for project managers to develop accurate cost and schedule estimates for large, complex software development projects. None of the approaches or tools available today can estimate the true cost of software with any high degree of accuracy early in a project. This paper provides an approach that utilizes a software development process simulation model that considers and conveys the level of uncertainty that exists when developing an initial estimate. A NASA project will be analyzed using simulation and data from the Software Engineering Laboratory to show the benefits of such an approach.
Directory of Open Access Journals (Sweden)
C. M. R. Mateo
2017-10-01
Full Text Available Global-scale river models (GRMs are core tools for providing consistent estimates of global flood hazard, especially in data-scarce regions. Due to former limitations in computational power and input datasets, most GRMs have been developed to use simplified representations of flow physics and run at coarse spatial resolutions. With increasing computational power and improved datasets, the application of GRMs to finer resolutions is becoming a reality. To support development in this direction, the suitability of GRMs for application to finer resolutions needs to be assessed. This study investigates the impacts of spatial resolution and flow connectivity representation on the predictive capability of a GRM, CaMa-Flood, in simulating the 2011 extreme flood in Thailand. Analyses show that when single downstream connectivity (SDC is assumed, simulation results deteriorate with finer spatial resolution; Nash–Sutcliffe efficiency coefficients decreased by more than 50 % between simulation results at 10 km resolution and 1 km resolution. When multiple downstream connectivity (MDC is represented, simulation results slightly improve with finer spatial resolution. The SDC simulations result in excessive backflows on very flat floodplains due to the restrictive flow directions at finer resolutions. MDC channels attenuated these effects by maintaining flow connectivity and flow capacity between floodplains in varying spatial resolutions. While a regional-scale flood was chosen as a test case, these findings should be universal and may have significant impacts on large- to global-scale simulations, especially in regions where mega deltas exist.These results demonstrate that a GRM can be used for higher resolution simulations of large-scale floods, provided that MDC in rivers and floodplains is adequately represented in the model structure.
Mateo, Cherry May R.; Yamazaki, Dai; Kim, Hyungjun; Champathong, Adisorn; Vaze, Jai; Oki, Taikan
2017-10-01
Global-scale river models (GRMs) are core tools for providing consistent estimates of global flood hazard, especially in data-scarce regions. Due to former limitations in computational power and input datasets, most GRMs have been developed to use simplified representations of flow physics and run at coarse spatial resolutions. With increasing computational power and improved datasets, the application of GRMs to finer resolutions is becoming a reality. To support development in this direction, the suitability of GRMs for application to finer resolutions needs to be assessed. This study investigates the impacts of spatial resolution and flow connectivity representation on the predictive capability of a GRM, CaMa-Flood, in simulating the 2011 extreme flood in Thailand. Analyses show that when single downstream connectivity (SDC) is assumed, simulation results deteriorate with finer spatial resolution; Nash-Sutcliffe efficiency coefficients decreased by more than 50 % between simulation results at 10 km resolution and 1 km resolution. When multiple downstream connectivity (MDC) is represented, simulation results slightly improve with finer spatial resolution. The SDC simulations result in excessive backflows on very flat floodplains due to the restrictive flow directions at finer resolutions. MDC channels attenuated these effects by maintaining flow connectivity and flow capacity between floodplains in varying spatial resolutions. While a regional-scale flood was chosen as a test case, these findings should be universal and may have significant impacts on large- to global-scale simulations, especially in regions where mega deltas exist.These results demonstrate that a GRM can be used for higher resolution simulations of large-scale floods, provided that MDC in rivers and floodplains is adequately represented in the model structure.
High-resolution global climate modelling: the UPSCALE project, a large-simulation campaign
Directory of Open Access Journals (Sweden)
M. S. Mizielinski
2014-08-01
Full Text Available The UPSCALE (UK on PRACE: weather-resolving Simulations of Climate for globAL Environmental risk project constructed and ran an ensemble of HadGEM3 (Hadley Centre Global Environment Model 3 atmosphere-only global climate simulations over the period 1985–2011, at resolutions of N512 (25 km, N216 (60 km and N96 (130 km as used in current global weather forecasting, seasonal prediction and climate modelling respectively. Alongside these present climate simulations a parallel ensemble looking at extremes of future climate was run, using a time-slice methodology to consider conditions at the end of this century. These simulations were primarily performed using a 144 million core hour, single year grant of computing time from PRACE (the Partnership for Advanced Computing in Europe in 2012, with additional resources supplied by the Natural Environment Research Council (NERC and the Met Office. Almost 400 terabytes of simulation data were generated on the HERMIT supercomputer at the High Performance Computing Center Stuttgart (HLRS, and transferred to the JASMIN super-data cluster provided by the Science and Technology Facilities Council Centre for Data Archival (STFC CEDA for analysis and storage. In this paper we describe the implementation of the project, present the technical challenges in terms of optimisation, data output, transfer and storage that such a project involves and include details of the model configuration and the composition of the UPSCALE data set. This data set is available for scientific analysis to allow assessment of the value of model resolution in both present and potential future climate conditions.
Torner, Benjamin; Konnigk, Lucas; Hallier, Sebastian; Kumar, Jitendra; Witte, Matthias; Wurm, Frank-Hendrik
2018-06-01
Numerical flow analysis (computational fluid dynamics) in combination with the prediction of blood damage is an important procedure to investigate the hemocompatibility of a blood pump, since blood trauma due to shear stresses remains a problem in these devices. Today, the numerical damage prediction is conducted using unsteady Reynolds-averaged Navier-Stokes simulations. Investigations with large eddy simulations are rarely being performed for blood pumps. Hence, the aim of the study is to examine the viscous shear stresses of a large eddy simulation in a blood pump and compare the results with an unsteady Reynolds-averaged Navier-Stokes simulation. The simulations were carried out at two operation points of a blood pump. The flow was simulated on a 100M element mesh for the large eddy simulation and a 20M element mesh for the unsteady Reynolds-averaged Navier-Stokes simulation. As a first step, the large eddy simulation was verified by analyzing internal dissipative losses within the pump. Then, the pump characteristics and mean and turbulent viscous shear stresses were compared between the two simulation methods. The verification showed that the large eddy simulation is able to reproduce the significant portion of dissipative losses, which is a global indication that the equivalent viscous shear stresses are adequately resolved. The comparison with the unsteady Reynolds-averaged Navier-Stokes simulation revealed that the hydraulic parameters were in agreement, but differences for the shear stresses were found. The results show the potential of the large eddy simulation as a high-quality comparative case to check the suitability of a chosen Reynolds-averaged Navier-Stokes setup and turbulence model. Furthermore, the results lead to suggest that large eddy simulations are superior to unsteady Reynolds-averaged Navier-Stokes simulations when instantaneous stresses are applied for the blood damage prediction.
Hybrid Large-Eddy/Reynolds-Averaged Simulation of a Supersonic Cavity Using VULCAN
Quinlan, Jesse; McDaniel, James; Baurle, Robert A.
2013-01-01
Simulations of a supersonic recessed-cavity flow are performed using a hybrid large-eddy/Reynolds-averaged simulation approach utilizing an inflow turbulence recycling procedure and hybridized inviscid flux scheme. Calorically perfect air enters a three-dimensional domain at a free stream Mach number of 2.92. Simulations are performed to assess grid sensitivity of the solution, efficacy of the turbulence recycling, and the effect of the shock sensor used with the hybridized inviscid flux scheme. Analysis of the turbulent boundary layer upstream of the rearward-facing step for each case indicates excellent agreement with theoretical predictions. Mean velocity and pressure results are compared to Reynolds-averaged simulations and experimental data for each case and indicate good agreement on the finest grid. Simulations are repeated on a coarsened grid, and results indicate strong grid density sensitivity. Simulations are performed with and without inflow turbulence recycling on the coarse grid to isolate the effect of the recycling procedure, which is demonstrably critical to capturing the relevant shear layer dynamics. Shock sensor formulations of Ducros and Larsson are found to predict mean flow statistics equally well.
A Parallel, Finite-Volume Algorithm for Large-Eddy Simulation of Turbulent Flows
Bui, Trong T.
1999-01-01
A parallel, finite-volume algorithm has been developed for large-eddy simulation (LES) of compressible turbulent flows. This algorithm includes piecewise linear least-square reconstruction, trilinear finite-element interpolation, Roe flux-difference splitting, and second-order MacCormack time marching. Parallel implementation is done using the message-passing programming model. In this paper, the numerical algorithm is described. To validate the numerical method for turbulence simulation, LES of fully developed turbulent flow in a square duct is performed for a Reynolds number of 320 based on the average friction velocity and the hydraulic diameter of the duct. Direct numerical simulation (DNS) results are available for this test case, and the accuracy of this algorithm for turbulence simulations can be ascertained by comparing the LES solutions with the DNS results. The effects of grid resolution, upwind numerical dissipation, and subgrid-scale dissipation on the accuracy of the LES are examined. Comparison with DNS results shows that the standard Roe flux-difference splitting dissipation adversely affects the accuracy of the turbulence simulation. For accurate turbulence simulations, only 3-5 percent of the standard Roe flux-difference splitting dissipation is needed.
Large-scale simulations of error-prone quantum computation devices
International Nuclear Information System (INIS)
Trieu, Doan Binh
2009-01-01
The theoretical concepts of quantum computation in the idealized and undisturbed case are well understood. However, in practice, all quantum computation devices do suffer from decoherence effects as well as from operational imprecisions. This work assesses the power of error-prone quantum computation devices using large-scale numerical simulations on parallel supercomputers. We present the Juelich Massively Parallel Ideal Quantum Computer Simulator (JUMPIQCS), that simulates a generic quantum computer on gate level. It comprises an error model for decoherence and operational errors. The robustness of various algorithms in the presence of noise has been analyzed. The simulation results show that for large system sizes and long computations it is imperative to actively correct errors by means of quantum error correction. We implemented the 5-, 7-, and 9-qubit quantum error correction codes. Our simulations confirm that using error-prone correction circuits with non-fault-tolerant quantum error correction will always fail, because more errors are introduced than being corrected. Fault-tolerant methods can overcome this problem, provided that the single qubit error rate is below a certain threshold. We incorporated fault-tolerant quantum error correction techniques into JUMPIQCS using Steane's 7-qubit code and determined this threshold numerically. Using the depolarizing channel as the source of decoherence, we find a threshold error rate of (5.2±0.2) x 10 -6 . For Gaussian distributed operational over-rotations the threshold lies at a standard deviation of 0.0431±0.0002. We can conclude that quantum error correction is especially well suited for the correction of operational imprecisions and systematic over-rotations. For realistic simulations of specific quantum computation devices we need to extend the generic model to dynamic simulations, i.e. time-dependent Hamiltonian simulations of realistic hardware models. We focus on today's most advanced technology, i
Large-scale simulations of error-prone quantum computation devices
Energy Technology Data Exchange (ETDEWEB)
Trieu, Doan Binh
2009-07-01
The theoretical concepts of quantum computation in the idealized and undisturbed case are well understood. However, in practice, all quantum computation devices do suffer from decoherence effects as well as from operational imprecisions. This work assesses the power of error-prone quantum computation devices using large-scale numerical simulations on parallel supercomputers. We present the Juelich Massively Parallel Ideal Quantum Computer Simulator (JUMPIQCS), that simulates a generic quantum computer on gate level. It comprises an error model for decoherence and operational errors. The robustness of various algorithms in the presence of noise has been analyzed. The simulation results show that for large system sizes and long computations it is imperative to actively correct errors by means of quantum error correction. We implemented the 5-, 7-, and 9-qubit quantum error correction codes. Our simulations confirm that using error-prone correction circuits with non-fault-tolerant quantum error correction will always fail, because more errors are introduced than being corrected. Fault-tolerant methods can overcome this problem, provided that the single qubit error rate is below a certain threshold. We incorporated fault-tolerant quantum error correction techniques into JUMPIQCS using Steane's 7-qubit code and determined this threshold numerically. Using the depolarizing channel as the source of decoherence, we find a threshold error rate of (5.2{+-}0.2) x 10{sup -6}. For Gaussian distributed operational over-rotations the threshold lies at a standard deviation of 0.0431{+-}0.0002. We can conclude that quantum error correction is especially well suited for the correction of operational imprecisions and systematic over-rotations. For realistic simulations of specific quantum computation devices we need to extend the generic model to dynamic simulations, i.e. time-dependent Hamiltonian simulations of realistic hardware models. We focus on today's most advanced
Orgel, Csilla; Kereszturi, Ákos; Váczi, Tamás; Groemer, Gernot; Sattler, Birgit
2014-02-01
Between 15 and 25 April 2011 in the framework of the PolAres programme of the Austrian Space Forum, a five-day field test of the Aouda.X spacesuit simulator was conducted at the Rio Tinto Mars-analogue site in southern Spain. The field crew was supported by a full-scale Mission Control Center (MCC) in Innsbruck, Austria. The field telemetry data were relayed to the MCC, enabling a Remote Science Support (RSS) team to study field data in near-real-time and adjust the flight planning in a flexible manner. We report on the experiences in the field of robotics, geophysics (Ground Penetrating Radar) and geology as well as life sciences in a simulated spaceflight operational environment. Extravehicular Activity (EVA) maps had been prepared using Google Earth and aerial images. The Rio Tinto mining area offers an excellent location for Mars analogue simulations. It is recognised as a terrestrial Mars analogue site because of the presence of jarosite and related sulphates, which have been identified by the NASA Mars Exploration Rover "Opportunity" in the El Capitan region of Meridiani Planum on Mars. The acidic, high ferric-sulphate content water of Rio Tinto is also considered as a possible analogue in astrobiology regarding the analysis of ferric sulphate related biochemical pathways and produced biomarkers. During our Mars simulation, 18 different types of soil and rock samples were collected by the spacesuit tester. The Raman results confirm the presence of minerals expected, such as jarosite, different Fe oxides and oxi-hydroxides, pyrite and complex Mg and Ca sulphates. Eight science experiments were conducted in the field. In this contribution first we list the important findings during the management and realisation of tests, and also a first summary of the scientific results. Based on these experiences suggestions for future analogue work are also summarised. We finish with recommendations for future field missions, including the preparation of the experiments
Commercial applications of large-scale Research and Development computer simulation technologies
International Nuclear Information System (INIS)
Kuok Mee Ling; Pascal Chen; Wen Ho Lee
1998-01-01
The potential commercial applications of two large-scale R and D computer simulation technologies are presented. One such technology is based on the numerical solution of the hydrodynamics equations, and is embodied in the two-dimensional Eulerian code EULE2D, which solves the hydrodynamic equations with various models for the equation of state (EOS), constitutive relations and fracture mechanics. EULE2D is an R and D code originally developed to design and analyze conventional munitions for anti-armor penetrations such as shaped charges, explosive formed projectiles, and kinetic energy rods. Simulated results agree very well with actual experiments. A commercial application presented here is the design and simulation of shaped charges for oil and gas well bore perforation. The other R and D simulation technology is based on the numerical solution of Maxwell's partial differential equations of electromagnetics in space and time, and is implemented in the three-dimensional code FDTD-SPICE, which solves Maxwell's equations in the time domain with finite-differences in the three spatial dimensions and calls SPICE for information when nonlinear active devices are involved. The FDTD method has been used in the radar cross-section modeling of military aircrafts and many other electromagnetic phenomena. The coupling of FDTD method with SPICE, a popular circuit and device simulation program, provides a powerful tool for the simulation and design of microwave and millimeter-wave circuits containing nonlinear active semiconductor devices. A commercial application of FDTD-SPICE presented here is the simulation of a two-element active antenna system. The simulation results and the experimental measurements are in excellent agreement. (Author)
Large-scale simulation of ductile fracture process of microstructured materials
International Nuclear Information System (INIS)
Tian Rong; Wang Chaowei
2011-01-01
The promise of computational science in the extreme-scale computing era is to reduce and decompose macroscopic complexities into microscopic simplicities with the expense of high spatial and temporal resolution of computing. In materials science and engineering, the direct combination of 3D microstructure data sets and 3D large-scale simulations provides unique opportunity for the development of a comprehensive understanding of nano/microstructure-property relationships in order to systematically design materials with specific desired properties. In the paper, we present a framework simulating the ductile fracture process zone in microstructural detail. The experimentally reconstructed microstructural data set is directly embedded into a FE mesh model to improve the simulation fidelity of microstructure effects on fracture toughness. To the best of our knowledge, it is for the first time that the linking of fracture toughness to multiscale microstructures in a realistic 3D numerical model in a direct manner is accomplished. (author)
Real-world-time simulation of memory consolidation in a large-scale cerebellar model
Directory of Open Access Journals (Sweden)
Masato eGosui
2016-03-01
Full Text Available We report development of a large-scale spiking network model of thecerebellum composed of more than 1 million neurons. The model isimplemented on graphics processing units (GPUs, which are dedicatedhardware for parallel computing. Using 4 GPUs simultaneously, we achieve realtime simulation, in which computer simulation ofcerebellar activity for 1 sec completes within 1 sec in thereal-world time, with temporal resolution of 1 msec.This allows us to carry out a very long-term computer simulationof cerebellar activity in a practical time with millisecond temporalresolution. Using the model, we carry out computer simulationof long-term gain adaptation of optokinetic response (OKR eye movementsfor 5 days aimed to study the neural mechanisms of posttraining memoryconsolidation. The simulation results are consistent with animal experimentsand our theory of posttraining memory consolidation. These resultssuggest that realtime computing provides a useful means to studya very slow neural process such as memory consolidation in the brain.
Random number generators for large-scale parallel Monte Carlo simulations on FPGA
Lin, Y.; Wang, F.; Liu, B.
2018-05-01
Through parallelization, field programmable gate array (FPGA) can achieve unprecedented speeds in large-scale parallel Monte Carlo (LPMC) simulations. FPGA presents both new constraints and new opportunities for the implementations of random number generators (RNGs), which are key elements of any Monte Carlo (MC) simulation system. Using empirical and application based tests, this study evaluates all of the four RNGs used in previous FPGA based MC studies and newly proposed FPGA implementations for two well-known high-quality RNGs that are suitable for LPMC studies on FPGA. One of the newly proposed FPGA implementations: a parallel version of additive lagged Fibonacci generator (Parallel ALFG) is found to be the best among the evaluated RNGs in fulfilling the needs of LPMC simulations on FPGA.
Halo Models of Large Scale Structure and Reliability of Cosmological N-Body Simulations
Directory of Open Access Journals (Sweden)
José Gaite
2013-05-01
Full Text Available Halo models of the large scale structure of the Universe are critically examined, focusing on the definition of halos as smooth distributions of cold dark matter. This definition is essentially based on the results of cosmological N-body simulations. By a careful analysis of the standard assumptions of halo models and N-body simulations and by taking into account previous studies of self-similarity of the cosmic web structure, we conclude that N-body cosmological simulations are not fully reliable in the range of scales where halos appear. Therefore, to have a consistent definition of halos is necessary either to define them as entities of arbitrary size with a grainy rather than smooth structure or to define their size in terms of small-scale baryonic physics.
Plasmonic resonances of nanoparticles from large-scale quantum mechanical simulations
Zhang, Xu; Xiang, Hongping; Zhang, Mingliang; Lu, Gang
2017-09-01
Plasmonic resonance of metallic nanoparticles results from coherent motion of its conduction electrons, driven by incident light. For the nanoparticles less than 10 nm in diameter, localized surface plasmonic resonances become sensitive to the quantum nature of the conduction electrons. Unfortunately, quantum mechanical simulations based on time-dependent Kohn-Sham density functional theory are computationally too expensive to tackle metal particles larger than 2 nm. Herein, we introduce the recently developed time-dependent orbital-free density functional theory (TD-OFDFT) approach which enables large-scale quantum mechanical simulations of plasmonic responses of metallic nanostructures. Using TD-OFDFT, we have performed quantum mechanical simulations to understand size-dependent plasmonic response of Na nanoparticles and plasmonic responses in Na nanoparticle dimers and trimers. An outlook of future development of the TD-OFDFT method is also presented.
DEFF Research Database (Denmark)
Chivaee, Hamid Sarlak; Sørensen, Jens Nørkær; Mikkelsen, Robert Flemming
2012-01-01
Large eddy simulation (LES) of flow in a wind farm is studied in neutral as well as thermally stratified atmospheric boundary layer (ABL). An approach has been practiced to simulate the flow in a fully developed wind farm boundary layer. The approach is based on the Immersed Boundary Method (IBM......) and involves implementation of an arbitrary prescribed initial boundary layer (See [1]). A prescribed initial boundary layer profile is enforced through the computational domain using body forces to maintain a desired flow field. The body forces are then stored and applied on the domain through the simulation...... and the boundary layer shape will be modified due to the interaction of the turbine wakes and buoyancy contributions. The implemented method is capable of capturing the most important features of wakes of wind farms [1] while having the advantage of resolving the wall layer with a coarser grid than typically...
International Nuclear Information System (INIS)
Winter, L.H.L.; Chakraborty, D.P.; Waes, P.F.G.M.
1988-01-01
Two image intensifier tubes have recently been introduced whose large imaging area makes them suitable for chest imaging (Phillips Pulmodiagnost TLX slit II and Siemens TX 57 large entrance field II). Both modalities present a 10 x 10-cm hard copy image to the radiologist. A receiver operating characteristic (ROC) curve study with simulated interstitial disease was performed to compare the image quality of these image intensifiers with conventional chest images. The relative ranking in terms of decreasing ROC areas was Siemens, conventional, and Philips. Compared with conventional imaging, none of the differences in ROC curve area were statistically significant at the 5% level
Evaluation of simulated-LOCA tests that produced large fuel cladding ballooning
International Nuclear Information System (INIS)
Powers, D.A.; Meyer, R.O.
1979-02-01
A description is given of the NRC review and evaluation of simulated-LOCA tests that produced large axially extended ballooing in Zircaloy fuel cladding. Technical summaries are presented on the likelihood of the transient that was used in the tests, the effects of temperature variations on strain localization, and the results of other similar experiments. It is concluded that (a) the large axially extended deformations were an artifact of the experimental technique, (b) current NRC licensing positions are not invalidated by this new information, and (c) no new research programs are needed to study this phenomenon
DEFF Research Database (Denmark)
Markert, Frank; Kozine, Igor
2012-01-01
in applying DES models to the analysis of large infrastructures for refueling stations and water supply. Two case studies are described which are concerned with the inherently safer supply and storage of hydrogen at refueling stations and an established drinking water supply system of a large metropolitan...... area, respectively. For both, the simulation aims at identifying points of potential improvement from the reliability point of view. This allows setting up a list of activities and safety measures to reduce risk and possible losses and mitigate the consequences of accidents. Based on the cases...
Large-eddy simulation of plume dispersion within regular arrays of cubic buildings
Nakayama, H.; Jurcakova, K.; Nagai, H.
2011-04-01
There is a potential problem that hazardous and flammable materials are accidentally or intentionally released within populated urban areas. For the assessment of human health hazard from toxic substances, the existence of high concentration peaks in a plume should be considered. For the safety analysis of flammable gas, certain critical threshold levels should be evaluated. Therefore, in such a situation, not only average levels but also instantaneous magnitudes of concentration should be accurately predicted. In this study, we perform Large-Eddy Simulation (LES) of plume dispersion within regular arrays of cubic buildings with large obstacle densities and investigate the influence of the building arrangement on the characteristics of mean and fluctuating concentrations.
Cardall, Christian Y.; Budiardja, Reuben D.
2018-01-01
The large-scale computer simulation of a system of physical fields governed by partial differential equations requires some means of approximating the mathematical limit of continuity. For example, conservation laws are often treated with a 'finite-volume' approach in which space is partitioned into a large number of small 'cells,' with fluxes through cell faces providing an intuitive discretization modeled on the mathematical definition of the divergence operator. Here we describe and make available Fortran 2003 classes furnishing extensible object-oriented implementations of simple meshes and the evolution of generic conserved currents thereon, along with individual 'unit test' programs and larger example problems demonstrating their use. These classes inaugurate the Mathematics division of our developing astrophysics simulation code GENASIS (Gen eral A strophysical Si mulation S ystem), which will be expanded over time to include additional meshing options, mathematical operations, solver types, and solver variations appropriate for many multiphysics applications.
Large Scale Beam-beam Simulations for the CERN LHC using Distributed Computing
Herr, Werner; McIntosh, E; Schmidt, F
2006-01-01
We report on a large scale simulation of beam-beam effects for the CERN Large Hadron Collider (LHC). The stability of particles which experience head-on and long-range beam-beam effects was investigated for different optical configurations and machine imperfections. To cover the interesting parameter space required computing resources not available at CERN. The necessary resources were available in the LHC@home project, based on the BOINC platform. At present, this project makes more than 60000 hosts available for distributed computing. We shall discuss our experience using this system during a simulation campaign of more than six months and describe the tools and procedures necessary to ensure consistent results. The results from this extended study are presented and future plans are discussed.
A high-resolution code for large eddy simulation of incompressible turbulent boundary layer flows
Cheng, Wan
2014-03-01
We describe a framework for large eddy simulation (LES) of incompressible turbulent boundary layers over a flat plate. This framework uses a fractional-step method with fourth-order finite difference on a staggered mesh. We present several laminar examples to establish the fourth-order accuracy and energy conservation property of the code. Furthermore, we implement a recycling method to generate turbulent inflow. We use the stretched spiral vortex subgrid-scale model and virtual wall model to simulate the turbulent boundary layer flow. We find that the case with Reθ ≈ 2.5 × 105 agrees well with available experimental measurements of wall friction, streamwise velocity profiles and turbulent intensities. We demonstrate that for cases with extremely large Reynolds numbers (Reθ = 1012), the present LES can reasonably predict the flow with a coarse mesh. The parallel implementation of the LES code demonstrates reasonable scaling on O(103) cores. © 2013 Elsevier Ltd.
Discontinuous Galerkin methodology for Large-Eddy Simulations of wind turbine airfoils
DEFF Research Database (Denmark)
Frére, A.; Sørensen, Niels N.; Hillewaert, K.
2016-01-01
This paper aims at evaluating the potential of the Discontinuous Galerkin (DG) methodology for Large-Eddy Simulation (LES) of wind turbine airfoils. The DG method has shown high accuracy, excellent scalability and capacity to handle unstructured meshes. It is however not used in the wind energy...... sector yet. The present study aims at evaluating this methodology on an application which is relevant for that sector and focuses on blade section aerodynamics characterization. To be pertinent for large wind turbines, the simulations would need to be at low Mach numbers (M ≤ 0.3) where compressible...... at low and high Reynolds numbers and compares the results to state-of-the-art models used in industry, namely the panel method (XFOIL with boundary layer modeling) and Reynolds Averaged Navier-Stokes (RANS). At low Reynolds number (Re = 6 × 104), involving laminar boundary layer separation and transition...
Review of Dynamic Modeling and Simulation of Large Scale Belt Conveyor System
He, Qing; Li, Hong
Belt conveyor is one of the most important devices to transport bulk-solid material for long distance. Dynamic analysis is the key to decide whether the design is rational in technique, safe and reliable in running, feasible in economy. It is very important to study dynamic properties, improve efficiency and productivity, guarantee conveyor safe, reliable and stable running. The dynamic researches and applications of large scale belt conveyor are discussed. The main research topics, the state-of-the-art of dynamic researches on belt conveyor are analyzed. The main future works focus on dynamic analysis, modeling and simulation of main components and whole system, nonlinear modeling, simulation and vibration analysis of large scale conveyor system.
Plebe, Alice; Grasso, Giorgio
2016-12-01
This paper describes a system developed for the simulation of flames inside an open-source 3D computer graphic software, Blender, with the aim of analyzing in virtual reality scenarios of hazards in large-scale industrial plants. The advantages of Blender are of rendering at high resolution the very complex structure of large industrial plants, and of embedding a physical engine based on smoothed particle hydrodynamics. This particle system is used to evolve a simulated fire. The interaction of this fire with the components of the plant is computed using polyhedron separation distance, adopting a Voronoi-based strategy that optimizes the number of feature distance computations. Results on a real oil and gas refining industry are presented.
Large eddy simulation of mixing between hot and cold sodium flows - comparison with experiments
Energy Technology Data Exchange (ETDEWEB)
Simoneau, J.P.; Noe, H.; Menant, B.
1995-09-01
The large eddy simulation is becoming a potential powerful tool for the calculation of turbulent flows. In nuclear liquid metal cooled fast reactors, the knowledge of the turbulence characteristics is of great interest for the prediction and the analysis of thermal stripping phenomena. The objective of this paper is to give a contribution in the evaluation of the large eddy simulation technique is an individual case. The problem chosen is the case of the mixing between hot and cold sodium flows. The computations are compared with available sodium tests. This study shows acceptable qualitative results but the simple model used is not able to predict the turbulence characteristics. More complex models including larger domains around the fluctuating zone and fluctuating boundary conditions could be necessary. Validation works are continuing.
A dynamic globalization model for large eddy simulation of complex turbulent flow
Energy Technology Data Exchange (ETDEWEB)
Choi, Hae Cheon; Park, No Ma; Kim, Jin Seok [Seoul National Univ., Seoul (Korea, Republic of)
2005-07-01
A dynamic subgrid-scale model is proposed for large eddy simulation of turbulent flows in complex geometry. The eddy viscosity model by Vreman [Phys. Fluids, 16, 3670 (2004)] is considered as a base model. A priori tests with the original Vreman model show that it predicts the correct profile of subgrid-scale dissipation in turbulent channel flow but the optimal model coefficient is far from universal. Dynamic procedures of determining the model coefficient are proposed based on the 'global equilibrium' between the subgrid-scale dissipation and viscous dissipation. An important feature of the proposed procedures is that the model coefficient determined is globally constant in space but varies only in time. Large eddy simulations with the present dynamic model are conducted for forced isotropic turbulence, turbulent channel flow and flow over a sphere, showing excellent agreements with previous results.
Molecular dynamics simulations of sputtering of organic overlayers by slow, large clusters
International Nuclear Information System (INIS)
Rzeznik, L.; Czerwinski, B.; Garrison, B.J.; Winograd, N.; Postawa, Z.
2008-01-01
The ion-stimulated desorption of organic molecules by impact of large and slow clusters is examined using molecular dynamics (MDs) computer simulations. The investigated system, represented by a monolayer of benzene deposited on Ag{1 1 1}, is irradiated with projectiles composed of thousands of noble gas atoms having a kinetic energy of 0.1-20 eV/atom. The sputtering yield of molecular species and the kinetic energy distributions are analyzed and compared to the results obtain for PS4 overlayer. The simulations demonstrate quite clearly that the physics of ejection by large and slow clusters is distinct from the ejection events stimulated by the popular SIMS clusters, like C 60 , Au 3 and SF 5 at tens of keV energies.
Large eddy simulation of particulate flow inside a differentially heated cavity
Energy Technology Data Exchange (ETDEWEB)
Bosshard, Christoph, E-mail: christoph.bosshard@a3.epfl.ch [Paul Scherrer Institut, Laboratory for Thermalhydraulics (LTH), 5232 Villigen PSI (Switzerland); Dehbi, Abdelouahab, E-mail: abdel.dehbi@psi.ch [Paul Scherrer Institut, Laboratory for Thermalhydraulics (LTH), 5232 Villigen PSI (Switzerland); Deville, Michel, E-mail: michel.deville@epfl.ch [École Polytechnique Fédérale de Lausanne, STI-DO, Station 12, 1015 Lausanne (Switzerland); Leriche, Emmanuel, E-mail: emmanuel.leriche@univ-lille1.fr [Université de Lille I, Laboratoire de Mécanique de Lille, Avenue Paul Langevin, Cité Scientifique, F-59655 Villeneuve d’Ascq Cédex (France); Soldati, Alfredo, E-mail: soldati@uniud.it [Dipartimento di Energetica e Macchine and Centro Interdipartimentale di Fluidodinamica e Idraulica, Universitá degli Studi di Udine, Udine (Italy)
2014-02-15
Highlights: • Nuclear accident leads to airborne radioactive particles in containment atmosphere. • Large eddy simulation with particles in differentially heated cavity is carried out. • LES results show negligible differences with direct numerical simulation. • Four different particle sets with diameters from 10 μm to 35 μm are tracked. • Particle removal dominated by gravity settling and turbophoresis is negligible. - Abstract: In nuclear safety, some severe accident scenarios lead to the presence of fission products in aerosol form in the closed containment atmosphere. It is important to understand the particle depletion process to estimate the risk of a release of radioactivity to the environment should a containment break occur. As a model for the containment, we use the three-dimensional differentially heated cavity problem. The differentially heated cavity is a cubical box with a hot wall and a cold wall on vertical opposite sides. On the other walls of the cube we have adiabatic boundary conditions. For the velocity field the no-slip boundary condition is applied. The flow of the air in the cavity is described by the Boussinesq equations. The method used to simulate the turbulent flow is the large eddy simulation (LES) where the dynamics of the large eddies is resolved by the computational grid and the small eddies are modelled by the introduction of subgrid scale quantities using a filter function. Particle trajectories are computed using the Lagrangian particle tracking method, including the relevant forces (drag, gravity, thermophoresis). Four different sets with each set containing one million particles and diameters of 10 μm, 15 μm, 25 μm and 35 μm are simulated. Simulation results for the flow field and particle sizes from 15 μm to 35 μm are compared to previous results from direct numerical simulation (DNS). The integration time of the LES is three times longer and the smallest particles have been simulated only in the LES. Particle
Beaton, Kara H.; Chappell, Steven P.; Abercromby, Andrew F. J.; Lim, Darlene S. S.
2018-01-01
The Biologic Analog Science Associated with Lava Terrains (BASALT) project is a four-year program dedicated to iteratively designing, implementing, and evaluating concepts of operations (ConOps) and supporting capabilities to enable and enhance scientific exploration for future human Mars missions. The BASALT project has incorporated three field deployments during which real (non-simulated) biological and geochemical field science have been conducted at two high-fidelity Mars analog locations under simulated Mars mission conditions, including communication delays and data transmission limitations. BASALT's primary Science objective has been to extract basaltic samples for the purpose of investigating how microbial communities and habitability correlate with the physical and geochemical characteristics of chemically altered basalt environments. Field sites include the active East Rift Zone on the Big Island of Hawai'i, reminiscent of early Mars when basaltic volcanism and interaction with water were widespread, and the dormant eastern Snake River Plain in Idaho, similar to present-day Mars where basaltic volcanism is rare and most evidence for volcano-driven hydrothermal activity is relict. BASALT's primary Science Operations objective has been to investigate exploration ConOps and capabilities that facilitate scientific return during human-robotic exploration under Mars mission constraints. Each field deployment has consisted of ten extravehicular activities (EVAs) on the volcanic flows in which crews of two extravehicular and two intravehicular crewmembers conducted the field science while communicating across time delay and under bandwidth constraints with an Earth-based Mission Support Center (MSC) comprised of expert scientists and operators. Communication latencies of 5 and 15 min one-way light time and low (0.512 Mb/s uplink, 1.54 Mb/s downlink) and high (5.0 Mb/s uplink, 10.0 Mb/s downlink) bandwidth conditions were evaluated. EVA crewmembers communicated
Application of parallel computing techniques to a large-scale reservoir simulation
International Nuclear Information System (INIS)
Zhang, Keni; Wu, Yu-Shu; Ding, Chris; Pruess, Karsten
2001-01-01
Even with the continual advances made in both computational algorithms and computer hardware used in reservoir modeling studies, large-scale simulation of fluid and heat flow in heterogeneous reservoirs remains a challenge. The problem commonly arises from intensive computational requirement for detailed modeling investigations of real-world reservoirs. This paper presents the application of a massive parallel-computing version of the TOUGH2 code developed for performing large-scale field simulations. As an application example, the parallelized TOUGH2 code is applied to develop a three-dimensional unsaturated-zone numerical model simulating flow of moisture, gas, and heat in the unsaturated zone of Yucca Mountain, Nevada, a potential repository for high-level radioactive waste. The modeling approach employs refined spatial discretization to represent the heterogeneous fractured tuffs of the system, using more than a million 3-D gridblocks. The problem of two-phase flow and heat transfer within the model domain leads to a total of 3,226,566 linear equations to be solved per Newton iteration. The simulation is conducted on a Cray T3E-900, a distributed-memory massively parallel computer. Simulation results indicate that the parallel computing technique, as implemented in the TOUGH2 code, is very efficient. The reliability and accuracy of the model results have been demonstrated by comparing them to those of small-scale (coarse-grid) models. These comparisons show that simulation results obtained with the refined grid provide more detailed predictions of the future flow conditions at the site, aiding in the assessment of proposed repository performance
Large-Eddy Simulation of Flow and Pollutant Transport in Urban Street Canyons with Ground Heating
Li, Xian-Xiang; Britter, Rex E.; Koh, Tieh Yong; Norford, Leslie Keith; Liu, Chun-Ho; Entekhabi, Dara; Leung, Dennis Y. C.
2009-01-01
Our study employed large-eddy simulation (LES) based on a one-equation subgrid-scale model to investigate the flow field and pollutant dispersion characteristics inside urban street canyons. Unstable thermal stratification was produced by heating the ground of the street canyon. Using the Boussinesq approximation, thermal buoyancy forces were taken into account in both the Navier–Stokes equations and the transport equation for subgrid-scale turbulent kinetic energy (TKE). The LESs were valida...
Nonlinear δf Simulation Studies of Intense Charged Particle Beams with Large Temperature Anisotropy
International Nuclear Information System (INIS)
Startsev, Edward A.; Davidson, Ronald C.; Qin, Hong
2002-01-01
In this paper, a 3-D nonlinear perturbative particle simulation code (BEST) [H. Qin, R.C. Davidson and W.W. Lee, Physical Review Special Topics on Accelerators and Beams 3 (2000) 084401] is used to systematically study the stability properties of intense nonneutral charged particle beams with large temperature anisotropy (T perpendicularb >> T parallelb ). The most unstable modes are identified, and their eigenfrequencies, radial mode structure, and nonlinear dynamics are determined for axisymmetric perturbations with ∂/∂θ = 0
Large-scale derived flood frequency analysis based on continuous simulation
Dung Nguyen, Viet; Hundecha, Yeshewatesfa; Guse, Björn; Vorogushyn, Sergiy; Merz, Bruno
2016-04-01
There is an increasing need for spatially consistent flood risk assessments at the regional scale (several 100.000 km2), in particular in the insurance industry and for national risk reduction strategies. However, most large-scale flood risk assessments are composed of smaller-scale assessments and show spatial inconsistencies. To overcome this deficit, a large-scale flood model composed of a weather generator and catchments models was developed reflecting the spatially inherent heterogeneity. The weather generator is a multisite and multivariate stochastic model capable of generating synthetic meteorological fields (precipitation, temperature, etc.) at daily resolution for the regional scale. These fields respect the observed autocorrelation, spatial correlation and co-variance between the variables. They are used as input into catchment models. A long-term simulation of this combined system enables to derive very long discharge series at many catchment locations serving as a basic for spatially consistent flood risk estimates at the regional scale. This combined model was set up and validated for major river catchments in Germany. The weather generator was trained by 53-year observation data at 528 stations covering not only the complete Germany but also parts of France, Switzerland, Czech Republic and Australia with the aggregated spatial scale of 443,931 km2. 10.000 years of daily meteorological fields for the study area were generated. Likewise, rainfall-runoff simulations with SWIM were performed for the entire Elbe, Rhine, Weser, Donau and Ems catchments. The validation results illustrate a good performance of the combined system, as the simulated flood magnitudes and frequencies agree well with the observed flood data. Based on continuous simulation this model chain is then used to estimate flood quantiles for the whole Germany including upstream headwater catchments in neighbouring countries. This continuous large scale approach overcomes the several
Simple Model for Simulating Characteristics of River Flow Velocity in Large Scale
Directory of Open Access Journals (Sweden)
Husin Alatas
2015-01-01
Full Text Available We propose a simple computer based phenomenological model to simulate the characteristics of river flow velocity in large scale. We use shuttle radar tomography mission based digital elevation model in grid form to define the terrain of catchment area. The model relies on mass-momentum conservation law and modified equation of motion of falling body in inclined plane. We assume inelastic collision occurs at every junction of two river branches to describe the dynamics of merged flow velocity.
Large Eddy Simulations of the Flow in a Three-Dimensional Ventilated Room
DEFF Research Database (Denmark)
Davidson, Lars; Nielsen, Peter V.
We have done Large Eddy Simulations (LES) of the flow in a three-dimensional ventilated room. A finite volume method is used with a collocated grid arrangement. The momentum equations are solved with an explicit method using central differencing for all terms. The pressure is obtained from a Pois...... a Poisson equation, which is solved with a conjugate gradient method. For the discretization in time we use the Adam-Bashfourth scheme, which is second-order accurate....
Analyzing large data sets from XGC1 magnetic fusion simulations using apache spark
Energy Technology Data Exchange (ETDEWEB)
Churchill, R. Michael [Princeton Plasma Physics Lab. (PPPL), Princeton, NJ (United States)
2016-11-21
Apache Spark is explored as a tool for analyzing large data sets from the magnetic fusion simulation code XGCI. Implementation details of Apache Spark on the NERSC Edison supercomputer are discussed, including binary file reading, and parameter setup. Here, an unsupervised machine learning algorithm, k-means clustering, is applied to XGCI particle distribution function data, showing that highly turbulent spatial regions do not have common coherent structures, but rather broad, ring-like structures in velocity space.
International Nuclear Information System (INIS)
Lee, D.J.; Brown, D.J.
1982-01-01
The effects of large scale processing on the properties of cemented simulant sodium nitrate waste have been investigated. Leach tests have been performed on full-size drums, cores and laboratory samples of cement formulations containing Ordinary Portland Cement (OPC), Sulphate Resisting Portland Cement (SRPC) and a blended cement (90% ground granulated blast furnace slag/10% OPC). In addition, development of the cement hydration exotherms with time and the temperature distribution in 220 dm 3 samples have been followed. (author)
Li, Jun; Fu, Siyao; He, Haibo; Jia, Hongfei; Li, Yanzhong; Guo, Yi
2015-11-01
Large-scale regional evacuation is an important part of national security emergency response plan. Large commercial shopping area, as the typical service system, its emergency evacuation is one of the hot research topics. A systematic methodology based on Cellular Automata with the Dynamic Floor Field and event driven model has been proposed, and the methodology has been examined within context of a case study involving the evacuation within a commercial shopping mall. Pedestrians walking is based on Cellular Automata and event driven model. In this paper, the event driven model is adopted to simulate the pedestrian movement patterns, the simulation process is divided into normal situation and emergency evacuation. The model is composed of four layers: environment layer, customer layer, clerk layer and trajectory layer. For the simulation of movement route of pedestrians, the model takes into account purchase intention of customers and density of pedestrians. Based on evacuation model of Cellular Automata with Dynamic Floor Field and event driven model, we can reflect behavior characteristics of customers and clerks at the situations of normal and emergency evacuation. The distribution of individual evacuation time as a function of initial positions and the dynamics of the evacuation process is studied. Our results indicate that the evacuation model using the combination of Cellular Automata with Dynamic Floor Field and event driven scheduling can be used to simulate the evacuation of pedestrian flows in indoor areas with complicated surroundings and to investigate the layout of shopping mall.
Large Eddy Simulation of the Diurnal Cycle in Southeast Pacific Stratocumulus
Energy Technology Data Exchange (ETDEWEB)
Caldwell, P; Bretherton, C
2008-03-03
This paper describes a series of 6 day large eddy simulations of a deep, sometimes drizzling stratocumulus-topped boundary layer based on forcings from the East Pacific Investigation of Climate (EPIC) 2001 field campaign. The base simulation was found to reproduce the observed mean boundary layer properties quite well. The diurnal cycle of liquid water path was also well captured, although good agreement appears to result partially from compensating errors in the diurnal cycles of cloud base and cloud top due to overentrainment around midday. At other times of the day, entrainment is found to be proportional to the vertically-integrated buoyancy flux. Model stratification matches observations well; turbulence profiles suggest that the boundary layer is always at least somewhat decoupled. Model drizzle appears to be too sensitive to liquid water path and subcloud evaporation appears to be too weak. Removing the diurnal cycle of subsidence had little effect on simulated cloud albedo. Simulations with changed droplet concentration and drizzle susceptibility showed large liquid water path differences at night, but differences were quite small at midday. Droplet concentration also had a significant impact on entrainment, primarily through droplet sedimentation feedback rather than through drizzle processes.
Large Eddy Simulation of turbulent flow in wire wrapped fuel pin bundles cooled by sodium
International Nuclear Information System (INIS)
Saxena, Aakanksha; Cadiou, Thierry; Bieder, Ulrich; Viazzo, Stephane
2013-06-01
The objective of the study is to understand the thermal hydraulics in a core sub-assembly with liquid sodium as coolant by performing detailed numerical simulations. The passage for the coolant flow between the fuel rods is maintained by thin wires wrapped around the rods. The contact point between the fuel pin and the spacer wire is the region of creation of hot spots and a cyclic variation of temperature in hot spots can adversely affect the mechanical properties of the clad due to the phenomena like thermal stripping. The current status quo provides two different models to perform the numerical simulations, namely Reynolds Averaged Navier-Stokes (RANS) and Large Eddy Simulation (LES). The two models differ in the extent of modelling used to close the Navier-Stokes equations. LES is a filtered approach where the large scale of motions are explicitly resolved while the small scale motions are modelled whereas RANS is a time averaging approach where all scale of motions are modelled. Thus LES involves less modelling as compared to RANS and so the results are comparatively more accurate. An attempt has been made to use the LES model. The simulations have been performed using the code Trio-U (developed by CEA). The turbulent statistics of the flow and thermal quantities are calculated. Finally the goal is to obtain the frequency of temperature oscillations at the region of hot spots near the spacer wire. (authors)
Large-eddy simulation of wind turbine wake interactions on locally refined Cartesian grids
Angelidis, Dionysios; Sotiropoulos, Fotis
2014-11-01
Performing high-fidelity numerical simulations of turbulent flow in wind farms remains a challenging issue mainly because of the large computational resources required to accurately simulate the turbine wakes and turbine/turbine interactions. The discretization of the governing equations on structured grids for mesoscale calculations may not be the most efficient approach for resolving the large disparity of spatial scales. A 3D Cartesian grid refinement method enabling the efficient coupling of the Actuator Line Model (ALM) with locally refined unstructured Cartesian grids adapted to accurately resolve tip vortices and multi-turbine interactions, is presented. Second order schemes are employed for the discretization of the incompressible Navier-Stokes equations in a hybrid staggered/non-staggered formulation coupled with a fractional step method that ensures the satisfaction of local mass conservation to machine zero. The current approach enables multi-resolution LES of turbulent flow in multi-turbine wind farms. The numerical simulations are in good agreement with experimental measurements and are able to resolve the rich dynamics of turbine wakes on grids containing only a small fraction of the grid nodes that would be required in simulations without local mesh refinement. This material is based upon work supported by the Department of Energy under Award Number DE-EE0005482 and the National Science Foundation under Award number NSF PFI:BIC 1318201.
Numerical simulation of pseudoelastic shape memory alloys using the large time increment method
Gu, Xiaojun; Zhang, Weihong; Zaki, Wael; Moumni, Ziad
2017-04-01
The paper presents a numerical implementation of the large time increment (LATIN) method for the simulation of shape memory alloys (SMAs) in the pseudoelastic range. The method was initially proposed as an alternative to the conventional incremental approach for the integration of nonlinear constitutive models. It is adapted here for the simulation of pseudoelastic SMA behavior using the Zaki-Moumni model and is shown to be especially useful in situations where the phase transformation process presents little or lack of hardening. In these situations, a slight stress variation in a load increment can result in large variations of strain and local state variables, which may lead to difficulties in numerical convergence. In contrast to the conventional incremental method, the LATIN method solve the global equilibrium and local consistency conditions sequentially for the entire loading path. The achieved solution must satisfy the conditions of static and kinematic admissibility and consistency simultaneously after several iterations. 3D numerical implementation is accomplished using an implicit algorithm and is then used for finite element simulation using the software Abaqus. Computational tests demonstrate the ability of this approach to simulate SMAs presenting flat phase transformation plateaus and subjected to complex loading cases, such as the quasi-static behavior of a stent structure. Some numerical results are contrasted to those obtained using step-by-step incremental integration.
General-relativistic Large-eddy Simulations of Binary Neutron Star Mergers
Energy Technology Data Exchange (ETDEWEB)
Radice, David, E-mail: dradice@astro.princeton.edu [Institute for Advanced Study, 1 Einstein Drive, Princeton, NJ 08540 (United States)
2017-03-20
The flow inside remnants of binary neutron star (NS) mergers is expected to be turbulent, because of magnetohydrodynamics instability activated at scales too small to be resolved in simulations. To study the large-scale impact of these instabilities, we develop a new formalism, based on the large-eddy simulation technique, for the modeling of subgrid-scale turbulent transport in general relativity. We apply it, for the first time, to the simulation of the late-inspiral and merger of two NSs. We find that turbulence can significantly affect the structure and survival time of the merger remnant, as well as its gravitational-wave (GW) and neutrino emissions. The former will be relevant for GW observation of merging NSs. The latter will affect the composition of the outflow driven by the merger and might influence its nucleosynthetic yields. The accretion rate after black hole formation is also affected. Nevertheless, we find that, for the most likely values of the turbulence mixing efficiency, these effects are relatively small and the GW signal will be affected only weakly by the turbulence. Thus, our simulations provide a first validation of all existing post-merger GW models.
Large-scale Validation of AMIP II Land-surface Simulations: Preliminary Results for Ten Models
Energy Technology Data Exchange (ETDEWEB)
Phillips, T J; Henderson-Sellers, A; Irannejad, P; McGuffie, K; Zhang, H
2005-12-01
This report summarizes initial findings of a large-scale validation of the land-surface simulations of ten atmospheric general circulation models that are entries in phase II of the Atmospheric Model Intercomparison Project (AMIP II). This validation is conducted by AMIP Diagnostic Subproject 12 on Land-surface Processes and Parameterizations, which is focusing on putative relationships between the continental climate simulations and the associated models' land-surface schemes. The selected models typify the diversity of representations of land-surface climate that are currently implemented by the global modeling community. The current dearth of global-scale terrestrial observations makes exacting validation of AMIP II continental simulations impractical. Thus, selected land-surface processes of the models are compared with several alternative validation data sets, which include merged in-situ/satellite products, climate reanalyses, and off-line simulations of land-surface schemes that are driven by observed forcings. The aggregated spatio-temporal differences between each simulated process and a chosen reference data set then are quantified by means of root-mean-square error statistics; the differences among alternative validation data sets are similarly quantified as an estimate of the current observational uncertainty in the selected land-surface process. Examples of these metrics are displayed for land-surface air temperature, precipitation, and the latent and sensible heat fluxes. It is found that the simulations of surface air temperature, when aggregated over all land and seasons, agree most closely with the chosen reference data, while the simulations of precipitation agree least. In the latter case, there also is considerable inter-model scatter in the error statistics, with the reanalyses estimates of precipitation resembling the AMIP II simulations more than to the chosen reference data. In aggregate, the simulations of land-surface latent and
Pedroza, Claudia; Han, Weilu; Thanh Truong, Van Thi; Green, Charles; Tyson, Jon E
2018-01-01
One of the main advantages of Bayesian analyses of clinical trials is their ability to formally incorporate skepticism about large treatment effects through the use of informative priors. We conducted a simulation study to assess the performance of informative normal, Student- t, and beta distributions in estimating relative risk (RR) or odds ratio (OR) for binary outcomes. Simulation scenarios varied the prior standard deviation (SD; level of skepticism of large treatment effects), outcome rate in the control group, true treatment effect, and sample size. We compared the priors with regards to bias, mean squared error (MSE), and coverage of 95% credible intervals. Simulation results show that the prior SD influenced the posterior to a greater degree than the particular distributional form of the prior. For RR, priors with a 95% interval of 0.50-2.0 performed well in terms of bias, MSE, and coverage under most scenarios. For OR, priors with a wider 95% interval of 0.23-4.35 had good performance. We recommend the use of informative priors that exclude implausibly large treatment effects in analyses of clinical trials, particularly for major outcomes such as mortality.
Large Eddy Simulation of Unstably Stratified Turbulent Flow over Urban-Like Building Arrays
Directory of Open Access Journals (Sweden)
Bobin Wang
2013-01-01
Full Text Available Thermal instability induced by solar radiation is the most common condition of urban atmosphere in daytime. Compared to researches under neutral conditions, only a few numerical works studied the unstable urban boundary layer and the effect of buoyancy force is unclear. In this paper, unstably stratified turbulent boundary layer flow over three-dimensional urban-like building arrays with ground heating is simulated. Large eddy simulation is applied to capture main turbulence structures and the effect of buoyancy force on turbulence can be investigated. Lagrangian dynamic subgrid scale model is used for complex flow together with a wall function, taking into account the large pressure gradient near buildings. The numerical model and method are verified with the results measured in wind tunnel experiment. The simulated results satisfy well with the experiment in mean velocity and temperature, as well as turbulent intensities. Mean flow structure inside canopy layer varies with thermal instability, while no large secondary vortex is observed. Turbulent intensities are enhanced, as buoyancy force contributes to the production of turbulent kinetic energy.
Sever, G.; Collis, S. M.; Ghate, V. P.
2017-12-01
Three-dimensional numerical experiments are performed to explore the mechanical and thermal impacts of Graciosa Island on the sampling of oceanic airflow and cloud evolution. Ideal and real configurations of flow and terrain are planned using high-resolution, large-eddy resolving (e.g., Δ cold-pool formation upstream of an ideal two-kilometer island, with von Kármán like vortices propagation downstream. Although the peak height of Graciosa is less than half kilometer, the Azores island chain has a mountain over 2 km, which may be leading to more complex flow patterns when simulations are extended to a larger domain. Preliminary idealized low-resolution moist simulations indicate that the cloud field is impacted due to the presence of the island. Longer simulations that are performed to capture diurnal evolution of island boundary layer show distinct land/sea breeze formations under quiescent flow conditions. Further numerical experiments are planned to extend moist simulations to include realistic atmospheric profiles and observations of surface fluxes coupled with radiative effects. This work is intended to produce a useful simulation framework coupled with instruments to guide airborne and ground sampling strategies during the ACE-ENA field campaign which is aimed to better characterize marine boundary layer clouds.
International Nuclear Information System (INIS)
Guo Huan; Liu Yangang; Daum, Peter H; Senum, Gunnar I; Tao, W-K
2008-01-01
We simulated a marine stratus deck sampled during the Marine Stratus/Stratocumulus Experiment (MASE) with a three-dimensional large eddy simulation (LES) model at different model resolutions. Various characteristics of the vertical velocity from the model simulations were evaluated against those derived from the corresponding aircraft in situ observations, focusing on standard deviation, skewness, kurtosis, probability density function (PDF), power spectrum, and structure function. Our results show that although the LES model captures reasonably well the lower-order moments (e.g., horizontal averages and standard deviations), it fails to simulate many aspects of the higher-order moments, such as kurtosis, especially near cloud base and cloud top. Further investigations of the PDFs, power spectra, and structure functions reveal that compared to the observations, the model generally underestimates relatively strong variations on small scales. The results also suggest that increasing the model resolutions improves the agreements between the model results and the observations in virtually all of the properties that we examined. Furthermore, the results indicate that a vertical grid size <10 m is necessary for accurately simulating even the standard-deviation profile, posing new challenges to computer resources.
Large-scale introduction of wind power stations in the Swedish grid: a simulation study
Energy Technology Data Exchange (ETDEWEB)
Larsson, L
1978-08-01
This report describes a simulation study on the factors to be considered if wind power were to be introduced to the south Swedish power grid on a large scale. The simulations are based upon a heuristic power generation planning model, developed for the purpose. The heuristic technique reflects the actual running strategies of a big power company with suitable accuracy. All simulations refer to certain typical days in 1976 to which all wind data and system characteristics are related. The installed amount of wind power will not be subject to optimization. All differences between planned and real wind power generation is equalized by regulation of the hydro power. The simulations made differ according to how the installed amount of wind power is handled in the power generation planning. The simulations indicate that the power system examined could well bear an introduction of wind power up to a level of 20% of the total power installed. This result is of course valid only for the days examined and does not necessarily apply to the present day structure of the system.
Large eddy simulation for predicting turbulent heat transfer in gas turbines.
Tafti, Danesh K; He, Long; Nagendra, K
2014-08-13
Blade cooling technology will play a critical role in the next generation of propulsion and power generation gas turbines. Accurate prediction of blade metal temperature can avoid the use of excessive compressed bypass air and allow higher turbine inlet temperature, increasing fuel efficiency and decreasing emissions. Large eddy simulation (LES) has been established to predict heat transfer coefficients with good accuracy under various non-canonical flows, but is still limited to relatively simple geometries and low Reynolds numbers. It is envisioned that the projected increase in computational power combined with a drop in price-to-performance ratio will make system-level simulations using LES in complex blade geometries at engine conditions accessible to the design process in the coming one to two decades. In making this possible, two key challenges are addressed in this paper: working with complex intricate blade geometries and simulating high-Reynolds-number (Re) flows. It is proposed to use the immersed boundary method (IBM) combined with LES wall functions. A ribbed duct at Re=20 000 is simulated using the IBM, and a two-pass ribbed duct is simulated at Re=100 000 with and without rotation (rotation number Ro=0.2) using LES with wall functions. The results validate that the IBM is a viable alternative to body-conforming grids and that LES with wall functions reproduces experimental results at a much lower computational cost. © 2014 The Author(s) Published by the Royal Society. All rights reserved.
Ajami, H.; Sharma, A.; Lakshmi, V.
2017-12-01
Application of semi-distributed hydrologic modeling frameworks is a viable alternative to fully distributed hyper-resolution hydrologic models due to computational efficiency and resolving fine-scale spatial structure of hydrologic fluxes and states. However, fidelity of semi-distributed model simulations is impacted by (1) formulation of hydrologic response units (HRUs), and (2) aggregation of catchment properties for formulating simulation elements. Here, we evaluate the performance of a recently developed Soil Moisture and Runoff simulation Toolkit (SMART) for large catchment scale simulations. In SMART, topologically connected HRUs are delineated using thresholds obtained from topographic and geomorphic analysis of a catchment, and simulation elements are equivalent cross sections (ECS) representative of a hillslope in first order sub-basins. Earlier investigations have shown that formulation of ECSs at the scale of a first order sub-basin reduces computational time significantly without compromising simulation accuracy. However, the implementation of this approach has not been fully explored for catchment scale simulations. To assess SMART performance, we set-up the model over the Little Washita watershed in Oklahoma. Model evaluations using in-situ soil moisture observations show satisfactory model performance. In addition, we evaluated the performance of a number of soil moisture disaggregation schemes recently developed to provide spatially explicit soil moisture outputs at fine scale resolution. Our results illustrate that the statistical disaggregation scheme performs significantly better than the methods based on topographic data. Future work is focused on assessing the performance of SMART using remotely sensed soil moisture observations using spatially based model evaluation metrics.
Tang, G.; Bartlein, P. J.
2012-08-01
Satellite-based data, such as vegetation type and fractional vegetation cover, are widely used in hydrologic models to prescribe the vegetation state in a study region. Dynamic global vegetation models (DGVM) simulate land surface hydrology. Incorporation of satellite-based data into a DGVM may enhance a model's ability to simulate land surface hydrology by reducing the task of model parameterization and providing distributed information on land characteristics. The objectives of this study are to (i) modify a DGVM for simulating land surface water balances; (ii) evaluate the modified model in simulating actual evapotranspiration (ET), soil moisture, and surface runoff at regional or watershed scales; and (iii) gain insight into the ability of both the original and modified model to simulate large spatial scale land surface hydrology. To achieve these objectives, we introduce the "LPJ-hydrology" (LH) model which incorporates satellite-based data into the Lund-Potsdam-Jena (LPJ) DGVM. To evaluate the model we ran LH using historical (1981-2006) climate data and satellite-based land covers at 2.5 arc-min grid cells for the conterminous US and for the entire world using coarser climate and land cover data. We evaluated the simulated ET, soil moisture, and surface runoff using a set of observed or simulated data at different spatial scales. Our results demonstrate that spatial patterns of LH-simulated annual ET and surface runoff are in accordance with previously published data for the US; LH-modeled monthly stream flow for 12 major rivers in the US was consistent with observed values respectively during the years 1981-2006 (R2 > 0.46, p 0.52). The modeled mean annual discharges for 10 major rivers worldwide also agreed well (differences day method for snowmelt computation, the addition of the solar radiation effect on snowmelt enabled LH to better simulate monthly stream flow in winter and early spring for rivers located at mid-to-high latitudes. In addition, LH
Massively parallel Monte Carlo. Experiences running nuclear simulations on a large condor cluster
International Nuclear Information System (INIS)
Tickner, James; O'Dwyer, Joel; Roach, Greg; Uher, Josef; Hitchen, Greg
2010-01-01
The trivially-parallel nature of Monte Carlo (MC) simulations make them ideally suited for running on a distributed, heterogeneous computing environment. We report on the setup and operation of a large, cycle-harvesting Condor computer cluster, used to run MC simulations of nuclear instruments ('jobs') on approximately 4,500 desktop PCs. Successful operation must balance the competing goals of maximizing the availability of machines for running jobs whilst minimizing the impact on users' PC performance. This requires classification of jobs according to anticipated run-time and priority and careful optimization of the parameters used to control job allocation to host machines. To maximize use of a large Condor cluster, we have created a powerful suite of tools to handle job submission and analysis, as the manual creation, submission and evaluation of large numbers (hundred to thousands) of jobs would be too arduous. We describe some of the key aspects of this suite, which has been interfaced to the well-known MCNP and EGSnrc nuclear codes and our in-house PHOTON optical MC code. We report on our practical experiences of operating our Condor cluster and present examples of several large-scale instrument design problems that have been solved using this tool. (author)
Life as an emergent phenomenon: studies from a large-scale boid simulation and web data
Ikegami, Takashi; Mototake, Yoh-ichi; Kobori, Shintaro; Oka, Mizuki; Hashimoto, Yasuhiro
2017-11-01
A large group with a special structure can become the mother of emergence. We discuss this hypothesis in relation to large-scale boid simulations and web data. In the boid swarm simulations, the nucleation, organization and collapse dynamics were found to be more diverse in larger flocks than in smaller flocks. In the second analysis, large web data, consisting of shared photos with descriptive tags, tended to group together users with similar tendencies, allowing the network to develop a core-periphery structure. We show that the generation rate of novel tags and their usage frequencies are high in the higher-order cliques. In this case, novelty is not considered to arise randomly; rather, it is generated as a result of a large and structured network. We contextualize these results in terms of adjacent possible theory and as a new way to understand collective intelligence. We argue that excessive information and material flow can become a source of innovation. This article is part of the themed issue 'Reconceptualizing the origins of life'.
Statistical Analysis of Large Simulated Yield Datasets for Studying Climate Effects
Makowski, David; Asseng, Senthold; Ewert, Frank; Bassu, Simona; Durand, Jean-Louis; Martre, Pierre; Adam, Myriam; Aggarwal, Pramod K.; Angulo, Carlos; Baron, Chritian;
2015-01-01
Many studies have been carried out during the last decade to study the effect of climate change on crop yields and other key crop characteristics. In these studies, one or several crop models were used to simulate crop growth and development for different climate scenarios that correspond to different projections of atmospheric CO2 concentration, temperature, and rainfall changes (Semenov et al., 1996; Tubiello and Ewert, 2002; White et al., 2011). The Agricultural Model Intercomparison and Improvement Project (AgMIP; Rosenzweig et al., 2013) builds on these studies with the goal of using an ensemble of multiple crop models in order to assess effects of climate change scenarios for several crops in contrasting environments. These studies generate large datasets, including thousands of simulated crop yield data. They include series of yield values obtained by combining several crop models with different climate scenarios that are defined by several climatic variables (temperature, CO2, rainfall, etc.). Such datasets potentially provide useful information on the possible effects of different climate change scenarios on crop yields. However, it is sometimes difficult to analyze these datasets and to summarize them in a useful way due to their structural complexity; simulated yield data can differ among contrasting climate scenarios, sites, and crop models. Another issue is that it is not straightforward to extrapolate the results obtained for the scenarios to alternative climate change scenarios not initially included in the simulation protocols. Additional dynamic crop model simulations for new climate change scenarios are an option but this approach is costly, especially when a large number of crop models are used to generate the simulated data, as in AgMIP. Statistical models have been used to analyze responses of measured yield data to climate variables in past studies (Lobell et al., 2011), but the use of a statistical model to analyze yields simulated by complex
Large eddy simulations of coal jet flame ignition using the direct quadrature method of moments
Pedel, Julien
The Direct Quadrature Method of Moments (DQMOM) was implemented in the Large Eddy Simulation (LES) tool ARCHES to model coal particles. LES coupled with DQMOM was first applied to nonreacting particle-laden turbulent jets. Simulation results were compared to experimental data and accurately modeled a wide range of particle behaviors, such as particle jet waviness, spreading, break up, particle clustering and segregation, in different configurations. Simulations also accurately predicted the mean axial velocity along the centerline for both the gas phase and the solid phase, thus demonstrating the validity of the approach to model particles in turbulent flows. LES was then applied to the prediction of pulverized coal flame ignition. The stability of an oxy-coal flame as a function of changing primary gas composition (CO2 and O2) was first investigated. Flame stability was measured using optical measurements of the flame standoff distance in a 40 kW pilot facility. Large Eddy Simulations (LES) of the facility provided valuable insight into the experimentally observed data and the importance of factors such as heterogeneous reactions, radiation or wall temperature. The effects of three parameters on the flame stand-off distance were studied and simulation predictions were compared to experimental data using the data collaboration method. An additional validation study of the ARCHES LES tool was then performed on an air-fired pulverized coal jet flame ignited by a preheated gas flow. The simulation results were compared qualitatively and quantitatively to experimental observations for different inlet stoichiometric ratios. LES simulations were able to capture the various combustion regimes observed during flame ignition and to accurately model the flame stand-off distance sensitivity to the stoichiometric ratio. Gas temperature and coal burnout predictions were also examined and showed good agreement with experimental data. Overall, this research shows that high
Gilchrist, J. J.; Jordan, T. H.; Shaw, B. E.; Milner, K. R.; Richards-Dinger, K. B.; Dieterich, J. H.
2017-12-01
Within the SCEC Collaboratory for Interseismic Simulation and Modeling (CISM), we are developing physics-based forecasting models for earthquake ruptures in California. We employ the 3D boundary element code RSQSim (Rate-State Earthquake Simulator of Dieterich & Richards-Dinger, 2010) to generate synthetic catalogs with tens of millions of events that span up to a million years each. This code models rupture nucleation by rate- and state-dependent friction and Coulomb stress transfer in complex, fully interacting fault systems. The Uniform California Earthquake Rupture Forecast Version 3 (UCERF3) fault and deformation models are used to specify the fault geometry and long-term slip rates. We have employed the Blue Waters supercomputer to generate long catalogs of simulated California seismicity from which we calculate the forecasting statistics for large events. We have performed probabilistic seismic hazard analysis with RSQSim catalogs that were calibrated with system-wide parameters and found a remarkably good agreement with UCERF3 (Milner et al., this meeting). We build on this analysis, comparing the conditional probabilities of sequences of large events from RSQSim and UCERF3. In making these comparisons, we consider the epistemic uncertainties associated with the RSQSim parameters (e.g., rate- and state-frictional parameters), as well as the effects of model-tuning (e.g., adjusting the RSQSim parameters to match UCERF3 recurrence rates). The comparisons illustrate how physics-based rupture simulators might assist forecasters in understanding the short-term hazards of large aftershocks and multi-event sequences associated with complex, multi-fault ruptures.
Large Eddy Simulation of a thermal mixing tee in order to assess the thermal fatigue
International Nuclear Information System (INIS)
Galpin, J.; Simoneau, J.P.
2011-01-01
Highlights: → In this study, we perform a Large Eddy Simulation of a mixing tee, for which experimental thermal statistics are available. → A special methodology has been set up for comparing properly the fluctuations with the experiment. → A comparison between the Smagorinsky and the structure-function sub-grid scale model is achieved out. → Slight better predictions are obtained with the structure-function model. → The possibility to reduce the computational domain by prescribing synthetic turbulence at the inlet is tested. First results are encouraging and underline the advantage of considering this technique instead of a standard noise at the entrance of the domain. - Abstract: The present paper deals with thermal fatigue phenomenon, and more particularly with the numerical simulation using Large Eddy Simulation technique of a mixing tee, for which experimental thermal statistics are available. The sensitivity to the sub-grid scale closure is first evaluated by comparing the experimental statistics with the numerical results obtained via both the Smagorinsky and the structure-function models. Because of a difference of temporal resolution between the experiment and the simulation, the direct comparison of the fluctuations is not possible. Therefore, a methodology based on filtering the numerical results is proposed in order to achieve a proper comparison. The comparison of the numerical results with the experiment suggests that slight better predictions are obtained with the structure-function model even if the dependency of the results to the sub-grid scale model is low. Then, the possibility to reduce the fluid computational domain by prescribing synthetic turbulence at the inlet is tested. First results are encouraging and underline the advantage of considering this technique instead of a standard noise at the entrance of the domain. All the simulations are conducted with the commercial CFD code STAR-CD.
Simulations and measurements of beam loss patterns at the CERN Large Hadron Collider
Bruce, R.; Assmann, R. W.; Boccone, V.; Bracco, C.; Brugger, M.; Cauchi, M.; Cerutti, F.; Deboy, D.; Ferrari, A.; Lari, L.; Marsili, A.; Mereghetti, A.; Mirarchi, D.; Quaranta, E.; Redaelli, S.; Robert-Demolaize, G.; Rossi, A.; Salvachua, B.; Skordis, E.; Tambasco, C.; Valentino, G.; Weiler, T.; Vlachoudis, V.; Wollmann, D.
2014-08-01
The CERN Large Hadron Collider (LHC) is designed to collide proton beams of unprecedented energy, in order to extend the frontiers of high-energy particle physics. During the first very successful running period in 2010-2013, the LHC was routinely storing protons at 3.5-4 TeV with a total beam energy of up to 146 MJ, and even higher stored energies are foreseen in the future. This puts extraordinary demands on the control of beam losses. An uncontrolled loss of even a tiny fraction of the beam could cause a superconducting magnet to undergo a transition into a normal-conducting state, or in the worst case cause material damage. Hence a multistage collimation system has been installed in order to safely intercept high-amplitude beam protons before they are lost elsewhere. To guarantee adequate protection from the collimators, a detailed theoretical understanding is needed. This article presents results of numerical simulations of the distribution of beam losses around the LHC that have leaked out of the collimation system. The studies include tracking of protons through the fields of more than 5000 magnets in the 27 km LHC ring over hundreds of revolutions, and Monte Carlo simulations of particle-matter interactions both in collimators and machine elements being hit by escaping particles. The simulation results agree typically within a factor 2 with measurements of beam loss distributions from the previous LHC run. Considering the complex simulation, which must account for a very large number of unknown imperfections, and in view of the total losses around the ring spanning over 7 orders of magnitude, we consider this an excellent agreement. Our results give confidence in the simulation tools, which are used also for the design of future accelerators.
Simulations and measurements of beam loss patterns at the CERN Large Hadron Collider
Directory of Open Access Journals (Sweden)
R. Bruce
2014-08-01
Full Text Available The CERN Large Hadron Collider (LHC is designed to collide proton beams of unprecedented energy, in order to extend the frontiers of high-energy particle physics. During the first very successful running period in 2010–2013, the LHC was routinely storing protons at 3.5–4 TeV with a total beam energy of up to 146 MJ, and even higher stored energies are foreseen in the future. This puts extraordinary demands on the control of beam losses. An uncontrolled loss of even a tiny fraction of the beam could cause a superconducting magnet to undergo a transition into a normal-conducting state, or in the worst case cause material damage. Hence a multistage collimation system has been installed in order to safely intercept high-amplitude beam protons before they are lost elsewhere. To guarantee adequate protection from the collimators, a detailed theoretical understanding is needed. This article presents results of numerical simulations of the distribution of beam losses around the LHC that have leaked out of the collimation system. The studies include tracking of protons through the fields of more than 5000 magnets in the 27 km LHC ring over hundreds of revolutions, and Monte Carlo simulations of particle-matter interactions both in collimators and machine elements being hit by escaping particles. The simulation results agree typically within a factor 2 with measurements of beam loss distributions from the previous LHC run. Considering the complex simulation, which must account for a very large number of unknown imperfections, and in view of the total losses around the ring spanning over 7 orders of magnitude, we consider this an excellent agreement. Our results give confidence in the simulation tools, which are used also for the design of future accelerators.
Scientific data management challenges, technology and deployment
Rotem, Doron
2010-01-01
Dealing with the volume, complexity, and diversity of data currently being generated by scientific experiments and simulations often causes scientists to waste productive time. Scientific Data Management: Challenges, Technology, and Deployment describes cutting-edge technologies and solutions for managing and analyzing vast amounts of data, helping scientists focus on their scientific goals. The book begins with coverage of efficient storage systems, discussing how to write and read large volumes of data without slowing the simulation, analysis, or visualization processes. It then focuses on the efficient data movement and management of storage spaces and explores emerging database systems for scientific data. The book also addresses how to best organize data for analysis purposes, how to effectively conduct searches over large datasets, how to successfully automate multistep scientific process workflows, and how to automatically collect metadata and lineage information. This book provides a comprehensive u...
A method of orbital analysis for large-scale first-principles simulations
International Nuclear Information System (INIS)
Ohwaki, Tsukuru; Otani, Minoru; Ozaki, Taisuke
2014-01-01
An efficient method of calculating the natural bond orbitals (NBOs) based on a truncation of the entire density matrix of a whole system is presented for large-scale density functional theory calculations. The method recovers an orbital picture for O(N) electronic structure methods which directly evaluate the density matrix without using Kohn-Sham orbitals, thus enabling quantitative analysis of chemical reactions in large-scale systems in the language of localized Lewis-type chemical bonds. With the density matrix calculated by either an exact diagonalization or O(N) method, the computational cost is O(1) for the calculation of NBOs associated with a local region where a chemical reaction takes place. As an illustration of the method, we demonstrate how an electronic structure in a local region of interest can be analyzed by NBOs in a large-scale first-principles molecular dynamics simulation for a liquid electrolyte bulk model (propylene carbonate + LiBF 4 )
Copy of Using Emulation and Simulation to Understand the Large-Scale Behavior of the Internet.
Energy Technology Data Exchange (ETDEWEB)
Adalsteinsson, Helgi; Armstrong, Robert C.; Chiang, Ken; Gentile, Ann C.; Lloyd, Levi; Minnich, Ronald G.; Vanderveen, Keith; Van Randwyk, Jamie A; Rudish, Don W.
2008-10-01
We report on the work done in the late-start LDRDUsing Emulation and Simulation toUnderstand the Large-Scale Behavior of the Internet. We describe the creation of a researchplatform that emulates many thousands of machines to be used for the study of large-scale inter-net behavior. We describe a proof-of-concept simple attack we performed in this environment.We describe the successful capture of a Storm bot and, from the study of the bot and furtherliterature search, establish large-scale aspects we seek to understand via emulation of Storm onour research platform in possible follow-on work. Finally, we discuss possible future work.3
Valuation of large variable annuity portfolios: Monte Carlo simulation and synthetic datasets
Directory of Open Access Journals (Sweden)
Gan Guojun
2017-12-01
Full Text Available Metamodeling techniques have recently been proposed to address the computational issues related to the valuation of large portfolios of variable annuity contracts. However, it is extremely diffcult, if not impossible, for researchers to obtain real datasets frominsurance companies in order to test their metamodeling techniques on such real datasets and publish the results in academic journals. To facilitate the development and dissemination of research related to the effcient valuation of large variable annuity portfolios, this paper creates a large synthetic portfolio of variable annuity contracts based on the properties of real portfolios of variable annuities and implements a simple Monte Carlo simulation engine for valuing the synthetic portfolio. In addition, this paper presents fair market values and Greeks for the synthetic portfolio of variable annuity contracts that are important quantities for managing the financial risks associated with variable annuities. The resulting datasets can be used by researchers to test and compare the performance of various metamodeling techniques.
Planetary Structures And Simulations Of Large-scale Impacts On Mars
Swift, Damian; El-Dasher, B.
2009-09-01
The impact of large meteroids is a possible cause for isolated orogeny on bodies devoid of tectonic activity. On Mars, there is a significant, but not perfect, correlation between large, isolated volcanoes and antipodal impact craters. On Mercury and the Moon, brecciated terrain and other unusual surface features can be found at the antipodes of large impact sites. On Earth, there is a moderate correlation between long-lived mantle hotspots at opposite sides of the planet, with meteoroid impact suggested as a possible cause. If induced by impacts, the mechanisms of orogeny and volcanism thus appear to vary between these bodies, presumably because of differences in internal structure. Continuum mechanics (hydrocode) simulations have been used to investigate the response of planetary bodies to impacts, requiring assumptions about the structure of the body: its composition and temperature profile, and the constitutive properties (equation of state, strength, viscosity) of the components. We are able to predict theoretically and test experimentally the constitutive properties of matter under planetary conditions, with reasonable accuracy. To provide a reference series of simulations, we have constructed self-consistent planetary structures using simplified compositions (Fe core and basalt-like mantle), which turn out to agree surprisingly well with the moments of inertia. We have performed simulations of large-scale impacts, studying the transmission of energy to the antipodes. For Mars, significant antipodal heating to depths of a few tens of kilometers was predicted from compression waves transmitted through the mantle. Such heating is a mechanism for volcanism on Mars, possibly in conjunction with crustal cracking induced by surface waves. This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
Cloud-enabled large-scale land surface model simulations with the NASA Land Information System
Duffy, D.; Vaughan, G.; Clark, M. P.; Peters-Lidard, C. D.; Nijssen, B.; Nearing, G. S.; Rheingrover, S.; Kumar, S.; Geiger, J. V.
2017-12-01
Developed by the Hydrological Sciences Laboratory at NASA Goddard Space Flight Center (GSFC), the Land Information System (LIS) is a high-performance software framework for terrestrial hydrology modeling and data assimilation. LIS provides the ability to integrate satellite and ground-based observational products and advanced modeling algorithms to extract land surface states and fluxes. Through a partnership with the National Center for Atmospheric Research (NCAR) and the University of Washington, the LIS model is currently being extended to include the Structure for Unifying Multiple Modeling Alternatives (SUMMA). With the addition of SUMMA in LIS, meaningful simulations containing a large multi-model ensemble will be enabled and can provide advanced probabilistic continental-domain modeling capabilities at spatial scales relevant for water managers. The resulting LIS/SUMMA application framework is difficult for non-experts to install due to the large amount of dependencies on specific versions of operating systems, libraries, and compilers. This has created a significant barrier to entry for domain scientists that are interested in using the software on their own systems or in the cloud. In addition, the requirement to support multiple run time environments across the LIS community has created a significant burden on the NASA team. To overcome these challenges, LIS/SUMMA has been deployed using Linux containers, which allows for an entire software package along with all dependences to be installed within a working runtime environment, and Kubernetes, which orchestrates the deployment of a cluster of containers. Within a cloud environment, users can now easily create a cluster of virtual machines and run large-scale LIS/SUMMA simulations. Installations that have taken weeks and months can now be performed in minutes of time. This presentation will discuss the steps required to create a cloud-enabled large-scale simulation, present examples of its use, and
Zilberter, Ilya Alexandrovich
In this work, a hybrid Large Eddy Simulation / Reynolds-Averaged Navier Stokes (LES/RANS) turbulence model is applied to simulate two flows relevant to directed energy applications. The flow solver blends the Menter Baseline turbulence closure near solid boundaries with a Lenormand-type subgrid model in the free-stream with a blending function that employs the ratio of estimated inner and outer turbulent length scales. A Mach 2.2 mixing nozzle/diffuser system representative of a gas laser is simulated under a range of exit pressures to assess the ability of the model to predict the dynamics of the shock train. The simulation captures the location of the shock train responsible for pressure recovery but under-predicts the rate of pressure increase. Predicted turbulence production at the wall is found to be highly sensitive to the behavior of the RANS turbulence model. A Mach 2.3, high-Reynolds number, three-dimensional cavity flow is also simulated in order to compute the wavefront aberrations of an optical beam passing thorough the cavity. The cavity geometry is modeled using an immersed boundary method, and an auxiliary flat plate simulation is performed to replicate the effects of the wind-tunnel boundary layer on the computed optical path difference. Pressure spectra extracted on the cavity walls agree with empirical predictions based on Rossiter's formula. Proper orthogonal modes of the wavefront aberrations in a beam originating from the cavity center agree well with experimental data despite uncertainty about in flow turbulence levels and boundary layer thicknesses over the wind tunnel window. Dynamic mode decomposition of a planar wavefront spanning the cavity reveals that wavefront distortions are driven by shear layer oscillations at the Rossiter frequencies; these disturbances create eddy shocklets that propagate into the free-stream, creating additional optical wavefront distortion.
Homogeneous SPC/E water nucleation in large molecular dynamics simulations.
Angélil, Raymond; Diemand, Jürg; Tanaka, Kyoko K; Tanaka, Hidekazu
2015-08-14
We perform direct large molecular dynamics simulations of homogeneous SPC/E water nucleation, using up to ∼ 4 ⋅ 10(6) molecules. Our large system sizes allow us to measure extremely low and accurate nucleation rates, down to ∼ 10(19) cm(-3) s(-1), helping close the gap between experimentally measured rates ∼ 10(17) cm(-3) s(-1). We are also able to precisely measure size distributions, sticking efficiencies, cluster temperatures, and cluster internal densities. We introduce a new functional form to implement the Yasuoka-Matsumoto nucleation rate measurement technique (threshold method). Comparison to nucleation models shows that classical nucleation theory over-estimates nucleation rates by a few orders of magnitude. The semi-phenomenological nucleation model does better, under-predicting rates by at worst a factor of 24. Unlike what has been observed in Lennard-Jones simulations, post-critical clusters have temperatures consistent with the run average temperature. Also, we observe that post-critical clusters have densities very slightly higher, ∼ 5%, than bulk liquid. We re-calibrate a Hale-type J vs. S scaling relation using both experimental and simulation data, finding remarkable consistency in over 30 orders of magnitude in the nucleation rate range and 180 K in the temperature range.
Representative elements: A step to large-scale fracture system simulation
International Nuclear Information System (INIS)
Clemo, T.M.
1987-01-01
Large-scale simulation of flow and transport in fractured media requires the development of a technique to represent the effect of a large number of fractures. Representative elements are used as a tool to model a subset of a fracture system as a single distributed entity. Representative elements are part of a modeling concept called dual permeability. Dual permeability modeling combines discrete fracture simulation of the most important fractures with the distributed modeling of the less important fracture of a fracture system. This study investigates the use of stochastic analysis to determine properties of representative elements. Given an assumption of fully developed laminar flow, the net fracture conductivities and hence flow velocities can be determined from descriptive statistics of fracture spacing, orientation, aperture, and extent. The distribution of physical characteristics about their mean leads to a distribution of the associated conductivities. The variance of hydraulic conductivity induces dispersion into the transport process. Simple fracture systems are treated to demonstrate the usefulness of stochastic analysis. Explicit equations for conductivity of an element are developed and the dispersion characteristics are shown. Explicit formulation of the hydraulic conductivity and transport dispersion reveals the dependence of these important characteristics on the parameters used to describe the fracture system. Understanding these dependencies will help to focus efforts to identify the characteristics of fracture systems. Simulations of stochastically generated fracture sets do not provide this explicit functional dependence on the fracture system parameters. 12 refs., 6 figs
Large Eddy Simulation of Vertical Axis Wind Turbine wakes; Part II: effects of inflow turbulence
Duponcheel, Matthieu; Chatelain, Philippe; Caprace, Denis-Gabriel; Winckelmans, Gregoire
2017-11-01
The aerodynamics of Vertical Axis Wind Turbines (VAWTs) is inherently unsteady, which leads to vorticity shedding mechanisms due to both the lift distribution along the blade and its time evolution. Large-scale, fine-resolution Large Eddy Simulations of the flow past Vertical Axis Wind Turbines have been performed using a state-of-the-art Vortex Particle-Mesh (VPM) method combined with immersed lifting lines. Inflow turbulence with a prescribed turbulence intensity (TI) is injected at the inlet of the simulation from a precomputed synthetic turbulence field obtained using the Mann algorithm. The wake of a standard, medium-solidity, H-shaped machine is simulated for several TI levels. The complex wake development is captured in details and over long distances: from the blades to the near wake coherent vortices, then through the transitional ones to the fully developed turbulent far wake. Mean flow and turbulence statistics are computed over more than 10 diameters downstream of the machine. The sensitivity of the wake topology and decay to the TI level is assessed.
Topology of Large-Scale Structure by Galaxy Type: Hydrodynamic Simulations
Gott, J. Richard, III; Cen, Renyue; Ostriker, Jeremiah P.
1996-07-01
The topology of large-scale structure is studied as a function of galaxy type using the genus statistic. In hydrodynamical cosmological cold dark matter simulations, galaxies form on caustic surfaces (Zeldovich pancakes) and then slowly drain onto filaments and clusters. The earliest forming galaxies in the simulations (defined as "ellipticals") are thus seen at the present epoch preferentially in clusters (tending toward a meatball topology), while the latest forming galaxies (defined as "spirals") are seen currently in a spongelike topology. The topology is measured by the genus (number of "doughnut" holes minus number of isolated regions) of the smoothed density-contour surfaces. The measured genus curve for all galaxies as a function of density obeys approximately the theoretical curve expected for random- phase initial conditions, but the early-forming elliptical galaxies show a shift toward a meatball topology relative to the late-forming spirals. Simulations using standard biasing schemes fail to show such an effect. Large observational samples separated by galaxy type could be used to test for this effect.
Sudjai, W.; Juntasaro, V.; Juttijudata, V.
2018-01-01
The accuracy of predicting turbulence induced secondary flows is crucially important in many industrial applications such as turbine blade internal cooling passages in a gas turbine and fuel rod bundles in a nuclear reactor. A straight square duct is popularly used to reveal the characteristic of turbulence induced secondary flows which consists of two counter rotating vortices distributed in each corner of the duct. For a rotating duct, the flow can be divided into the pressure side and the suction side. The turbulence induced secondary flows are converted to the Coriolis force driven two large circulations with a pair of additional vortices on the pressure wall due to the rotational effect. In this paper, the Large Eddy Simulation (LES) of turbulence induced secondary flows in a straight square duct is performed using the ANSYS FLUENT CFD software. A dynamic kinetic energy subgrid-scale model is used to describe the three-dimensional incompressible turbulent flows in the stationary and the rotating straight square ducts. The Reynolds number based on the friction velocity and the hydraulic diameter is 300 with the various rotation numbers for the rotating cases. The flow is assumed fully developed by imposing the constant pressure gradient in the streamwise direction. For the rotating cases, the rotational axis is placed perpendicular to the streamwise direction. The simulation results on the secondary flows and the turbulent statistics are found to be in good agreement with the available Direct Numerical Simulation (DNS) data. Finally, the details of the Coriolis effects are discussed.
International Nuclear Information System (INIS)
Ng, Cho; Akcelik, Volkan; Candel, Arno; Chen, Sheng; Ge, Lixin; Kabel, Andreas; Lee, Lie-Quan; Li, Zenghai; Prudencio, Ernesto; Schussman, Greg; Uplenchwar1, Ravi; Xiao1, Liling; Ko1, Kwok; Austin, T.; Cary, J.R.; Ovtchinnikov, S.; Smith, D.N.; Werner, G.R.; Bellantoni, L.; TechX Corp.; Fermilab
2008-01-01
SciDAC1, with its support for the 'Advanced Computing for 21st Century Accelerator Science and Technology' (AST) project, witnessed dramatic advances in electromagnetic (EM) simulations for the design and optimization of important accelerators across the Office of Science. In SciDAC2, EM simulations continue to play an important role in the 'Community Petascale Project for Accelerator Science and Simulation' (ComPASS), through close collaborations with SciDAC CETs/Institutes in computational science. Existing codes will be improved and new multi-physics tools will be developed to model large accelerator systems with unprecedented realism and high accuracy using computing resources at petascale. These tools aim at targeting the most challenging problems facing the ComPASS project. Supported by advances in computational science research, they have been successfully applied to the International Linear Collider (ILC) and the Large Hadron Collider (LHC) in High Energy Physics (HEP), the JLab 12-GeV Upgrade in Nuclear Physics (NP), as well as the Spallation Neutron Source (SNS) and the Linac Coherent Light Source (LCLS) in Basic Energy Sciences (BES)
FDTD simulation of microwave sintering in large (500/4000 liter) multimode cavities
Energy Technology Data Exchange (ETDEWEB)
Subirats, M.; Iskander, M.F.; White, M.J. [Univ. of Utah, Salt Lake City, UT (United States). Electrical Engineering Dept.; Kiggans, J. [Oak Ridge National Lab., TN (United States)
1996-12-31
To help develop large-scale microwave-sintering processes and to explore the feasibility of the commercial utilization of this technology, the authors used the recently developed multi-grid 3D Finite-Difference Time-Domain (FDTD) code and the 3D Finite-Difference Heat-Transfer (FDHT) code to determine the electromagnetic (EM) fields, the microwave power deposition, and temperature-distribution patterns in layers of samples processed in large-scale multimode microwave cavities. This paper presents results obtained from the simulation of realistic sintering experiments carried out in both 500 and 4,000 liter furnaces operating at 2.45 GHz. The ceramic ware being sintered is placed inside a cubical crucible box made of rectangular plates of various ceramic materials with various electrical and thermal properties. The crucible box can accommodate up to 5 layers of ceramic samples with 16 to 20 cup-like samples per layer. Simulation results provided guidelines regarding selection of crucible-box materials, crucible-box geometry, number of layers, shelf material between layers, and the fraction volume of the load vs. that of the furnace. Results from the FDTD and FDHT simulations will be presented and various tradeoffs involved in designing an effective microwave-processing system will be compared graphically.
International Nuclear Information System (INIS)
Ng, C; Akcelik, V; Candel, A; Chen, S; Ge, L; Kabel, A; Lee, Lie-Quan; Li, Z; Prudencio, E; Schussman, G; Uplenchwar, R; Xiao, L; Ko, K; Austin, T; Cary, J R; Ovtchinnikov, S; Smith, D N; Werner, G R; Bellantoni, L
2008-01-01
SciDAC-1, with its support for the 'Advanced Computing for 21st Century Accelerator Science and Technology' project, witnessed dramatic advances in electromagnetic (EM) simulations for the design and optimization of important accelerators across the Office of Science. In SciDAC2, EM simulations continue to play an important role in the 'Community Petascale Project for Accelerator Science and Simulation' (ComPASS), through close collaborations with SciDAC Centers and Insitutes in computational science. Existing codes will be improved and new multi-physics tools will be developed to model large accelerator systems with unprecedented realism and high accuracy using computing resources at petascale. These tools aim at targeting the most challenging problems facing the ComPASS project. Supported by advances in computational science research, they have been successfully applied to the International Linear Collider and the Large Hadron Collider in high energy physics, the JLab 12-GeV Upgrade in nuclear physics, and the Spallation Neutron Source and the Linac Coherent Light Source in basic energy sciences
Unsteady adjoint for large eddy simulation of a coupled turbine stator-rotor system
Talnikar, Chaitanya; Wang, Qiqi; Laskowski, Gregory
2016-11-01
Unsteady fluid flow simulations like large eddy simulation are crucial in capturing key physics in turbomachinery applications like separation and wake formation in flow over a turbine vane with a downstream blade. To determine how sensitive the design objectives of the coupled system are to control parameters, an unsteady adjoint is needed. It enables the computation of the gradient of an objective with respect to a large number of inputs in a computationally efficient manner. In this paper we present unsteady adjoint solutions for a coupled turbine stator-rotor system. As the transonic fluid flows over the stator vane, the boundary layer transitions to turbulence. The turbulent wake then impinges on the rotor blades, causing early separation. This coupled system exhibits chaotic dynamics which causes conventional adjoint solutions to diverge exponentially, resulting in the corruption of the sensitivities obtained from the adjoint solutions for long-time simulations. In this presentation, adjoint solutions for aerothermal objectives are obtained through a localized adjoint viscosity injection method which aims to stabilize the adjoint solution and maintain accurate sensitivities. Preliminary results obtained from the supercomputer Mira will be shown in the presentation.
Large-Eddy Simulation of Internal Flow through Human Vocal Folds
Lasota, Martin; Šidlof, Petr
2018-06-01
The phonatory process occurs when air is expelled from the lungs through the glottis and the pressure drop causes flow-induced oscillations of the vocal folds. The flow fields created in phonation are highly unsteady and the coherent vortex structures are also generated. For accuracy it is essential to compute on humanlike computational domain and appropriate mathematical model. The work deals with numerical simulation of air flow within the space between plicae vocales and plicae vestibulares. In addition to the dynamic width of the rima glottidis, where the sound is generated, there are lateral ventriculus laryngis and sacculus laryngis included in the computational domain as well. The paper presents the results from OpenFOAM which are obtained with a large-eddy simulation using second-order finite volume discretization of incompressible Navier-Stokes equations. Large-eddy simulations with different subgrid scale models are executed on structured mesh. In these cases are used only the subgrid scale models which model turbulence via turbulent viscosity and Boussinesq approximation in subglottal and supraglottal area in larynx.
Large-Eddy Simulation of Internal Flow through Human Vocal Folds
Directory of Open Access Journals (Sweden)
Lasota Martin
2018-01-01
Full Text Available The phonatory process occurs when air is expelled from the lungs through the glottis and the pressure drop causes flow-induced oscillations of the vocal folds. The flow fields created in phonation are highly unsteady and the coherent vortex structures are also generated. For accuracy it is essential to compute on humanlike computational domain and appropriate mathematical model. The work deals with numerical simulation of air flow within the space between plicae vocales and plicae vestibulares. In addition to the dynamic width of the rima glottidis, where the sound is generated, there are lateral ventriculus laryngis and sacculus laryngis included in the computational domain as well. The paper presents the results from OpenFOAM which are obtained with a large-eddy simulation using second-order finite volume discretization of incompressible Navier-Stokes equations. Large-eddy simulations with different subgrid scale models are executed on structured mesh. In these cases are used only the subgrid scale models which model turbulence via turbulent viscosity and Boussinesq approximation in subglottal and supraglottal area in larynx.
World, We Have Problems: Simulation for Large Complex, Risky Projects, and Events
Elfrey, Priscilla
2010-01-01
Prior to a spacewalk during the NASA STS/129 mission in November 2009, Columbia Broadcasting System (CBS) correspondent William Harwood reported astronauts, "were awakened again", as they had been the day previously. Fearing something not properly connected was causing a leak, the crew, both on the ground and in space, stopped and checked everything. The alarm proved false. The crew did complete its work ahead of schedule, but the incident reminds us that correctly connecting hundreds and thousands of entities, subsystems and systems, finding leaks, loosening stuck valves, and adding replacements to very large complex systems over time does not occur magically. Everywhere major projects present similar pressures. Lives are at - risk. Responsibility is heavy. Large natural and human-created disasters introduce parallel difficulties as people work across boundaries their countries, disciplines, languages, and cultures with known immediate dangers as well as the unexpected. NASA has long accepted that when humans have to go where humans cannot go that simulation is the sole solution. The Agency uses simulation to achieve consensus, reduce ambiguity and uncertainty, understand problems, make decisions, support design, do planning and troubleshooting, as well as for operations, training, testing, and evaluation. Simulation is at the heart of all such complex systems, products, projects, programs, and events. Difficult, hazardous short and, especially, long-term activities have a persistent need for simulation from the first insight into a possibly workable idea or answer until the final report perhaps beyond our lifetime is put in the archive. With simulation we create a common mental model, try-out breakdowns of machinery or teamwork, and find opportunity for improvement. Lifecycle simulation proves to be increasingly important as risks and consequences intensify. Across the world, disasters are increasing. We anticipate more of them, as the results of global warming
Simulation of a Large Wildfire in a Coupled Fire-Atmosphere Model
Directory of Open Access Journals (Sweden)
Jean-Baptiste Filippi
2018-06-01
Full Text Available The Aullene fire devastated more than 3000 ha of Mediterranean maquis and pine forest in July 2009. The simulation of combustion processes, as well as atmospheric dynamics represents a challenge for such scenarios because of the various involved scales, from the scale of the individual flames to the larger regional scale. A coupled approach between the Meso-NH (Meso-scale Non-Hydrostatic atmospheric model running in LES (Large Eddy Simulation mode and the ForeFire fire spread model is proposed for predicting fine- to large-scale effects of this extreme wildfire, showing that such simulation is possible in a reasonable time using current supercomputers. The coupling involves the surface wind to drive the fire, while heat from combustion and water vapor fluxes are injected into the atmosphere at each atmospheric time step. To be representative of the phenomenon, a sub-meter resolution was used for the simulation of the fire front, while atmospheric simulations were performed with nested grids from 2400-m to 50-m resolution. Simulations were run with or without feedback from the fire to the atmospheric model, or without coupling from the atmosphere to the fire. In the two-way mode, the burnt area was reproduced with a good degree of realism at the local scale, where an acceleration in the valley wind and over sloping terrain pushed the fire line to locations in accordance with fire passing point observations. At the regional scale, the simulated fire plume compares well with the satellite image. The study explores the strong fire-atmosphere interactions leading to intense convective updrafts extending above the boundary layer, significant downdrafts behind the fire line in the upper plume, and horizontal wind speeds feeding strong inflow into the base of the convective updrafts. The fire-induced dynamics is induced by strong near-surface sensible heat fluxes reaching maximum values of 240 kW m − 2 . The dynamical production of turbulent kinetic
Helman, E. Udi
This dissertation conducts research into the large-scale simulation of oligopolistic competition in wholesale electricity markets. The dissertation has two parts. Part I is an examination of the structure and properties of several spatial, or network, equilibrium models of oligopolistic electricity markets formulated as mixed linear complementarity problems (LCP). Part II is a large-scale application of such models to the electricity system that encompasses most of the United States east of the Rocky Mountains, the Eastern Interconnection. Part I consists of Chapters 1 to 6. The models developed in this part continue research into mixed LCP models of oligopolistic electricity markets initiated by Hobbs [67] and subsequently developed by Metzler [87] and Metzler, Hobbs and Pang [88]. Hobbs' central contribution is a network market model with Cournot competition in generation and a price-taking spatial arbitrage firm that eliminates spatial price discrimination by the Cournot firms. In one variant, the solution to this model is shown to be equivalent to the "no arbitrage" condition in a "pool" market, in which a Regional Transmission Operator optimizes spot sales such that the congestion price between two locations is exactly equivalent to the difference in the energy prices at those locations (commonly known as locational marginal pricing). Extensions to this model are presented in Chapters 5 and 6. One of these is a market model with a profit-maximizing arbitrage firm. This model is structured as a mathematical program with equilibrium constraints (MPEC), but due to the linearity of its constraints, can be solved as a mixed LCP. Part II consists of Chapters 7 to 12. The core of these chapters is a large-scale simulation of the U.S. Eastern Interconnection applying one of the Cournot competition with arbitrage models. This is the first oligopolistic equilibrium market model to encompass the full Eastern Interconnection with a realistic network representation (using
Marvel-ous Dwarfs: Results from Four Heroically Large Simulated Volumes of Dwarf Galaxies
Munshi, Ferah; Brooks, Alyson; Weisz, Daniel; Bellovary, Jillian; Christensen, Charlotte
2018-01-01
We present results from high resolution, fully cosmological simulations of cosmic sheets that contain many dwarf galaxies. Together, they create the largest collection of simulated dwarf galaxies to date, with z=0 stellar masses comparable to the LMC or smaller. In total, we have simulated almost 100 luminous dwarf galaxies, forming a sample of simulated dwarfs which span a wide range of physical (stellar and halo mass) and evolutionary properties (merger history). We show how they can be calibrated against a wealth of observations of nearby galaxies including star formation histories, HI masses and kinematics, as well as stellar metallicities. We present preliminary results answering the following key questions: What is the slope of the stellar mass function at extremely low masses? Do halos with HI and no stars exist? What is the scatter in the stellar to halo mass relationship as a function of dwarf mass? What drives the scatter? With this large suite, we are beginning to statistically characterize dwarf galaxies and identify the types and numbers of outliers to expect.
On the effect of numerical errors in large eddy simulations of turbulent flows
International Nuclear Information System (INIS)
Kravchenko, A.G.; Moin, P.
1997-01-01
Aliased and dealiased numerical simulations of a turbulent channel flow are performed using spectral and finite difference methods. Analytical and numerical studies show that aliasing errors are more destructive for spectral and high-order finite-difference calculations than for low-order finite-difference simulations. Numerical errors have different effects for different forms of the nonlinear terms in the Navier-Stokes equations. For divergence and convective forms, spectral methods are energy-conserving only if dealiasing is performed. For skew-symmetric and rotational forms, both spectral and finite-difference methods are energy-conserving even in the presence of aliasing errors. It is shown that discrepancies between the results of dealiased spectral and standard nondialiased finite-difference methods are due to both aliasing and truncation errors with the latter being the leading source of differences. The relative importance of aliasing and truncation errors as compared to subgrid scale model terms in large eddy simulations is analyzed and discussed. For low-order finite-difference simulations, truncation errors can exceed the magnitude of the subgrid scale term. 25 refs., 17 figs., 1 tab
Energy Technology Data Exchange (ETDEWEB)
Quon, Eliot; Churchfield, Matthew; Cheung, Lawrence; Kern, Stefan
2017-02-01
This paper details the development of an aeroelastic wind plant model with large-eddy simulation (LES). The chosen LES solver is the Simulator for Wind Farm Applications (SOWFA) based on the OpenFOAM framework, coupled to NREL's comprehensive aeroelastic analysis tool, FAST. An atmospheric boundary layer (ABL) precursor simulation was constructed based on assessments of meteorological tower, lidar, and radar data over a 3-hour window. This precursor was tuned to the specific atmospheric conditions that occurred both prior to and during the measurement campaign, enabling capture of a night-to-day transition in the turbulent ABL. In the absence of height-varying temperature measurements, spatially averaged radar data were sufficient to characterize the atmospheric stability of the wind plant in terms of the shear profile, and near-ground temperature sensors provided a reasonable estimate of the ground heating rate describing the morning transition. A full aeroelastic simulation was then performed for a subset of turbines within the wind plant, driven by the precursor. Analysis of two turbines within the array, one directly waked by the other, demonstrated good agreement with measured time-averaged loads.
Development of a Wind Plant Large-Eddy Simulation with Measurement-Driven Atmospheric Inflow
Energy Technology Data Exchange (ETDEWEB)
Quon, Eliot W.; Churchfield, Matthew J.; Cheung, Lawrence; Kern, Stefan
2017-01-09
This paper details the development of an aeroelastic wind plant model with large-eddy simulation (LES). The chosen LES solver is the Simulator for Wind Farm Applications (SOWFA) based on the OpenFOAM framework, coupled to NREL's comprehensive aeroelastic analysis tool, FAST. An atmospheric boundary layer (ABL) precursor simulation was constructed based on assessments of meteorological tower, lidar, and radar data over a 3-hour window. This precursor was tuned to the specific atmospheric conditions that occurred both prior to and during the measurement campaign, enabling capture of a night-to-day transition in the turbulent ABL. In the absence of height-varying temperature measurements, spatially averaged radar data were sufficient to characterize the atmospheric stability of the wind plant in terms of the shear profile, and near-ground temperature sensors provided a reasonable estimate of the ground heating rate describing the morning transition. A full aeroelastic simulation was then performed for a subset of turbines within the wind plant, driven by the precursor. Analysis of two turbines within the array, one directly waked by the other, demonstrated good agreement with measured time-averaged loads.
A regularized vortex-particle mesh method for large eddy simulation
Spietz, H. J.; Walther, J. H.; Hejlesen, M. M.
2017-11-01
We present recent developments of the remeshed vortex particle-mesh method for simulating incompressible fluid flow. The presented method relies on a parallel higher-order FFT based solver for the Poisson equation. Arbitrary high order is achieved through regularization of singular Green's function solutions to the Poisson equation and recently we have derived novel high order solutions for a mixture of open and periodic domains. With this approach the simulated variables may formally be viewed as the approximate solution to the filtered Navier Stokes equations, hence we use the method for Large Eddy Simulation by including a dynamic subfilter-scale model based on test-filters compatible with the aforementioned regularization functions. Further the subfilter-scale model uses Lagrangian averaging, which is a natural candidate in light of the Lagrangian nature of vortex particle methods. A multiresolution variation of the method is applied to simulate the benchmark problem of the flow past a square cylinder at Re = 22000 and the obtained results are compared to results from the literature.
rpe v5: an emulator for reduced floating-point precision in large numerical simulations
Dawson, Andrew; Düben, Peter D.
2017-06-01
This paper describes the rpe (reduced-precision emulator) library which has the capability to emulate the use of arbitrary reduced floating-point precision within large numerical models written in Fortran. The rpe software allows model developers to test how reduced floating-point precision affects the result of their simulations without having to make extensive code changes or port the model onto specialized hardware. The software can be used to identify parts of a program that are problematic for numerical precision and to guide changes to the program to allow a stronger reduction in precision.The development of rpe was motivated by the strong demand for more computing power. If numerical precision can be reduced for an application under consideration while still achieving results of acceptable quality, computational cost can be reduced, since a reduction in numerical precision may allow an increase in performance or a reduction in power consumption. For simulations with weather and climate models, savings due to a reduction in precision could be reinvested to allow model simulations at higher spatial resolution or complexity, or to increase the number of ensemble members to improve predictions. rpe was developed with a particular focus on the community of weather and climate modelling, but the software could be used with numerical simulations from other domains.
Large Eddy Simulations of a Bottom Boundary Layer Under a Shallow Geostrophic Front
Bateman, S. P.; Simeonov, J.; Calantoni, J.
2017-12-01
The unstratified surf zone and the stratified shelf waters are often separated by dynamic fronts that can strongly impact the character of the Ekman bottom boundary layer. Here, we use large eddy simulations to study the turbulent bottom boundary layer associated with a geostrophic current on a stratified shelf of uniform depth. The simulations are initialized with a spatially uniform vertical shear that is in geostrophic balance with a pressure gradient due to a linear horizontal temperature variation. Superposed on the temperature front is a stable vertical temperature gradient. As turbulence develops near the bottom, the turbulence-induced mixing gradually erodes the initial uniform temperature stratification and a well-mixed layer grows in height until the turbulence becomes fully developed. The simulations provide the spatial distribution of the turbulent dissipation and the Reynolds stresses in the fully developed boundary layer. We vary the initial linear stratification and investigate its effect on the height of the bottom boundary layer and the turbulence statistics. The results are compared to previous models and simulations of stratified bottom Ekman layers.
Energy Technology Data Exchange (ETDEWEB)
Lorteau, Mathieu, E-mail: mathieu.lorteau@onera.fr; Cléro, Franck, E-mail: franck.clero@onera.fr; Vuillot, François, E-mail: francois.vuillot@onera.fr [Onera–The French Aerospace Lab, F-92322 Châtillon (France)
2015-07-15
In the framework of jet noise computation, a numerical simulation of a subsonic turbulent hot jet is performed using large-eddy simulation. A geometrical tripping is used in order to trigger the turbulence at the nozzle exit. In a first part, the validity of the simulation is assessed by comparison with experimental measurements. The mean and rms velocity fields show good agreement, so do the azimuthal composition of the near pressure field and the far field spectra. Discrepancies remain close to the nozzle exit which lead to a limited overestimation of the pressure levels in both near and far fields, especially near the 90{sup ∘} angular sector. Two point correlation analyses are then applied to the data obtained from the simulation. These enable to link the downstream acoustic radiation, which is the main direction of radiation, to pressure waves developing in the shear layer and propagating toward the potential core end. The intermittency of the downstream acoustic radiation is evidenced and related to the coherent structures developing in the shear layer.
Evaluation of sub grid scale and local wall models in Large-eddy simulations of separated flow
Sam Ali Al; Szasz Robert; Revstedt Johan
2015-01-01
The performance of the Sub Grid Scale models is studied by simulating a separated flow over a wavy channel. The first and second order statistical moments of the resolved velocities obtained by using Large-Eddy simulations at different mesh resolutions are compared with Direct Numerical Simulations data. The effectiveness of modeling the wall stresses by using local log-law is then tested on a relatively coarse grid. The results exhibit a good agreement between highly-resolved Large Eddy Simu...
Tian, Yang; Liu, Zhilin; Li, Xiaoqian; Zhang, Lihua; Li, Ruiqing; Jiang, Ripeng; Dong, Fang
2018-05-01
Ultrasonic sonotrodes play an essential role in transmitting power ultrasound into the large-scale metallic casting. However, cavitation erosion considerably impairs the in-service performance of ultrasonic sonotrodes, leading to marginal microstructural refinement. In this work, the cavitation erosion behaviour of ultrasonic sonotrodes in large-scale castings was explored using the industry-level experiments of Al alloy cylindrical ingots (i.e. 630 mm in diameter and 6000 mm in length). When introducing power ultrasound, severe cavitation erosion was found to reproducibly occur at some specific positions on ultrasonic sonotrodes. However, there is no cavitation erosion present on the ultrasonic sonotrodes that were not driven by electric generator. Vibratory examination showed cavitation erosion depended on the vibration state of ultrasonic sonotrodes. Moreover, a finite element (FE) model was developed to simulate the evolution and distribution of acoustic pressure in 3-D solidification volume. FE simulation results confirmed that significant dynamic interaction between sonotrodes and melts only happened at some specific positions corresponding to severe cavitation erosion. This work will allow for developing more advanced ultrasonic sonotrodes with better cavitation erosion-resistance, in particular for large-scale castings, from the perspectives of ultrasonic physics and mechanical design. Copyright © 2018 Elsevier B.V. All rights reserved.
Experiments and Large-Eddy Simulations of acoustically forced bluff-body flows
Energy Technology Data Exchange (ETDEWEB)
Ayache, S.; Dawson, J.R.; Triantafyllidis, A. [Department of Engineering, University of Cambridge (United Kingdom); Balachandran, R. [Department of Mechanical Engineering, University College London (United Kingdom); Mastorakos, E., E-mail: em257@eng.cam.ac.u [Department of Engineering, University of Cambridge (United Kingdom)
2010-10-15
The isothermal air flow behind an enclosed axisymmetric bluff body, with the incoming flow being forced by a loudspeaker at a single frequency and with large amplitude, has been explored with high data-rate Laser-Doppler Anemometry measurements and Large-Eddy Simulations. The comparison between experiment and simulations allows a quantification of the accuracy of LES for turbulent flows with periodicity and the results provide insights into the structure of flows relevant to combustors undergoing self-excited oscillations. At low forcing frequencies, the whole flow pulsates with the incoming flow, although at a phase lag that depends on spatial location. At high forcing frequencies, vortices are shed from the bluff body and the recirculation zone, as a whole, pulsates less. Despite the fact that the incoming flow has an oscillation that is virtually monochromatic, the velocity spectra show peaks at various harmonics, whose relative magnitudes vary with location. A sub-harmonic peak is also observed inside the recirculation zone possibly caused by merging of the shed vortices. The phase-averaged turbulent fluctuations show large temporal and spatial variations. The LES reproduces reasonably accurately the experimental findings in terms of phase-averaged mean and r.m.s. velocities, vortex formation, and spectral peaks.
Experiments and Large-Eddy Simulations of acoustically forced bluff-body flows
International Nuclear Information System (INIS)
Ayache, S.; Dawson, J.R.; Triantafyllidis, A.; Balachandran, R.; Mastorakos, E.
2010-01-01
The isothermal air flow behind an enclosed axisymmetric bluff body, with the incoming flow being forced by a loudspeaker at a single frequency and with large amplitude, has been explored with high data-rate Laser-Doppler Anemometry measurements and Large-Eddy Simulations. The comparison between experiment and simulations allows a quantification of the accuracy of LES for turbulent flows with periodicity and the results provide insights into the structure of flows relevant to combustors undergoing self-excited oscillations. At low forcing frequencies, the whole flow pulsates with the incoming flow, although at a phase lag that depends on spatial location. At high forcing frequencies, vortices are shed from the bluff body and the recirculation zone, as a whole, pulsates less. Despite the fact that the incoming flow has an oscillation that is virtually monochromatic, the velocity spectra show peaks at various harmonics, whose relative magnitudes vary with location. A sub-harmonic peak is also observed inside the recirculation zone possibly caused by merging of the shed vortices. The phase-averaged turbulent fluctuations show large temporal and spatial variations. The LES reproduces reasonably accurately the experimental findings in terms of phase-averaged mean and r.m.s. velocities, vortex formation, and spectral peaks.
Large-scale and Long-duration Simulation of a Multi-stage Eruptive Solar Event
Jiang, chaowei; Hu, Qiang; Wu, S. T.
2015-04-01
We employ a data-driven 3D MHD active region evolution model by using the Conservation Element and Solution Element (CESE) numerical method. This newly developed model retains the full MHD effects, allowing time-dependent boundary conditions and time evolution studies. The time-dependent simulation is driven by measured vector magnetograms and the method of MHD characteristics on the bottom boundary. We have applied the model to investigate the coronal magnetic field evolution of AR11283 which was characterized by a pre-existing sigmoid structure in the core region and multiple eruptions, both in relatively small and large scales. We have succeeded in producing the core magnetic field structure and the subsequent eruptions of flux-rope structures (see https://dl.dropboxusercontent.com/u/96898685/large.mp4 for an animation) as the measured vector magnetograms on the bottom boundary evolve in time with constant flux emergence. The whole process, lasting for about an hour in real time, compares well with the corresponding SDO/AIA and coronagraph imaging observations. From these results, we show the capability of the model, largely data-driven, that is able to simulate complex, topological, and highly dynamic active region evolutions. (We acknowledge partial support of NSF grants AGS 1153323 and AGS 1062050, and data support from SDO/HMI and AIA teams).
A Steam Jet Plume Simulation in a Large Bulk Space with a System Code MARS
International Nuclear Information System (INIS)
Bae, Sung Won; Chung, Bub Dong
2006-01-01
From May 2002, the OECD-SETH group has launched the PANDA Project in order to provide an experimental data base for a multi-dimensional code assessment. OECD-SETH group expects the PANDA Project will meet the increasing needs for adequate experimental data for a 3D distribution of relevant variables like the temperature, velocity and steam-air concentrations that are measured with a sufficient resolution and accuracy. The scope of the PANDA Project is the mixture stratification and mixing phenomena in a large bulk space. Total of 24 test series are still being performed in PSI, Switzerland. The PANDA facility consists of 2 main large vessels and 1 connection pipe Within the large vessels, a steam injection nozzle and outlet vent are arranged for each test case. These tests are categorized into 3 modes, i.e. the high momentum, near wall plume, and free plume tests. KAERI has also participated in the SETH group since 1997 so that the multi-dimensional capability of the MARS code could be assessed and developed. Test 17, the high steam jet injection test, has already been simulated by MARS and shows promising results. Now, the test 9 and 9bis cases which use a low speed horizontal steam jet flow have been simulated and investigated
Large Eddy and Interface Simulation (LEIS) of liquid entrainment in turbulent stratified flow
International Nuclear Information System (INIS)
Gulati, S.; Buongiorno, J.; Lakehal, D.
2011-01-01
Dryout of the liquid film on the fuel rods in BWR fuel assemblies leads to an abrupt decrease in heat transfer coefficient and can result in fuel failure. The process of mechanical mass transfer from the continuous liquid field into the continuous vapor field along the liquid-vapor interface is called entrainment and is the dominant depletion mechanism for the liquid film in annular flow. Using interface tracking methods combined with a Large Eddy Simulation approach, implemented in the Computational Multi-Fluid Dynamics (CMFD) code TransAT®, we are studying entrainment phenomena in BWR fuel assemblies. In this paper we report on the CMFD simulation approaches and the current validation effort for the code. (author)
Efficient Meshfree Large Deformation Simulation of Rainfall Induced Soil Slope Failure
Wang, Dongdong; Li, Ling
2010-05-01
An efficient Lagrangian Galerkin meshfree framework is presented for large deformation simulation of rainfall-induced soil slope failure. Detailed coupled soil-rainfall seepage equations are given for the proposed formulation. This nonlinear meshfree formulation is featured by the Lagrangian stabilized conforming nodal integration method where the low cost nature of nodal integration approach is kept and at the same time the numerical stability is maintained. The initiation and evolution of progressive failure in the soil slope is modeled by the coupled constitutive equations of isotropic damage and Drucker-Prager pressure-dependent plasticity. The gradient smoothing in the stabilized conforming integration also serves as a non-local regularization of material instability and consequently the present method is capable of effectively capture the shear band failure. The efficacy of the present method is demonstrated by simulating the rainfall-induced failure of two typical soil slopes.
Automatic Optimization for Large-Scale Real-Time Coastal Water Simulation
Directory of Open Access Journals (Sweden)
Shunli Wang
2016-01-01
Full Text Available We introduce an automatic optimization approach for the simulation of large-scale coastal water. To solve the singular problem of water waves obtained with the traditional model, a hybrid deep-shallow-water model is estimated by using an automatic coupling algorithm. It can handle arbitrary water depth and different underwater terrain. As a certain feature of coastal terrain, coastline is detected with the collision detection technology. Then, unnecessary water grid cells are simplified by the automatic simplification algorithm according to the depth. Finally, the model is calculated on Central Processing Unit (CPU and the simulation is implemented on Graphics Processing Unit (GPU. We show the effectiveness of our method with various results which achieve real-time rendering on consumer-level computer.
Energy Technology Data Exchange (ETDEWEB)
Lee, Kyongjun; Yang, Kyung-Soo [Inha University, Incheon (Korea, Republic of)
2017-04-15
Large Eddy simulation (LES) results of turbulent flow past a circular cylinder for the specified Reynolds numbers (Re = 63100, 126000, 252000) are presented. An immersed boundary method was employed to facilitate implementation of a circular cylinder in a Cartesian grid system. A dynamic subgrid-scale model, in which the model coefficient is dynamically determined by the current resolved flow field rather than assigned a prefixed constant, was implemented for accurate turbulence modeling. For better resolution near the cylinder surface and in the separated free-shear layers, a composite grid was used. Flow statistics including mean and rms values of force coefficients and Strouhal number of vortex shedding, are presented. Flow visualization using vorticity or Q contours are also shown. Our results are in better agreement with the MARIN measurements compared with RANS calculations reported in the previous ITTC workshop, confirming that LES is a more appropriate simulation methodology than a RANS approach to predict VIV for marine structures.
RTC simulations on large branched sewer systems with SmaRTControl.
de Korte, Kees; van Beest, Dick; van der Plaat, Marcel; de Graaf, Erno; Schaart, Niels
2009-01-01
In The Netherlands many large branched sewer systems exist. RTC can improve the performance of these systems. The objective of the universal algorithm of SmaRTControl is to improve the performance of the sewer system and the WWTP. The effect of RTC under rain weather flow conditions is simulated using a hydrological model with 19 drainage districts. The system related inefficiency coefficient (SIC) is introduced for assessment of the performance of sewer systems. The performance can be improved by RTC in combination with increased pumping capacities in the drainage districts, but without increasing the flow to the WWTP. Under dry weather flow conditions the flow to the WWTP can be equalized by storage of wastewater in the sewer system. It is concluded that SmaRTControl can improve the performance, that simulations are necessary and that SIC is an excellent parameter for assessment of the performance.
Jiang, Zhou; Xia, Zhenhua; Shi, Yipeng; Chen, Shiyi
2018-04-01
A fully developed spanwise rotating turbulent channel flow has been numerically investigated utilizing large-eddy simulation. Our focus is to assess the performances of the dynamic variants of eddy viscosity models, including dynamic Vreman's model (DVM), dynamic wall adapting local eddy viscosity (DWALE) model, dynamic σ (Dσ ) model, and the dynamic volumetric strain-stretching (DVSS) model, in this canonical flow. The results with dynamic Smagorinsky model (DSM) and direct numerical simulations (DNS) are used as references. Our results show that the DVM has a wrong asymptotic behavior in the near wall region, while the other three models can correctly predict it. In the high rotation case, the DWALE can get reliable mean velocity profile, but the turbulence intensities in the wall-normal and spanwise directions show clear deviations from DNS data. DVSS exhibits poor predictions on both the mean velocity profile and turbulence intensities. In all three cases, Dσ performs the best.
A Generic Discrete-Event Simulation Model for Outpatient Clinics in a Large Public Hospital
Directory of Open Access Journals (Sweden)
Waressara Weerawat
2013-01-01
Full Text Available The orthopedic outpatient department (OPD ward in a large Thai public hospital is modeled using Discrete-Event Stochastic (DES simulation. Key Performance Indicators (KPIs are used to measure effects across various clinical operations during different shifts throughout the day. By considering various KPIs such as wait times to see doctors, percentage of patients who can see a doctor within a target time frame, and the time that the last patient completes their doctor consultation, bottlenecks are identified and resource-critical clinics can be prioritized. The simulation model quantifies the chronic, high patient congestion that is prevalent amongst Thai public hospitals with very high patient-to-doctor ratios. Our model can be applied across five different OPD wards by modifying the model parameters. Throughout this work, we show how DES models can be used as decision-support tools for hospital management.
Dynamic subgrid scale model of large eddy simulation of cross bundle flows
International Nuclear Information System (INIS)
Hassan, Y.A.; Barsamian, H.R.
1996-01-01
The dynamic subgrid scale closure model of Germano et. al (1991) is used in the large eddy simulation code GUST for incompressible isothermal flows. Tube bundle geometries of staggered and non-staggered arrays are considered in deep bundle simulations. The advantage of the dynamic subgrid scale model is the exclusion of an input model coefficient. The model coefficient is evaluated dynamically for each nodal location in the flow domain. Dynamic subgrid scale results are obtained in the form of power spectral densities and flow visualization of turbulent characteristics. Comparisons are performed among the dynamic subgrid scale model, the Smagorinsky eddy viscosity model (that is used as the base model for the dynamic subgrid scale model) and available experimental data. Spectral results of the dynamic subgrid scale model correlate better with experimental data. Satisfactory turbulence characteristics are observed through flow visualization
One-Way Nested Large-Eddy Simulation over the Askervein Hill
Directory of Open Access Journals (Sweden)
James D. Doyle
2009-07-01
Full Text Available Large-eddy simulation (LES models have been used extensively to study atmospheric boundary layer turbulence over flat surfaces; however, LES applications over topography are less common. We evaluate the ability of an existing model – COAMPS^{®}-LES – to simulate flow over terrain using data from the Askervein Hill Project. A new approach is suggested for the treatment of the lateral boundaries using one-way grid nesting. LES wind profile and speed-up are compared with observations at various locations around the hill. The COAMPS-LES model performs generally well. This case could serve as a useful benchmark for evaluating LES models for applications over topography.
Computation of unsteady flow and aerodynamic noise of NACA0018 airfoil using large-eddy simulation
Energy Technology Data Exchange (ETDEWEB)
Kim, H.-J. [Department of Mechanical Engineering, Inha University, 253 Yonghyun-dong, Nam-gu, Incheon 402-751 (Korea, Republic of); Lee, S. [Department of Mechanical Engineering, Inha University, 253 Yonghyun-dong, Nam-gu, Incheon 402-751 (Korea, Republic of)]. E-mail: sbaelee@inha.ac.kr; Fujisawa, N. [Department of Mechanical and Production Engineering, Niigata University, 8050 Ikarashi-2, Niigata 950-2181 (Japan)
2006-04-15
The flow field around a symmetrical NACA airfoil in the uniform flow under generation of noise was numerically studied. The numerical simulation was carried out by a large-eddy simulation that employs a deductive dynamic model as the subgrid-scale model. The results at small angle of attack {alpha} = 3-6{sup o} indicate that the discrete frequency noise is generated when the separated laminar flow reattaches near the trailing edge of pressure side and the strong instability thereafter affects positive vortices shed near the trailing edge. The quasi-periodic behavior of negative vortex formation on the suction side is affected by the strength and the periodicity of positive vortices near the trailing edge. The computation using aero-acoustic analogy indicates the primary discrete peak at the Strouhal frequency (=2f . {delta}/U ) of 0.15 by the vortex shedding from the trailing edge, which is in a close agreement with the experiment.
Research on simulation of supercritical steam turbine system in large thermal power station
Zhou, Qiongyang
2018-04-01
In order to improve the stability and safety of supercritical steam turbine system operation in large thermal power station, the body of the steam turbine is modeled in this paper. And in accordance with the hierarchical modeling idea, the steam turbine body model, condensing system model, deaeration system model and regenerative system model are combined to build a simulation model of steam turbine system according to the connection relationship of each subsystem of steam turbine. Finally, the correctness of the model is verified by design and operation data of the 600MW supercritical unit. The results show that the maximum simulation error of the model is 2.15%, which meets the requirements of the engineering. This research provides a platform for the research on the variable operating conditions of the turbine system, and lays a foundation for the construction of the whole plant model of the thermal power plant.
Kreye, Phillip; Meon, Günter
2016-04-01
Complex concepts for the physically correct depiction of dominant processes in the hydrosphere are increasingly at the forefront of hydrological modelling. Many scientific issues in hydrological modelling demand for additional system variables besides a simulation of runoff only, such as groundwater recharge or soil moisture conditions. Models that include soil water simulations are either very simplified or require a high number of parameters. Against this backdrop there is a heightened demand of observations to be used to calibrate the model. A reasonable integration of groundwater data or remote sensing data in calibration procedures as well as the identifiability of physically plausible sets of parameters is subject to research in the field of hydrology. Since this data is often combined with conceptual models, the given interfaces are not suitable for such demands. Furthermore, the application of automated optimisation procedures is generally associated with conceptual models, whose (fast) computing times allow many iterations of the optimisation in an acceptable time frame. One of the main aims of this study is to reduce the discrepancy between scientific and practical applications in the field of hydrological modelling. Therefore, the soil model DYVESOM (DYnamic VEgetation SOil Model) was developed as one of the primary components of the hydrological modelling system PANTA RHEI. DYVESOMs structure provides the required interfaces for the calibrations made at runoff, satellite based soil moisture and groundwater level. The model considers spatial and temporal differentiated feedback of the development of the vegetation on the soil system. In addition, small scale heterogeneities of soil properties (subgrid-variability) are parameterized by variation of van Genuchten parameters depending on distribution functions. Different sets of parameters are operated simultaneously while interacting with each other. The developed soil model is innovative regarding concept
Numerical simulation of seismic wave propagation from land-excited large volume air-gun source
Cao, W.; Zhang, W.
2017-12-01
The land-excited large volume air-gun source can be used to study regional underground structures and to detect temporal velocity changes. The air-gun source is characterized by rich low frequency energy (from bubble oscillation, 2-8Hz) and high repeatability. It can be excited in rivers, reservoirs or man-made pool. Numerical simulation of the seismic wave propagation from the air-gun source helps to understand the energy partitioning and characteristics of the waveform records at stations. However, the effective energy recorded at a distance station is from the process of bubble oscillation, which can not be approximated by a single point source. We propose a method to simulate the seismic wave propagation from the land-excited large volume air-gun source by finite difference method. The process can be divided into three parts: bubble oscillation and source coupling, solid-fluid coupling and the propagation in the solid medium. For the first part, the wavelet of the bubble oscillation can be simulated by bubble model. We use wave injection method combining the bubble wavelet with elastic wave equation to achieve the source coupling. Then, the solid-fluid boundary condition is implemented along the water bottom. And the last part is the seismic wave propagation in the solid medium, which can be readily implemented by the finite difference method. Our method can get accuracy waveform of land-excited large volume air-gun source. Based on the above forward modeling technology, we analysis the effect of the excited P wave and the energy of converted S wave due to different water shapes. We study two land-excited large volume air-gun fields, one is Binchuan in Yunnan, and the other is Hutubi in Xinjiang. The station in Binchuan, Yunnan is located in a large irregular reservoir, the waveform records have a clear S wave. Nevertheless, the station in Hutubi, Xinjiang is located in a small man-made pool, the waveform records have very weak S wave. Better understanding of
A long-term, continuous simulation approach for large-scale flood risk assessments
Falter, Daniela; Schröter, Kai; Viet Dung, Nguyen; Vorogushyn, Sergiy; Hundecha, Yeshewatesfa; Kreibich, Heidi; Apel, Heiko; Merz, Bruno
2014-05-01
The Regional Flood Model (RFM) is a process based model cascade developed for flood risk assessments of large-scale basins. RFM consists of four model parts: the rainfall-runoff model SWIM, a 1D channel routing model, a 2D hinterland inundation model and the flood loss estimation model for residential buildings FLEMOps+r. The model cascade was recently undertaken a proof-of-concept study at the Elbe catchment (Germany) to demonstrate that flood risk assessments, based on a continuous simulation approach, including rainfall-runoff, hydrodynamic and damage estimation models, are feasible for large catchments. The results of this study indicated that uncertainties are significant, especially for hydrodynamic simulations. This was basically a consequence of low data quality and disregarding dike breaches. Therefore, RFM was applied with a refined hydraulic model setup for the Elbe tributary Mulde. The study area Mulde catchment comprises about 6,000 km2 and 380 river-km. The inclusion of more reliable information on overbank cross-sections and dikes considerably improved the results. For the application of RFM for flood risk assessments, long-term climate input data is needed to drive the model chain. This model input was provided by a multi-site, multi-variate weather generator that produces sets of synthetic meteorological data reproducing the current climate statistics. The data set comprises 100 realizations of 100 years of meteorological data. With the proposed continuous simulation approach of RFM, we simulated a virtual period of 10,000 years covering the entire flood risk chain including hydrological, 1D/2D hydrodynamic and flood damage estimation models. This provided a record of around 2.000 inundation events affecting the study area with spatially detailed information on inundation depths and damage to residential buildings on a resolution of 100 m. This serves as basis for a spatially consistent, flood risk assessment for the Mulde catchment presented in
Li, J.
2017-12-01
Large-watershed flood simulation and forecasting is very important for a distributed hydrological model in the application. There are some challenges including the model's spatial resolution effect, model performance and accuracy and so on. To cope with the challenge of the model's spatial resolution effect, different model resolution including 1000m*1000m, 600m*600m, 500m*500m, 400m*400m, 200m*200m were used to build the distributed hydrological model—Liuxihe model respectively. The purpose is to find which one is the best resolution for Liuxihe model in Large-watershed flood simulation and forecasting. This study sets up a physically based distributed hydrological model for flood forecasting of the Liujiang River basin in south China. Terrain data digital elevation model (DEM), soil type and land use type are downloaded from the website freely. The model parameters are optimized by using an improved Particle Swarm Optimization(PSO) algorithm; And parameter optimization could reduce the parameter uncertainty that exists for physically deriving model parameters. The different model resolution (200m*200m—1000m*1000m ) are proposed for modeling the Liujiang River basin flood with the Liuxihe model in this study. The best model's spatial resolution effect for flood simulation and forecasting is 200m*200m.And with the model's spatial resolution reduction, the model performance and accuracy also become worse and worse. When the model resolution is 1000m*1000m, the flood simulation and forecasting result is the worst, also the river channel divided based on this resolution is differs from the actual one. To keep the model with an acceptable performance, minimum model spatial resolution is needed. The suggested threshold model spatial resolution for modeling the Liujiang River basin flood is a 500m*500m grid cell, but the model spatial resolution with a 200m*200m grid cell is recommended in this study to keep the model at a best performance.
Load Balancing Scientific Applications
Energy Technology Data Exchange (ETDEWEB)
Pearce, Olga Tkachyshyn [Texas A & M Univ., College Station, TX (United States)
2014-12-01
The largest supercomputers have millions of independent processors, and concurrency levels are rapidly increasing. For ideal efficiency, developers of the simulations that run on these machines must ensure that computational work is evenly balanced among processors. Assigning work evenly is challenging because many large modern parallel codes simulate behavior of physical systems that evolve over time, and their workloads change over time. Furthermore, the cost of imbalanced load increases with scale because most large-scale scientific simulations today use a Single Program Multiple Data (SPMD) parallel programming model, and an increasing number of processors will wait for the slowest one at the synchronization points. To address load imbalance, many large-scale parallel applications use dynamic load balance algorithms to redistribute work evenly. The research objective of this dissertation is to develop methods to decide when and how to load balance the application, and to balance it effectively and affordably. We measure and evaluate the computational load of the application, and develop strategies to decide when and how to correct the imbalance. Depending on the simulation, a fast, local load balance algorithm may be suitable, or a more sophisticated and expensive algorithm may be required. We developed a model for comparison of load balance algorithms for a specific state of the simulation that enables the selection of a balancing algorithm that will minimize overall runtime.
Use of a large-scale rainfall simulator reveals novel insights into stemflow generation
Levia, D. F., Jr.; Iida, S. I.; Nanko, K.; Sun, X.; Shinohara, Y.; Sakai, N.
2017-12-01
Detailed knowledge of stemflow generation and its effects on both hydrological and biogoechemical cycling is important to achieve a holistic understanding of forest ecosystems. Field studies and a smaller set of experiments performed under laboratory conditions have increased our process-based knowledge of stemflow production. Building upon these earlier works, a large-scale rainfall simulator was employed to deepen our understanding of stemflow generation processes. The use of the large-scale rainfall simulator provides a unique opportunity to examine a range of rainfall intensities under constant conditions that are difficult under natural conditions due to the variable nature of rainfall intensities in the field. Stemflow generation and production was examined for three species- Cryptomeria japonica D. Don (Japanese cedar), Chamaecyparis obtusa (Siebold & Zucc.) Endl. (Japanese cypress), Zelkova serrata Thunb. (Japanese zelkova)- under both leafed and leafless conditions at several different rainfall intensities (15, 20, 30, 40, 50, and 100 mm h-1) using a large-scale rainfall simulator in National Research Institute for Earth Science and Disaster Resilience (Tsukuba, Japan). Stemflow production and rates and funneling ratios were examined in relation to both rainfall intensity and canopy structure. Preliminary results indicate a dynamic and complex response of the funneling ratios of individual trees to different rainfall intensities among the species examined. This is partly the result of different canopy structures, hydrophobicity of vegetative surfaces, and differential wet-up processes across species and rainfall intensities. This presentation delves into these differences and attempts to distill them into generalizable patterns, which can advance our theories of stemflow generation processes and ultimately permit better stewardship of forest resources. ________________ Funding note: This research was supported by JSPS Invitation Fellowship for Research in
Experimental validation of large-eddy simulation for swirling methane-air non-premixed combustion
Energy Technology Data Exchange (ETDEWEB)
Hu, L.Y.; Luo, Y.H.; Xu, C.S. [Shanghai Jiaotong Univ. (China). School of Mechanical Engineering; Zhou, L.X. [Tsinghua Univ., Beijing (China). Dept. of Engineering Mechanics
2013-07-01
Large-eddy simulation of swirling methane-air non-premixed combustion was carried out using a Smagorinsky-Lilly subgrid scale stress model and a presumed-PDF fast-chemistry combustion model. The LES statistical results are validated by PIV, temperature and species concentration measurements made by the present authors. The results indicate that in the present case the presumed-PDF fast-chemistry combustion model is a fairish one. The instantaneous vorticity and temperature maps show clearly the development and the interaction between coherent structures and combustion.
Large-eddy simulations of the non-reactive flow in the Sydney swirl burner
DEFF Research Database (Denmark)
Yang, Yang; Kær, Søren Knudsen
2012-01-01
results. In medium swirling case, there are two reverse-flow regions with a collar-like structure between them. The existence of strong unsteady structure, precessing vortex core, was proven. Coherent structures are detached from the instantaneous field. Q-criterion was used to visualize vorticity field...... with distinct clear structure of vortice tubes. Dominating spatial–temporal structures contained in different cross sections were extracted using proper orthogonal decomposition. In high swirling case, there is only one long reverse-flow region. In this paper, we proved the capability of a commercial CFD...... package in predicting complex flow field and presented the potential of large eddy simulation in understanding dynamics....
DEFF Research Database (Denmark)
Cavar, Dalibor; Meyer, Knud Erik
2011-01-01
A large eddy simulation (LES) study of turbulent non-equilibrium boundary layer flow over 2 D Bump, at comparatively low Reynolds number Reh = U∞h/ν = 1950, was conducted. A well-known LES issue of obtaining and sustaining turbulent flow inside the computational domain at such low Re, is addresse...... partially confirm a close interdependency between generation and evolution of internal layers and the abrupt changes in the skin friction, previously reported in the literature. © 2011 American Society of Mechanical Engineers....
Large Eddy Simulation of turbulent flows in compound channels with a finite element code
International Nuclear Information System (INIS)
Xavier, C.M.; Petry, A.P.; Moeller, S.V.
2011-01-01
This paper presents the numerical investigation of the developing flow in a compound channel formed by a rectangular main channel and a gap in one of the sidewalls. A three dimensional Large Eddy Simulation computational code with the classic Smagorinsky model is introduced, where the transient flow is modeled through the conservation equations of mass and momentum of a quasi-incompressible, isothermal continuous medium. Finite Element Method, Taylor-Galerkin scheme and linear hexahedrical elements are applied. Numerical results of velocity profile show the development of a shear layer in agreement with experimental results obtained with Pitot tube and hot wires. (author)
CELES: CUDA-accelerated simulation of electromagnetic scattering by large ensembles of spheres
Egel, Amos; Pattelli, Lorenzo; Mazzamuto, Giacomo; Wiersma, Diederik S.; Lemmer, Uli
2017-09-01
CELES is a freely available MATLAB toolbox to simulate light scattering by many spherical particles. Aiming at high computational performance, CELES leverages block-diagonal preconditioning, a lookup-table approach to evaluate costly functions and massively parallel execution on NVIDIA graphics processing units using the CUDA computing platform. The combination of these techniques allows to efficiently address large electrodynamic problems (>104 scatterers) on inexpensive consumer hardware. In this paper, we validate near- and far-field distributions against the well-established multi-sphere T-matrix (MSTM) code and discuss the convergence behavior for ensembles of different sizes, including an exemplary system comprising 105 particles.
International Nuclear Information System (INIS)
Gray, S.K.; Noid, D.W.; Sumpter, B.G.
1994-01-01
We test the suitability of a variety of explicit symplectic integrators for molecular dynamics calculations on Hamiltonian systems. These integrators are extremely simple algorithms with low memory requirements, and appear to be well suited for large scale simulations. We first apply all the methods to a simple test case using the ideas of Berendsen and van Gunsteren. We then use the integrators to generate long time trajectories of a 1000 unit polyethylene chain. Calculations are also performed with two popular but nonsymplectic integrators. The most efficient integrators of the set investigated are deduced. We also discuss certain variations on the basic symplectic integration technique
Aggregated Representation of Distribution Networks for Large-Scale Transmission Network Simulations
DEFF Research Database (Denmark)
Göksu, Ömer; Altin, Müfit; Sørensen, Poul Ejnar
2014-01-01
As a common practice of large-scale transmission network analysis the distribution networks have been represented as aggregated loads. However, with increasing share of distributed generation, especially wind and solar power, in the distribution networks, it became necessary to include...... the distributed generation within those analysis. In this paper a practical methodology to obtain aggregated behaviour of the distributed generation is proposed. The methodology, which is based on the use of the IEC standard wind turbine models, is applied on a benchmark distribution network via simulations....
Parallelization of a beam dynamics code and first large scale radio frequency quadrupole simulations
Directory of Open Access Journals (Sweden)
J. Xu
2007-01-01
Full Text Available The design and operation support of hadron (proton and heavy-ion linear accelerators require substantial use of beam dynamics simulation tools. The beam dynamics code TRACK has been originally developed at Argonne National Laboratory (ANL to fulfill the special requirements of the rare isotope accelerator (RIA accelerator systems. From the beginning, the code has been developed to make it useful in the three stages of a linear accelerator project, namely, the design, commissioning, and operation of the machine. To realize this concept, the code has unique features such as end-to-end simulations from the ion source to the final beam destination and automatic procedures for tuning of a multiple charge state heavy-ion beam. The TRACK code has become a general beam dynamics code for hadron linacs and has found wide applications worldwide. Until recently, the code has remained serial except for a simple parallelization used for the simulation of multiple seeds to study the machine errors. To speed up computation, the TRACK Poisson solver has been parallelized. This paper discusses different parallel models for solving the Poisson equation with the primary goal to extend the scalability of the code onto 1024 and more processors of the new generation of supercomputers known as BlueGene (BG/L. Domain decomposition techniques have been adapted and incorporated into the parallel version of the TRACK code. To demonstrate the new capabilities of the parallelized TRACK code, the dynamics of a 45 mA proton beam represented by 10^{8} particles has been simulated through the 325 MHz radio frequency quadrupole and initial accelerator section of the proposed FNAL proton driver. The results show the benefits and advantages of large-scale parallel computing in beam dynamics simulations.
WRF nested large-eddy simulations of deep convection during SEAC4RS
Heath, Nicholas K.; Fuelberg, Henry E.; Tanelli, Simone; Turk, F. Joseph; Lawson, R. Paul; Woods, Sarah; Freeman, Sean
2017-04-01
Large-eddy simulations (LES) and observations are often combined to increase our understanding and improve the simulation of deep convection. This study evaluates a nested LES method that uses the Weather Research and Forecasting (WRF) model and, specifically, tests whether the nested LES approach is useful for studying deep convection during a real-world case. The method was applied on 2 September 2013, a day of continental convection that occurred during the Studies of Emissions and Atmospheric Composition, Clouds and Climate Coupling by Regional Surveys (SEAC4RS) campaign. Mesoscale WRF output (1.35 km grid length) was used to drive a nested LES with 450 m grid spacing, which then drove a 150 m domain. Results reveal that the 450 m nested LES reasonably simulates observed reflectivity distributions and aircraft-observed in-cloud vertical velocities during the study period. However, when examining convective updrafts, reducing the grid spacing to 150 m worsened results. We find that the simulated updrafts in the 150 m run become too diluted by entrainment, thereby generating updrafts that are weaker than observed. Lastly, the 450 m simulation is combined with observations to study the processes forcing strong midlevel cloud/updraft edge downdrafts that were observed on 2 September. Results suggest that these strong downdrafts are forced by evaporative cooling due to mixing and by perturbation pressure forces acting to restore mass continuity around neighboring updrafts. We conclude that the WRF nested LES approach, with further development and evaluation, could potentially provide an effective method for studying deep convection in real-world cases.
Effect of grid resolution on large eddy simulation of wall-bounded turbulence
Rezaeiravesh, S.; Liefvendahl, M.
2018-05-01
The effect of grid resolution on a large eddy simulation (LES) of a wall-bounded turbulent flow is investigated. A channel flow simulation campaign involving a systematic variation of the streamwise (Δx) and spanwise (Δz) grid resolution is used for this purpose. The main friction-velocity-based Reynolds number investigated is 300. Near the walls, the grid cell size is determined by the frictional scaling, Δx+ and Δz+, and strongly anisotropic cells, with first Δy+ ˜ 1, thus aiming for the wall-resolving LES. Results are compared to direct numerical simulations, and several quality measures are investigated, including the error in the predicted mean friction velocity and the error in cross-channel profiles of flow statistics. To reduce the total number of channel flow simulations, techniques from the framework of uncertainty quantification are employed. In particular, a generalized polynomial chaos expansion (gPCE) is used to create metamodels for the errors over the allowed parameter ranges. The differing behavior of the different quality measures is demonstrated and analyzed. It is shown that friction velocity and profiles of the velocity and Reynolds stress tensor are most sensitive to Δz+, while the error in the turbulent kinetic energy is mostly influenced by Δx+. Recommendations for grid resolution requirements are given, together with the quantification of the resulting predictive accuracy. The sensitivity of the results to the subgrid-scale (SGS) model and varying Reynolds number is also investigated. All simulations are carried out with second-order accurate finite-volume-based solver OpenFOAM. It is shown that the choice of numerical scheme for the convective term significantly influences the error portraits. It is emphasized that the proposed methodology, involving the gPCE, can be applied to other modeling approaches, i.e., other numerical methods and the choice of SGS model.
GPU-Accelerated Sparse Matrix Solvers for Large-Scale Simulations, Phase II
National Aeronautics and Space Administration — At the heart of scientific computing and numerical analysis are linear algebra solvers. In scientific computing, the focus is on the partial differential equations...
International Nuclear Information System (INIS)
Ferreira, Maria da Conceição da Rocha
2018-01-01
The thesis reveals some aspects that concern the fear presented by a great part of society to the use of atomic energy. There are many studies pointing to the memories that refer to the terror of radioactive contamination or the destruction caused by atomic weapons, or even a controversial environmental view of the energy efficiency against climate warming. The object herein is the communication of scientific and technological risk, revealing the importance of journalism on the information given to a non specialized population. The premises concerning accidents caused by nuclear or radiological causes are that they have something beyond any other accident of technological causes when they are object of communication by mass media. The image of the bomb destruction can be a constant of terror on the apprehension of the risk in those accidents. French Discourse analysis is the theory support approached to search about the construction of the fear evolving the atomic energy, by analyzing some of the articles of mass communication media. The time selection, as a first cut, were the decades of 1980/1990, which were celebrated by the events of Chernobyl, worldwide, and the Cesium-137, in Goiania, Brazil. The same accidents are given a second cut on the celebration of their anniversaries, in cycles of up to 30 years, in a way of upgrading the production conditions of the discourses around them and their effects on the learning of society. The analysis was articulated between texts and images as discursive materials that have their own significations on the final effect of senses, which is, according to the methodology adopted, strongly affected by ideology. (author)
Simulation research on the process of large scale ship plane segmentation intelligent workshop
Xu, Peng; Liao, Liangchuang; Zhou, Chao; Xue, Rui; Fu, Wei
2017-04-01
Large scale ship plane segmentation intelligent workshop is a new thing, and there is no research work in related fields at home and abroad. The mode of production should be transformed by the existing industry 2.0 or part of industry 3.0, also transformed from "human brain analysis and judgment + machine manufacturing" to "machine analysis and judgment + machine manufacturing". In this transforming process, there are a great deal of tasks need to be determined on the aspects of management and technology, such as workshop structure evolution, development of intelligent equipment and changes in business model. Along with them is the reformation of the whole workshop. Process simulation in this project would verify general layout and process flow of large scale ship plane section intelligent workshop, also would analyze intelligent workshop working efficiency, which is significant to the next step of the transformation of plane segmentation intelligent workshop.
High-fidelity large eddy simulation for supersonic jet noise prediction
Aikens, Kurt M.
The problem of intense sound radiation from supersonic jets is a concern for both civil and military applications. As a result, many experimental and computational efforts are focused at evaluating possible noise suppression techniques. Large-eddy simulation (LES) is utilized in many computational studies to simulate the turbulent jet flowfield. Integral methods such as the Ffowcs Williams-Hawkings (FWH) method are then used for propagation of the sound waves to the farfield. Improving the accuracy of this two-step methodology and evaluating beveled converging-diverging nozzles for noise suppression are the main tasks of this work. First, a series of numerical experiments are undertaken to ensure adequate numerical accuracy of the FWH methodology. This includes an analysis of different treatments for the downstream integration surface: with or without including an end-cap, averaging over multiple end-caps, and including an approximate surface integral correction term. Secondly, shock-capturing methods based on characteristic filtering and adaptive spatial filtering are used to extend a highly-parallelizable multiblock subsonic LES code to enable simulations of supersonic jets. The code is based on high-order numerical methods for accurate prediction of the acoustic sources and propagation of the sound waves. Furthermore, this new code is more efficient than the legacy version, allows cylindrical multiblock topologies, and is capable of simulating nozzles with resolved turbulent boundary layers when coupled with an approximate turbulent inflow boundary condition. Even though such wall-resolved simulations are more physically accurate, their expense is often prohibitive. To make simulations more economical, a wall model is developed and implemented. The wall modeling methodology is validated for turbulent quasi-incompressible and compressible zero pressure gradient flat plate boundary layers, and for subsonic and supersonic jets. The supersonic code additions and the
Large-Eddy Simulation (LES of Spray Transients: Start and End of Injection Phenomena
Directory of Open Access Journals (Sweden)
Battistoni Michele
2016-01-01
Full Text Available This work reports investigations on Diesel spray transients, accounting for internal nozzle flow and needle motion, and demonstrates how seamless calculations of internal flow and external jet can be accomplished in a Large-Eddy Simulation (LES framework using an Eulerian mixture model. Sub-grid stresses are modeled with the Dynamic Structure (DS model, a non-viscosity based one-equation LES model. Two problems are studied with high level of spatial and temporal resolution. The first one concerns an End-Of-Injection (EOI case where gas ingestion, cavitation, and dribble formation are resolved. The second case is a Start-Of-Injection (SOI simulation that aims at analyzing the effect of residual gas trapped inside the injector sac on spray penetration and rate of fuel injection. Simulation results are compared against experiments carried out at Argonne National Laboratory (ANL using synchrotron X-ray. A mesh sensitivity analysis is conducted to assess the quality of the LES approach by evaluating the resolved turbulent kinetic energy budget and comparing the outcomes with a length-scale resolution index. LES of both EOI and SOI processes have been carried out on a single hole Diesel injector, providing insights in to the physics of the processes, with internal and external flow details, and linking the phenomena at the end of an injection event to those at the start of a new injection. Concerning the EOI, the model predicts ligament formation and gas ingestion, as observed experimentally, and the amount of residual gas in the nozzle sac matches with the available data. The fast dynamics of the process is described in detail. The simulation provides unique insights into the physics at the EOI. Similarly, the SOI simulation shows how gas is ejected first, and liquid fuel starts being injected with a delay. The simulation starts from a very low needle lift and is able to predict the actual Rate-Of-Injection (ROI and jet penetration, based only on the
International Nuclear Information System (INIS)
1973-01-01
One of the main aims of the IAEA is to foster the exchange of scientific and technical information and one of the main ways of doing this is to convene international scientific meetings. They range from large international conferences bringing together several hundred scientists, smaller symposia attended by an average of 150 to 250 participants and seminars designed to instruct rather than inform, to smaller panels and study groups of 10 to 30 experts brought together to advise on a particular programme or to develop a set of regulations. The topics of these meetings cover every part of the Agency's activities and form a backbone of many of its programmes. (author)
Forecasting wildland fire behavior using high-resolution large-eddy simulations
Munoz-Esparza, D.; Kosovic, B.; Jimenez, P. A.; Anderson, A.; DeCastro, A.; Brown, B.
2017-12-01
Wildland fires are responsible for large socio-economic impacts. Fires affect the environment, damage structures, threaten lives, cause health issues, and involve large suppression costs. These impacts can be mitigated via accurate fire spread forecast to inform the incident management team. To this end, the state of Colorado is funding the development of the Colorado Fire Prediction System (CO-FPS). The system is based on the Weather Research and Forecasting (WRF) model enhanced with a fire behavior module (WRF-Fire). Realistic representation of wildland fire behavior requires explicit representation of small scale weather phenomena to properly account for coupled atmosphere-wildfire interactions. Moreover, transport and dispersion of biomass burning emissions from wildfires is controlled by turbulent processes in the atmospheric boundary layer, which are difficult to parameterize and typically lead to large errors when simplified source estimation and injection height methods are used. Therefore, we utilize turbulence-resolving large-eddy simulations at a resolution of 111 m to forecast fire spread and smoke distribution using a coupled atmosphere-wildfire model. This presentation will describe our improvements to the level-set based fire-spread algorithm in WRF-Fire and an evaluation of the operational system using 12 wildfire events that occurred in Colorado in 2016, as well as other historical fires. In addition, the benefits of explicit representation of turbulence for smoke transport and dispersion will be demonstrated.
Piloted simulator study of allowable time delays in large-airplane response
Grantham, William D.; Bert T.?aetingas, Stephen A.dings with ran; Bert T.?aetingas, Stephen A.dings with ran
1987-01-01
A piloted simulation was performed to determine the permissible time delay and phase shift in the flight control system of a specific large transport-type airplane. The study was conducted with a six degree of freedom ground-based simulator and a math model similar to an advanced wide-body jet transport. Time delays in discrete and lagged form were incorporated into the longitudinal, lateral, and directional control systems of the airplane. Three experienced pilots flew simulated approaches and landings with random localizer and glide slope offsets during instrument tracking as their principal evaluation task. Results of the present study suggest a level 1 (satisfactory) handling qualities limit for the effective time delay of 0.15 sec in both the pitch and roll axes, as opposed to a 0.10-sec limit of the present specification (MIL-F-8785C) for both axes. Also, the present results suggest a level 2 (acceptable but unsatisfactory) handling qualities limit for an effective time delay of 0.82 sec and 0.57 sec for the pitch and roll axes, respectively, as opposed to 0.20 sec of the present specifications for both axes. In the area of phase shift between cockpit input and control surface deflection,the results of this study, flown in turbulent air, suggest less severe phase shift limitations for the approach and landing task-approximately 50 deg. in pitch and 40 deg. in roll - as opposed to 15 deg. of the present specifications for both axes.
Time-domain hybrid method for simulating large amplitude motions of ships advancing in waves
Directory of Open Access Journals (Sweden)
Shukui Liu
2011-03-01
Full Text Available Typical results obtained by a newly developed, nonlinear time domain hybrid method for simulating large amplitude motions of ships advancing with constant forward speed in waves are presented. The method is hybrid in the way of combining a time-domain transient Green function method and a Rankine source method. The present approach employs a simple double integration algorithm with respect to time to simulate the free-surface boundary condition. During the simulation, the diffraction and radiation forces are computed by pressure integration over the mean wetted surface, whereas the incident wave and hydrostatic restoring forces/moments are calculated on the instantaneously wetted surface of the hull. Typical numerical results of application of the method to the seakeeping performance of a standard containership, namely the ITTC S175, are herein presented. Comparisons have been made between the results from the present method, the frequency domain 3D panel method (NEWDRIFT of NTUA-SDL and available experimental data and good agreement has been observed for all studied cases between the results of the present method and comparable other data.
Large Eddy Simulation of a cooling impinging jet to a turbulent crossflow
Georgiou, Michail; Papalexandris, Miltiadis
2015-11-01
In this talk we report on Large Eddy Simulations of a cooling impinging jet to a turbulent channel flow. The impinging jet enters the turbulent stream in an oblique direction. This type of flow is relevant to the so-called ``Pressurized Thermal Shock'' phenomenon that can occur in pressurized water reactors. First we elaborate on issues related to the set-up of the simulations of the flow of interest such as, imposition of turbulent inflows, choice of subgrid-scale model and others. Also, the issue of the commutator error due to the anisotropy of the spatial cut-off filter induced by non-uniform grids is being discussed. In the second part of the talk we present results of our simulations. In particular, we focus on the high-shear and recirculation zones that are developed and on the characteristics of the temperature field. The budget for the mean kinetic energy of the resolved-scale turbulent velocity fluctuations is also discussed and analyzed. Financial support has been provided by Bel V, a subsidiary of the Federal Agency for Nuclear Control of Belgium.
On the properties of energy stable flux reconstruction schemes for implicit large eddy simulation
Vermeire, B. C.; Vincent, P. E.
2016-12-01
We begin by investigating the stability, order of accuracy, and dispersion and dissipation characteristics of the extended range of energy stable flux reconstruction (E-ESFR) schemes in the context of implicit large eddy simulation (ILES). We proceed to demonstrate that subsets of the E-ESFR schemes are more stable than collocation nodal discontinuous Galerkin methods recovered with the flux reconstruction approach (FRDG) for marginally-resolved ILES simulations of the Taylor-Green vortex. These schemes are shown to have reduced dissipation and dispersion errors relative to FRDG schemes of the same polynomial degree and, simultaneously, have increased Courant-Friedrichs-Lewy (CFL) limits. Finally, we simulate turbulent flow over an SD7003 aerofoil using two of the most stable E-ESFR schemes identified by the aforementioned Taylor-Green vortex experiments. Results demonstrate that subsets of E-ESFR schemes appear more stable than the commonly used FRDG method, have increased CFL limits, and are suitable for ILES of complex turbulent flows on unstructured grids.
Investigation of Turbulent Tip Leakage Vortex in an Axial Water Jet Pump with Large Eddy Simulation
Hah, Chunill; Katz, Joseph
2012-01-01
Detailed steady and unsteady numerical studies were performed to investigate tip clearance flow in an axial water jet pump. The primary objective is to understand physics of unsteady tip clearance flow, unsteady tip leakage vortex, and cavitation inception in an axial water jet pump. Steady pressure field and resulting steady tip leakage vortex from a steady flow analysis do not seem to explain measured cavitation inception correctly. The measured flow field near the tip is unsteady and measured cavitation inception is highly transient. Flow visualization with cavitation bubbles shows that the leakage vortex is oscillating significantly and many intermittent vortex ropes are present between the suction side of the blade and the tip leakage core vortex. Although the flow field is highly transient, the overall flow structure is stable and a characteristic frequency seems to exist. To capture relevant flow physics as much as possible, a Reynolds-averaged Navier-Stokes (RANS) calculation and a Large Eddy Simulation (LES) were applied for the current investigation. The present study reveals that several vortices from the tip leakage vortex system cross the tip gap of the adjacent blade periodically. Sudden changes in local pressure field inside tip gap due to these vortices create vortex ropes. The instantaneous pressure filed inside the tip gap is drastically different from that of the steady flow simulation. Unsteady flow simulation which can calculate unsteady vortex motion is necessary to calculate cavitation inception accurately even at design flow condition in such a water jet pump.
Hot air impingement on a flat plate using Large Eddy Simulation (LES) technique
Plengsa-ard, C.; Kaewbumrung, M.
2018-01-01
Impinging hot gas jets to a flat plate generate very high heat transfer coefficients in the impingement zone. The magnitude of heat transfer prediction near the stagnation point is important and accurate heat flux distribution are needed. This research studies on heat transfer and flow field resulting from a single hot air impinging wall. The simulation is carried out using computational fluid dynamics (CFD) commercial code FLUENT. Large Eddy Simulation (LES) approach with a subgrid-scale Smagorinsky-Lilly model is present. The classical Werner-Wengle wall model is used to compute the predicted results of velocity and temperature near walls. The Smagorinsky constant in the turbulence model is set to 0.1 and is kept constant throughout the investigation. The hot gas jet impingement on the flat plate with a constant surface temperature is chosen to validate the predicted heat flux results with experimental data. The jet Reynolds number is equal to 20,000 and a fixed jet-to-plate spacing of H/D = 2.0. Nusselt number on the impingement surface is calculated. As predicted by the wall model, the instantaneous computed Nusselt number agree fairly well with experimental data. The largest values of calculated Nusselt number are near the stagnation point and decrease monotonically in the wall jet region. Also, the contour plots of instantaneous values of wall heat flux on a flat plate are captured by LES simulation.
Large Eddy Simulation of Entropy Generation in a Turbulent Mixing Layer
Sheikhi, Reza H.; Safari, Mehdi; Hadi, Fatemeh
2013-11-01
Entropy transport equation is considered in large eddy simulation (LES) of turbulent flows. The irreversible entropy generation in this equation provides a more general description of subgrid scale (SGS) dissipation due to heat conduction, mass diffusion and viscosity effects. A new methodology is developed, termed the entropy filtered density function (En-FDF), to account for all individual entropy generation effects in turbulent flows. The En-FDF represents the joint probability density function of entropy, frequency, velocity and scalar fields within the SGS. An exact transport equation is developed for the En-FDF, which is modeled by a system of stochastic differential equations, incorporating the second law of thermodynamics. The modeled En-FDF transport equation is solved by a Lagrangian Monte Carlo method. The methodology is employed to simulate a turbulent mixing layer involving transport of passive scalars and entropy. Various modes of entropy generation are obtained from the En-FDF and analyzed. Predictions are assessed against data generated by direct numerical simulation (DNS). The En-FDF predictions are in good agreements with the DNS data.
Direct and large eddy simulations of a bottom Ekman layer under an external stratification
Energy Technology Data Exchange (ETDEWEB)
Taylor, John R. [Department of Mechanical and Aerospace Engineering, University of California, San Diego La Jolla, CA 92093 (United States); Sarkar, Sutanu [Department of Mechanical and Aerospace Engineering, University of California, San Diego La Jolla, CA 92093 (United States)], E-mail: sarkar@ucsd.edu
2008-06-15
A steady Ekman layer with a thermally stratified outer flow and an adiabatic boundary condition at the lower wall is studied using direct numerical simulation (DNS) and large eddy simulation (LES). An initially linear temperature profile is mixed by turbulence near the wall, and a stable thermocline forms above the mixed layer. The thickness of the mixed layer is reduced by the outer layer stratification. Observations from the DNS are used to evaluate the performance of the LES model and to examine the resolution requirements. A resolved LES and a near-wall model LES (NWM-LES) both compare reasonably well with the DNS when the thermal field is treated as a passive scalar. When buoyancy effects are included, the LES mean velocity and temperature profiles also agree well with the DNS. However, the NWM-LES does not sufficiently account for the overturning scales responsible for entrainment at the top of the mixed layer. As a result, the turbulent heat flux and the rate of change of the mixed layer temperature are significantly underestimated in the NWM-LES. In order to accurately simulate the boundary layer growth, the motions responsible for entrainment must either be resolved or more accurately represented in improved subgrid-scale models.
Direct and large eddy simulations of a bottom Ekman layer under an external stratification
International Nuclear Information System (INIS)
Taylor, John R.; Sarkar, Sutanu
2008-01-01
A steady Ekman layer with a thermally stratified outer flow and an adiabatic boundary condition at the lower wall is studied using direct numerical simulation (DNS) and large eddy simulation (LES). An initially linear temperature profile is mixed by turbulence near the wall, and a stable thermocline forms above the mixed layer. The thickness of the mixed layer is reduced by the outer layer stratification. Observations from the DNS are used to evaluate the performance of the LES model and to examine the resolution requirements. A resolved LES and a near-wall model LES (NWM-LES) both compare reasonably well with the DNS when the thermal field is treated as a passive scalar. When buoyancy effects are included, the LES mean velocity and temperature profiles also agree well with the DNS. However, the NWM-LES does not sufficiently account for the overturning scales responsible for entrainment at the top of the mixed layer. As a result, the turbulent heat flux and the rate of change of the mixed layer temperature are significantly underestimated in the NWM-LES. In order to accurately simulate the boundary layer growth, the motions responsible for entrainment must either be resolved or more accurately represented in improved subgrid-scale models
Stochastic four-way coupling of gas-solid flows for Large Eddy Simulations
Curran, Thomas; Denner, Fabian; van Wachem, Berend
2017-11-01
The interaction of solid particles with turbulence has for long been a topic of interest for predicting the behavior of industrially relevant flows. For the turbulent fluid phase, Large Eddy Simulation (LES) methods are widely used for their low computational cost, leaving only the sub-grid scales (SGS) of turbulence to be modelled. Although LES has seen great success in predicting the behavior of turbulent single-phase flows, the development of LES for turbulent gas-solid flows is still in its infancy. This contribution aims at constructing a model to describe the four-way coupling of particles in an LES framework, by considering the role particles play in the transport of turbulent kinetic energy across the scales. Firstly, a stochastic model reconstructing the sub-grid velocities for the particle tracking is presented. Secondly, to solve particle-particle interaction, most models involve a deterministic treatment of the collisions. We finally introduce a stochastic model for estimating the collision probability. All results are validated against fully resolved DNS-DPS simulations. The final goal of this contribution is to propose a global stochastic method adapted to two-phase LES simulation where the number of particles considered can be significantly increased. Financial support from PetroBras is gratefully acknowledged.
Energy Technology Data Exchange (ETDEWEB)
Jang, Yong Jun; Kim, Jin Ho; Park, Sung Huk; Koo, Dong Hoe [Korea Railroad Research Institute, Uiwang (Korea, Republic of)
2015-11-15
Large eddy simulation (LES) method is applied to systematically investigate the cooling fluid flow and the temperature distribution under the operating of air conditioning in the deeply underground subway station. The Shin-Gum-Ho subway station in Seoul which is the 8{sup th} floor and 43.6 m deep is selected for this analysis. The entire station is covered for simulation. The ventilation mode for air conditioning is kept as ordinary state. Different operating conditions for Platform screen door (PSD) are applied. First one is PSD is completely close and second one is PSD is regularly open and close which imitate the actual circumstances in the platform. The ventilation diffusers are modeled as 95 square shapes in the lobby and 222 squares in the platform. The temperature variations and flow behaviors are numerically simulated after operating of air conditioning for the whole station and the calculated results are compared with experimental data. LES method solves the momentum and thermal equations. Werner-Wengle wall law is applied to viscous sub layers for near wall resolution. The total grid numbers are 7.5 million and the whole domain is divided to 22 blocks. Multi blocks are computed in parallel using MPI. The results show the temperature difference in the platform between PSD-close and PSD-regularly open and close cases is 3-4 .deg. C.
International Nuclear Information System (INIS)
Jang, Yong Jun; Kim, Jin Ho; Park, Sung Huk; Koo, Dong Hoe
2015-01-01
Large eddy simulation (LES) method is applied to systematically investigate the cooling fluid flow and the temperature distribution under the operating of air conditioning in the deeply underground subway station. The Shin-Gum-Ho subway station in Seoul which is the 8"t"h floor and 43.6 m deep is selected for this analysis. The entire station is covered for simulation. The ventilation mode for air conditioning is kept as ordinary state. Different operating conditions for Platform screen door (PSD) are applied. First one is PSD is completely close and second one is PSD is regularly open and close which imitate the actual circumstances in the platform. The ventilation diffusers are modeled as 95 square shapes in the lobby and 222 squares in the platform. The temperature variations and flow behaviors are numerically simulated after operating of air conditioning for the whole station and the calculated results are compared with experimental data. LES method solves the momentum and thermal equations. Werner-Wengle wall law is applied to viscous sub layers for near wall resolution. The total grid numbers are 7.5 million and the whole domain is divided to 22 blocks. Multi blocks are computed in parallel using MPI. The results show the temperature difference in the platform between PSD-close and PSD-regularly open and close cases is 3-4 .deg. C
Large eddy simulation of new subgrid scale model for three-dimensional bundle flows
International Nuclear Information System (INIS)
Barsamian, H.R.; Hassan, Y.A.
2004-01-01
Having led to increased inefficiencies and power plant shutdowns fluid flow induced vibrations within heat exchangers are of great concern due to tube fretting-wear or fatigue failures. Historically, scaling law and measurement accuracy problems were encountered for experimental analysis at considerable effort and expense. However, supercomputers and accurate numerical methods have provided reliable results and substantial decrease in cost. In this investigation Large Eddy Simulation has been successfully used to simulate turbulent flow by the numeric solution of the incompressible, isothermal, single phase Navier-Stokes equations. The eddy viscosity model and a new subgrid scale model have been utilized to model the smaller eddies in the flow domain. A triangular array flow field was considered and numerical simulations were performed in two- and three-dimensional fields, and were compared to experimental findings. Results show good agreement of the numerical findings to that of the experimental, and solutions obtained with the new subgrid scale model represent better energy dissipation for the smaller eddies. (author)
Simulation-optimization of large agro-hydrosystems using a decomposition approach
Schuetze, Niels; Grundmann, Jens
2014-05-01
In this contribution a stochastic simulation-optimization framework for decision support for optimal planning and operation of water supply of large agro-hydrosystems is presented. It is based on a decomposition solution strategy which allows for (i) the usage of numerical process models together with efficient Monte Carlo simulations for a reliable estimation of higher quantiles of the minimum agricultural water demand for full and deficit irrigation strategies at small scale (farm level), and (ii) the utilization of the optimization results at small scale for solving water resources management problems at regional scale. As a secondary result of several simulation-optimization runs at the smaller scale stochastic crop-water production functions (SCWPF) for different crops are derived which can be used as a basic tool for assessing the impact of climate variability on risk for potential yield. In addition, microeconomic impacts of climate change and the vulnerability of the agro-ecological systems are evaluated. The developed methodology is demonstrated through its application on a real-world case study for the South Al-Batinah region in the Sultanate of Oman where a coastal aquifer is affected by saltwater intrusion due to excessive groundwater withdrawal for irrigated agriculture.
Implementation of a roughness element to trip transition in large-eddy simulation
Boudet, J.; Monier, J.-F.; Gao, F.
2015-02-01
In aerodynamics, the laminar or turbulent regime of a boundary layer has a strong influence on friction or heat transfer. In practical applications, it is sometimes necessary to trip the transition to turbulent, and a common way is by use of a roughness element ( e.g. a step) on the wall. The present paper is concerned with the numerical implementation of such a trip in large-eddy simulations. The study is carried out on a flat-plate boundary layer configuration, with Reynolds number Rex=1.3×106. First, this work brings the opportunity to introduce a practical methodology to assess convergence in large-eddy simulations. Second, concerning the trip implementation, a volume source term is proposed and is shown to yield a smoother and faster transition than a grid step. Moreover, it is easier to implement and more adaptable. Finally, two subgrid-scale models are tested: the WALE model of Nicoud and Ducros ( Flow Turbul. Combust., vol. 62, 1999) and the shear-improved Smagorinsky model of Lévêque et al. ( J. Fluid Mech., vol. 570, 2007). Both models allow transition, but the former appears to yield a faster transition and a better prediction of friction in the turbulent regime.
Large-eddy simulations of the non-reactive flow in the Sydney swirl burner
International Nuclear Information System (INIS)
Yang Yang; Kær, Søren Knudsen
2012-01-01
Highlights: ► Rational mesh and grid system for LES are discussed. ► Validated results are provided and discrepancy of mean radial velocity component is discussed. ► Flow structures are identified using vorticity field. ► We performed POD on cross sections to assist in understanding of coherent structures. - Abstract: This paper presents a numerical investigation using large-eddy simulation. Two isothermal cases from the Sydney swirling flame database with different swirl numbers were tested. Rational grid system and mesh details were presented firstly. Validations showed overall good agreement in time averaged results. In medium swirling case, there are two reverse-flow regions with a collar-like structure between them. The existence of strong unsteady structure, precessing vortex core, was proven. Coherent structures are detached from the instantaneous field. Q-criterion was used to visualize vorticity field with distinct clear structure of vortice tubes. Dominating spatial–temporal structures contained in different cross sections were extracted using proper orthogonal decomposition. In high swirling case, there is only one long reverse-flow region. In this paper, we proved the capability of a commercial CFD package in predicting complex flow field and presented the potential of large eddy simulation in understanding dynamics.
A Simulation of Rainwater Harvesting Design and Demand-Side Controls for Large Hospitals
Directory of Open Access Journals (Sweden)
Lawrence V. Fulton
2018-05-01
Full Text Available Inpatient health buildings in the United States are the most intensive users of water among large commercial buildings. Large facilities (greater than 1 million square feet consume an average of 90 million gallons per building per year. The distribution and treatment of water imposes a significant electrical power demand, which may be the single largest energy requirement for various states. Supply and demand-side solutions are needed, particularly in arid and semi-arid regions where water is scarce. This study uses continuous simulations based on 71 years of historical data to estimate how rainwater harvesting systems and demand-side interventions (e.g., low-flow devices, xeriscaping would offset the demand for externally-provided water sources in a semi-arid region. Simulations from time series models are used to generate alternative rainfall models to account for potential non-stationarity and volatility. Results demonstrate that hospital external water consumption might be reduced by approximately 25% using conservative assumptions and depending on the design of experiment parameters associated with rainfall capture area, building size, holding tank specifications, and conservation efforts.
Simulated effects of host fish distribution on juvenile unionid mussel dispersal in a large river
Daraio, J.A.; Weber, L.J.; Zigler, S.J.; Newton, T.J.; Nestler, J.M.
2012-01-01
Larval mussels (Family Unionidae) are obligate parasites on fish, and after excystment from their host, as juveniles, they are transported with flow. We know relatively little about the mechanisms that affect dispersal and subsequent settlement of juvenile mussels in large rivers. We used a three-dimensional hydrodynamic model of a reach of the Upper Mississippi River with stochastic Lagrangian particle tracking to simulate juvenile dispersal. Sensitivity analyses were used to determine the importance of excystment location in two-dimensional space (lateral and longitudinal) and to assess the effects of vertical location (depth in the water column) on dispersal distances and juvenile settling distributions. In our simulations, greater than 50% of juveniles mussels settled on the river bottom within 500 m of their point of excystment, regardless of the vertical location of the fish in the water column. Dispersal distances were most variable in environments with higher velocity and high gradients in velocity, such as along channel margins, near the channel bed, or where effects of river bed morphology caused large changes in hydraulics. Dispersal distance was greater and variance was greater when juvenile excystment occurred in areas where vertical velocity (w) was positive (indicating an upward velocity) than when w was negative. Juvenile dispersal distance is likely to be more variable for mussels species whose hosts inhabit areas with steeper velocity gradients (e.g. channel margins) than a host that generally inhabits low-flow environments (e.g. impounded areas).
Energy Technology Data Exchange (ETDEWEB)
Chakravarty, A.; Emanuel, A.S.; Bernath, J.A. [Chevron Petroleum Technology Company, LaHabra, CA (United States)
1997-08-01
The application of streamtube models for reservoir simulation has an extensive history in the oil industry. Although these models are strictly applicable only to fields under voidage balance, they have proved to be useful in a large number of fields provided that there is no solution gas evolution and production. These models combine the benefit of very fast computational time with the practical ability to model a large reservoir over the course of its history. These models do not, however, directly incorporate the detailed geological information that recent experience has taught is important. This paper presents a technique for mapping the saturation information contained in a history matched streamtube model onto a detailed geostatistically derived finite difference grid. With this technique, the saturation information in a streamtube model, data that is actually statistical in nature, can be identified with actual physical locations in a field and a picture of the remaining oil saturation can be determined. Alternatively, the streamtube model can be used to simulate the early development history of a field and the saturation data then used to initialize detailed late time finite difference models. The proposed method is presented through an example application to the Ninian reservoir. This reservoir, located in the North Sea (UK), is a heterogeneous sandstone characterized by a line drive waterflood, with about 160 wells, and a 16 year history. The reservoir was satisfactorily history matched and mapped for remaining oil saturation. A comparison to 3-D seismic survey and recently drilled wells have provided preliminary verification.
An Examination of Parameters Affecting Large Eddy Simulations of Flow Past a Square Cylinder
Mankbadi, M. R.; Georgiadis, N. J.
2014-01-01
Separated flow over a bluff body is analyzed via large eddy simulations. The turbulent flow around a square cylinder features a variety of complex flow phenomena such as highly unsteady vortical structures, reverse flow in the near wall region, and wake turbulence. The formation of spanwise vortices is often times artificially suppressed in computations by either insufficient depth or a coarse spanwise resolution. As the resolution is refined and the domain extended, the artificial turbulent energy exchange between spanwise and streamwise turbulence is eliminated within the wake region. A parametric study is performed highlighting the effects of spanwise vortices where the spanwise computational domain's resolution and depth are varied. For Re=22,000, the mean and turbulent statistics computed from the numerical large eddy simulations (NLES) are in good agreement with experimental data. Von-Karman shedding is observed in the wake of the cylinder. Mesh independence is illustrated by comparing a mesh resolution of 2 million to 16 million. Sensitivities to time stepping were minimized and sampling frequency sensitivities were nonpresent. While increasing the spanwise depth and resolution can be costly, this practice was found to be necessary to eliminating the artificial turbulent energy exchange.
Large eddy simulations of isothermal confined swirling flow in an industrial gas-turbine
International Nuclear Information System (INIS)
Bulat, G.; Jones, W.P.; Navarro-Martinez, S.
2015-01-01
Highlights: • We conduct a large eddy simulation of an industrial gas turbine. • The results are compared with measurements obtained under isothermal conditions. • The method reproduces the observed precessing vortex and central vortex cores. • The profiles of mean and rms velocities are found to be captured to a good accuracy. - Abstract: The paper describes the results of a computational study of the strongly swirling isothermal flow in the combustion chamber of an industrial gas turbine. The flow field characteristics are computed using large eddy simulation in conjunction with a dynamic version of the Smagorinsky model for the sub-grid-scale stresses. Grid refinement studies demonstrate that the results are essentially grid independent. The LES results are compared with an extensive set of measurements and the agreement with these is overall good. The method is shown to be capable of reproducing the observed precessing vortex and central vortex cores and the profiles of mean and rms velocities are found to be captured to a good accuracy. The overall flow structure is shown to be virtually independent of Reynolds number
Large eddy simulation of Loss of Vacuum Accident in STARDUST facility
International Nuclear Information System (INIS)
Benedetti, Miriam; Gaudio, Pasquale; Lupelli, Ivan; Malizia, Andrea; Porfiri, Maria Teresa; Richetta, Maria
2013-01-01
Highlights: ► Fusion safety, plasma material interaction. ► Numerical and experimental data comparison to analyze the consequences of Loss of Vacuum Accident that can provoke dust mobilization inside the Vacuum Vessel of the Nuclear Fusion Reactor ITER-like. -- Abstract: The development of computational fluid dynamic (CFD) models of air ingress into the vacuum vessel (VV) represents an important issue concerning the safety analysis of nuclear fusion devices, in particular in the field of dust mobilization. The present work deals with the large eddy simulations (LES) of fluid dynamic fields during a vessel filling at near vacuum conditions to support the safety study of Loss of Vacuum Accidents (LOVA) events triggered by air income. The model's results are compared to the experimental data provided by STARDUST facility at different pressurization rates (100 Pa/s, 300 Pa/s and 500 Pa/s). Simulation's results compare favorably with experimental data, demonstrating the possibility of implementing LES in large vacuum systems as tokamaks
International Nuclear Information System (INIS)
Patil, Sunil; Tafti, Danesh
2012-01-01
Highlights: ► Large eddy simulation. ► Wall layer modeling. ► Synthetic inlet turbulence. ► Swirl flows. - Abstract: Large eddy simulations of complex high Reynolds number flows are carried out with the near wall region being modeled with a zonal two layer model. A novel formulation for solving the turbulent boundary layer equation for the effective tangential velocity in a generalized co-ordinate system is presented and applied in the near wall zonal treatment. This formulation reduces the computational time in the inner layer significantly compared to the conventional two layer formulations present in the literature and is most suitable for complex geometries involving body fitted structured and unstructured meshes. The cost effectiveness and accuracy of the proposed wall model, used with the synthetic eddy method (SEM) to generate inlet turbulence, is investigated in turbulent channel flow, flow over a backward facing step, and confined swirling flows at moderately high Reynolds numbers. Predictions are compared with available DNS, experimental LDV data, as well as wall resolved LES. In all cases, there is at least an order of magnitude reduction in computational cost with no significant loss in prediction accuracy.
New simulation capabilities of electron clouds in ion beams with large tune depression
International Nuclear Information System (INIS)
Vay, J.-L.; Furman, M.A.; Seidl, P.A.
2007-01-01
We have developed a new, comprehensive set of simulation tools aimed at modeling the interaction of intense ion beams and electron clouds (e-clouds). The set contains the 3-D accelerator PIC code WARP and the 2-D 'slice' e-cloud code POSINST [M. Furman, this workshop, paper TUAX05], as well as a merger of the two, augmented by new modules for impact ionization and neutral gas generation. The new capability runs on workstations or parallel supercomputers and contains advanced features such as mesh refinement, disparate adaptive time stepping, and a new 'drift-Lorentz' particle mover for tracking charged particles in magnetic fields using large time steps. It is being applied to the modeling of ion beams (1 MeV, 180 mA, K+) for heavy ion inertial fusion and warm dense matter studies, as they interact with electron clouds in the High-Current Experiment (HCX) [experimental results discussed by A. Molvik, this workshop, paper THAW02]. We describe the capabilities and present recent simulation results with detailed comparisons against the HCX experiment, as well as their application (in a different regime) to the modeling of e-clouds in the Large Hadron Collider (LHC). (author)
New simulation capabilities of electron clouds in ion beams with large tune depression
International Nuclear Information System (INIS)
Lawrence Livermore National Laboratory
2006-01-01
We have developed a new, comprehensive set of simulation tools aimed at modeling the interaction of intense ion beams and electron clouds (e-clouds). The set contains the 3-D accelerator PIC code WARP and the 2-D ''slice'' e-cloud code POSINST [M. Furman, this workshop, paper TUAX05], as well as a merger of the two, augmented by new modules for impact ionization and neutral gas generation. The new capability runs on workstations or parallel supercomputers and contains advanced features such as mesh refinement, disparate adaptive time stepping, and a new ''drift-Lorentz'' particle mover for tracking charged particles in magnetic fields using large time steps. It is being applied to the modeling of ion beams (1 MeV, 180 mA, K+) for heavy ion inertial fusion and warm dense matter studies, as they interact with electron clouds in the High-Current Experiment (HCX) [experimental results discussed by A. Molvik, this workshop, paper THAW02]. We describe the capabilities and present recent simulation results with detailed comparisons against the HCX experiment, as well as their application (in a different regime) to the modeling of e-clouds in the Large Hadron Collider (LHC)
New simulation capabilities of electron clouds in ion beams with large tune depression
International Nuclear Information System (INIS)
Vay, J.L.; Furman, M.A.; Seidl, P.A.; Cohen, R.H.; Friedman, A.; Grote, D.P.; Kireeff-Covo, M.; Molvik, A.W.; Stoltz, P.H.; Veitzer, S.; Verboncoeur, J.P.
2006-01-01
The authors have developed a new, comprehensive set of simulation tools aimed at modeling the interaction of intense ion beams and electron clouds (e-clouds). The set contains the 3-D accelerator PIC code WARP and the 2-D ''slice'' e-cloud code POSINST, as well as a merger of the two, augmented by new modules for impact ionization and neutral gas generation. The new capability runs on workstations or parallel supercomputers and contains advanced features such as mesh refinement, disparate adaptive time stepping, and a new ''drift-Lorentz'' particle mover for tracking charged particles in magnetic fields using large time steps. It is being applied to the modeling of ion beams (1 MeV, 180 mA, K+) for heavy ion inertial fusion and warm dense matter studies, as they interact with electron clouds in the High-Current Experiment (HCX). They describe the capabilities and present recent simulation results with detailed comparisons against the HCX experiment, as well as their application (in a different regime) to the modeling of e-clouds in the Large Hadron Collider (LHC)
Testing of Large-Scale ICV Glasses with Hanford LAW Simulant
Energy Technology Data Exchange (ETDEWEB)
Hrma, Pavel R.; Kim, Dong-Sang; Vienna, John D.; Matyas, Josef; Smith, Donald E.; Schweiger, Michael J.; Yeager, John D.
2005-03-01
Preliminary glass compositions for immobilizing Hanford low-activity waste (LAW) by the in-container vitrification (ICV) process were initially fabricated at crucible- and engineering-scale, including simulants and actual (radioactive) LAW. Glasses were characterized for vapor hydration test (VHT) and product consistency test (PCT) responses and crystallinity (both quenched and slow-cooled samples). Selected glasses were tested for toxicity characteristic leach procedure (TCLP) responses, viscosity, and electrical conductivity. This testing showed that glasses with LAW loading of 20 mass% can be made readily and meet all product constraints by a far margin. Glasses with over 22 mass% Na2O can be made to meet all other product quality and process constraints. Large-scale testing was performed at the AMEC, Geomelt Division facility in Richland. Three tests were conducted using simulated LAW with increasing loadings of 12, 17, and 20 mass% Na2O. Glass samples were taken from the test products in a manner to represent the full expected range of product performance. These samples were characterized for composition, density, crystalline and non-crystalline phase assemblage, and durability using the VHT, PCT, and TCLP tests. The results, presented in this report, show that the AMEC ICV product with meets all waste form requirements with a large margin. These results provide strong evidence that the Hanford LAW can be successfully vitrified by the ICV technology and can meet all the constraints related to product quality. The economic feasibility of the ICV technology can be further enhanced by subsequent optimization.
Test-particle simulations of SEP propagation in IMF with large-scale fluctuations
Kelly, J.; Dalla, S.; Laitinen, T.
2012-11-01
The results of full-orbit test-particle simulations of SEPs propagating through an IMF which exhibits large-scale fluctuations are presented. A variety of propagation conditions are simulated - scatter-free, and scattering with mean free path, λ, of 0.3 and 2.0 AU - and the cross-field transport of SEPs is investigated. When calculating cross-field displacements the Parker spiral geometry is accounted for and the role of magnetic field expansion is taken into account. It is found that transport across the magnetic field is enhanced in the λ =0.3 AU and λ =2 AU cases, compared to the scatter-free case, with the λ =2 AU case in particular containing outlying particles that had strayed a large distance across the IMF. Outliers are catergorized by means of Chauvenet's criterion and it is found that typically between 1 and 2% of the population falls within this category. The ratio of latitudinal to longitudinal diffusion coefficient perpendicular to the magnetic field is typically 0.2, suggesting that transport in latitude is less efficient.
International Nuclear Information System (INIS)
Figueroa, Aldo; Meunier, Patrice; Villermaux, Emmanuel; Cuevas, Sergio; Ramos, Eduardo
2014-01-01
We present a combination of experiment, theory, and modelling on laminar mixing at large Péclet number. The flow is produced by oscillating electromagnetic forces in a thin electrolytic fluid layer, leading to oscillating dipoles, quadrupoles, octopoles, and disordered flows. The numerical simulations are based on the Diffusive Strip Method (DSM) which was recently introduced (P. Meunier and E. Villermaux, “The diffusive strip method for scalar mixing in two-dimensions,” J. Fluid Mech. 662, 134–172 (2010)) to solve the advection-diffusion problem by combining Lagrangian techniques and theoretical modelling of the diffusion. Numerical simulations obtained with the DSM are in reasonable agreement with quantitative dye visualization experiments of the scalar fields. A theoretical model based on log-normal Probability Density Functions (PDFs) of stretching factors, characteristic of homogeneous turbulence in the Batchelor regime, allows to predict the PDFs of scalar in agreement with numerical and experimental results. This model also indicates that the PDFs of scalar are asymptotically close to log-normal at late stages, except for the large concentration levels which correspond to low stretching factors
Large-Eddy Simulation of the Aerodynamic and Aeroacoustic Performance of a Ventilation Fan
Directory of Open Access Journals (Sweden)
Stefano Bianchi
2013-01-01
Full Text Available There are controversial requirements involved in developing numerical methodologies in order to compute the flow in industrial fans. The full resolution of turbulence spectrum in such high-Reynolds number flow configurations entails unreasonably expensive computational costs. The authors applied the study to a large unidirectional axial flow fan unit for tunnel ventilation to operate in the forward direction under ambient conditions. This delivered cooling air to the tunnel under routine operation, or hot gases at 400∘C under emergency conditions in the event of a tunnel fire. The simulations were carried out using the open source code OpenFOAM, within which they implemented a very large eddy simulation (VLES based on one-equation SGS model to solve a transport equation for the modelled (subgrid turbulent kinetic energy. This subgrid turbulence model improvement is a remedial strategy in VLES of high-Reynolds number industrial flows which are able to tackle the turbulence spectrum’s well-known insufficient resolution. The VLES of the industrial fan permits detecting the unsteady topology of the rotor flow. This paper explores the evolution of secondary flow phenomena and speculates on its influence on the actual load capability when operating at peak-pressure condition. Predicted noise emissions, in terms of sound pressure level spectra, are also compared with experimental results and found to agree within the uncertainty of the measurements.
Shimojo, Fuyuki; Hattori, Shinnosuke; Kalia, Rajiv K.; Kunaseth, Manaschai; Mou, Weiwei; Nakano, Aiichiro; Nomura, Ken-ichi; Ohmura, Satoshi; Rajak, Pankaj; Shimamura, Kohei; Vashishta, Priya
2014-05-01
We introduce an extension of the divide-and-conquer (DC) algorithmic paradigm called divide-conquer-recombine (DCR) to perform large quantum molecular dynamics (QMD) simulations on massively parallel supercomputers, in which interatomic forces are computed quantum mechanically in the framework of density functional theory (DFT). In DCR, the DC phase constructs globally informed, overlapping local-domain solutions, which in the recombine phase are synthesized into a global solution encompassing large spatiotemporal scales. For the DC phase, we design a lean divide-and-conquer (LDC) DFT algorithm, which significantly reduces the prefactor of the O(N) computational cost for N electrons by applying a density-adaptive boundary condition at the peripheries of the DC domains. Our globally scalable and locally efficient solver is based on a hybrid real-reciprocal space approach that combines: (1) a highly scalable real-space multigrid to represent the global charge density; and (2) a numerically efficient plane-wave basis for local electronic wave functions and charge density within each domain. Hybrid space-band decomposition is used to implement the LDC-DFT algorithm on parallel computers. A benchmark test on an IBM Blue Gene/Q computer exhibits an isogranular parallel efficiency of 0.984 on 786 432 cores for a 50.3 × 106-atom SiC system. As a test of production runs, LDC-DFT-based QMD simulation involving 16 661 atoms is performed on the Blue Gene/Q to study on-demand production of hydrogen gas from water using LiAl alloy particles. As an example of the recombine phase, LDC-DFT electronic structures are used as a basis set to describe global photoexcitation dynamics with nonadiabatic QMD (NAQMD) and kinetic Monte Carlo (KMC) methods. The NAQMD simulations are based on the linear response time-dependent density functional theory to describe electronic excited states and a surface-hopping approach to describe transitions between the excited states. A series of techniques
International Nuclear Information System (INIS)
Shimojo, Fuyuki; Hattori, Shinnosuke; Kalia, Rajiv K.; Mou, Weiwei; Nakano, Aiichiro; Nomura, Ken-ichi; Rajak, Pankaj; Vashishta, Priya; Kunaseth, Manaschai; Ohmura, Satoshi; Shimamura, Kohei
2014-01-01
We introduce an extension of the divide-and-conquer (DC) algorithmic paradigm called divide-conquer-recombine (DCR) to perform large quantum molecular dynamics (QMD) simulations on massively parallel supercomputers, in which interatomic forces are computed quantum mechanically in the framework of density functional theory (DFT). In DCR, the DC phase constructs globally informed, overlapping local-domain solutions, which in the recombine phase are synthesized into a global solution encompassing large spatiotemporal scales. For the DC phase, we design a lean divide-and-conquer (LDC) DFT algorithm, which significantly reduces the prefactor of the O(N) computational cost for N electrons by applying a density-adaptive boundary condition at the peripheries of the DC domains. Our globally scalable and locally efficient solver is based on a hybrid real-reciprocal space approach that combines: (1) a highly scalable real-space multigrid to represent the global charge density; and (2) a numerically efficient plane-wave basis for local electronic wave functions and charge density within each domain. Hybrid space-band decomposition is used to implement the LDC-DFT algorithm on parallel computers. A benchmark test on an IBM Blue Gene/Q computer exhibits an isogranular parallel efficiency of 0.984 on 786 432 cores for a 50.3 × 10 6 -atom SiC system. As a test of production runs, LDC-DFT-based QMD simulation involving 16 661 atoms is performed on the Blue Gene/Q to study on-demand production of hydrogen gas from water using LiAl alloy particles. As an example of the recombine phase, LDC-DFT electronic structures are used as a basis set to describe global photoexcitation dynamics with nonadiabatic QMD (NAQMD) and kinetic Monte Carlo (KMC) methods. The NAQMD simulations are based on the linear response time-dependent density functional theory to describe electronic excited states and a surface-hopping approach to describe transitions between the excited states. A series of
Validation of CALMET/CALPUFF models simulations around a large power plant stack
Energy Technology Data Exchange (ETDEWEB)
Hernandez-Garces, A.; Souto, J. A.; Rodriguez, A.; Saavedra, S.; Casares, J. J.
2015-07-01
Calmest/CALPUFF modeling system is frequently used in the study of atmospheric processes and pollution, and several validation tests were performed until now; nevertheless, most of them were based on experiments with a large compilation of surface and aloft meteorological measurements, rarely available. At the same time, the use of a large operational smokestack as tracer/pollutant source is not usual. In this work, first CALMET meteorological diagnostic model is nested to WRF meteorological prognostic model simulations (3x3 km{sup 2} horizontal resolution) over a complex terrain and coastal domain at NW Spain, covering 100x100 km{sup 2}, with a coal-fired power plant emitting SO{sub 2}. Simulations were performed during three different periods when SO{sub 2} hourly glc peaks were observed. NCEP reanalysis were applied as initial and boundary conditions. Yong Sei University-Pleim-Chang (YSU) PBL scheme was selected in the WRF model to provide the best input to three different CALMET horizontal resolutions, 1x1 km{sup 2}, 0.5x0.5 km{sup 2}, and 0.2x0.2 km{sup 2}. The best results, very similar between them, were achieved using the last two resolutions; therefore, the 0.5x0.5 km{sup 2} resolution was selected to test different CALMET meteorological inputs, using several combinations of WRF outputs and/or surface and upper-air measurements available in the simulation domain. With respect to meteorological aloft models output, CALMET PBL depth estimations are very similar to PBL depth estimations using upper-air measurements (rawinsondes), and significantly better than WRF PBL depth results. Regarding surface models surface output, the available meteorological sites were divided in two groups, one to provide meteorological input to CALMET (when applied), and another to models validation. Comparing WRF and CALMET outputs against surface measurements (from sites for models validation) the lowest RMSE was achieved using as CALMET input dataset WRF output combined with
Validation of CALMET/CALPUFF models simulations around a large power plant stack
Energy Technology Data Exchange (ETDEWEB)
Hernandez-Garces, A.; Souto Rodriguez, J.A.; Saavedra, S.; Casares, J.J.
2015-07-01
CALMET/CALPUFF modeling system is frequently used in the study of atmospheric processes and pollution, and several validation tests were performed until now; nevertheless, most of them were based on experiments with a large compilation of surface and aloft meteorological measurements, rarely available. At the same time, the use of a large operational smokestack as tracer/pollutant source is not usual. In this work, first CALMET meteorological diagnostic model is nested to WRF meteorological prognostic model simulations (3x3 km2 horizontal resolution) over a complex terrain and coastal domain at NW Spain, covering 100x100 km2 , with a coal-fired power plant emitting SO2. Simulations were performed during three different periods when SO2 hourly glc peaks were observed. NCEP reanalysis were applied as initial and boundary conditions. Yong Sei University-Pleim-Chang (YSU) PBL scheme was selected in the WRF model to provide the best input to three different CALMET horizontal resolutions, 1x1 km2 , 0.5x0.5 km2 , and 0.2x0.2 km2. The best results, very similar between them, were achieved using the last two resolutions; therefore, the 0.5x0.5 km2 resolution was selected to test different CALMET meteorological inputs, using several combinations of WRF outputs and/or surface and upper-air measurements available in the simulation domain. With respect to meteorological aloft models output, CALMET PBL depth estimations are very similar to PBL depth estimations using upper-air measurements (rawinsondes), and significantly better than WRF PBL depth results. Regarding surface models surface output, the available meteorological sites were divided in two groups, one to provide meteorological input to CALMET (when applied), and another to models validation. Comparing WRF and CALMET outputs against surface measurements (from sites for models validation) the lowest RMSE was achieved using as CALMET input dataset WRF output combined with surface measurements (from sites for CALMET model
Validation of CALMET/CALPUFF models simulations around a large power plant stack
International Nuclear Information System (INIS)
Hernandez-Garces, A.; Souto, J. A.; Rodriguez, A.; Saavedra, S.; Casares, J. J.
2015-01-01
Calmest/CALPUFF modeling system is frequently used in the study of atmospheric processes and pollution, and several validation tests were performed until now; nevertheless, most of them were based on experiments with a large compilation of surface and aloft meteorological measurements, rarely available. At the same time, the use of a large operational smokestack as tracer/pollutant source is not usual. In this work, first CALMET meteorological diagnostic model is nested to WRF meteorological prognostic model simulations (3x3 km 2 horizontal resolution) over a complex terrain and coastal domain at NW Spain, covering 100x100 km 2 , with a coal-fired power plant emitting SO 2 . Simulations were performed during three different periods when SO 2 hourly glc peaks were observed. NCEP reanalysis were applied as initial and boundary conditions. Yong Sei University-Pleim-Chang (YSU) PBL scheme was selected in the WRF model to provide the best input to three different CALMET horizontal resolutions, 1x1 km 2 , 0.5x0.5 km 2 , and 0.2x0.2 km 2 . The best results, very similar between them, were achieved using the last two resolutions; therefore, the 0.5x0.5 km 2 resolution was selected to test different CALMET meteorological inputs, using several combinations of WRF outputs and/or surface and upper-air measurements available in the simulation domain. With respect to meteorological aloft models output, CALMET PBL depth estimations are very similar to PBL depth estimations using upper-air measurements (rawinsondes), and significantly better than WRF PBL depth results. Regarding surface models surface output, the available meteorological sites were divided in two groups, one to provide meteorological input to CALMET (when applied), and another to models validation. Comparing WRF and CALMET outputs against surface measurements (from sites for models validation) the lowest RMSE was achieved using as CALMET input dataset WRF output combined with surface measurements (from sites for
Investigation of Numerical Dissipation in Classical and Implicit Large Eddy Simulations
Directory of Open Access Journals (Sweden)
Moutassem El Rafei
2017-12-01
Full Text Available The quantitative measure of dissipative properties of different numerical schemes is crucial to computational methods in the field of aerospace applications. Therefore, the objective of the present study is to examine the resolving power of Monotonic Upwind Scheme for Conservation Laws (MUSCL scheme with three different slope limiters: one second-order and two third-order used within the framework of Implicit Large Eddy Simulations (ILES. The performance of the dynamic Smagorinsky subgrid-scale model used in the classical Large Eddy Simulation (LES approach is examined. The assessment of these schemes is of significant importance to understand the numerical dissipation that could affect the accuracy of the numerical solution. A modified equation analysis has been employed to the convective term of the fully-compressible Navier–Stokes equations to formulate an analytical expression of truncation error for the second-order upwind scheme. The contribution of second-order partial derivatives in the expression of truncation error showed that the effect of this numerical error could not be neglected compared to the total kinetic energy dissipation rate. Transitions from laminar to turbulent flow are visualized considering the inviscid Taylor–Green Vortex (TGV test-case. The evolution in time of volumetrically-averaged kinetic energy and kinetic energy dissipation rate have been monitored for all numerical schemes and all grid levels. The dissipation mechanism has been compared to Direct Numerical Simulation (DNS data found in the literature at different Reynolds numbers. We found that the resolving power and the symmetry breaking property are enhanced with finer grid resolutions. The production of vorticity has been observed in terms of enstrophy and effective viscosity. The instantaneous kinetic energy spectrum has been computed using a three-dimensional Fast Fourier Transform (FFT. All combinations of numerical methods produce a k − 4 spectrum
Directory of Open Access Journals (Sweden)
G. Tang
2012-08-01
Full Text Available Satellite-based data, such as vegetation type and fractional vegetation cover, are widely used in hydrologic models to prescribe the vegetation state in a study region. Dynamic global vegetation models (DGVM simulate land surface hydrology. Incorporation of satellite-based data into a DGVM may enhance a model's ability to simulate land surface hydrology by reducing the task of model parameterization and providing distributed information on land characteristics. The objectives of this study are to (i modify a DGVM for simulating land surface water balances; (ii evaluate the modified model in simulating actual evapotranspiration (ET, soil moisture, and surface runoff at regional or watershed scales; and (iii gain insight into the ability of both the original and modified model to simulate large spatial scale land surface hydrology. To achieve these objectives, we introduce the "LPJ-hydrology" (LH model which incorporates satellite-based data into the Lund-Potsdam-Jena (LPJ DGVM. To evaluate the model we ran LH using historical (1981–2006 climate data and satellite-based land covers at 2.5 arc-min grid cells for the conterminous US and for the entire world using coarser climate and land cover data. We evaluated the simulated ET, soil moisture, and surface runoff using a set of observed or simulated data at different spatial scales. Our results demonstrate that spatial patterns of LH-simulated annual ET and surface runoff are in accordance with previously published data for the US; LH-modeled monthly stream flow for 12 major rivers in the US was consistent with observed values respectively during the years 1981–2006 (R^{2} > 0.46, p < 0.01; Nash-Sutcliffe Coefficient > 0.52. The modeled mean annual discharges for 10 major rivers worldwide also agreed well (differences < 15% with observed values for these rivers. Compared to a degree-day method for snowmelt computation, the addition of the solar radiation effect on snowmelt