Frutos-Alfaro, Francisco
2015-01-01
A program to generate codes in Fortran and C of the full Magnetohydrodynamic equations is shown. The program used the free computer algebra system software REDUCE. This software has a package called EXCALC, which is an exterior calculus program. The advantage of this program is that it can be modified to include another complex metric or spacetime. The output of this program is modified by means of a LINUX script which creates a new REDUCE program to manipulate the MHD equations to obtain a code that can be used as a seed for a MHD code for numerical applications. As an example, we present part of output of our programs for Cartesian coordinates and how to do the discretization.
Frutos-Alfaro, Francisco; Carboni-Mendez, Rodrigo
2015-01-01
A program to generate codes in Fortran and C of the full Magnetohydrodynamic equations is shown. The program used the free computer algebra system software REDUCE. This software has a package called EXCALC, which is an exterior calculus program. The advantage of this program is that it can be modified to include another complex metric or spacetime. The output of this program is modified by means of a LINUX script which creates a new REDUCE program to manipulate the MHD equations to obtain a c...
Skála, J.; Baruffa, F.; Büchner, J.; Rampp, M.
2015-08-01
Context. The numerical simulation of turbulence and flows in almost ideal astrophysical plasmas with large Reynolds numbers motivates the implementation of magnetohydrodynamical (MHD) computer codes with low resistivity. They need to be computationally efficient and scale well with large numbers of CPU cores, allow obtaining a high grid resolution over large simulation domains, and be easily and modularly extensible, for instance, to new initial and boundary conditions. Aims: Our aims are the implementation, optimization, and verification of a computationally efficient, highly scalable, and easily extensible low-dissipative MHD simulation code for the numerical investigation of the dynamics of astrophysical plasmas with large Reynolds numbers in three dimensions (3D). Methods: The new GOEMHD3 code discretizes the ideal part of the MHD equations using a fast and efficient leap-frog scheme that is second-order accurate in space and time and whose initial and boundary conditions can easily be modified. For the investigation of diffusive and dissipative processes the corresponding terms are discretized by a DuFort-Frankel scheme. To always fulfill the Courant-Friedrichs-Lewy stability criterion, the time step of the code is adapted dynamically. Numerically induced local oscillations are suppressed by explicit, externally controlled diffusion terms. Non-equidistant grids are implemented, which enhance the spatial resolution, where needed. GOEMHD3 is parallelized based on the hybrid MPI-OpenMP programing paradigm, adopting a standard two-dimensional domain-decomposition approach. Results: The ideal part of the equation solver is verified by performing numerical tests of the evolution of the well-understood Kelvin-Helmholtz instability and of Orszag-Tang vortices. The accuracy of solving the (resistive) induction equation is tested by simulating the decay of a cylindrical current column. Furthermore, we show that the computational performance of the code scales very
The CHEASE code for toroidal MHD equilibria
Luetjens, H. [Ecole Polytechnique, 91 - Palaiseau (France). Centre de Physique Theorique; Bondeson, A. [Chalmers Univ. of Technology, Goeteborg (Sweden). Inst. for Electromagnetic Field Theory and Plasma Physics; Sauter, O. [ITER-San Diego, La Jolla, CA (United States)
1996-03-01
CHEASE solves the Grad-Shafranov equation for the MHD equilibrium of a Tokamak-like plasma with pressure and current profiles specified by analytic forms or sets of data points. Equilibria marginally stable to ballooning modes or with a prescribed fraction of bootstrap current can be computed. The code provides a mapping to magnetic flux coordinates, suitable for MHD stability calculations or global wave propagation studies. The code computes equilibrium quantities for the stability codes ERATO, MARS, PEST, NOVA-W and XTOR and for the global wave propagation codes LION and PENN. The two-dimensional MHD equilibrium (Grad-Shafranov) equation is solved in variational form. The discretization uses bicubic Hermite finite elements with continuous first order derivates for the poloidal flux function {Psi}. The nonlinearity of the problem is handled by Picard iteration. The mapping to flux coordinates is carried out with a method which conserves the accuracy of the cubic finite elements. The code uses routines from the CRAY libsci.a program library. However, all these routines are included in the CHEASE package itself. If CHEASE computes equilibrium quantities for MARS with fast Fourier transforms, the NAG library is required. CHEASE is written in standard FORTRAN-77, except for the use of the input facility NAMELIST. CHEASE uses variable names with up to 8 characters, and therefore violates the ANSI standard. CHEASE transfers plot quantities through an external disk file to a plot program named PCHEASE using the UNIRAS or the NCAR plot package. (author) figs., tabs., 34 refs.
MAGNETOHYDRODYNAMIC EQUATIONS (MHD GENERATION CODE
Francisco Frutos Alfaro
2017-04-01
Full Text Available A program to generate codes in Fortran and C of the full magnetohydrodynamic equations is shown. The program uses the free computer algebra system software REDUCE. This software has a package called EXCALC, which is an exterior calculus program. The advantage of this program is that it can be modified to include another complex metric or spacetime. The output of this program is modified by means of a LINUX script which creates a new REDUCE program to manipulate the magnetohydrodynamic equations to obtain a code that can be used as a seed for a magnetohydrodynamic code for numerical applications. As an example, we present part of the output of our programs for Cartesian coordinates and how to do the discretization.
MHD computations for stellarators
Johnson, J.L.
1985-12-01
Considerable progress has been made in the development of computational techniques for studying the magnetohydrodynamic equilibrium and stability properties of three-dimensional configurations. Several different approaches have evolved to the point where comparison of results determined with different techniques shows good agreement. 55 refs., 7 figs.
Porting a Hall MHD Code to a Graphic Processing Unit
Dorelli, John C.
2011-01-01
We present our experience porting a Hall MHD code to a Graphics Processing Unit (GPU). The code is a 2nd order accurate MUSCL-Hancock scheme which makes use of an HLL Riemann solver to compute numerical fluxes and second-order finite differences to compute the Hall contribution to the electric field. The divergence of the magnetic field is controlled with Dedner?s hyperbolic divergence cleaning method. Preliminary benchmark tests indicate a speedup (relative to a single Nehalem core) of 58x for a double precision calculation. We discuss scaling issues which arise when distributing work across multiple GPUs in a CPU-GPU cluster.
Passive stabilization in a linear MHD stability code
Todd, A.M.M.
1980-03-01
Utilizing a Galerkin procedure to calculate the vacuum contribution to the ideal MHD Lagrangian, the implementation of realistic boundary conditions are described in a linear stability code. The procedure permits calculation of the effect of arbitrary conducting structure on ideal MHD instabilities, as opposed to the prior use of an encircling shell. The passive stabilization of conducting coils on the tokamak vertical instability is calculated within the PEST code and gives excellent agreement with 2-D time dependent simulations of PDX.
Modified NASA-Lewis Chemical Equilibrium Code for MHD applications
Sacks, R. A.; Geyer, H. K.; Grammel, S. J.; Doss, E. D.
1979-12-01
A substantially modified version of the NASA-Lewis Chemical Equilibrium Code has recently been developed. The modifications were designed to extend the power and convenience of the Code as a tool for performing combustor analysis for MHD systems studies. This report describes the effect of the programming details from a user point of view, but does not describe the Code in detail.
A New MHD Code with Adaptive Mesh Refinement and Parallelization for Astrophysics
Jiang, R L; Chen, P F
2012-01-01
A new code, named MAP, is written in Fortran language for magnetohydrodynamics (MHD) calculation with the adaptive mesh refinement (AMR) and Message Passing Interface (MPI) parallelization. There are several optional numerical schemes for computing the MHD part, namely, modified Mac Cormack Scheme (MMC), Lax-Friedrichs scheme (LF) and weighted essentially non-oscillatory (WENO) scheme. All of them are second order, two-step, component-wise schemes for hyperbolic conservative equations. The total variation diminishing (TVD) limiters and approximate Riemann solvers are also equipped. A high resolution can be achieved by the hierarchical block-structured AMR mesh. We use the extended generalized Lagrange multiplier (EGLM) MHD equations to reduce the non-divergence free error produced by the scheme in the magnetic induction equation. The numerical algorithms for the non-ideal terms, e.g., the resistivity and the thermal conduction, are also equipped in the MAP code. The details of the AMR and MPI algorithms are d...
Linear MHD stability studies with the STARWALL code
Merkel, P
2015-01-01
The STARWALL/CAS3D/OPTIM code package is a powerful tool to study the linear MHD stability of 3D, ideal equilibria in the presence of multiply-connected ideal and/or resistive conducting structures, and their feedback stabilization by external currents. Robust feedback stabilization of resistive wall modes can be modelled with the OPTIM code. Resistive MHD studies are possible combining STARWALL with the linear, resistive 2D CASTOR code as well as nonlinear MHD simulations combining STARWALL with the JOREK code. In the present paper, a detailed description of the STARWALL code is given and some of its applications are presented to demonstrate the methods used. Conducting structures are treated in the thin wall approximation and depending on their complexity they are discretized by a spectral method or by triangular finite elements. As an example, a configuration is considered consisting of an ideal plasma surrounded by a vacuum domain containing a resistive wall and bounded by an external wall. Ideal linear M...
A Fast MHD Code for Gravitationally Stratified Media using Graphical Processing Units: SMAUG
M. K. Griffiths; V. Fedun; R.Erdélyi
2015-03-01
Parallelization techniques have been exploited most successfully by the gaming/graphics industry with the adoption of graphical processing units (GPUs), possessing hundreds of processor cores. The opportunity has been recognized by the computational sciences and engineering communities, who have recently harnessed successfully the numerical performance of GPUs. For example, parallel magnetohydrodynamic (MHD) algorithms are important for numerical modelling of highly inhomogeneous solar, astrophysical and geophysical plasmas. Here, we describe the implementation of SMAUG, the Sheffield Magnetohydrodynamics Algorithm Using GPUs. SMAUG is a 1–3D MHD code capable of modelling magnetized and gravitationally stratified plasma. The objective of this paper is to present the numerical methods and techniques used for porting the code to this novel and highly parallel compute architecture. The methods employed are justified by the performance benchmarks and validation results demonstrating that the code successfully simulates the physics for a range of test scenarios including a full 3D realistic model of wave propagation in the solar atmosphere.
Global and Kinetic MHD Simulation by the Gpic-MHD Code
Hiroshi NAITOU; Yusuke YAMADA; Kenji KAJIWARA; Wei-li LEE; Shinji TOKUDA; Masatoshi YAGI
2011-01-01
In order to implement large-scale and high-beta tokamak simulation, a new algorithm of the electromagnetic gyrokinetic PIC （particle-in-cell） code was proposed and installed on the Gpic-MHD code [Gyrokinetic PIC code for magnetohydrodynamic （MHD） simulation]. In the new algorithm, the vorticity equation and the generalized Ohm＇s law along the magnetic field are derived from the basic equations of the gyrokinetic Vlasov, Poisson, and Ampere system and are used to describe the spatio-temporal evolution of the field quantities of the electrostatic potential φ and the longitudinal component of the vector potential Az. The basic algorithm is equivalent to solving the reduced-MHD-type equations with kinetic corrections, in which MHD physics related to Alfven modes are well described. The estimation of perturbed electron pressure from particle dynamics is dominant, while the effects of other moments are negligible. Another advantage of the algorithm is that the longitudinal induced electric field, ETz = -δAz/δt, is explicitly estimated by the generalized Ohm＇s law and used in the equations of motion. Furthermore, the particle velocities along the magnetic field are used （vz-formulation） instead of generalized momentums （pz-formulation）, hence there is no problem of ＇cancellation＇, which would otherwise appear when Az is estimated from the Ampere＇s law in the pz-formulation. The successful simulation of the collisionless internal kink mode by the new Gpic-MHD with realistic values of the large-scale and high-beta tokamaks revealed the usefulness of the new algorithm.
CAFE: A NEW RELATIVISTIC MHD CODE
Lora-Clavijo, F. D.; Cruz-Osorio, A. [Instituto de Astronomía, Universidad Nacional Autónoma de México, AP 70-264, Distrito Federal 04510, México (Mexico); Guzmán, F. S., E-mail: fdlora@astro.unam.mx, E-mail: aosorio@astro.unam.mx, E-mail: guzman@ifm.umich.mx [Instituto de Física y Matemáticas, Universidad Michoacana de San Nicolás de Hidalgo. Edificio C-3, Cd. Universitaria, 58040 Morelia, Michoacán, México (Mexico)
2015-06-22
We introduce CAFE, a new independent code designed to solve the equations of relativistic ideal magnetohydrodynamics (RMHD) in three dimensions. We present the standard tests for an RMHD code and for the relativistic hydrodynamics regime because we have not reported them before. The tests include the one-dimensional Riemann problems related to blast waves, head-on collisions of streams, and states with transverse velocities, with and without magnetic field, which is aligned or transverse, constant or discontinuous across the initial discontinuity. Among the two-dimensional (2D) and 3D tests without magnetic field, we include the 2D Riemann problem, a one-dimensional shock tube along a diagonal, the high-speed Emery wind tunnel, the Kelvin–Helmholtz (KH) instability, a set of jets, and a 3D spherical blast wave, whereas in the presence of a magnetic field we show the magnetic rotor, the cylindrical explosion, a case of Kelvin–Helmholtz instability, and a 3D magnetic field advection loop. The code uses high-resolution shock-capturing methods, and we present the error analysis for a combination that uses the Harten, Lax, van Leer, and Einfeldt (HLLE) flux formula combined with a linear, piecewise parabolic method and fifth-order weighted essentially nonoscillatory reconstructors. We use the flux-constrained transport and the divergence cleaning methods to control the divergence-free magnetic field constraint.
CAFE: A New Relativistic MHD Code
Lora-Clavijo, F D; Guzman, F S
2014-01-01
We present CAFE, a new independent code designed to solve the equations of Relativistic ideal Magnetohydrodynamics (RMHD) in 3D. We present the standard tests for a RMHD code and for the Relativistic Hydrodynamics (RMD) regime since we have not reported them before. The tests include the 1D Riemann problems related to blast waves, head-on collision of streams and states with transverse velocities, with and without magnetic field, which is aligned or transverse, constant or discontinuous across the initial discontinuity. Among the 2D tests, without magnetic field we include the 2D Riemann problem, the high speed Emery wind tunnel, the Kelvin-Helmholtz instability test and a set of jets, whereas in the presence of a magnetic field we show the magnetic rotor, the cylindrical explosion and the Kelvin-Helmholtz instability. The code uses High Resolution Shock Capturing methods and as a standard set up we present the error analysis with a simple combination that uses the HLLE flux formula combined with linear, PPM ...
CAFE: A New Relativistic MHD Code
Lora-Clavijo, F. D.; Cruz-Osorio, A.; Guzmán, F. S.
2015-06-01
We introduce CAFE, a new independent code designed to solve the equations of relativistic ideal magnetohydrodynamics (RMHD) in three dimensions. We present the standard tests for an RMHD code and for the relativistic hydrodynamics regime because we have not reported them before. The tests include the one-dimensional Riemann problems related to blast waves, head-on collisions of streams, and states with transverse velocities, with and without magnetic field, which is aligned or transverse, constant or discontinuous across the initial discontinuity. Among the two-dimensional (2D) and 3D tests without magnetic field, we include the 2D Riemann problem, a one-dimensional shock tube along a diagonal, the high-speed Emery wind tunnel, the Kelvin-Helmholtz (KH) instability, a set of jets, and a 3D spherical blast wave, whereas in the presence of a magnetic field we show the magnetic rotor, the cylindrical explosion, a case of Kelvin-Helmholtz instability, and a 3D magnetic field advection loop. The code uses high-resolution shock-capturing methods, and we present the error analysis for a combination that uses the Harten, Lax, van Leer, and Einfeldt (HLLE) flux formula combined with a linear, piecewise parabolic method and fifth-order weighted essentially nonoscillatory reconstructors. We use the flux-constrained transport and the divergence cleaning methods to control the divergence-free magnetic field constraint.
Extension of the MURaM radiative MHD code for coronal simulations
Rempel, Matthias
2016-01-01
We present a new version of the MURaM radiative MHD code that allows for simulations spanning from the upper convection zone into the solar corona. We implemented the relevant coronal physics in terms of optically thin radiative loss, field aligned heat conduction and an equilibrium ionization equation of state. We artificially limit the coronal Alfv{\\'e}n and heat conduction speeds to computationally manageable values using an approximation to semi-relativistic MHD with an artificially reduced speed of light (Boris correction). We present example solutions ranging from quiet to active Sun in order to verify the validity of our approach. We quantify the role of numerical diffusivity for the effective coronal heating. We find that the (numerical) magnetic Prandtl number determines the ratio of resistive to viscous heating and that owing to the very large magnetic Prandtl number of the solar corona, heating is expected to happen predominantly through viscous dissipation. We find that reasonable solutions can be...
The 3D MHD code GOEMHD3 for large-Reynolds-number astrophysical plasmas
Skála, J; Büchner, J; Rampp, M
2014-01-01
The numerical simulation of turbulence and flows in almost ideal, large-Reynolds-number astrophysical plasmas motivates the implementation of almost conservative MHD computer codes. They should efficiently calculate, use highly parallelized schemes scaling well with large numbers of CPU cores, allows to obtain a high grid resolution over large simulation domains and which can easily be adapted to new computer architectures as well as to new initial and boundary conditions, allow modular extensions. The new massively parallel simulation code GOEMHD3 enables efficient and fast simulations of almost ideal, large-Reynolds-number astrophysical plasma flows, well resolved and on huge grids covering large domains. Its abilities are validated by major tests of ideal and weakly dissipative plasma phenomena. The high resolution ($2048^3$ grid points) simulation of a large part of the solar corona above an observed active region proved the excellent parallel scalability of the code using more than 30.000 processor cores...
Relativistic modeling capabilities in PERSEUS extended MHD simulation code for HED plasmas
Hamlin, Nathaniel D., E-mail: nh322@cornell.edu [438 Rhodes Hall, Cornell University, Ithaca, NY, 14853 (United States); Seyler, Charles E., E-mail: ces7@cornell.edu [Cornell University, Ithaca, NY, 14853 (United States)
2014-12-15
We discuss the incorporation of relativistic modeling capabilities into the PERSEUS extended MHD simulation code for high-energy-density (HED) plasmas, and present the latest hybrid X-pinch simulation results. The use of fully relativistic equations enables the model to remain self-consistent in simulations of such relativistic phenomena as X-pinches and laser-plasma interactions. By suitable formulation of the relativistic generalized Ohm’s law as an evolution equation, we have reduced the recovery of primitive variables, a major technical challenge in relativistic codes, to a straightforward algebraic computation. Our code recovers expected results in the non-relativistic limit, and reveals new physics in the modeling of electron beam acceleration following an X-pinch. Through the use of a relaxation scheme, relativistic PERSEUS is able to handle nine orders of magnitude in density variation, making it the first fluid code, to our knowledge, that can simulate relativistic HED plasmas.
Li, Xujing; Zheng, Weiying
2016-10-01
A new parallel code based on discontinuous Galerkin (DG) method for hyperbolic conservation laws on three dimensional unstructured meshes is developed recently. This code can be used for simulations of MHD equations, which are very important in magnetic confined plasma research. The main challenges in MHD simulations in fusion include the complex geometry of the configurations, such as plasma in tokamaks, the possibly discontinuous solutions and large scale computing. Our new developed code is based on three dimensional unstructured meshes, i.e. tetrahedron. This makes the code flexible to arbitrary geometries. Second order polynomials are used on each element and HWENO type limiter are applied. The accuracy tests show that our scheme reaches the desired three order accuracy and the nonlinear shock test demonstrate that our code can capture the sharp shock transitions. Moreover, One of the advantages of DG compared with the classical finite element methods is that the matrices solved are localized on each element, making it easy for parallelization. Several simulations including the kink instabilities in toroidal geometry will be present here. Chinese National Magnetic Confinement Fusion Science Program 2015GB110003.
A new MHD code with adaptive mesh refinement and parallelization for astrophysics
Jiang, R.-L.; Fang, C.; Chen, P.-F.
2012-08-01
A new code, named MAP, is written in FORTRAN language for magnetohydrodynamics (MHD) simulations with the adaptive mesh refinement (AMR) and Message Passing Interface (MPI) parallelization. There are several optional numerical schemes for computing the MHD part, namely, modified Mac Cormack Scheme (MMC), Lax-Friedrichs scheme (LF), and weighted essentially non-oscillatory (WENO) scheme. All of them are second-order, two-step, component-wise schemes for hyperbolic conservative equations. The total variation diminishing (TVD) limiters and approximate Riemann solvers are also equipped. A high resolution can be achieved by the hierarchical block-structured AMR mesh. We use the extended generalized Lagrange multiplier (EGLM) MHD equations to reduce the non-divergence free error produced by the scheme in the magnetic induction equation. The numerical algorithms for the non-ideal terms, e.g., the resistivity and the thermal conduction, are also equipped in the code. The details of the AMR and MPI algorithms are described in the paper.
VisAn MHD: a toolbox in Matlab for MHD computer model data visualisation and analysis
P. Daum
2007-03-01
Full Text Available Among the many challenges facing modern space physics today is the need for a visualisation and analysis package which can examine the results from the diversity of numerical and empirical computer models as well as observational data. Magnetohydrodynamic (MHD models represent the latest numerical models of the complex Earth's space environment and have the unique ability to span the enormous distances present in the magnetosphere from several hundred kilometres to several thousand kilometres above the Earth surface. This feature enables scientist to study complex structures of processes where otherwise only point measurements from satellites or ground-based instruments are available. Only by combining these observational data and the MHD simulations it is possible to enlarge the scope of the point-to-point observations and to fill the gaps left by measurements in order to get a full 3-D representation of the processes in our geospace environment. In this paper we introduce the VisAn MHD toolbox for Matlab as a tool for the visualisation and analysis of observational data and MHD simulations. We have created an easy to use tool which is capable of highly sophisticated visualisations and data analysis of the results from a diverse set of MHD models in combination with in situ measurements from satellites and ground-based instruments. The toolbox is being released under an open-source licensing agreement to facilitate and encourage community use and contribution.
Extension of the MURaM Radiative MHD Code for Coronal Simulations
Rempel, M.
2017-01-01
We present a new version of the MURaM radiative magnetohydrodynamics (MHD) code that allows for simulations spanning from the upper convection zone into the solar corona. We implement the relevant coronal physics in terms of optically thin radiative loss, field aligned heat conduction, and an equilibrium ionization equation of state. We artificially limit the coronal Alfvén and heat conduction speeds to computationally manageable values using an approximation to semi-relativistic MHD with an artificially reduced speed of light (Boris correction). We present example solutions ranging from quiet to active Sun in order to verify the validity of our approach. We quantify the role of numerical diffusivity for the effective coronal heating. We find that the (numerical) magnetic Prandtl number determines the ratio of resistive to viscous heating and that owing to the very large magnetic Prandtl number of the solar corona, heating is expected to happen predominantly through viscous dissipation. We find that reasonable solutions can be obtained with values of the reduced speed of light just marginally larger than the maximum sound speed. Overall this leads to a fully explicit code that can compute the time evolution of the solar corona in response to photospheric driving using numerical time steps not much smaller than 0.1 s. Numerical simulations of the coronal response to flux emergence covering a time span of a few days are well within reach using this approach.
FOI-PERFECT code: 3D relaxation MHD modeling and Applications
Wang, Gang-Hua; Duan, Shu-Chao; Comutational Physics Team Team
2016-10-01
One of the challenges in numerical simulations of electromagnetically driven high energy density (HED) systems is the existence of vacuum region. FOI-PERFECT code adopts a full relaxation magnetohydrodynamic (MHD) model. The electromagnetic part of the conventional model adopts the magnetic diffusion approximation. The vacuum region is approximated by artificially increasing the resistivity. On one hand the phase/group velocity is superluminal and hence non-physical in the vacuum region, on the other hand a diffusion equation with large diffusion coefficient can only be solved by implicit scheme which is difficult to be parallelized and converge. A better alternative is to solve the full electromagnetic equations. Maxwell's equations coupled with the constitutive equation, generalized Ohm's law, constitute a relaxation model. The dispersion relation is given to show its transition from electromagnetic propagation in vacuum to resistive MHD in plasma in a natural way. The phase and group velocities are finite for this system. A better time stepping is adopted to give a 3rd full order convergence in time domain without the stiff relaxation term restriction. Therefore it is convenient for explicit & parallel computations. Some numerical results of FOI-PERFECT code are also given. Project supported by the National Natural Science Foundation of China (Grant No. 11571293) And Foundation of China Academy of Engineering Physics (Grant No. 2015B0201023).
Reuter, K.; Jenko, F.; Forest, C. B.; Bayliss, R. A.
2008-08-01
A parallel implementation of a nonlinear pseudo-spectral MHD code for the simulation of turbulent dynamos in spherical geometry is reported. It employs a dual domain decomposition technique in both real and spectral space. It is shown that this method shows nearly ideal scaling going up to 128 CPUs on Beowulf-type clusters with fast interconnect. Furthermore, the potential of exploiting single precision arithmetic on standard x86 processors is examined. It is pointed out that the MHD code thereby achieves a maximum speedup of 1.7, whereas the validity of the computations is still granted. The combination of both measures will allow for the direct numerical simulation of highly turbulent cases ( 1500
WhiskyMHD: Numerical Code for General Relativistic Magnetohydrodynamics
Baiotti, Luca; Giacomazzo, Bruno; Hawke, Ian; et al.
2010-10-01
Whisky is a code to evolve the equations of general relativistic hydrodynamics (GRHD) and magnetohydrodynamics (GRMHD) in 3D Cartesian coordinates on a curved dynamical background. It was originally developed by and for members of the EU Network on Sources of Gravitational Radiation and is based on the Cactus Computational Toolkit. Whisky can also implement adaptive mesh refinement (AMR) if compiled together with Carpet. Whisky has grown from earlier codes such as GR3D and GRAstro_Hydro, but has been rewritten to take advantage of some of the latest research performed here in the EU. The motivation behind Whisky is to compute gravitational radiation waveforms for systems that involve matter. Examples would include the merger of a binary system containing a neutron star, which are expected to be reasonably common in the universe and expected to produce substantial amounts of radiation. Other possible sources are given in the projects list.
Further validation of liquid metal MHD code for unstructured grid based on OpenFOAM
Feng, Jingchao; Chen, Hongli, E-mail: hlchen1@ustc.edu.cn; He, Qingyun; Ye, Minyou
2015-11-15
Highlights: • Specific correction scheme has been adopted to revise the calculating result for non-orthogonal meshes. • The developed MHD code based on OpenFOAM platform has been validated by benchmark cases under uniform and non-uniform magnetic field in round and rectangular ducts. • ALEX experimental results have been used to validate the MHD code based on OpenFOAM. - Abstract: In fusion liquid metal blankets, complex geometries involving contractions, expansions, bends, manifolds are very common. The characteristics of liquid metal flow in these geometries are significant. In order to extend the magnetohydrodynamic (MHD) solver developed on OpenFOAM platform to be applied in the complex geometry, the MHD solver based on unstructured meshes has been implemented. The adoption of non-orthogonal correction techniques in the solver makes it possible to process the non-orthogonal meshes in complex geometries. The present paper focused on the validation of the code under critical conditions. An analytical solution benchmark case and two experimental benchmark cases were conducted to validate the code. Benchmark case I is MHD flow in a circular pipe with arbitrary electric conductivity of the walls in a uniform magnetic field. Benchmark cases II and III are experimental cases of 3D laminar steady MHD flow under fringing magnetic field. In all these cases, the numerical results match well with the benchmark cases.
Summers, R.M.; Cole, R.K. Jr.; Smith, R.C.; Stuart, D.S.; Thompson, S.L. [Sandia National Labs., Albuquerque, NM (United States); Hodge, S.A.; Hyman, C.R.; Sanders, R.L. [Oak Ridge National Lab., TN (United States)
1995-03-01
MELCOR is a fully integrated, engineering-level computer code that models the progression of severe accidents in light water reactor nuclear power plants. MELCOR is being developed at Sandia National Laboratories for the U.S. Nuclear Regulatory Commission as a second-generation plant risk assessment tool and the successor to the Source Term Code Package. A broad spectrum of severe accident phenomena in both boiling and pressurized water reactors is treated in MELCOR in a unified framework. These include: thermal-hydraulic response in the reactor coolant system, reactor cavity, containment, and confinement buildings; core heatup, degradation, and relocation; core-concrete attack; hydrogen production, transport, and combustion; fission product release and transport; and the impact of engineered safety features on thermal-hydraulic and radionuclide behavior. Current uses of MELCOR include estimation of severe accident source terms and their sensitivities and uncertainties in a variety of applications. This publication of the MELCOR computer code manuals corresponds to MELCOR 1.8.3, released to users in August, 1994. Volume 1 contains a primer that describes MELCOR`s phenomenological scope, organization (by package), and documentation. The remainder of Volume 1 contains the MELCOR Users Guides, which provide the input instructions and guidelines for each package. Volume 2 contains the MELCOR Reference Manuals, which describe the phenomenological models that have been implemented in each package.
Divergence-free MHD Simulations with the HERACLES Code
Vides J.
2013-12-01
Full Text Available Numerical simulations of the magnetohydrodynamics (MHD equations have played a significant role in plasma research over the years. The need of obtaining physical and stable solutions to these equations has led to the development of several schemes, all requiring to satisfy and preserve the divergence constraint of the magnetic field numerically. In this paper, we aim to show the importance of maintaining this constraint numerically. We investigate in particular the hyperbolic divergence cleaning technique applied to the ideal MHD equations on a collocated grid and compare it to the constrained transport technique that uses a staggered grid to maintain the property. The methods are implemented in the software HERACLES and several numerical tests are presented, where the robustness and accuracy of the different schemes can be directly compared.
On the measurements of numerical viscosity and resistivity in Eulerian MHD codes
Rembiasz, Tomasz; Cerdá-Durán, Pablo; Aloy, Miguel-Ángel; Müller, Ewald
2016-01-01
We propose a simple ansatz for estimating the value of the numerical resistivity and the numerical viscosity of any Eulerian MHD code. We test this ansatz with the help of simulations of the propagation of (magneto)sonic waves, Alfven waves, and the tearing mode instability using the MHD code Aenus. By comparing the simu- lation results with analytical solutions of the resistive-viscous MHD equations and an empirical ansatz for the growth rate of tearing modes we measure the numerical viscosity and resistivity of Aenus. The comparison shows that the fast-magnetosonic speed and wavelength are the characteristic velocity and length, respectively, of the aforementioned (relatively simple) systems. We also determine the dependance of the numerical viscosity and resistivity on the time integration method, the spatial reconstruction scheme and (to a lesser extent) the Riemann solver employed in the simulations. From the measured results we infer the numerical resolution (as a function of the spatial reconstruction ...
Hirshman, S. P.; Shafer, M. W.; Seal, S. K.; Canik, J. M.
2016-04-01
> The SIESTA magnetohydrodynamic (MHD) equilibrium code has been used to compute a sequence of ideally stable equilibria resulting from numerical variation of the helical resonant magnetic perturbation (RMP) applied to an axisymmetric DIII-D plasma equilibrium. Increasing the perturbation strength at the dominant , resonant surface leads to lower MHD energies and increases in the equilibrium island widths at the (and sidebands) surfaces, in agreement with theoretical expectations. Island overlap at large perturbation strengths leads to stochastic magnetic fields which correlate well with the experimentally inferred field structure. The magnitude and spatial phase (around the dominant rational surfaces) of the resonant (shielding) component of the parallel current are shown to change qualitatively with the magnetic island topology.
Network coding for computing: Linear codes
Appuswamy, Rathinakumar; Karamchandani, Nikhil; Zeger, Kenneth
2011-01-01
In network coding it is known that linear codes are sufficient to achieve the coding capacity in multicast networks and that they are not sufficient in general to achieve the coding capacity in non-multicast networks. In network computing, Rai, Dey, and Shenvi have recently shown that linear codes are not sufficient in general for solvability of multi-receiver networks with scalar linear target functions. We study single receiver networks where the receiver node demands a target function of the source messages. We show that linear codes may provide a computing capacity advantage over routing only when the receiver demands a `linearly-reducible' target function. % Many known target functions including the arithmetic sum, minimum, and maximum are not linearly-reducible. Thus, the use of non-linear codes is essential in order to obtain a computing capacity advantage over routing if the receiver demands a target function that is not linearly-reducible. We also show that if a target function is linearly-reducible,...
Newtonian CAFE: a new ideal MHD code to study the solar atmosphere
González-Avilés, J. J.; Cruz-Osorio, A.; Lora-Clavijo, F. D.; Guzmán, F. S.
2015-12-01
We present a new code designed to solve the equations of classical ideal magnetohydrodynamics (MHD) in three dimensions, submitted to a constant gravitational field. The purpose of the code centres on the analysis of solar phenomena within the photosphere-corona region. We present 1D and 2D standard tests to demonstrate the quality of the numerical results obtained with our code. As solar tests we present the transverse oscillations of Alfvénic pulses in coronal loops using a 2.5D model, and as 3D tests we present the propagation of impulsively generated MHD-gravity waves and vortices in the solar atmosphere. The code is based on high-resolution shock-capturing methods, uses the Harten-Lax-van Leer-Einfeldt (HLLE) flux formula combined with Minmod, MC, and WENO5 reconstructors. The divergence free magnetic field constraint is controlled using the Flux Constrained Transport method.
Newtonian CAFE: a new ideal MHD code to study the solar atmosphere
Gonzalez-Aviles, J J; Lora-Clavijo, F D; Guzman, F S
2015-01-01
We present a new code designed to solve the equations of classical ideal magneto-hydrodynamics (MHD) in three dimensions, submitted to a constant gravitational field. The purpose of the code centers on the analysis of solar phenomena within the photosphere-corona region. We present 1D and 2D standard tests to demonstrate the quality of the numerical results obtained with our code. As solar tests we present the transverse oscillations of Alfvenic pulses in coronal loops using a 2.5D model, and as 3D tests we present the propagation of impulsively generated MHD-gravity waves and vortices in the solar atmosphere. The code is based on high-resolution shock-capturing methods, uses the HLLE flux formula combined with Minmod, MC and WENO5 reconstructors. The divergence free magnetic field constraint is controlled using the Flux Constrained Transport method.
A numerical code for a three-dimensional magnetospheric MHD equilibrium model
Voigt, G.-H.
1992-01-01
Two dimensional and three dimensional MHD equilibrium models were begun for Earth's magnetosphere. The original proposal was motivated by realizing that global, purely data based models of Earth's magnetosphere are inadequate for studying the underlying plasma physical principles according to which the magnetosphere evolves on the quasi-static convection time scale. Complex numerical grid generation schemes were established for a 3-D Poisson solver, and a robust Grad-Shafranov solver was coded for high beta MHD equilibria. Thus, the effects were calculated of both the magnetopause geometry and boundary conditions on the magnetotail current distribution.
Large Scale Earth’s Bow Shock with Northern IMF as Simulated by PIC Code in Parallel with MHD Model
Suleiman Baraka
2016-06-01
In this paper, we propose a 3D kinetic model (particle-in-cell, PIC) for the description of the large scale Earth’s bow shock. The proposed version is stable and does not require huge or extensive computer resources. Because PIC simulations work with scaled plasma and field parameters, we also propose to validate our code by comparing its results with the available MHD simulations under same scaled solar wind (SW) and (IMF) conditions. We report new results from the two models. In both codes the Earth’s bow shock position is found to be $\\approx 14.8 R_{{\\rm E}}$ along the Sun–Earth line, and $\\approx 29 R_{{\\rm E}}$ on the dusk side. Those findings are consistent with past in situ observations. Both simulations reproduce the theoretical jump conditions at the shock. However, the PIC code density and temperature distributions are inflated and slightly shifted sunward when compared to the MHD results. Kinetic electron motions and reflected ions upstream may cause this sunward shift. Species distributions in the foreshock region are depicted within the transition of the shock (measured $\\approx$2$c/\\omega_{pi}$ for $ \\Theta_{Bn}=90^{\\circ}$ and $M_{{\\rm MS}} = 4.7 $) and in the downstream. The size of the foot jump in the magnetic field at the shock is measured to be ($1.7 c/ \\omega_{pi} $). In the foreshocked region, the thermal velocity is found equal to 213 km $s^{−1}$ at $15R_{{\\rm E}}$ and is equal to $63 km s^{-1}$ at $12 R_{{\\rm E}}$ (magnetosheath region). Despite the large cell size of the current version of the PIC code, it is powerful to retain macrostructure of planets magnetospheres in very short time, thus it can be used for pedagogical test purposes. It is also likely complementary with MHD to deepen our understanding of the large scale magnetosphere.
Boss, Alan P
2013-01-01
Magnetic fields are important contributers to the dynamics of collapsing molecular cloud cores, and can have a major effect on whether collapse results in a single protostar or fragmentation into a binary or multiple protostar system. New models are presented of the collapse of magnetic cloud cores using the adaptive mesh refinement (AMR) code Enzo2.0. The code was used to calculate the ideal magnetohydrodynamics (MHD) of initially spherical, uniform density and rotation clouds with density perturbations, i.e., the Boss and Bodenheimer (1979) standard isothermal test case for three dimensional (3D) hydrodynamics (HD) codes. After first verifying that Enzo reproduces the binary fragmentation expected for the non-magnetic test case, a large set of models was computed with varied initial magnetic field strengths and directions with respect to the cloud core axis of rotation (parallel or perpendicular), density perturbation amplitudes, and equations of state. Three significantly different outcomes resulted: (1) c...
Shapiro, Wilbur
1996-01-01
This is an overview of new and updated industrial codes for seal design and testing. GCYLT (gas cylindrical seals -- turbulent), SPIRALI (spiral-groove seals -- incompressible), KTK (knife to knife) Labyrinth Seal Code, and DYSEAL (dynamic seal analysis) are covered. CGYLT uses G-factors for Poiseuille and Couette turbulence coefficients. SPIRALI is updated to include turbulence and inertia, but maintains the narrow groove theory. KTK labyrinth seal code handles straight or stepped seals. And DYSEAL provides dynamics for the seal geometry.
Stabilization of the SIESTA MHD Equilibrium Code Using Rapid Cholesky Factorization
Hirshman, S. P.; D'Azevedo, E. A.; Seal, S. K.
2016-10-01
The SIESTA MHD equilibrium code solves the discretized nonlinear MHD force F ≡ J X B - ∇p for a 3D plasma which may contain islands and stochastic regions. At each nonlinear evolution step, it solves a set of linearized MHD equations which can be written r ≡ Ax - b = 0, where A is the linearized MHD Hessian matrix. When the solution norm | x| is small enough, the nonlinear force norm will be close to the linearized force norm | r| 0 obtained using preconditioned GMRES. In many cases, this procedure works well and leads to a vanishing nonlinear residual (equilibrium) after several iterations in SIESTA. In some cases, however, | x|>1 results and the SIESTA code has to be restarted to obtain nonlinear convergence. In order to make SIESTA more robust and avoid such restarts, we have implemented a new rapid QR factorization of the Hessian which allows us to rapidly and accurately solve the least-squares problem AT r = 0, subject to the condition | x|QR method is based on a pairwise row factorization of the tri-diagonal Hessian. It provides a complete Cholesky factorization while preserving the memory allocation of A. This work was supported by the U.S. D.O.E. contract DE-AC05-00OR22725.
Computer controlled MHD power consolidation and pulse-generation system
Johnson, R.
The major goal of this project is to establish the feasibility of a power conversion technology which will permit the direct synthesis of computer programmable pulse power. Feasibility will be established in this project by demonstration of direct synthesis of commercial frequency power by means of computer control. The power input to the conversion system is assumed to be a magnetohydrodynamic (MHD) Faraday connected generator which may be viewed as a multi-terminal d.c. source. This consolidation/inversion process is referred to subsequently as Pulse-Amplitude-Synthesis-and-Control (PASC). A secondary goal is to deliver a controller subsystem consisting of a computer, software, and computer interface board which can serve as one of the building blocks for a possible Phase 2 prototype system. This report covers the initial six months portion of the project and includes discussions on the following areas: (1) selection of a control computer with software tool kit for development of the PASC controller contract requirement; (2) problem formulation considerations for simulation of the PASC technique on digital computers; (3) initial simulation results for the PASC transformer, including simulation results obtained using SPICE and the INTEG program; (4) a survey of available gate-turn-off (GTO's), power semiconductors, power field effect transistors (PFET's), and fiber optics signal cabling and transducers.
Large Scale Earth's Bow Shock with Northern IMF as simulated by PIC code in parallel with MHD model
Baraka, Suleiman M
2016-01-01
In this paper, we propose a 3D kinetic model (Particle-in-Cell PIC ) for the description of the large scale Earth's bow shock. The proposed version is stable and does not require huge or extensive computer resources. Because PIC simulations work with scaled plasma and field parameters, we also propose to validate our code by comparing its results with the available MHD simulations under same scaled Solar wind ( SW ) and ( IMF ) conditions. We report new results from the two models. In both codes the Earth's bow shock position is found to be ~14.8 RE along the Sun-Earth line, and ~ 29 RE on the dusk side. Those findings are consistent with past in situ observations. Both simulations reproduce the theoretical jump conditions at the shock. However, the PIC code density and temperature distributions are inflated and slightly shifted sunward when compared to the MHD results. Kinetic electron motions and reflected ions upstream may cause this sunward shift. Species distributions in the foreshock region are depicted...
Progress on accelerated calculation of 3D MHD equilibrium with the PIES code
Raburn, Daniel; Reiman, Allan; Monticello, Donald
2016-10-01
Continuing progress has been made in accelerating the 3D MHD equilibrium code, PIES, using an external numerical wrapper. The PIES code (Princeton Iterative Equilibrium Solver) is capable of calculating 3D MHD equilibria with islands. The numerical wrapper has been demonstrated to greatly improve the rate of convergence in numerous cases corresponding to equilibria in the TFTR device where magnetic islands are present; the numerical wrapper makes use of a Jacobian-free Newton-Krylov solver along with adaptive preconditioning and a sophisticated subspace-restricted Levenberg backtracking algorithm. The wrapper has recently been improved by automation which combines the preexisting backtracking algorithm with insights gained from the stability of the Picard algorithm traditionally used with PIES. Improved progress logging and stopping criteria have also been incorporated in to the numerical wrapper.
FARGO3D: A new GPU-oriented MHD code
Benítez-Llambay, Pablo
2016-01-01
We present the FARGO3D code, recently publicly released. It is a magnetohydrodynamics code developed with special emphasis on protoplanetary disks physics and planet-disk interactions, and parallelized with MPI. The hydrodynamics algorithms are based on finite difference upwind, dimensionally split methods. The magnetohydrodynamics algorithms consist of the constrained transport method to preserve the divergence-free property of the magnetic field to machine accuracy, coupled to a method of characteristics for the evaluation of electromotive forces and Lorentz forces. Orbital advection is implemented, and an N-body solver is included to simulate planets or stars interacting with the gas. We present our implementation in detail and present a number of widely known tests for comparison purposes. One strength of FARGO3D is that it can run on both "Graphical Processing Units" (GPUs) or "Central Processing unit" (CPUs), achieving large speed up with respect to CPU cores. We describe our implementation choices, whi...
Viscous, resistive MHD stability computed by spectral techniques
Dahlburg, R. B.; Zang, T. A.; Montgomery, D.; Hussaini, M. Y.
1983-01-01
Expansions in Chebyshev polynomials are used to study the linear stability of one dimensional magnetohydrodynamic (MHD) quasi-equilibria, in the presence of finite resistivity and viscosity. The method is modeled on the one used by Orszag in accurate computation of solutions of the Orr-Sommerfeld equation. Two Reynolds like numbers involving Alfven speeds, length scales, kinematic viscosity, and magnetic diffusivity govern the stability boundaries, which are determined by the geometric mean of the two Reynolds like numbers. Marginal stability curves, growth rates versus Reynolds like numbers, and growth rates versus parallel wave numbers are exhibited. A numerical result which appears general is that instability was found to be associated with inflection points in the current profile, though no general analytical proof has emerged. It is possible that nonlinear subcritical three dimensional instabilities may exist, similar to those in Poiseuille and Couette flow.
Computer controlled MHD power consolidation and pulse generation system
Johnson, R.; Marcotte, K.; Donnelly, M.
1990-01-01
The major goal of this research project is to establish the feasibility of a power conversion technology which will permit the direct synthesis of computer programmable pulse power. Feasibility has been established in this project by demonstration of direct synthesis of commercial frequency power by means of computer control. The power input to the conversion system is assumed to be a Faraday connected MHD generator which may be viewed as a multi-terminal dc source and is simulated for the purpose of this demonstration by a set of dc power supplies. This consolidation/inversion (CI), process will be referred to subsequently as Pulse Amplitude Synthesis and Control (PASC). A secondary goal is to deliver a controller subsystem consisting of a computer, software, and computer interface board which can serve as one of the building blocks for a possible phase II prototype system. This report period work summarizes the accomplishments and covers the high points of the two year project. 6 refs., 41 figs.
FARGO3D: A NEW GPU-ORIENTED MHD CODE
Benitez-Llambay, Pablo [Instituto de Astronomía Teórica y Experimental, Observatorio Astronónomico, Universidad Nacional de Córdoba. Laprida 854, X5000BGR, Córdoba (Argentina); Masset, Frédéric S., E-mail: pbllambay@oac.unc.edu.ar, E-mail: masset@icf.unam.mx [Instituto de Ciencias Físicas, Universidad Nacional Autónoma de México (UNAM), Apdo. Postal 48-3,62251-Cuernavaca, Morelos (Mexico)
2016-03-15
We present the FARGO3D code, recently publicly released. It is a magnetohydrodynamics code developed with special emphasis on the physics of protoplanetary disks and planet–disk interactions, and parallelized with MPI. The hydrodynamics algorithms are based on finite-difference upwind, dimensionally split methods. The magnetohydrodynamics algorithms consist of the constrained transport method to preserve the divergence-free property of the magnetic field to machine accuracy, coupled to a method of characteristics for the evaluation of electromotive forces and Lorentz forces. Orbital advection is implemented, and an N-body solver is included to simulate planets or stars interacting with the gas. We present our implementation in detail and present a number of widely known tests for comparison purposes. One strength of FARGO3D is that it can run on either graphical processing units (GPUs) or central processing units (CPUs), achieving large speed-up with respect to CPU cores. We describe our implementation choices, which allow a user with no prior knowledge of GPU programming to develop new routines for CPUs, and have them translated automatically for GPUs.
On the Measurements of Numerical Viscosity and Resistivity in Eulerian MHD Codes
Rembiasz, Tomasz; Obergaulinger, Martin; Cerdá-Durán, Pablo; Aloy, Miguel-Ángel; Müller, Ewald
2017-06-01
We propose a simple ansatz for estimating the value of the numerical resistivity and the numerical viscosity of any Eulerian MHD code. We test this ansatz with the help of simulations of the propagation of (magneto)sonic waves, Alfvén waves, and the tearing mode (TM) instability using the MHD code Aenus. By comparing the simulation results with analytical solutions of the resistive-viscous MHD equations and an empirical ansatz for the growth rate of TMs, we measure the numerical viscosity and resistivity of Aenus. The comparison shows that the fast magnetosonic speed and wavelength are the characteristic velocity and length, respectively, of the aforementioned (relatively simple) systems. We also determine the dependence of the numerical viscosity and resistivity on the time integration method, the spatial reconstruction scheme and (to a lesser extent) the Riemann solver employed in the simulations. From the measured results, we infer the numerical resolution (as a function of the spatial reconstruction method) required to properly resolve the growth and saturation level of the magnetic field amplified by the magnetorotational instability in the post-collapsed core of massive stars. Our results show that it is most advantageous to resort to ultra-high-order methods (e.g., the ninth-order monotonicity-preserving method) to tackle this problem properly, in particular, in three-dimensional simulations.
Alfvén ionization in an MHD-gas interactions code
Wilson, A. D.; Diver, D. A.
2016-07-01
A numerical model of partially ionized plasmas is developed in order to capture their evolving ionization fractions as a result of Alfvén ionization (AI). The mechanism of, and the parameter regime necessary for, AI is discussed and an expression for the AI rate based on fluid parameters, from a gas-MHD model, is derived. This AI term is added to an existing MHD-gas interactions' code, and the result is a linear, 2D, two-fluid model that includes momentum transfer between charged and neutral species as well as an ionization rate that depends on the velocity fields of both fluids. The dynamics of waves propagating through such a partially ionized plasma are investigated, and it is found that AI has a significant influence on the fluid dynamics as well as both the local and global ionization fraction.
Laser-Plasma Modeling Using PERSEUS Extended-MHD Simulation Code for HED Plasmas
Hamlin, Nathaniel; Seyler, Charles
2016-10-01
We discuss the use of the PERSEUS extended-MHD simulation code for high-energy-density (HED) plasmas in modeling laser-plasma interactions in relativistic and nonrelativistic regimes. By formulating the fluid equations as a relaxation system in which the current is semi-implicitly time-advanced using the Generalized Ohm's Law, PERSEUS enables modeling of two-fluid phenomena in dense plasmas without the need to resolve the smallest electron length and time scales. For relativistic and nonrelativistic laser-target interactions, we have validated a cycle-averaged absorption (CAA) laser driver model against the direct approach of driving the electromagnetic fields. The CAA model refers to driving the radiation energy and flux rather than the fields, and using hyperbolic radiative transport, coupled to the plasma equations via energy source terms, to model absorption and propagation of the radiation. CAA has the advantage of not requiring adequate grid resolution of each laser wavelength, so that the system can span many wavelengths without requiring prohibitive CPU time. For several laser-target problems, we compare existing MHD results to extended-MHD results generated using PERSEUS with the CAA model, and examine effects arising from Hall physics. This work is supported by the National Nuclear Security Administration stewardship sciences academic program under Department of Energy cooperative agreements DE-FOA-0001153 and DE-NA0001836.
A novel code for numerical 3-D MHD studies of CME expansion
J. Kleimann
2009-03-01
Full Text Available A recent third-order, essentially non-oscillatory central scheme to advance the equations of single-fluid magnetohydrodynamics (MHD in time has been implemented into a new numerical code. This code operates on a 3-D Cartesian, non-staggered grid, and is able to handle shock-like gradients without producing spurious oscillations.
To demonstrate the suitability of our code for the simulation of coronal mass ejections (CMEs and similar heliospheric transients, we present selected results from test cases and perform studies of the solar wind expansion during phases of minimum solar activity. We can demonstrate convergence of the system into a stable Parker-like steady state for both hydrodynamic and MHD winds. The model is subsequently applied to expansion studies of CME-like plasma bubbles, and their evolution is monitored until a stationary state similar to the initial one is achieved. In spite of the model's (current simplicity, we can confirm the CME's nearly self-similar evolution close to the Sun, thus highlighting the importance of detailed modelling especially at small heliospheric radii.
Additionally, alternative methods to implement boundary conditions at the coronal base, as well as strategies to ensure a solenoidal magnetic field, are discussed and evaluated.
CosmosDG: An hp-adaptive Discontinuous Galerkin Code for Hyper-resolved Relativistic MHD
Anninos, Peter; Bryant, Colton; Fragile, P. Chris; Holgado, A. Miguel; Lau, Cheuk; Nemergut, Daniel
2017-08-01
We have extended Cosmos++, a multidimensional unstructured adaptive mesh code for solving the covariant Newtonian and general relativistic radiation magnetohydrodynamic (MHD) equations, to accommodate both discrete finite volume and arbitrarily high-order finite element structures. The new finite element implementation, called CosmosDG, is based on a discontinuous Galerkin (DG) formulation, using both entropy-based artificial viscosity and slope limiting procedures for the regularization of shocks. High-order multistage forward Euler and strong-stability preserving Runge-Kutta time integration options complement high-order spatial discretization. We have also added flexibility in the code infrastructure allowing for both adaptive mesh and adaptive basis order refinement to be performed separately or simultaneously in a local (cell-by-cell) manner. We discuss in this report the DG formulation and present tests demonstrating the robustness, accuracy, and convergence of our numerical methods applied to special and general relativistic MHD, although we note that an equivalent capability currently also exists in CosmosDG for Newtonian systems.
Benchmarking a hybrid MHD/kinetic code with C-2 experimental data
Magee, Richard; Clary, Ryan; Dettrick, Sean; Korepanov, Sergey; Onofri, Marco; Smirnov, Artem; TAE Team
2013-10-01
The C-2 device creates field-reversed configuration (FRC) plasmas via the dynamic merging of two compact toroids and heated with neutral beams. Simulations of these plasmas are performed with Q2D - a hybrid MHD/Monte Carlo code that evolves the plasma according to the resistive MHD equations and treats the neutral beam injected fast ions as a minority kinetic species. Recent Q2D runs have resulted in testable predictions, namely that the axial profile of the fast ions is double-peaked, and charge-exchange neutrals are localized in pitch-angle. In some simulations, the fast particle population can induce magnetic fluctuations. These fluctuations are largest in the radial component, have a characteristic frequency approximately equal to the fast ion bounce frequency (f ~ 150 kHz), and a broad k spectrum. These fluctuations have the beneficial effect of smoothing out the double-peaked axial fast ion density profile, resulting in an increased fast ion density at the mid-plane. We will present results from a benchmarking study to quantitatively compare the results of Q2D runs to existing C-2 experimental data.
Marx, Alain; Lütjens, Hinrich
2017-03-01
A hybrid MPI/OpenMP parallel version of the XTOR-2F code [Lütjens and Luciani, J. Comput. Phys. 229 (2010) 8130] solving the two-fluid MHD equations in full tokamak geometry by means of an iterative Newton-Krylov matrix-free method has been developed. The present work shows that the code has been parallelized significantly despite the numerical profile of the problem solved by XTOR-2F, i.e. a discretization with pseudo-spectral representations in all angular directions, the stiffness of the two-fluid stability problem in tokamaks, and the use of a direct LU decomposition to invert the physical pre-conditioner at every Krylov iteration of the solver. The execution time of the parallelized version is an order of magnitude smaller than the sequential one for low resolution cases, with an increasing speedup when the discretization mesh is refined. Moreover, it allows to perform simulations with higher resolutions, previously forbidden because of memory limitations.
An MHD Code for the Study of Magnetic Structures in the Solar Wind
Allred, J. C.; MacNeice, P. J.
2015-01-01
We have developed a 2.5D MHD code designed to study how the solar wind influences the evolution of transient events in the solar corona and inner heliosphere. The code includes thermal conduction, coronal heating and radiative cooling. Thermal conduction is assumed to be magnetic field-aligned in the inner corona and transitions to a collisionless formulation in the outer corona. We have developed a stable method to handle field-aligned conduction around magnetic null points. The inner boundary is placed in the upper transition region, and the mass flux across the boundary is determined from 1D field-aligned characteristics and a 'radiative energy balance' condition. The 2.5D nature of this code makes it ideal for parameter studies not yet possible with 3D codes. We have made this code publicly available as a tool for the community. To this end we have developed a graphical interface to aid in the selection of appropriate options and a graphical interface that can process and visualize the data produced by the simulation. As an example, we show a simulation of a dipole field stretched into a helmet streamer by the solar wind. Plasmoids periodically erupt from the streamer, and we perform a parameter study of how the frequency and location of these eruptions changed in response to different levels of coronal heating. As a further example, we show the solar wind stretching a compact multi-polar flux system. This flux system will be used to study breakout coronal mass ejections in the presence of the solar wind.
Physics codes on parallel computers
Eltgroth, P.G.
1985-12-04
An effort is under way to develop physics codes which realize the potential of parallel machines. A new explicit algorithm for the computation of hydrodynamics has been developed which avoids global synchronization entirely. The approach, called the Independent Time Step Method (ITSM), allows each zone to advance at its own pace, determined by local information. The method, coded in FORTRAN, has demonstrated parallelism of greater than 20 on the Denelcor HEP machine. ITSM can also be used to replace current implicit treatments of problems involving diffusion and heat conduction. Four different approaches toward work distribution have been investigated and implemented for the one-dimensional code on the Denelcor HEP. They are ''self-scheduled'', an ASKFOR monitor, a ''queue of queues'' monitor, and a distributed ASKFOR monitor. The self-scheduled approach shows the lowest overhead but the poorest speedup. The distributed ASKFOR monitor shows the best speedup and the lowest execution times on the tested problems. 2 refs., 3 figs.
Structure and computation of two-dimensional incompressible extended MHD
Grasso, D; Abdelhamid, H M; Morrison, P J
2016-01-01
A comprehensive study of a reduced version of Lust's equations, the extended magnetohydrodynamic (XMHD) model obtained from the two-fluid theory for electrons and ions with the enforcement of quasineutrality, is given. Starting from the Hamiltonian structure of the fully three-dimensional theory, a Hamiltonian two-dimensional incompressible four-field model is derived. In this way energy conservation along with four families of Casimir invariants are naturally obtained. The construction facilitates various limits leading to the Hamiltonian forms of Hall, inertial, and ideal MHD, with their conserved energies and Casimir invariants. Basic linear theory of the four-field model is treated, and the growth rate for collisionless reconnection is obtained. Results from nonlinear simulations of collisionless tearing are presented and interpreted using, in particular normal fields, a product of the Hamiltonian theory that gives rise to simplified equations of motion.
Structure and computation of two-dimensional incompressible extended MHD
Grasso, D.; Tassi, E.; Abdelhamid, H. M.; Morrison, P. J.
2017-01-01
A comprehensive study of the extended magnetohydrodynamic model obtained from the two-fluid theory for electrons and ions with the enforcement of quasineutrality is given. Starting from the Hamiltonian structure of the fully three-dimensional theory, a Hamiltonian two-dimensional incompressible four-field model is derived. In this way, the energy conservation along with four families of Casimir invariants is naturally obtained. The construction facilitates various limits leading to the Hamiltonian forms of Hall, inertial, and ideal MHD, with their conserved energies and Casimir invariants. Basic linear theory of the four-field model is treated, and the growth rate for collisionless reconnection is obtained. Results from nonlinear simulations of collisionless tearing are presented and interpreted using, in particular, normal fields, a product of the Hamiltonian theory that gives rise to simplified equations of motion.
PLUTO code for computational Astrophysics: News and Developments
Tzeferacos, P.; Mignone, A.
2012-01-01
We present an overview on recent developments and functionalities available with the PLUTO code for astrophysical fluid dynamics. The recent extension of the code to a conservative finite difference formulation and high order spatial discretization of the compressible equations of magneto-hydrodynamics (MHD), complementary to its finite volume approach, allows for a highly accurate treatment of smooth flows, while avoiding loss of accuracy near smooth extrema and providing sharp non-oscillatory transitions at discontinuities. Among the novel features, we present alternative, fully explicit treatments to include non-ideal dissipative processes (namely viscosity, resistivity and anisotropic thermal conduction), that do not suffer from the usual timestep limitation of explicit time stepping. These methods, offsprings of the multistep Runge-Kutta family that use a Chebyshev polynomial recursion, are competitive substitutes of computationally expensive implicit schemes that involve sparse matrix inversion. Several multi-dimensional benchmarks and appli-cations assess the potential of PLUTO to efficiently handle many astrophysical problems.
Nakamura, Y.; Matsumoto, T.; Wakatani, M. [Kyoto Univ. (Japan). Plasma Physics Lab.; Galkin, S.A.; Drozdov, V.V.; Martynov, A.A.; Poshekhonov, Yu.Yu. [Keldysh Institute of Applied Mathematics, Moscow (Russian Federation); Ichiguchi, K. [National Institute for Fusion Science, Nagoya (Japan); Garcia, L. [Universidad Carlos III de Madrid (Spain); Carreras, B.A. [Oak Ridge National Lab., TN (United States)] [and others
1995-04-01
A particular configuration of the LHD stellarator with an unusually flat pressure profile has been chosen to be a test case for comparison of the MHD stability property predictions of different three-dimensional and averaged codes for the purpose of code comparison and validation. In particular, two relatively localized instabilities, the fastest growing modes with toroidal mode number n = 2 and n = 3 were studied using several different codes, with the good agreement that has been found providing justification for the use of any of them for equilibria of the type considered.
Characterizing Video Coding Computing in Conference Systems
Tuquerres, G.
2000-01-01
In this paper, a number of coding operations is provided for computing continuous data streams, in particular, video streams. A coding capability of the operations is expressed by a pyramidal structure in which coding processes and requirements of a distributed information system are represented. Th
Kolb, C.E.; Yousefian, V.; Wormhoudt, J.; Haimes, R.; Martinez-Sanchez, M.; Kerrebrock, J.L.
1978-01-30
Research has included theoretical modeling of important plasma chemical effects such as: conductivity reductions due to condensed slag/electron interactions; conductivity and generator efficiency reductions due to the formation of slag-related negative ion species; and the loss of alkali seed due to chemical combination with condensed slag. A summary of the major conclusions in each of these areas is presented. A major output of the modeling effort has been the development of an MHD plasma chemistry core flow model. This model has been formulated into a computer program designated the PACKAGE code (Plasma Analysis, Chemical Kinetics, And Generator Efficiency). The PACKAGE code is designed to calculate the effect of coal rank, ash percentage, ash composition, air preheat temperatures, equivalence ratio, and various generator channel parameters on the overall efficiency of open-cycle, coal-fired MHD generators. A complete description of the PACKAGE code and a preliminary version of the PACKAGE user's manual are included. A laboratory measurements program involving direct, mass spectrometric sampling of the positive and negative ions formed in a one atmosphere coal combustion plasma was also completed during the contract's initial phase. The relative ion concentrations formed in a plasma due to the methane augmented combustion of pulverized Montana Rosebud coal with potassium carbonate seed and preheated air are summarized. Positive ions measured include K/sup +/, KO/sup +/, Na/sup +/, Rb/sup +/, Cs/sup +/, and CsO/sup +/, while negative ions identified include PO/sub 3//sup -/, PO/sub 2//sup -/, BO/sub 2//sup -/, OH/sup -/, SH/sup -/, and probably HCrO/sub 3/, HMoO/sub 4//sup -/, and HWO/sub 3//sup -/. Comparison of the measurements with PACKAGE code predictions are presented. Preliminary design considerations for a mass spectrometric sampling probe capable of characterizing coal combustion plasmas from full scale combustors and flow trains are presented
Computer Code for Nanostructure Simulation
Filikhin, Igor; Vlahovic, Branislav
2009-01-01
Due to their small size, nanostructures can have stress and thermal gradients that are larger than any macroscopic analogue. These gradients can lead to specific regions that are susceptible to failure via processes such as plastic deformation by dislocation emission, chemical debonding, and interfacial alloying. A program has been developed that rigorously simulates and predicts optoelectronic properties of nanostructures of virtually any geometrical complexity and material composition. It can be used in simulations of energy level structure, wave functions, density of states of spatially configured phonon-coupled electrons, excitons in quantum dots, quantum rings, quantum ring complexes, and more. The code can be used to calculate stress distributions and thermal transport properties for a variety of nanostructures and interfaces, transport and scattering at nanoscale interfaces and surfaces under various stress states, and alloy compositional gradients. The code allows users to perform modeling of charge transport processes through quantum-dot (QD) arrays as functions of inter-dot distance, array order versus disorder, QD orientation, shape, size, and chemical composition for applications in photovoltaics and physical properties of QD-based biochemical sensors. The code can be used to study the hot exciton formation/relation dynamics in arrays of QDs of different shapes and sizes at different temperatures. It also can be used to understand the relation among the deposition parameters and inherent stresses, strain deformation, heat flow, and failure of nanostructures.
Cloud Computing for Complex Performance Codes.
Appel, Gordon John [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Hadgu, Teklu [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Klein, Brandon Thorin [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Miner, John Gifford [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)
2017-02-01
This report describes the use of cloud computing services for running complex public domain performance assessment problems. The work consisted of two phases: Phase 1 was to demonstrate complex codes, on several differently configured servers, could run and compute trivial small scale problems in a commercial cloud infrastructure. Phase 2 focused on proving non-trivial large scale problems could be computed in the commercial cloud environment. The cloud computing effort was successfully applied using codes of interest to the geohydrology and nuclear waste disposal modeling community.
Adaptive Mesh Computations with the PLUTO Code for Astrophysical Fluid Dynamics
Mignone, A.; Zanni, C.
2012-07-01
We present an overview of the current version of the PLUTO code for numerical simulations of astrophysical fluid flows over block-structured adaptive grids. The code preserves its modular framework for the solution of the classical or relativistic magnetohydrodynamics (MHD) equations while exploiting the distributed infrastructure of the Chombo library for multidimensional adaptive mesh refinement (AMR) parallel computations. Equations are evolved in time using an explicit second-order, dimensionally unsplit time stepping scheme based on a cell-centered discretization of the flow variables. Efficiency and robustness are shown through multidimensional benchmarks and applications to problems of astrophysical relevance.
Gender codes why women are leaving computing
Misa, Thomas J
2010-01-01
The computing profession is facing a serious gender crisis. Women are abandoning the computing field at an alarming rate. Fewer are entering the profession than anytime in the past twenty-five years, while too many are leaving the field in mid-career. With a maximum of insight and a minimum of jargon, Gender Codes explains the complex social and cultural processes at work in gender and computing today. Edited by Thomas Misa and featuring a Foreword by Linda Shafer, Chair of the IEEE Computer Society Press, this insightful collection of essays explores the persisting gender imbalance in computing and presents a clear course of action for turning things around.
Computer Security: is your code sane?
Stefan Lueders, Computer Security Team
2015-01-01
How many of us write code? Software? Programs? Scripts? How many of us are properly trained in this and how well do we do it? Do we write functional, clean and correct code, without flaws, bugs and vulnerabilities*? In other words: are our codes sane? Figuring out weaknesses is not that easy (see our quiz in an earlier Bulletin article). Therefore, in order to improve the sanity of your code, prevent common pit-falls, and avoid the bugs and vulnerabilities that can crash your code, or – worse – that can be misused and exploited by attackers, the CERN Computer Security team has reviewed its recommendations for checking the security compliance of your code. “Static Code Analysers” are stand-alone programs that can be run on top of your software stack, regardless of whether it uses Java, C/C++, Perl, PHP, Python, etc. These analysers identify weaknesses and inconsistencies including: employing undeclared variables; expressions resu...
Incompressible face seals: Computer code IFACE
Artiles, Antonio
1994-01-01
Capabilities of the computer code IFACE are given in viewgraph format. These include: two dimensional, incompressible, isoviscous flow; rotation of both rotor and housing; roughness in both rotor and housing; arbitrary film thickness distribution, including steps, pockets, and tapers; three degrees of freedom; dynamic coefficients; prescribed force and moments; pocket pressures or orifice size; turbulence, Couette and Poiseuille flow; cavitation; and inertia pressure drops at inlets to film.
Computing Challenges in Coded Mask Imaging
Skinner, Gerald
2009-01-01
This slide presaentation reviews the complications and challenges in developing computer systems for Coded Mask Imaging telescopes. The coded mask technique is used when there is no other way to create the telescope, (i.e., when there are wide fields of view, high energies for focusing or low energies for the Compton/Tracker Techniques and very good angular resolution.) The coded mask telescope is described, and the mask is reviewed. The coded Masks for the INTErnational Gamma-Ray Astrophysics Laboratory (INTEGRAL) instruments are shown, and a chart showing the types of position sensitive detectors used for the coded mask telescopes is also reviewed. Slides describe the mechanism of recovering an image from the masked pattern. The correlation with the mask pattern is described. The Matrix approach is reviewed, and other approaches to image reconstruction are described. Included in the presentation is a review of the Energetic X-ray Imaging Survey Telescope (EXIST) / High Energy Telescope (HET), with information about the mission, the operation of the telescope, comparison of the EXIST/HET with the SWIFT/BAT and details of the design of the EXIST/HET.
New developments in the Saphire computer codes
Russell, K.D.; Wood, S.T.; Kvarfordt, K.J. [Idaho Engineering Lab., Idaho Falls, ID (United States)] [and others
1996-03-01
The Systems Analysis Programs for Hands-on Integrated Reliability Evaluations (SAPHIRE) refers to a suite of computer programs that were developed to create and analyze a probabilistic risk assessment (PRA) of a nuclear power plant. Many recent enhancements to this suite of codes have been made. This presentation will provide an overview of these features and capabilities. The presentation will include a discussion of the new GEM module. This module greatly reduces and simplifies the work necessary to use the SAPHIRE code in event assessment applications. An overview of the features provided in the new Windows version will also be provided. This version is a full Windows 32-bit implementation and offers many new and exciting features. [A separate computer demonstration was held to allow interested participants to get a preview of these features.] The new capabilities that have been added since version 5.0 will be covered. Some of these major new features include the ability to store an unlimited number of basic events, gates, systems, sequences, etc.; the addition of improved reporting capabilities to allow the user to generate and {open_quotes}scroll{close_quotes} through custom reports; the addition of multi-variable importance measures; and the simplification of the user interface. Although originally designed as a PRA Level 1 suite of codes, capabilities have recently been added to SAPHIRE to allow the user to apply the code in Level 2 analyses. These features will be discussed in detail during the presentation. The modifications and capabilities added to this version of SAPHIRE significantly extend the code in many important areas. Together, these extensions represent a major step forward in PC-based risk analysis tools. This presentation provides a current up-to-date status of these important PRA analysis tools.
A New Code for Numerical Simulation of MHD Astrophysical Flows With Chemistry
Kulikov, Igor; Protasov, Viktor
2016-01-01
The new code for numerical simulation of magnetic hydrodynamical astrophysical flows with consideration of chemical reactions is given in the paper. At the heart of the code - the new original low-dissipation numerical method based on a combination of operator splitting approach and piecewise-parabolic method on the local stencil. The details of the numerical method are described; the main tests and the scheme of parallel implementation are shown. The chemodynamics of the hydrogen while the turbulent formation of molecular clouds is modeled.
SALE: Safeguards Analytical Laboratory Evaluation computer code
Carroll, D.J.; Bush, W.J.; Dolan, C.A.
1976-09-01
The Safeguards Analytical Laboratory Evaluation (SALE) program implements an industry-wide quality control and evaluation system aimed at identifying and reducing analytical chemical measurement errors. Samples of well-characterized materials are distributed to laboratory participants at periodic intervals for determination of uranium or plutonium concentration and isotopic distributions. The results of these determinations are statistically-evaluated, and each participant is informed of the accuracy and precision of his results in a timely manner. The SALE computer code which produces the report is designed to facilitate rapid transmission of this information in order that meaningful quality control will be provided. Various statistical techniques comprise the output of the SALE computer code. Assuming an unbalanced nested design, an analysis of variance is performed in subroutine NEST resulting in a test of significance for time and analyst effects. A trend test is performed in subroutine TREND. Microfilm plots are obtained from subroutine CUMPLT. Within-laboratory standard deviations are calculated in the main program or subroutine VAREST, and between-laboratory standard deviations are calculated in SBLV. Other statistical tests are also performed. Up to 1,500 pieces of data for each nuclear material sampled by 75 (or fewer) laboratories may be analyzed with this code. The input deck necessary to run the program is shown, and input parameters are discussed in detail. Printed output and microfilm plot output are described. Output from a typical SALE run is included as a sample problem.
Neutron spectrum unfolding using computer code SAIPS
Karim, S
1999-01-01
The main objective of this project was to study the neutron energy spectrum at rabbit station-1 in Pakistan Research Reactor (PARR-I). To do so, multiple foils activation method was used to get the saturated activities. The computer code SAIPS was used to unfold the neutron spectra from the measured reaction rates. Of the three built in codes in SAIPS, only SANDI and WINDOWS were used. Contribution of thermal part of the spectra was observed to be higher than the fast one. It was found that the WINDOWS gave smooth spectra while SANDII spectra have violet oscillations in the resonance region. The uncertainties in the WINDOWS results are higher than those of SANDII. The results show reasonable agreement with the published results.
From Coding to Computational Thinking and Back
DePryck, K.
2016-01-01
Presentation of Dr. Koen DePryck in the Computational Thinking Session in TEEM 2016 Conference, held in the University of Salamanca (Spain), Nov 2-4, 2016. Introducing coding in the curriculum at an early age is considered a long-term investment in bridging the skills gap between the technology demands of the labour market and the availability of people to fill them. The keys to success include moving from mere literacy to active control – not only at the level of learners but also ...
Spiking network simulation code for petascale computers
Kunkel, Susanne; Schmidt, Maximilian; Eppler, Jochen M.; Plesser, Hans E.; Masumoto, Gen; Igarashi, Jun; Ishii, Shin; Fukai, Tomoki; Morrison, Abigail; Diesmann, Markus; Helias, Moritz
2014-01-01
Brain-scale networks exhibit a breathtaking heterogeneity in the dynamical properties and parameters of their constituents. At cellular resolution, the entities of theory are neurons and synapses and over the past decade researchers have learned to manage the heterogeneity of neurons and synapses with efficient data structures. Already early parallel simulation codes stored synapses in a distributed fashion such that a synapse solely consumes memory on the compute node harboring the target neuron. As petaflop computers with some 100,000 nodes become increasingly available for neuroscience, new challenges arise for neuronal network simulation software: Each neuron contacts on the order of 10,000 other neurons and thus has targets only on a fraction of all compute nodes; furthermore, for any given source neuron, at most a single synapse is typically created on any compute node. From the viewpoint of an individual compute node, the heterogeneity in the synaptic target lists thus collapses along two dimensions: the dimension of the types of synapses and the dimension of the number of synapses of a given type. Here we present a data structure taking advantage of this double collapse using metaprogramming techniques. After introducing the relevant scaling scenario for brain-scale simulations, we quantitatively discuss the performance on two supercomputers. We show that the novel architecture scales to the largest petascale supercomputers available today. PMID:25346682
Spiking network simulation code for petascale computers
Susanne eKunkel
2014-10-01
Full Text Available Brain-scale networks exhibit a breathtaking heterogeneity in the dynamical properties and parameters of their constituents. At cellular resolution, the entities of theory are neurons and synapses and over the past decade researchers have learned to manage the heterogeneity of neurons and synapses with efficient data structures. Already early parallel simulation codes stored synapses in a distributed fashion such that a synapse solely consumes memory on the compute node harboring the target neuron. As petaflop computers with some 100,000 nodes become increasingly available for neuroscience, new challenges arise for neuronal network simulation software: Each neuron contacts on the order of 10,000 other neurons and thus has targets only on a fraction of all compute nodes; furthermore, for any given source neuron, at most a single synapse is typically created on any compute node. From the viewpoint of an individual compute node, the heterogeneity in the synaptic target lists thus collapses along two dimensions: the dimension of the types of synapses and the dimension of the number of synapses of a given type. Here we present a data structure taking advantage of this double collapse using metaprogramming techniques. After introducing the relevant scaling scenario for brain-scale simulations, we quantitatively discuss the performance on two supercomputers. We show that the novel architecture scales to the largest petascale supercomputers available today.
Boss, A P
2014-01-01
We present the results of a large suite of three-dimensional (3D) models of the collapse of magnetic molecular cloud cores using the adaptive mesh refinement (AMR) code Enzo2.2 in the ideal magnetohydrodynamics (MHD) approximation. The cloud cores are initially either prolate or oblate, centrally condensed clouds with masses of 1.73 or 2.73 $M_\\odot$, respectively. The radial density profiles are Gaussian, with central densities 20 times higher than boundary densities. A barotropic equation of state is used to represent the transition from low density, isothermal phases, to high density, optically thick phases. The initial magnetic field strength ranges from 6.3 to 100 $\\mu$G, corresponding to clouds that are strongly to marginally supercritical, respectively, in terms of the mass to magnetic flux ratio. The magnetic field is initially uniform and aligned with the clouds' rotation axes, with initial ratios of rotational to gravitational energy ranging from $10^{-4}$ to 0.1. Two significantly different outcome...
MHD simulations on an unstructured mesh
Strauss, H.R. [New York Univ., NY (United States); Park, W.; Belova, E.; Fu, G.Y. [Princeton Univ., NJ (United States). Plasma Physics Lab.; Longcope, D.W. [Univ. of Montana, Missoula, MT (United States); Sugiyama, L.E. [Massachusetts Inst. of Tech., Cambridge, MA (United States)
1998-12-31
Two reasons for using an unstructured computational mesh are adaptivity, and alignment with arbitrarily shaped boundaries. Two codes which use finite element discretization on an unstructured mesh are described. FEM3D solves 2D and 3D RMHD using an adaptive grid. MH3D++, which incorporates methods of FEM3D into the MH3D generalized MHD code, can be used with shaped boundaries, which might be 3D.
Superimposed Code Theorectic Analysis of DNA Codes and DNA Computing
2010-03-01
Bounds for DNA Codes Based on Fibonacci Ensembles of DNA Sequences ”, 2008 IEEE Proceedings of International Symposium on Information Theory, pp. 2292...5, June 2008, pp. 525-34. 32 28. A. Macula, et al., “Random Coding Bounds for DNA Codes Based on Fibonacci Ensembles of DNA Sequences ”, 2008...combinatorial method of bio-memory design and detection that encodes item or process information as numerical sequences represented in DNA. ComDMem is a
Development of probabilistic internal dosimetry computer code
Noh, Siwan; Kwon, Tae-Eun; Lee, Jai-Ki
2017-02-01
Internal radiation dose assessment involves biokinetic models, the corresponding parameters, measured data, and many assumptions. Every component considered in the internal dose assessment has its own uncertainty, which is propagated in the intake activity and internal dose estimates. For research or scientific purposes, and for retrospective dose reconstruction for accident scenarios occurring in workplaces having a large quantity of unsealed radionuclides, such as nuclear power plants, nuclear fuel cycle facilities, and facilities in which nuclear medicine is practiced, a quantitative uncertainty assessment of the internal dose is often required. However, no calculation tools or computer codes that incorporate all the relevant processes and their corresponding uncertainties, i.e., from the measured data to the committed dose, are available. Thus, the objective of the present study is to develop an integrated probabilistic internal-dose-assessment computer code. First, the uncertainty components in internal dosimetry are identified, and quantitative uncertainty data are collected. Then, an uncertainty database is established for each component. In order to propagate these uncertainties in an internal dose assessment, a probabilistic internal-dose-assessment system that employs the Bayesian and Monte Carlo methods. Based on the developed system, we developed a probabilistic internal-dose-assessment code by using MATLAB so as to estimate the dose distributions from the measured data with uncertainty. Using the developed code, we calculated the internal dose distribution and statistical values ( e.g. the 2.5th, 5th, median, 95th, and 97.5th percentiles) for three sample scenarios. On the basis of the distributions, we performed a sensitivity analysis to determine the influence of each component on the resulting dose in order to identify the major component of the uncertainty in a bioassay. The results of this study can be applied to various situations. In cases of
ICAN Computer Code Adapted for Building Materials
Murthy, Pappu L. N.
1997-01-01
The NASA Lewis Research Center has been involved in developing composite micromechanics and macromechanics theories over the last three decades. These activities have resulted in several composite mechanics theories and structural analysis codes whose applications range from material behavior design and analysis to structural component response. One of these computer codes, the Integrated Composite Analyzer (ICAN), is designed primarily to address issues related to designing polymer matrix composites and predicting their properties - including hygral, thermal, and mechanical load effects. Recently, under a cost-sharing cooperative agreement with a Fortune 500 corporation, Master Builders Inc., ICAN was adapted to analyze building materials. The high costs and technical difficulties involved with the fabrication of continuous-fiber-reinforced composites sometimes limit their use. Particulate-reinforced composites can be thought of as a viable alternative. They are as easily processed to near-net shape as monolithic materials, yet have the improved stiffness, strength, and fracture toughness that is characteristic of continuous-fiber-reinforced composites. For example, particlereinforced metal-matrix composites show great potential for a variety of automotive applications, such as disk brake rotors, connecting rods, cylinder liners, and other hightemperature applications. Building materials, such as concrete, can be thought of as one of the oldest materials in this category of multiphase, particle-reinforced materials. The adaptation of ICAN to analyze particle-reinforced composite materials involved the development of new micromechanics-based theories. A derivative of the ICAN code, ICAN/PART, was developed and delivered to Master Builders Inc. as a part of the cooperative activity.
A surface code quantum computer in silicon.
Hill, Charles D; Peretz, Eldad; Hile, Samuel J; House, Matthew G; Fuechsle, Martin; Rogge, Sven; Simmons, Michelle Y; Hollenberg, Lloyd C L
2015-10-01
The exceptionally long quantum coherence times of phosphorus donor nuclear spin qubits in silicon, coupled with the proven scalability of silicon-based nano-electronics, make them attractive candidates for large-scale quantum computing. However, the high threshold of topological quantum error correction can only be captured in a two-dimensional array of qubits operating synchronously and in parallel-posing formidable fabrication and control challenges. We present an architecture that addresses these problems through a novel shared-control paradigm that is particularly suited to the natural uniformity of the phosphorus donor nuclear spin qubit states and electronic confinement. The architecture comprises a two-dimensional lattice of donor qubits sandwiched between two vertically separated control layers forming a mutually perpendicular crisscross gate array. Shared-control lines facilitate loading/unloading of single electrons to specific donors, thereby activating multiple qubits in parallel across the array on which the required operations for surface code quantum error correction are carried out by global spin control. The complexities of independent qubit control, wave function engineering, and ad hoc quantum interconnects are explicitly avoided. With many of the basic elements of fabrication and control based on demonstrated techniques and with simulated quantum operation below the surface code error threshold, the architecture represents a new pathway for large-scale quantum information processing in silicon and potentially in other qubit systems where uniformity can be exploited.
Srinath Vadlamani; Scott Kruger; Travis Austin
2008-06-19
Extended magnetohydrodynamic (MHD) codes are used to model the large, slow-growing instabilities that are projected to limit the performance of International Thermonuclear Experimental Reactor (ITER). The multiscale nature of the extended MHD equations requires an implicit approach. The current linear solvers needed for the implicit algorithm scale poorly because the resultant matrices are so ill-conditioned. A new solver is needed, especially one that scales to the petascale. The most successful scalable parallel processor solvers to date are multigrid solvers. Applying multigrid techniques to a set of equations whose fundamental modes are dispersive waves is a promising solution to CEMM problems. For the Phase 1, we implemented multigrid preconditioners from the HYPRE project of the Center for Applied Scientific Computing at LLNL via PETSc of the DOE SciDAC TOPS for the real matrix systems of the extended MHD code NIMROD which is a one of the primary modeling codes of the OFES-funded Center for Extended Magnetohydrodynamic Modeling (CEMM) SciDAC. We implemented the multigrid solvers on the fusion test problem that allows for real matrix systems with success, and in the process learned about the details of NIMROD data structures and the difficulties of inverting NIMROD operators. The further success of this project will allow for efficient usage of future petascale computers at the National Leadership Facilities: Oak Ridge National Laboratory, Argonne National Laboratory, and National Energy Research Scientific Computing Center. The project will be a collaborative effort between computational plasma physicists and applied mathematicians at Tech-X Corporation, applied mathematicians Front Range Scientific Computations, Inc. (who are collaborators on the HYPRE project), and other computational plasma physicists involved with the CEMM project.
Eslinger, Paul W.; Aaberg, Rosanne L.; Lopresti, Charles A.; Miley, Terri B.; Nichols, William E.; Strenge, Dennis L.
2004-09-14
This document contains detailed user instructions for a suite of utility codes developed for Rev. 1 of the Systems Assessment Capability. The suite of computer codes for Rev. 1 of Systems Assessment Capability performs many functions.
40 CFR 194.23 - Models and computer codes.
2010-07-01
... 40 Protection of Environment 24 2010-07-01 2010-07-01 false Models and computer codes. 194.23... General Requirements § 194.23 Models and computer codes. (a) Any compliance application shall include: (1... obtain stable solutions; (iv) Computer models accurately implement the numerical models; i.e.,...
On computations for thermal radiation in MHD channel flow with heat and mass transfer.
Hayat, T; Awais, M; Alsaedi, A; Safdar, Ambreen
2014-01-01
This study examines the simultaneous effects of heat and mass transfer on the three-dimensional boundary layer flow of viscous fluid between two infinite parallel plates. Magnetohydrodynamic (MHD) and thermal radiation effects are present. The governing problems are first modeled and then solved by homotopy analysis method (HAM). Influence of several embedded parameters on the velocity, concentration and temperature fields are described.
Alfvén eigenmode evolution computed with the VENUS and KINX codes for the ITER baseline scenario
Isaev, M. Yu., E-mail: isaev-my@nrcki.ru [National Research Center Kurchatov Institute (Russian Federation); Medvedev, S. Yu. [Russian Academy of Sciences, Keldysh Institute (Russian Federation); Cooper, W. A. [Suisse Plasma Centre (Switzerland)
2017-02-15
A new application of the VENUS code is described, which computes alpha particle orbits in the perturbed electromagnetic fields and its resonant interaction with the toroidal Alfvén eigenmodes (TAEs) for the ITER device. The ITER baseline scenario with Q = 10 and the plasma toroidal current of 15 MA is considered as the most important and relevant for the International Tokamak Physics Activity group on energetic particles (ITPA-EP). For this scenario, typical unstable TAE-modes with the toroidal index n = 20 have been predicted that are localized in the plasma core near the surface with safety factor q = 1. The spatial structure of ballooning and antiballooning modes has been computed with the ideal MHD code KINX. The linear growth rates and the saturation levels taking into account the damping effects and the different mode frequencies have been calculated with the VENUS code for both ballooning and antiballooning TAE-modes.
Alfvén eigenmode evolution computed with the VENUS and KINX codes for the ITER baseline scenario
Isaev, M. Yu.; Medvedev, S. Yu.; Cooper, W. A.
2017-02-01
A new application of the VENUS code is described, which computes alpha particle orbits in the perturbed electromagnetic fields and its resonant interaction with the toroidal Alfvén eigenmodes (TAEs) for the ITER device. The ITER baseline scenario with Q = 10 and the plasma toroidal current of 15 MA is considered as the most important and relevant for the International Tokamak Physics Activity group on energetic particles (ITPA-EP). For this scenario, typical unstable TAE-modes with the toroidal index n = 20 have been predicted that are localized in the plasma core near the surface with safety factor q = 1. The spatial structure of ballooning and antiballooning modes has been computed with the ideal MHD code KINX. The linear growth rates and the saturation levels taking into account the damping effects and the different mode frequencies have been calculated with the VENUS code for both ballooning and antiballooning TAE-modes.
Hanford Meteorological Station computer codes: Volume 6, The SFC computer code
Andrews, G.L.; Buck, J.W.
1987-11-01
Each hour the Hanford Meteorological Station (HMS), operated by Pacific Northwest Laboratory, records and archives weather observations. Hourly surface weather observations consist of weather phenomena such as cloud type and coverage; dry bulb, wet bulb, and dew point temperatures; relative humidity; atmospheric pressure; and wind speed and direction. The SFC computer code is used to archive those weather observations and apply quality assurance checks to the data. This code accesses an input file, which contains the previous archive's date and hour and an output file, which contains surface observations for the current day. As part of the program, a data entry form consisting of 24 fields must be filled in. The information on the form is appended to the daily file, which provides an archive for the hourly surface observations.
Private Computing and Mobile Code Systems
Cartrysse, K.
2005-01-01
This thesis' objective is to provide privacy to mobile code. A practical example of mobile code is a mobile software agent that performs a task on behalf of its user. The agent travels over the network and is executed at different locations of which beforehand it is not known whether or not these ca
Reducing Computational Overhead of Network Coding with Intrinsic Information Conveying
Heide, Janus; Zhang, Qi; Pedersen, Morten V.;
is RLNC (Random Linear Network Coding) and the goal is to reduce the amount of coding operations both at the coding and decoding node, and at the same time remove the need for dedicated signaling messages. In a traditional RLNC system, coding operation takes up significant computational resources and adds......This paper investigated the possibility of intrinsic information conveying in network coding systems. The information is embedded into the coding vector by constructing the vector based on a set of predefined rules. This information can subsequently be retrieved by any receiver. The starting point...
CASTOR: Normal-mode analysis of resistive MHD plasmas
Kerner, W.; Goedbloed, J. P.; Huysmans, G. T. A.; Poedts, S.; Schwarz, E.
1998-01-01
The CASTOR (complex Alfven spectrum of toroidal plasmas) code computes the entire spectrum of normal-modes in resistive MHD for general tokamak configurations. The applied Galerkin method, in conjunction with a Fourier finite-element discretisation, leads to a large scale eigenvalue problem A (x)
Computer codes for birds of North America
US Fish and Wildlife Service, Department of the Interior — Purpose of paper was to provide a more useful way to provide codes for all North American species, thus making the list useful for virtually all projects concerning...
Maget, P.; Huysmans, G. T. A.; Lütjens, H.; Ottaviani, M.; Moreau, Ph; Ségui, J.-L.
2009-06-01
Attempts to run non-inductive plasma discharges on Tore Supra sometimes fail due to the triggering of magneto-hydro-dynamic (MHD) instabilities that saturate at a large amplitude, producing degraded confinement and loss of wave driven fast electrons (the so-called MHD regime (Maget et al 2005 Nucl. Fusion 45 69-80)). In this paper we investigate the transition to this soft (in the sense of non-disruptive) MHD limit from experimental observations, and compare it with non-linear code predictions. Such a comparison suggests that different non-linear regimes, with periodic relaxations or saturation, are correctly understood. However, successful non-inductive discharges without detectable magnetic island at q = 2 cannot be reproduced if realistic transport coefficients are used in the computation. Additional physics seems mandatory for explaining these discharges, such as diamagnetic effects, that could also justify cases of abrupt transition to the MHD regime.
Batet, L.; Mas de les Valls, E.; Sedano, L. A.
2012-07-01
In the context of regenerating sheaths for fusion reactors, the CFD simulations of liquid metal channels (ML) are essential to know the phenomenology and obtain relevant information for design as: ML thermal gain, to know the thermal efficiency of the component, existence of hot spots, to define the materials to use, existence of flow inversion, etc. Apart from design parameters there are others, bridge parameter, required as inputs into system code. In this work shown GREENER/T4F capabilities for obtaining both parameters with a CFD tool based on open source OpenFOAM.
PORPST: A statistical postprocessor for the PORMC computer code
Eslinger, P.W.; Didier, B.T. (Pacific Northwest Lab., Richland, WA (United States))
1991-06-01
This report describes the theory underlying the PORPST code and gives details for using the code. The PORPST code is designed to do statistical postprocessing on files written by the PORMC computer code. The data written by PORMC are summarized in terms of means, variances, standard deviations, or statistical distributions. In addition, the PORPST code provides for plotting of the results, either internal to the code or through use of the CONTOUR3 postprocessor. Section 2.0 discusses the mathematical basis of the code, and Section 3.0 discusses the code structure. Section 4.0 describes the free-format point command language. Section 5.0 describes in detail the commands to run the program. Section 6.0 provides an example program run, and Section 7.0 provides the references. 11 refs., 1 fig., 17 tabs.
Optimization of KINETICS Chemical Computation Code
Donastorg, Cristina
2012-01-01
NASA JPL has been creating a code in FORTRAN called KINETICS to model the chemistry of planetary atmospheres. Recently there has been an effort to introduce Message Passing Interface (MPI) into the code so as to cut down the run time of the program. There has been some implementation of MPI into KINETICS; however, the code could still be more efficient than it currently is. One way to increase efficiency is to send only certain variables to all the processes when an MPI subroutine is called and to gather only certain variables when the subroutine is finished. Therefore, all the variables that are used in three of the main subroutines needed to be investigated. Because of the sheer amount of code that there is to comb through this task was given as a ten-week project. I have been able to create flowcharts outlining the subroutines, common blocks, and functions used within the three main subroutines. From these flowcharts I created tables outlining the variables used in each block and important information about each. All this information will be used to determine how to run MPI in KINETICS in the most efficient way possible.
Codes of Ethics for Computing at Russian Institutions and Universities.
Pourciau, Lester J.; Spain, Victoria, Ed.
1997-01-01
To determine the degree to which Russian institutions and universities have formulated and promulgated codes of ethics or policies for acceptable computer use, the author examined Russian institution and university home pages. Lists home pages examined, 10 commandments for computer ethics from the Computer Ethics Institute, and a policy statement…
Continuous Materiality: Through a Hierarchy of Computational Codes
Jichen Zhu
2008-01-01
Full Text Available The legacy of Cartesian dualism inherent in linguistic theory deeply influences current views on the relation between natural language, computer code, and the physical world. However, the oversimplified distinction between mind and body falls short of capturing the complex interaction between the material and the immaterial. In this paper, we posit a hierarchy of codes to delineate a wide spectrum of continuous materiality. Our research suggests that diagrams in architecture provide a valuable analog for approaching computer code in emergent digital systems. After commenting on ways that Cartesian dualism continues to haunt discussions of code, we turn our attention to diagrams and design morphology. Finally we notice the implications a material understanding of code bears for further research on the relation between human cognition and digital code. Our discussion concludes by noticing several areas that we have projected for ongoing research.
Structural Computer Code Evaluation. Volume I
1976-11-01
Rivlin model for large strains. Other exanmples are given in Reference 5. Hypoelasticity A hypoelastic material is one in which the components of...remains is the application of these codes to specific rocket nozzle problems and the evaluation of their capabilities to model modern nozzle mraterial...behavior. Further work may also require the development of appropriate material property data or new material models to adequately characterize these
Quantum computation with Turaev-Viro codes
Koenig, Robert; Reichardt, Ben W
2010-01-01
The Turaev-Viro invariant for a closed 3-manifold is defined as the contraction of a certain tensor network. The tensors correspond to tetrahedra in a triangulation of the manifold, with values determined by a fixed spherical category. For a manifold with boundary, the tensor network has free indices that can be associated to qudits, and its contraction gives the coefficients of a quantum error-correcting code. The code has local stabilizers determined by Levin and Wen. For example, applied to the genus-one handlebody using the Z_2 category, this construction yields the well-known toric code. For other categories, such as the Fibonacci category, the construction realizes a non-abelian anyon model over a discrete lattice. By studying braid group representations acting on equivalence classes of colored ribbon graphs embedded in a punctured sphere, we identify the anyons, and give a simple recipe for mapping fusion basis states of the doubled category to ribbon graphs. We explain how suitable initial states can ...
Lattice Boltzmann method fundamentals and engineering applications with computer codes
Mohamad, A A
2014-01-01
Introducing the Lattice Boltzmann Method in a readable manner, this book provides detailed examples with complete computer codes. It avoids the most complicated mathematics and physics without scarifying the basic fundamentals of the method.
Tuning complex computer code to data
Cox, D.; Park, J.S.; Sacks, J.; Singer, C.
1992-01-01
The problem of estimating parameters in a complex computer simulator of a nuclear fusion reactor from an experimental database is treated. Practical limitations do not permit a standard statistical analysis using nonlinear regression methodology. The assumption that the function giving the true theoretical predictions is a realization of a Gaussian stochastic process provides a statistical method for combining information from relatively few computer runs with information from the experimental database and making inferences on the parameters.
Computer aided power flow software engineering and code generation
Bacher, R. [Swiss Federal Inst. of Tech., Zuerich (Switzerland)
1996-02-01
In this paper a software engineering concept is described which permits the automatic solution of a non-linear set of network equations. The power flow equation set can be seen as a defined subset of a network equation set. The automated solution process is the numerical Newton-Raphson solution process of the power flow equations where the key code parts are the numeric mismatch and the numeric Jacobian term computation. It is shown that both the Jacobian and the mismatch term source code can be automatically generated in a conventional language such as Fortran or C. Thereby one starts from a high level, symbolic language with automatic differentiation and code generation facilities. As a result of this software engineering process an efficient, very high quality newton-Raphson solution code is generated which allows easier implementation of network equation model enhancements and easier code maintenance as compared to hand-coded Fortran or C code.
Computer aided power flow software engineering and code generation
Bacher, R. [Swiss Federal Inst. of Tech., Zuerich (Switzerland)
1995-12-31
In this paper a software engineering concept is described which permits the automatic solution of a non-linear set of network equations. The power flow equation set can be seen as a defined subset of a network equation set. The automated solution process is the numerical Newton-Raphson solution process of the power flow equations where the key code parts are the numeric mismatch and the numeric Jacobian term computation. It is shown that both the Jacobian and the mismatch term source code can be automatically generated in a conventional language such as Fortran or C. Thereby one starts from a high level, symbolic language with automatic differentiation and code generation facilities. As a result of this software engineering process an efficient, very high quality Newton-Raphson solution code is generated which allows easier implementation of network equation model enhancements and easier code maintenance as compared to hand-coded Fortran or C code.
APC: A New Code for Atmospheric Polarization Computations
Korkin, Sergey V.; Lyapustin, Alexei I.; Rozanov, Vladimir V.
2014-01-01
A new polarized radiative transfer code Atmospheric Polarization Computations (APC) is described. The code is based on separation of the diffuse light field into anisotropic and smooth (regular) parts. The anisotropic part is computed analytically. The smooth regular part is computed numerically using the discrete ordinates method. Vertical stratification of the atmosphere, common types of bidirectional surface reflection and scattering by spherical particles or spheroids are included. A particular consideration is given to computation of the bidirectional polarization distribution function (BPDF) of the waved ocean surface.
Proceedings of the workshop on nonlinear MHD and extended MHD
NONE
1998-12-01
Nonlinear MHD simulations have proven their value in interpreting experimental results over the years. As magnetic fusion experiments reach higher performance regimes, more sophisticated experimental diagnostics coupled with ever expanding computer capabilities have increased both the need for and the feasibility of nonlinear global simulations using models more realistic than regular ideal and resistive MHD. Such extended-MHD nonlinear simulations have already begun to produce useful results. These studies are expected to lead to ever more comprehensive simulation models in the future and to play a vital role in fully understanding fusion plasmas. Topics include the following: (1) current state of nonlinear MHD and extended-MHD simulations; (2) comparisons to experimental data; (3) discussions between experimentalists and theorists; (4) /equations for extended-MHD models, kinetic-based closures; and (5) paths toward more comprehensive simulation models, etc. Selected papers have been indexed separately for inclusion in the Energy Science and Technology Database.
A three dimensional MHD model of the earth's magnetosphere
Wu, C. C.; Walker, R. J.; Dawson, J. M.
1981-01-01
The results of a global MHD calculation of the steady state solar wind interaction with a dipole magnetic field are presented. The computer code used, being much faster than previous codes, makes it possible to increase the number of grid points in the system by an order of magnitude. The resulting model qualitatively reproduces many of the observed features of the quiet time magnetosphere including the bow shock, magnetopause, and plasma sheet.
Neutron noise computation using panda deterministic code
Humbert, Ph. [CEA Bruyeres le Chatel (France)
2003-07-01
PANDA is a general purpose discrete ordinates neutron transport code with deterministic and non deterministic applications. In this paper we consider the adaptation of PANDA to stochastic neutron counting problems. More specifically we consider the first two moments of the count number probability distribution. In a first part we will recall the equations for the single neutron and source induced count number moments with the corresponding expression for the excess of relative variance or Feynman function. In a second part we discuss the numerical solution of these inhomogeneous adjoint time dependent transport coupled equations with discrete ordinate methods. Finally, numerical applications are presented in the third part. (author)
Leenaarts, J; Hansteen, V; van der Voort, L Rouppe
2009-01-01
Interpretation of imagery of the solar chromosphere in the widely used \\CaIIIR infrared line is hampered by its complex, three-dimensional and non-LTE formation. Forward modelling is required to aid understanding. We use a 3D non-LTE radiative transfer code to compute synthetic \\CaIIIR images from a radiation-MHD simulation of the solar atmosphere spanning from the convection zone to the corona. We compare the simulation with observations obtained with the CRISP filter at the Swedish 1--m Solar Telescope. We find that the simulation reproduces dark patches in the blue line wing caused by Doppler shifts, brightenings in the line core caused by upward-propagating shocks and thin dark elongated structures in the line core that form the interface between upward and downward gas motion in the chromosphere. The synthetic line core is narrower than the observed one, indicating that the sun exhibits both more vigorous large-scale dynamics as well as small scale motions that are not resolved within the simulation, pre...
Computer code for intraply hybrid composite design
Chamis, C. C.; Sinclair, J. H.
1981-01-01
A computer program has been developed and is described herein for intraply hybrid composite design (INHYD). The program includes several composite micromechanics theories, intraply hybrid composite theories and a hygrothermomechanical theory. These theories provide INHYD with considerable flexibility and capability which the user can exercise through several available options. Key features and capabilities of INHYD are illustrated through selected samples.
MHD and heat transfer benchmark problems for liquid metal flow in rectangular ducts. Final paper
Sidorenkov, S.I. [D.V. Efremov Scientific Research Inst. of Electrophysical Apparatus, St. Petersburg (Russian Federation); Hua, T.Q. [Argonne National Lab., IL (United States); Araseki, Hideo [Central Research Inst. of Electric Power Industry, Tokyo (Japan)
1994-07-01
Liquid metal cooling systems of a self-cooled blanket in a tokamak reactor will likely include channels of rectangular cross section where liquid metal is circulated in the presence of strong magnetic fields. MHD pressure drop, velocity distribution and heat transfer characteristics are important issues in the engineering design considerations. Computer codes for the reliable solution of three-dimensional MHD flow problems are needed for fusion relevant conditions. This paper describes four benchmark problems to validate magnetohydrodynamic (MHD) and heat transfer computer codes. The problems include rectangular duct geometry with uniform and nonuniform magnetic fields, with and without surface heat flux, and various rectangular cross sections. Two of the problems are based on experiments. Participants in this benchmarking activity come from three countries: The Russian Federation, The United States, and Japan. The solution methods to the problems are described. Results from the different computer codes are presented and compared.
Computer vision cracks the leaf code.
Wilf, Peter; Zhang, Shengping; Chikkerur, Sharat; Little, Stefan A; Wing, Scott L; Serre, Thomas
2016-03-22
Understanding the extremely variable, complex shape and venation characters of angiosperm leaves is one of the most challenging problems in botany. Machine learning offers opportunities to analyze large numbers of specimens, to discover novel leaf features of angiosperm clades that may have phylogenetic significance, and to use those characters to classify unknowns. Previous computer vision approaches have primarily focused on leaf identification at the species level. It remains an open question whether learning and classification are possible among major evolutionary groups such as families and orders, which usually contain hundreds to thousands of species each and exhibit many times the foliar variation of individual species. Here, we tested whether a computer vision algorithm could use a database of 7,597 leaf images from 2,001 genera to learn features of botanical families and orders, then classify novel images. The images are of cleared leaves, specimens that are chemically bleached, then stained to reveal venation. Machine learning was used to learn a codebook of visual elements representing leaf shape and venation patterns. The resulting automated system learned to classify images into families and orders with a success rate many times greater than chance. Of direct botanical interest, the responses of diagnostic features can be visualized on leaf images as heat maps, which are likely to prompt recognition and evolutionary interpretation of a wealth of novel morphological characters. With assistance from computer vision, leaves are poised to make numerous new contributions to systematic and paleobotanical studies.
HUDU: The Hanford Unified Dose Utility computer code
Scherpelz, R.I.
1991-02-01
The Hanford Unified Dose Utility (HUDU) computer program was developed to provide rapid initial assessment of radiological emergency situations. The HUDU code uses a straight-line Gaussian atmospheric dispersion model to estimate the transport of radionuclides released from an accident site. For dose points on the plume centerline, it calculates internal doses due to inhalation and external doses due to exposure to the plume. The program incorporates a number of features unique to the Hanford Site (operated by the US Department of Energy), including a library of source terms derived from various facilities' safety analysis reports. The HUDU code was designed to run on an IBM-PC or compatible personal computer. The user interface was designed for fast and easy operation with minimal user training. The theoretical basis and mathematical models used in the HUDU computer code are described, as are the computer code itself and the data libraries used. Detailed instructions for operating the code are also included. Appendices to the report contain descriptions of the program modules, listings of HUDU's data library, and descriptions of the verification tests that were run as part of the code development. 14 refs., 19 figs., 2 tabs.
Computer code applicability assessment for the advanced Candu reactor
Wren, D.J.; Langman, V.J.; Popov, N.; Snell, V.G. [Atomic Energy of Canada Ltd (Canada)
2004-07-01
AECL Technologies, the 100%-owned US subsidiary of Atomic Energy of Canada Ltd. (AECL), is currently the proponents of a pre-licensing review of the Advanced Candu Reactor (ACR) with the United States Nuclear Regulatory Commission (NRC). A key focus topic for this pre-application review is the NRC acceptance of the computer codes used in the safety analysis of the ACR. These codes have been developed and their predictions compared against experimental results over extended periods of time in Canada. These codes have also undergone formal validation in the 1990's. In support of this formal validation effort AECL has developed, implemented and currently maintains a Software Quality Assurance program (SQA) to ensure that its analytical, scientific and design computer codes meet the required standards for software used in safety analyses. This paper discusses the SQA program used to develop, qualify and maintain the computer codes used in ACR safety analysis, including the current program underway to confirm the applicability of these computer codes for use in ACR safety analyses. (authors)
Experimental methodology for computational fluid dynamics code validation
Aeschliman, D.P.; Oberkampf, W.L.
1997-09-01
Validation of Computational Fluid Dynamics (CFD) codes is an essential element of the code development process. Typically, CFD code validation is accomplished through comparison of computed results to previously published experimental data that were obtained for some other purpose, unrelated to code validation. As a result, it is a near certainty that not all of the information required by the code, particularly the boundary conditions, will be available. The common approach is therefore unsatisfactory, and a different method is required. This paper describes a methodology developed specifically for experimental validation of CFD codes. The methodology requires teamwork and cooperation between code developers and experimentalists throughout the validation process, and takes advantage of certain synergisms between CFD and experiment. The methodology employs a novel uncertainty analysis technique which helps to define the experimental plan for code validation wind tunnel experiments, and to distinguish between and quantify various types of experimental error. The methodology is demonstrated with an example of surface pressure measurements over a model of varying geometrical complexity in laminar, hypersonic, near perfect gas, 3-dimensional flow.
Computer Security: better code, fewer problems
Stefan Lueders, Computer Security Team
2016-01-01
The origin of many security incidents is negligence or unintentional mistakes made by web developers or programmers. In the rush to complete the work, due to skewed priorities, or just to ignorance, basic security principles can be omitted or forgotten. The resulting vulnerabilities lie dormant until the evil side spots them and decides to hit hard. Computer security incidents in the past have put CERN’s reputation at risk due to websites being defaced with negative messages about the Organization, hash files of passwords being extracted, restricted data exposed… And it all started with a little bit of negligence! If you check out the Top 10 web development blunders, you will see that the most prevalent mistakes are: Not filtering input, e.g. accepting “<“ or “>” in input fields even if only a number is expected. Not validating that input: you expect a birth date? So why accept letters? &...
A three-dimensional magnetostatics computer code for insertion devices.
Chubar, O; Elleaume, P; Chavanne, J
1998-05-01
RADIA is a three-dimensional magnetostatics computer code optimized for the design of undulators and wigglers. It solves boundary magnetostatics problems with magnetized and current-carrying volumes using the boundary integral approach. The magnetized volumes can be arbitrary polyhedrons with non-linear (iron) or linear anisotropic (permanent magnet) characteristics. The current-carrying elements can be straight or curved blocks with rectangular cross sections. Boundary conditions are simulated by the technique of mirroring. Analytical formulae used for the computation of the field produced by a magnetized volume of a polyhedron shape are detailed. The RADIA code is written in object-oriented C++ and interfaced to Mathematica [Mathematica is a registered trademark of Wolfram Research, Inc.]. The code outperforms currently available finite-element packages with respect to the CPU time of the solver and accuracy of the field integral estimations. An application of the code to the case of a wedge-pole undulator is presented.
Low Computational Complexity Network Coding For Mobile Networks
Heide, Janus
2012-01-01
Network Coding (NC) is a technique that can provide benefits in many types of networks, some examples from wireless networks are: In relay networks, either the physical or the data link layer, to reduce the number of transmissions. In reliable multicast, to reduce the amount of signaling and enable...... cooperation among receivers. In meshed networks, to simplify routing schemes and to increase robustness toward node failures. This thesis deals with implementation issues of one NC technique namely Random Linear Network Coding (RLNC) which can be described as a highly decentralized non-deterministic intra......-flow coding technique. One of the key challenges of this technique is its inherent computational complexity which can lead to high computational load and energy consumption in particular on the mobile platforms that are the target platform in this work. To increase the coding throughput several...
Recent applications of the transonic wing analysis computer code, TWING
Subramanian, N. R.; Holst, T. L.; Thomas, S. D.
1982-01-01
An evaluation of the transonic-wing-analysis computer code TWING is given. TWING utilizes a fully implicit approximate factorization iteration scheme to solve the full potential equation in conservative form. A numerical elliptic-solver grid-generation scheme is used to generate the required finite-difference mesh. Several wing configurations were analyzed, and the limits of applicability of this code was evaluated. Comparisons of computed results were made with available experimental data. Results indicate that the code is robust, accurate (when significant viscous effects are not present), and efficient. TWING generally produces solutions an order of magnitude faster than other conservative full potential codes using successive-line overrelaxation. The present method is applicable to a wide range of isolated wing configurations including high-aspect-ratio transport wings and low-aspect-ratio, high-sweep, fighter configurations.
The PLUTO Code for Adaptive Mesh Computations in Astrophysical Fluid Dynamics
Mignone, A.; Zanni, C.; Tzeferacos, P.; van Straalen, B.; Colella, P.; Bodo, G.
2012-01-01
We present a description of the adaptive mesh refinement (AMR) implementation of the PLUTO code for solving the equations of classical and special relativistic magnetohydrodynamics (MHD and RMHD). The current release exploits, in addition to the static grid version of the code, the distributed infrastructure of the CHOMBO library for multidimensional parallel computations over block-structured, adaptively refined grids. We employ a conservative finite-volume approach where primary flow quantities are discretized at the cell center in a dimensionally unsplit fashion using the Corner Transport Upwind method. Time stepping relies on a characteristic tracing step where piecewise parabolic method, weighted essentially non-oscillatory, or slope-limited linear interpolation schemes can be handily adopted. A characteristic decomposition-free version of the scheme is also illustrated. The solenoidal condition of the magnetic field is enforced by augmenting the equations with a generalized Lagrange multiplier providing propagation and damping of divergence errors through a mixed hyperbolic/parabolic explicit cleaning step. Among the novel features, we describe an extension of the scheme to include non-ideal dissipative processes, such as viscosity, resistivity, and anisotropic thermal conduction without operator splitting. Finally, we illustrate an efficient treatment of point-local, potentially stiff source terms over hierarchical nested grids by taking advantage of the adaptivity in time. Several multidimensional benchmarks and applications to problems of astrophysical relevance assess the potentiality of the AMR version of PLUTO in resolving flow features separated by large spatial and temporal disparities.
THE PLUTO CODE FOR ADAPTIVE MESH COMPUTATIONS IN ASTROPHYSICAL FLUID DYNAMICS
Mignone, A.; Tzeferacos, P. [Dipartimento di Fisica Generale, Universita di Torino, via Pietro Giuria 1, 10125 Torino (Italy); Zanni, C.; Bodo, G. [INAF, Osservatorio Astronomico di Torino, Strada Osservatorio 20, Pino Torinese 10025 (Italy); Van Straalen, B.; Colella, P. [Lawrence Berkeley National Laboratory, 1 Cyclotron Road, MS 50A-1148, Berkeley, CA 94720 (United States)
2012-01-01
We present a description of the adaptive mesh refinement (AMR) implementation of the PLUTO code for solving the equations of classical and special relativistic magnetohydrodynamics (MHD and RMHD). The current release exploits, in addition to the static grid version of the code, the distributed infrastructure of the CHOMBO library for multidimensional parallel computations over block-structured, adaptively refined grids. We employ a conservative finite-volume approach where primary flow quantities are discretized at the cell center in a dimensionally unsplit fashion using the Corner Transport Upwind method. Time stepping relies on a characteristic tracing step where piecewise parabolic method, weighted essentially non-oscillatory, or slope-limited linear interpolation schemes can be handily adopted. A characteristic decomposition-free version of the scheme is also illustrated. The solenoidal condition of the magnetic field is enforced by augmenting the equations with a generalized Lagrange multiplier providing propagation and damping of divergence errors through a mixed hyperbolic/parabolic explicit cleaning step. Among the novel features, we describe an extension of the scheme to include non-ideal dissipative processes, such as viscosity, resistivity, and anisotropic thermal conduction without operator splitting. Finally, we illustrate an efficient treatment of point-local, potentially stiff source terms over hierarchical nested grids by taking advantage of the adaptivity in time. Several multidimensional benchmarks and applications to problems of astrophysical relevance assess the potentiality of the AMR version of PLUTO in resolving flow features separated by large spatial and temporal disparities.
An advanced implicit solver for MHD
Udrea, Bogdan
A new implicit algorithm has been developed for the solution of the time-dependent, viscous and resistive single fluid magnetohydrodynamic (MHD) equations. The algorithm is based on an approximate Riemann solver for the hyperbolic fluxes and central differencing applied on a staggered grid for the parabolic fluxes. The algorithm employs a locally aligned coordinate system that allows the solution to the Riemann problems to be solved in a natural direction, normal to cell interfaces. The result is an original scheme that is robust and reduces the complexity of the flux formulas. The evaluation of the parabolic fluxes is also implemented using a locally aligned coordinate system, this time on the staggered grid. The implicit formulation employed by WARP3 is a two level scheme that was applied for the first time to the single fluid MHD model. The flux Jacobians that appear in the implicit scheme are evaluated numerically. The linear system that results from the implicit discretization is solved using a robust symmetric Gauss-Seidel method. The code has an explicit mode capability so that implementation and test of new algorithms or new physics can be performed in this simpler mode. Last but not least the code was designed and written to run on parallel computers so that complex, high resolution runs can be per formed in hours rather than days. The code has been benchmarked against analytical and experimental gas dynamics and MHD results. The benchmarks consisted of one-dimensional Riemann problems and diffusion dominated problems, two-dimensional supersonic flow over a wedge, axisymmetric magnetoplasmadynamic (MPD) thruster simulation and three-dimensional supersonic flow over intersecting wedges and spheromak stability simulation. The code has been proven to be robust and the results of the simulations showed excellent agreement with analytical and experimental results. Parallel performance studies showed that the code performs as expected when run on parallel
FLASH: A finite element computer code for variably saturated flow
Baca, R.G.; Magnuson, S.O.
1992-05-01
A numerical model was developed for use in performance assessment studies at the INEL. The numerical model, referred to as the FLASH computer code, is designed to simulate two-dimensional fluid flow in fractured-porous media. The code is specifically designed to model variably saturated flow in an arid site vadose zone and saturated flow in an unconfined aquifer. In addition, the code also has the capability to simulate heat conduction in the vadose zone. This report presents the following: description of the conceptual frame-work and mathematical theory; derivations of the finite element techniques and algorithms; computational examples that illustrate the capability of the code; and input instructions for the general use of the code. The FLASH computer code is aimed at providing environmental scientists at the INEL with a predictive tool for the subsurface water pathway. This numerical model is expected to be widely used in performance assessments for: (1) the Remedial Investigation/Feasibility Study process and (2) compliance studies required by the US Department of Energy Order 5820.2A.
Free-boundary ideal MHD stability of W7-X divertor equilibria
Nührenberg, C.
2016-07-01
Plasma configurations describing the stellarator experiment Wendelstein 7-X (W7-X) are computationally established taking into account the geometry of the test-divertor unit and the high-heat-flux divertor which will be installed in the vacuum chamber of the device (Gasparotto et al 2014 Fusion Eng. Des. 89 2121). These plasma equilibria are computationally studied for their global ideal magnetohydrodynamic (MHD) stability properties. Results from the ideal MHD stability code cas3d (Nührenberg 1996 Phys. Plasmas 3 2401), stability limits, spatial structures and growth rates are presented for free-boundary perturbations. The work focusses on the exploration of MHD unstable regions of the W7-X configuration space, thereby providing information for future experiments in W7-X aiming at an assessment of the role of ideal MHD in stellarator confinement.
Highly Optimized Code Generation for Stencil Codes with Computation Reuse for GPUs
Wen-Jing Ma; Kan Gao; Guo-Ping Long
2016-01-01
Computation reuse is known as an effective optimization technique. However, due to the complexity of modern GPU architectures, there is yet not enough understanding regarding the intriguing implications of the interplay of compu-tation reuse and hardware specifics on application performance. In this paper, we propose an automatic code generator for a class of stencil codes with inherent computation reuse on GPUs. For such applications, the proper reuse of intermediate results, combined with careful register and on-chip local memory usage, has profound implications on performance. Current state of the art does not address this problem in depth, partially due to the lack of a good program representation that can expose all potential computation reuse. In this paper, we leverage the computation overlap graph (COG), a simple representation of data dependence and data reuse with “element view”, to expose potential reuse opportunities. Using COG, we propose a portable code generation and tuning framework for GPUs. Compared with current state-of-the-art code generators, our experimental results show up to 56.7%performance improvement on modern GPUs such as NVIDIA C2050.
Parallelization of Finite Element Analysis Codes Using Heterogeneous Distributed Computing
Ozguner, Fusun
1996-01-01
Performance gains in computer design are quickly consumed as users seek to analyze larger problems to a higher degree of accuracy. Innovative computational methods, such as parallel and distributed computing, seek to multiply the power of existing hardware technology to satisfy the computational demands of large applications. In the early stages of this project, experiments were performed using two large, coarse-grained applications, CSTEM and METCAN. These applications were parallelized on an Intel iPSC/860 hypercube. It was found that the overall speedup was very low, due to large, inherently sequential code segments present in the applications. The overall execution time T(sub par), of the application is dependent on these sequential segments. If these segments make up a significant fraction of the overall code, the application will have a poor speedup measure.
Prodeto, a computer code for probabilistic fatigue design
Braam, H. [ECN-Solar and Wind Energy, Petten (Netherlands); Christensen, C.J.; Thoegersen, M.L. [Risoe National Lab., Roskilde (Denmark); Ronold, K.O. [Det Norske Veritas, Hoevik (Norway)
1999-03-01
A computer code for structural relibility analyses of wind turbine rotor blades subjected to fatigue loading is presented. With pre-processors that can transform measured and theoretically predicted load series to load range distributions by rain-flow counting and with a family of generic distribution models for parametric representation of these distribution this computer program is available for carying through probabilistic fatigue analyses of rotor blades. (au)
A Parametric Study of Extended-MHD Drift Tearing
King, Jacob R
2014-01-01
The linear drift-tearing mode is analyzed for different regimes of the plasma-$\\beta$, ion-skin-depth parameter space with an unreduced, extended-MHD model. New dispersion relations are found at moderate plasma $\\beta$ and previous drift-tearing results are classified as applicable at small plasma $\\beta$. The drift stabilization of the mode in the regimes varies from non-existent/weak to complete. As the diamagnetic-drift frequency is proportional to the plasma $\\beta$, verification exercises with unreduced, extended-MHD models in the small plasma-$\\beta$ regimes are impractical. The new dispersion relations in the moderate plasma-$\\beta$ regimes are used to verify the extended-MHD implementation of the NIMROD code [C. R. Sovinec et al., J. Comput. Phys. 195, 355 (2004)]. Given the small boundary-layer skin depth, discussion of the validity of the first-order finite-Larmour-radius model is presented.
Methods and computer codes for nuclear systems calculations
B P Kochurov; A P Knyazev; A Yu Kwaretzkheli
2007-02-01
Some numerical methods for reactor cell, sub-critical systems and 3D models of nuclear reactors are presented. The methods are developed for steady states and space–time calculations. Computer code TRIFON solves space-energy problem in (, ) systems of finite height and calculates heterogeneous few-group matrix parameters of reactor cells. These parameters are used as input data in the computer code SHERHAN solving the 3D heterogeneous reactor equation for steady states and 3D space–time neutron processes simulation. Modification of TRIFON was developed for the simulation of space–time processes in sub-critical systems with external sources. An option of SHERHAN code for the system with external sources is under development.
Computer code for double beta decay QRPA based calculations
Barbero, C. A.; Mariano, A. [Departamento de Física, Facultad de Ciencias Exactas, Universidad Nacional de La Plata, La Plata, Argentina and Instituto de Física La Plata, CONICET, La Plata (Argentina); Krmpotić, F. [Instituto de Física La Plata, CONICET, La Plata, Argentina and Instituto de Física Teórica, Universidade Estadual Paulista, São Paulo (Brazil); Samana, A. R.; Ferreira, V. dos Santos [Departamento de Ciências Exatas e Tecnológicas, Universidade Estadual de Santa Cruz, BA (Brazil); Bertulani, C. A. [Department of Physics, Texas A and M University-Commerce, Commerce, TX (United States)
2014-11-11
The computer code developed by our group some years ago for the evaluation of nuclear matrix elements, within the QRPA and PQRPA nuclear structure models, involved in neutrino-nucleus reactions, muon capture and β{sup ±} processes, is extended to include also the nuclear double beta decay.
Plagiarism Detection Algorithm for Source Code in Computer Science Education
Liu, Xin; Xu, Chan; Ouyang, Boyu
2015-01-01
Nowadays, computer programming is getting more necessary in the course of program design in college education. However, the trick of plagiarizing plus a little modification exists among some students' home works. It's not easy for teachers to judge if there's plagiarizing in source code or not. Traditional detection algorithms cannot fit this…
Connecting Neural Coding to Number Cognition: A Computational Account
Prather, Richard W.
2012-01-01
The current study presents a series of computational simulations that demonstrate how the neural coding of numerical magnitude may influence number cognition and development. This includes behavioral phenomena cataloged in cognitive literature such as the development of numerical estimation and operational momentum. Though neural research has…
General review of the MOSTAS computer code for wind turbines
Dungundji, J.; Wendell, J. H.
1981-01-01
The MOSTAS computer code for wind turbine analysis is reviewed, and techniques and methods used in its analyses are described. Impressions of its strengths and weakness, and recommendations for its application, modification, and further development are made. Basic techniques used in wind turbine stability and response analyses for systems with constant and periodic coefficients are reviewed.
Cosmological AMR MHD with Enzo
Xu, Hao [Los Alamos National Laboratory; Li, Hui [Los Alamos National Laboratory; Li, Shengtai [Los Alamos National Laboratory
2009-01-01
In this work, we present EnzoMHD, the extension of the cosmological code Enzoto include magnetic fields. We use the hyperbolic solver of Li et al. (2008) for the computation of interface fluxes. We use constrained transport methods of Balsara & Spicer (1999) and Gardiner & Stone (2005) to advance the induction equation, the reconstruction technique of Balsara (2001) to extend the Adaptive Mesh Refinement of Berger & Colella (1989) already used in Enzo, though formulated in a slightly different way for ease of implementation. This combination of methods preserves the divergence of the magnetic field to machine precision. We use operator splitting to include gravity and cosmological expansion. We then present a series of cosmological and non cosmologjcal tests problems to demonstrate the quality of solution resulting from this combination of solvers.
Computed radiography simulation using the Monte Carlo code MCNPX
Correa, S.C.A. [Programa de Engenharia Nuclear/COPPE, Universidade Federal do Rio de Janeiro, Ilha do Fundao, Caixa Postal 68509, 21945-970, Rio de Janeiro, RJ (Brazil); Centro Universitario Estadual da Zona Oeste (CCMAT)/UEZO, Av. Manuel Caldeira de Alvarenga, 1203, Campo Grande, 23070-200, Rio de Janeiro, RJ (Brazil); Souza, E.M. [Programa de Engenharia Nuclear/COPPE, Universidade Federal do Rio de Janeiro, Ilha do Fundao, Caixa Postal 68509, 21945-970, Rio de Janeiro, RJ (Brazil); Silva, A.X., E-mail: ademir@con.ufrj.b [PEN/COPPE-DNC/Poli CT, Universidade Federal do Rio de Janeiro, Ilha do Fundao, Caixa Postal 68509, 21945-970, Rio de Janeiro, RJ (Brazil); Cassiano, D.H. [Instituto de Radioprotecao e Dosimetria/CNEN Av. Salvador Allende, s/n, Recreio, 22780-160, Rio de Janeiro, RJ (Brazil); Lopes, R.T. [Programa de Engenharia Nuclear/COPPE, Universidade Federal do Rio de Janeiro, Ilha do Fundao, Caixa Postal 68509, 21945-970, Rio de Janeiro, RJ (Brazil)
2010-09-15
Simulating X-ray images has been of great interest in recent years as it makes possible an analysis of how X-ray images are affected owing to relevant operating parameters. In this paper, a procedure for simulating computed radiographic images using the Monte Carlo code MCNPX is proposed. The sensitivity curve of the BaFBr image plate detector as well as the characteristic noise of a 16-bit computed radiography system were considered during the methodology's development. The results obtained confirm that the proposed procedure for simulating computed radiographic images is satisfactory, as it allows obtaining results comparable with experimental data.
Fault-tolerant quantum computing with color codes
Landahl, Andrew J; Rice, Patrick R
2011-01-01
We present and analyze protocols for fault-tolerant quantum computing using color codes. We present circuit-level schemes for extracting the error syndrome of these codes fault-tolerantly. We further present an integer-program-based decoding algorithm for identifying the most likely error given the syndrome. We simulated our syndrome extraction and decoding algorithms against three physically-motivated noise models using Monte Carlo methods, and used the simulations to estimate the corresponding accuracy thresholds for fault-tolerant quantum error correction. We also used a self-avoiding walk analysis to lower-bound the accuracy threshold for two of these noise models. We present and analyze two architectures for fault-tolerantly computing with these codes: one with 2D arrays of qubits are stacked atop each other and one in a single 2D substrate. Our analysis demonstrates that color codes perform slightly better than Kitaev's surface codes when circuit details are ignored. When these details are considered, w...
New Parallel computing framework for radiation transport codes
Kostin, M.A.; /Michigan State U., NSCL; Mokhov, N.V.; /Fermilab; Niita, K.; /JAERI, Tokai
2010-09-01
A new parallel computing framework has been developed to use with general-purpose radiation transport codes. The framework was implemented as a C++ module that uses MPI for message passing. The module is significantly independent of radiation transport codes it can be used with, and is connected to the codes by means of a number of interface functions. The framework was integrated with the MARS15 code, and an effort is under way to deploy it in PHITS. Besides the parallel computing functionality, the framework offers a checkpoint facility that allows restarting calculations with a saved checkpoint file. The checkpoint facility can be used in single process calculations as well as in the parallel regime. Several checkpoint files can be merged into one thus combining results of several calculations. The framework also corrects some of the known problems with the scheduling and load balancing found in the original implementations of the parallel computing functionality in MARS15 and PHITS. The framework can be used efficiently on homogeneous systems and networks of workstations, where the interference from the other users is possible.
New Parallel computing framework for radiation transport codes
Kostin, M A; Niita, K
2012-01-01
A new parallel computing framework has been developed to use with general-purpose radiation transport codes. The framework was implemented as a C++ module that uses MPI for message passing. The module is significantly independent of radiation transport codes it can be used with, and is connected to the codes by means of a number of interface functions. The frame work was integrated with the MARS15 code, and an effort is under way to deploy it in PHITS. Besides the parallel computing functionality, the framework offers a checkpoint facility that allows restarting calculations with a saved checkpoint file. The checkpoint facility can be used in single process calculations as well as in the parallel regime. Several checkpoint files can be merged into one thus combining results of several calculations. The framework also corrects some of the known problems with the sch eduling and load balancing found in the original implementations of the parallel computing functionality in MARS15 and PHITS. The framework can be...
LMFBR models for the ORIGEN2 computer code
Croff, A.G.; McAdoo, J.W.; Bjerke, M.A.
1983-06-01
Reactor physics calculations have led to the development of nine liquid-metal fast breeder reactor (LMFBR) models for the ORIGEN2 computer code. Four of the models are based on the U-Pu fuel cycle, two are based on the Th-U-Pu fuel cycle, and three are based on the Th-/sup 233/U fuel cycle. The reactor models are based on cross sections taken directly from the reactor physics codes. Descriptions of the reactor models as well as values for the ORIGEN2 flux parameters THERM, RES, and FAST are given.
LMFBR models for the ORIGEN2 computer code
Croff, A.G.; McAdoo, J.W.; Bjerke, M.A.
1981-10-01
Reactor physics calculations have led to the development of nine liquid-metal fast breeder reactor (LMFBR) models for the ORIGEN2 computer code. Four of the models are based on the U-Pu fuel cycle, two are based on the Th-U-Pu fuel cycle, and three are based on the Th-/sup 238/U fuel cycle. The reactor models are based on cross sections taken directly from the reactor physics codes. Descriptions of the reactor models as well as values for the ORIGEN2 flux parameters THERM, RES, and FAST are given.
User's manual for HDR3 computer code
Arundale, C.J.
1982-10-01
A description of the HDR3 computer code and instructions for its use are provided. HDR3 calculates space heating costs for a hot dry rock (HDR) geothermal space heating system. The code also compares these costs to those of a specific oil heating system in use at the National Aeronautics and Space Administration Flight Center at Wallops Island, Virginia. HDR3 allows many HDR system parameters to be varied so that the user may examine various reservoir management schemes and may optimize reservoir design to suit a particular set of geophysical and economic parameters.
Picologlou, B.F.; Doss, E.D.; Geyer, H.K. (Argonne National Lab., IL (United States)); Sikes, W.C.; Ranellone, R.F. (Newport News Shipbuilding and Dry Dock Co., VA (United States))
1992-01-01
A two Tesla test facility was designed, built, and operated to investigate the performance of magnetohydrodynamic (MHD) seawater thrusters. The results of this investigation are used to validate a design oriented MHD thruster performance computer code. The thruster performance code consists of a one-dimensional MHD hydrodynamic model coupled to a two-dimensional electrical model. The code includes major loss mechanisms affecting the performance of the thruster. Among these losses are the joule dissipation losses, frictional losses, electrical end losses, and single electrode potential losses. The facility test loop, its components, and their design are presented in detail. Additionally, the test matrix and its rationale are discussed. Representative experimental results of the test program are presented, and are compared to pretest computer model predictions. Good agreement between predicted and measured data has served to validate the thruster performance computer models.
Computer codes for evaluation of control room habitability (HABIT)
Stage, S.A. [Pacific Northwest Lab., Richland, WA (United States)
1996-06-01
This report describes the Computer Codes for Evaluation of Control Room Habitability (HABIT). HABIT is a package of computer codes designed to be used for the evaluation of control room habitability in the event of an accidental release of toxic chemicals or radioactive materials. Given information about the design of a nuclear power plant, a scenario for the release of toxic chemicals or radionuclides, and information about the air flows and protection systems of the control room, HABIT can be used to estimate the chemical exposure or radiological dose to control room personnel. HABIT is an integrated package of several programs that previously needed to be run separately and required considerable user intervention. This report discusses the theoretical basis and physical assumptions made by each of the modules in HABIT and gives detailed information about the data entry windows. Sample runs are given for each of the modules. A brief section of programming notes is included. A set of computer disks will accompany this report if the report is ordered from the Energy Science and Technology Software Center. The disks contain the files needed to run HABIT on a personal computer running DOS. Source codes for the various HABIT routines are on the disks. Also included are input and output files for three demonstration runs.
War of ontology worlds: mathematics, computer code, or Esperanto?
Andrey Rzhetsky
2011-09-01
Full Text Available The use of structured knowledge representations-ontologies and terminologies-has become standard in biomedicine. Definitions of ontologies vary widely, as do the values and philosophies that underlie them. In seeking to make these views explicit, we conducted and summarized interviews with a dozen leading ontologists. Their views clustered into three broad perspectives that we summarize as mathematics, computer code, and Esperanto. Ontology as mathematics puts the ultimate premium on rigor and logic, symmetry and consistency of representation across scientific subfields, and the inclusion of only established, non-contradictory knowledge. Ontology as computer code focuses on utility and cultivates diversity, fitting ontologies to their purpose. Like computer languages C++, Prolog, and HTML, the code perspective holds that diverse applications warrant custom designed ontologies. Ontology as Esperanto focuses on facilitating cross-disciplinary communication, knowledge cross-referencing, and computation across datasets from diverse communities. We show how these views align with classical divides in science and suggest how a synthesis of their concerns could strengthen the next generation of biomedical ontologies.
War of Ontology Worlds: Mathematics, Computer Code, or Esperanto?
Rzhetsky, Andrey; Evans, James A.
2011-01-01
The use of structured knowledge representations—ontologies and terminologies—has become standard in biomedicine. Definitions of ontologies vary widely, as do the values and philosophies that underlie them. In seeking to make these views explicit, we conducted and summarized interviews with a dozen leading ontologists. Their views clustered into three broad perspectives that we summarize as mathematics, computer code, and Esperanto. Ontology as mathematics puts the ultimate premium on rigor and logic, symmetry and consistency of representation across scientific subfields, and the inclusion of only established, non-contradictory knowledge. Ontology as computer code focuses on utility and cultivates diversity, fitting ontologies to their purpose. Like computer languages C++, Prolog, and HTML, the code perspective holds that diverse applications warrant custom designed ontologies. Ontology as Esperanto focuses on facilitating cross-disciplinary communication, knowledge cross-referencing, and computation across datasets from diverse communities. We show how these views align with classical divides in science and suggest how a synthesis of their concerns could strengthen the next generation of biomedical ontologies. PMID:21980276
MHD Advanced Power Train Phase I, Final Report, Volume 7
A. R. Jones
1985-08-01
This appendix provides additional data in support of the MHD/Steam Power Plant Analyses reported in report Volume 5. The data is in the form of 3PA/SUMARY computer code printouts. The order of presentation in all four cases is as follows: (1) Overall Performance; (2) Component/Subsystem Information; (3) Plant Cost Accounts Summary; and (4) Plant Costing Details and Cost of Electricity.
Kako, T.; Watanabe, T. [eds.
1999-04-01
This is the proceeding of 'Study on Numerical Methods Related to Plasma Confinement' held in National Institute for Fusion Science. In this workshop, theoretical and numerical analyses of possible plasma equilibria with their stability properties are presented. These are also various talks on mathematical as well as numerical analyses related to the computational methods for fluid dynamics and plasma physics. The 14 papers are indexed individually. (J.P.N.)
Benchmarking of computer codes and approaches for modeling exposure scenarios
Seitz, R.R. [EG and G Idaho, Inc., Idaho Falls, ID (United States); Rittmann, P.D.; Wood, M.I. [Westinghouse Hanford Co., Richland, WA (United States); Cook, J.R. [Westinghouse Savannah River Co., Aiken, SC (United States)
1994-08-01
The US Department of Energy Headquarters established a performance assessment task team (PATT) to integrate the activities of DOE sites that are preparing performance assessments for the disposal of newly generated low-level waste. The PATT chartered a subteam with the task of comparing computer codes and exposure scenarios used for dose calculations in performance assessments. This report documents the efforts of the subteam. Computer codes considered in the comparison include GENII, PATHRAE-EPA, MICROSHIELD, and ISOSHLD. Calculations were also conducted using spreadsheets to provide a comparison at the most fundamental level. Calculations and modeling approaches are compared for unit radionuclide concentrations in water and soil for the ingestion, inhalation, and external dose pathways. Over 30 tables comparing inputs and results are provided.
Computational Complexity of Decoding Orthogonal Space-Time Block Codes
Ayanoglu, Ender; Karipidis, Eleftherios
2009-01-01
The computational complexity of optimum decoding for an orthogonal space-time block code G satisfying the orthogonality property that the Hermitian transpose of G multiplied by G is equal to a constant c times the sum of the squared symbols of the code times an identity matrix, where c is a positive integer is quantified. Four equivalent techniques of optimum decoding which have the same computational complexity are specified. Modifications to the basic formulation in special cases are calculated and illustrated by means of examples. This paper corrects and extends [1],[2], and unifies them with the results from the literature. In addition, a number of results from the literature are extended to the case c > 1.
Bragg optics computer codes for neutron scattering instrument design
Popovici, M.; Yelon, W.B.; Berliner, R.R. [Missouri Univ. Research Reactor, Columbia, MO (United States); Stoica, A.D. [Institute of Physics and Technology of Materials, Bucharest (Romania)
1997-09-01
Computer codes for neutron crystal spectrometer design, optimization and experiment planning are described. Phase space distributions, linewidths and absolute intensities are calculated by matrix methods in an extension of the Cooper-Nathans resolution function formalism. For modeling the Bragg reflection on bent crystals the lamellar approximation is used. Optimization is done by satisfying conditions of focusing in scattering and in real space, and by numerically maximizing figures of merit. Examples for three-axis and two-axis spectrometers are given.
General review of the MOSTAS computer code for wind turbines
Dugundji, J.; Wendell, J.H.
1981-06-01
The MOSTAS computer code for wind turbine analysis is reviewed, and the techniques and methods used in its analyses are described in some detail. Some impressions of its strengths and weaknesses, and some recommendations for its application, modification, and further development are made. Additionally, some basic techniques used in wind turbine stability and response analyses for systems with constant and periodic coefficients are reviewed in the Appendices.
Refactoring Android Java Code for On-Demand Computation Offloading
Zhang, Ying; Huang, Gang; Liu, Xuanzhe; Zhang, Wei; Zhang, Wei; Mei, Hong; Yang, Shunxiang
2012-01-01
International audience; Computation offloading is a promising way to improve the performance as well as reduce the battery energy consumption of a smartphone application by executing some part of the application on a remote server. Supporting such capability is not easy to smartphone app developers for 1) correctness: some codes, e.g. those for GPS, gravity and other sensors, can only run on the smartphone so that the developers have to identify which part of the application cannot be offload...
Methodology for computational fluid dynamics code verification/validation
Oberkampf, W.L.; Blottner, F.G.; Aeschliman, D.P.
1995-07-01
The issues of verification, calibration, and validation of computational fluid dynamics (CFD) codes has been receiving increasing levels of attention in the research literature and in engineering technology. Both CFD researchers and users of CFD codes are asking more critical and detailed questions concerning the accuracy, range of applicability, reliability and robustness of CFD codes and their predictions. This is a welcomed trend because it demonstrates that CFD is maturing from a research tool to the world of impacting engineering hardware and system design. In this environment, the broad issue of code quality assurance becomes paramount. However, the philosophy and methodology of building confidence in CFD code predictions has proven to be more difficult than many expected. A wide variety of physical modeling errors and discretization errors are discussed. Here, discretization errors refer to all errors caused by conversion of the original partial differential equations to algebraic equations, and their solution. Boundary conditions for both the partial differential equations and the discretized equations will be discussed. Contrasts are drawn between the assumptions and actual use of numerical method consistency and stability. Comments are also made concerning the existence and uniqueness of solutions for both the partial differential equations and the discrete equations. Various techniques are suggested for the detection and estimation of errors caused by physical modeling and discretization of the partial differential equations.
Problems in nonlinear resistive MHD
Turnbull, A.D.; Strait, E.J.; La Haye, R.J.; Chu, M.S.; Miller, R.L. [General Atomics, San Diego, CA (United States)
1998-12-31
Two experimentally relevant problems can relatively easily be tackled by nonlinear MHD codes. Both problems require plasma rotation in addition to the nonlinear mode coupling and full geometry already incorporated into the codes, but no additional physics seems to be crucial. These problems discussed here are: (1) nonlinear coupling and interaction of multiple MHD modes near the B limit and (2) nonlinear coupling of the m/n = 1/1 sawtooth mode with higher n gongs and development of seed islands outside q = 1.
Code manual for CONTAIN 2.0: A computer code for nuclear reactor containment analysis
Murata, K.K.; Williams, D.C.; Griffith, R.O.; Gido, R.G.; Tadios, E.L.; Davis, F.J.; Martinez, G.M.; Washington, K.E. [Sandia National Labs., Albuquerque, NM (United States); Tills, J. [J. Tills and Associates, Inc., Sandia Park, NM (United States)
1997-12-01
The CONTAIN 2.0 computer code is an integrated analysis tool used for predicting the physical conditions, chemical compositions, and distributions of radiological materials inside a containment building following the release of material from the primary system in a light-water reactor accident. It can also predict the source term to the environment. CONTAIN 2.0 is intended to replace the earlier CONTAIN 1.12, which was released in 1991. The purpose of this Code Manual is to provide full documentation of the features and models in CONTAIN 2.0. Besides complete descriptions of the models, this Code Manual provides a complete description of the input and output from the code. CONTAIN 2.0 is a highly flexible and modular code that can run problems that are either quite simple or highly complex. An important aspect of CONTAIN is that the interactions among thermal-hydraulic phenomena, aerosol behavior, and fission product behavior are taken into account. The code includes atmospheric models for steam/air thermodynamics, intercell flows, condensation/evaporation on structures and aerosols, aerosol behavior, and gas combustion. It also includes models for reactor cavity phenomena such as core-concrete interactions and coolant pool boiling. Heat conduction in structures, fission product decay and transport, radioactive decay heating, and the thermal-hydraulic and fission product decontamination effects of engineered safety features are also modeled. To the extent possible, the best available models for severe accident phenomena have been incorporated into CONTAIN, but it is intrinsic to the nature of accident analysis that significant uncertainty exists regarding numerous phenomena. In those cases, sensitivity studies can be performed with CONTAIN by means of user-specified input parameters. Thus, the code can be viewed as a tool designed to assist the knowledge reactor safety analyst in evaluating the consequences of specific modeling assumptions.
Improvement of level-1 PSA computer code package
Kim, Tae Woon; Park, C. K.; Kim, K. Y.; Han, S. H.; Jung, W. D.; Chang, S. C.; Yang, J. E.; Sung, T. Y.; Kang, D. I.; Park, J. H.; Lee, Y. H.; Kim, S. H.; Hwang, M. J.; Choi, S. Y.
1997-07-01
This year the fifth (final) year of the phase-I of the Government-sponsored Mid- and Long-term Nuclear Power Technology Development Project. The scope of this subproject titled on `The improvement of level-1 PSA Computer Codes` is divided into two main activities : (1) improvement of level-1 PSA methodology, (2) development of applications methodology of PSA techniques to operations and maintenance of nuclear power plant. Level-1 PSA code KIRAP is converted to PC-Windows environment. For the improvement of efficiency in performing PSA, the fast cutset generation algorithm and an analytical technique for handling logical loop in fault tree modeling are developed. Using about 30 foreign generic data sources, generic component reliability database (GDB) are developed considering dependency among source data. A computer program which handles dependency among data sources are also developed based on three stage bayesian updating technique. Common cause failure (CCF) analysis methods are reviewed and CCF database are established. Impact vectors can be estimated from this CCF database. A computer code, called MPRIDP, which handles CCF database are also developed. A CCF analysis reflecting plant-specific defensive strategy against CCF event is also performed. A risk monitor computer program, called Risk Monster, are being developed for the application to the operation and maintenance of nuclear power plant. The PSA application technique is applied to review the feasibility study of on-line maintenance and to the prioritization of in-service test (IST) of motor-operated valves (MOV). Finally, the root cause analysis (RCA) and reliability-centered maintenance (RCM) technologies are adopted and applied to the improvement of reliability of emergency diesel generators (EDG) of nuclear power plant. To help RCA and RCM analyses, two software programs are developed, which are EPIS and RAM Pro. (author). 129 refs., 20 tabs., 60 figs.
Computationally efficient sub-band coding of ECG signals.
Husøy, J H; Gjerde, T
1996-03-01
A data compression technique is presented for the compression of discrete time electrocardiogram (ECG) signals. The compression system is based on sub-band coding, a technique traditionally used for compressing speech and images. The sub-band coder employs quadrature mirror filter banks (QMF) with up to 32 critically sampled sub-bands. Both finite impulse response (FIR) and the more computationally efficient infinite impulse response (IIR) filter banks are considered as candidates in a complete ECG coding system. The sub-bands are threshold, quantized using uniform quantizers and run-length coded. The output of the run-length coder is further compressed by a Huffman coder. Extensive simulations indicate that 16 sub-bands are a suitable choice for this application. Furthermore, IIR filter banks are preferable due to their superiority in terms of computational efficiency. We conclude that the present scheme, which is suitable for real time implementation on a PC, can provide compression ratios between 5 and 15 without loss of clinical information.
Codes for Computationally Simple Channels: Explicit Constructions with Optimal Rate
Guruswami, Venkatesan
2010-01-01
In this paper, we consider coding schemes for computationally bounded channels, which can introduce an arbitrary set of errors as long as (a) the fraction of errors is bounded with high probability by a parameter p and (b) the process which adds the errors can be described by a sufficiently "simple" circuit. For three classes of channels, we provide explicit, efficiently encodable/decodable codes of optimal rate where only inefficiently decodable codes were previously known. In each case, we provide one encoder/decoder that works for every channel in the class. (1) Unique decoding for additive errors: We give the first construction of poly-time encodable/decodable codes for additive (a.k.a. oblivious) channels that achieve the Shannon capacity 1-H(p). Such channels capture binary symmetric errors and burst errors as special cases. (2) List-decoding for log-space channels: A space-S(n) channel reads and modifies the transmitted codeword as a stream, using at most S(n) bits of workspace on transmissions of n bi...
Compressing industrial computed tomography images by means of contour coding
Jiang, Haina; Zeng, Li
2013-10-01
An improved method for compressing industrial computed tomography (CT) images is presented. To have higher resolution and precision, the amount of industrial CT data has become larger and larger. Considering that industrial CT images are approximately piece-wise constant, we develop a compression method based on contour coding. The traditional contour-based method for compressing gray images usually needs two steps. The first is contour extraction and then compression, which is negative for compression efficiency. So we merge the Freeman encoding idea into an improved method for two-dimensional contours extraction (2-D-IMCE) to improve the compression efficiency. By exploiting the continuity and logical linking, preliminary contour codes are directly obtained simultaneously with the contour extraction. By that, the two steps of the traditional contour-based compression method are simplified into only one. Finally, Huffman coding is employed to further losslessly compress preliminary contour codes. Experimental results show that this method can obtain a good compression ratio as well as keeping satisfactory quality of compressed images.
Multicode comparison of selected source-term computer codes
Hermann, O.W.; Parks, C.V.; Renier, J.P.; Roddy, J.W.; Ashline, R.C.; Wilson, W.B.; LaBauve, R.J.
1989-04-01
This report summarizes the results of a study to assess the predictive capabilities of three radionuclide inventory/depletion computer codes, ORIGEN2, ORIGEN-S, and CINDER-2. The task was accomplished through a series of comparisons of their output for several light-water reactor (LWR) models (i.e., verification). Of the five cases chosen, two modeled typical boiling-water reactors (BWR) at burnups of 27.5 and 40 GWd/MTU and two represented typical pressurized-water reactors (PWR) at burnups of 33 and 50 GWd/MTU. In the fifth case, identical input data were used for each of the codes to examine the results of decay only and to show differences in nuclear decay constants and decay heat rates. Comparisons were made for several different characteristics (mass, radioactivity, and decay heat rate) for 52 radionuclides and for nine decay periods ranging from 30 d to 10,000 years. Only fission products and actinides were considered. The results are presented in comparative-ratio tables for each of the characteristics, decay periods, and cases. A brief summary description of each of the codes has been included. Of the more than 21,000 individual comparisons made for the three codes (taken two at a time), nearly half (45%) agreed to within 1%, and an additional 17% fell within the range of 1 to 5%. Approximately 8% of the comparison results disagreed by more than 30%. However, relatively good agreement was obtained for most of the radionuclides that are expected to contribute the greatest impact to waste disposal. Even though some defects have been noted, each of the codes in the comparison appears to produce respectable results. 12 figs., 12 tabs.
Code Verification of the HIGRAD Computational Fluid Dynamics Solver
Van Buren, Kendra L. [Los Alamos National Laboratory; Canfield, Jesse M. [Los Alamos National Laboratory; Hemez, Francois M. [Los Alamos National Laboratory; Sauer, Jeremy A. [Los Alamos National Laboratory
2012-05-04
The purpose of this report is to outline code and solution verification activities applied to HIGRAD, a Computational Fluid Dynamics (CFD) solver of the compressible Navier-Stokes equations developed at the Los Alamos National Laboratory, and used to simulate various phenomena such as the propagation of wildfires and atmospheric hydrodynamics. Code verification efforts, as described in this report, are an important first step to establish the credibility of numerical simulations. They provide evidence that the mathematical formulation is properly implemented without significant mistakes that would adversely impact the application of interest. Highly accurate analytical solutions are derived for four code verification test problems that exercise different aspects of the code. These test problems are referred to as: (i) the quiet start, (ii) the passive advection, (iii) the passive diffusion, and (iv) the piston-like problem. These problems are simulated using HIGRAD with different levels of mesh discretization and the numerical solutions are compared to their analytical counterparts. In addition, the rates of convergence are estimated to verify the numerical performance of the solver. The first three test problems produce numerical approximations as expected. The fourth test problem (piston-like) indicates the extent to which the code is able to simulate a 'mild' discontinuity, which is a condition that would typically be better handled by a Lagrangian formulation. The current investigation concludes that the numerical implementation of the solver performs as expected. The quality of solutions is sufficient to provide credible simulations of fluid flows around wind turbines. The main caveat associated to these findings is the low coverage provided by these four problems, and somewhat limited verification activities. A more comprehensive evaluation of HIGRAD may be beneficial for future studies.
Knowlton, Marie; Wetzel, Robin
2006-01-01
This study compared the length of text in English Braille American Edition, the Nemeth code, and the computer braille code with the Unified English Braille Code (UEBC)--also known as Unified English Braille (UEB). The findings indicate that differences in the length of text are dependent on the type of material that is transcribed and the grade…
pyro: A teaching code for computational astrophysical hydrodynamics
Zingale, Michael
2013-01-01
We describe pyro: a simple, freely-available code to aid students in learning the computational hydrodynamics methods widely used in astrophysics. pyro is written with simplicity and learning in mind and intended to allow students to experiment with various methods popular in the field, including those for advection, compressible and incompressible hydrodynamics, multigrid, and diffusion in a finite-volume framework. We show some of the test problems from pyro, describe its design philosophy, and suggest extensions for students to build their understanding of these methods.
pyro: A teaching code for computational astrophysical hydrodynamics
Zingale, M.
2014-10-01
We describe pyro: a simple, freely-available code to aid students in learning the computational hydrodynamics methods widely used in astrophysics. pyro is written with simplicity and learning in mind and intended to allow students to experiment with various methods popular in the field, including those for advection, compressible and incompressible hydrodynamics, multigrid, and diffusion in a finite-volume framework. We show some of the test problems from pyro, describe its design philosophy, and suggest extensions for students to build their understanding of these methods.
MHD Energy Bypass Scramjet Engine
Mehta, Unmeel B.; Bogdanoff, David W.; Park, Chul; Arnold, Jim (Technical Monitor)
2001-01-01
Revolutionary rather than evolutionary changes in propulsion systems are most likely to decrease cost of space transportation and to provide a global range capability. Hypersonic air-breathing propulsion is a revolutionary propulsion system. The performance of scramjet engines can be improved by the AJAX energy management concept. A magneto-hydro-dynamics (MHD) generator controls the flow and extracts flow energy in the engine inlet and a MHD accelerator downstream of the combustor accelerates the nozzle flow. A progress report toward developing the MHD technology is presented herein. Recent theoretical efforts are reviewed and ongoing experimental efforts are discussed. The latter efforts also include an ongoing collaboration between NASA, the US Air Force Research Laboratory, US industry, and Russian scientific organizations. Two of the critical technologies, the ionization of the air and the MHD accelerator, are briefly discussed. Examples of limiting the combustor entrance Mach number to a low supersonic value with a MHD energy bypass scheme are presented, demonstrating an improvement in scramjet performance. The results for a simplified design of an aerospace plane show that the specific impulse of the MHD-bypass system is better than the non-MHD system and typical rocket over a narrow region of flight speeds and design parameters. Equilibrium ionization and non-equilibrium ionization are discussed. The thermodynamic condition of air at the entrance of the engine inlet determines the method of ionization. The required external power for non-equilibrium ionization is computed. There have been many experiments in which electrical power generation has successfully been achieved by magneto-hydrodynamic (MHD) means. However, relatively few experiments have been made to date for the reverse case of achieving gas acceleration by the MHD means. An experiment in a shock tunnel is described in which MHD acceleration is investigated experimentally. MHD has several
Geometric plane shapes for computer-generated holographic engraving codes
Augier, Ángel G.; Rabal, Héctor; Sánchez, Raúl B.
2017-04-01
We report a new theoretical and experimental study on hologravures, as holographic computer-generated laser-engravings. A geometric theory of images based on the general principles of light ray behaviour is shown. The models used are also applicable for similar engravings obtained by any non-laser method, and the solutions allow for the analysis of particular situations, not only in the case of light reflection mode, but also in transmission mode geometry. This approach is a novel perspective allowing the three-dimensional (3D) design of engraved images for specific ends. We prove theoretically that plane curves of very general geometric shapes can be used to encode image information onto a two-dimensional (2D) engraving, showing notable influence on the behaviour of reconstructed images that appears as an exciting investigation topic, extending its applications. Several cases of code using particular curvilinear shapes are experimentally studied. The computer-generated objects are coded by using the chosen curve type, and engraved by a laser on a plane surface of suitable material. All images are recovered optically by adequate illumination. The pseudoscopic or orthoscopic character of these images is considered, and an appropriate interpretation is presented.
Evaluation of detonation energy from EXPLO5 computer code results
Suceska, M. [Brodarski Institute, Zagreb (Croatia). Marine Research and Special Technologies
1999-10-01
The detonation energies of several high explosives are evaluated from the results of chemical-equilibrium computer code named EXPLO5. Two methods of the evaluation of detonation energy are applied: (a) Direct evaluation from the internal energy of detonation products at the CJ point and the energy of shock compression of the detonation products, i.e. by equating the detonation energy and the heat of detonation, and (b) evaluation from the expansion isentrope of detonation products, applying the JWL model. These energies are compared to the energies computed from cylinder test derived JWL coefficients. It is found out that the detonation energies obtained directly from the energy of detonation products at the CJ point are uniformly to high (0.9445{+-}0.577 kJ/cm{sup 3}) while the detonation energies evaluated from the expansion isentrope, are in a considerable agreement (0.2072{+-}0.396 kJ/cm{sup 3}) with the energies calculated from cylinder test derived JWL coefficients. (orig.) [German] Die Detonationsenergien verschiedener Hochleistungssprengstoffe werden bewertet aus den Ergebnissen des Computer Codes fuer chemische Gleichgewichte genannt EXPLO5. Zwei Methoden wurden angewendet: (a) Direkte Bewertung aus der inneren Energie der Detonationsprodukte am CJ-Punkt und aus der Energie der Stosskompression der Detonationsprodukte, d.h. durch Gleichsetzung von Detonationsenergie und Detonationswaerme, (b) Auswertung durch die Expansions-Isentrope der Detonationsprodukte unter Anwendung des JWL-Modells. Diese Energien werden verglichen mit den berechneten Energien mit aus dem Zylindertest abgeleiteten JWL-Koeffizienten. Es wird gefunden, dass die Detonationsenergien, die direkt aus der Energie der Detonationsprodukte beim CJ-Punkt erhalten wurden, einheitlich zu hoch sind (0,9445{+-}0,577 kJ/cm{sup 3}), waehrend die aus der Expansions-Isentrope erhaltenen in guter Uebereinstimmung sind (0,2072{+-}0,396 kJ/cm{sup 3}) mit den berechneten Energien mit aus dem Zylindertest
Extended MHD Effects in High Energy Density Experiments
Seyler, Charles
2016-10-01
The MHD model is the workhorse for computational modeling of HEDP experiments. Plasma models are inheritably limited in scope, but MHD is expected to be a very good model for studying plasmas at the high densities attained in HEDP experiments. There are, however, important ways in which MHD fails to adequately describe the results, most notably due to the omission of the Hall term in the Ohm's law (a form of extended MHD or XMHD). This talk will discuss these failings by directly comparing simulations of MHD and XMHD for particularly relevant cases. The methodology is to simulate HEDP experiments using a Hall-MHD (HMHD) code based on a highly accurate and robust Discontinuous Galerkin method, and by comparison of HMHD to MHD draw conclusions about the impact of the Hall term. We focus on simulating two experimental pulsed power machines under various scenarios. We examine the MagLIF experiment on the Z-machine at Sandia National Laboratories and liner experiments on the COBRA machine at Cornell. For the MagLIF experiment we find that power flow in the feed leads to low density plasma ablation into the region surrounding the liner. The inflow of this plasma compresses axial magnetic flux onto the liner. In MHD this axial flux tends to resistively decay, whereas in HMHD a force-free current layer sustains the axial flux on the liner leading to a larger ratio of axial to azimuthal flux. During the liner compression the magneto-Rayleigh-Taylor instability leads to helical perturbations due to minimization of field line bending. Simulations of a cylindrical liner using the COBRA machine parameters can under certain conditions exhibit amplification of an axial field due to a force-free low-density current layer separated by some distance from the liner. This results in a configuration in which there is predominately axial field on the liner inside the current layer and azimuthal field outside the layer. We are currently attempting to experimentally verify the simulation
Good, Jonathon; Keenan, Sarah; Mishra, Punya
2016-01-01
The popular press is rife with examples of how students in the United States and around the globe are learning to program, make, and tinker. The Hour of Code, maker-education, and similar efforts are advocating that more students be exposed to principles found within computer science. We propose an expansion beyond simply teaching computational…
A computer code to simulate X-ray imaging techniques
Duvauchelle, Philippe E-mail: philippe.duvauchelle@insa-lyon.fr; Freud, Nicolas; Kaftandjian, Valerie; Babot, Daniel
2000-09-01
A computer code was developed to simulate the operation of radiographic, radioscopic or tomographic devices. The simulation is based on ray-tracing techniques and on the X-ray attenuation law. The use of computer-aided drawing (CAD) models enables simulations to be carried out with complex three-dimensional (3D) objects and the geometry of every component of the imaging chain, from the source to the detector, can be defined. Geometric unsharpness, for example, can be easily taken into account, even in complex configurations. Automatic translations or rotations of the object can be performed to simulate radioscopic or tomographic image acquisition. Simulations can be carried out with monochromatic or polychromatic beam spectra. This feature enables, for example, the beam hardening phenomenon to be dealt with or dual energy imaging techniques to be studied. The simulation principle is completely deterministic and consequently the computed images present no photon noise. Nevertheless, the variance of the signal associated with each pixel of the detector can be determined, which enables contrast-to-noise ratio (CNR) maps to be computed, in order to predict quantitatively the detectability of defects in the inspected object. The CNR is a relevant indicator for optimizing the experimental parameters. This paper provides several examples of simulated images that illustrate some of the rich possibilities offered by our software. Depending on the simulation type, the computation time order of magnitude can vary from 0.1 s (simple radiographic projection) up to several hours (3D tomography) on a PC, with a 400 MHz microprocessor. Our simulation tool proves to be useful in developing new specific applications, in choosing the most suitable components when designing a new testing chain, and in saving time by reducing the number of experimental tests.
Reasoning with Computer Code: a new Mathematical Logic
Pissanetzky, Sergio
2013-01-01
A logic is a mathematical model of knowledge used to study how we reason, how we describe the world, and how we infer the conclusions that determine our behavior. The logic presented here is natural. It has been experimentally observed, not designed. It represents knowledge as a causal set, includes a new type of inference based on the minimization of an action functional, and generates its own semantics, making it unnecessary to prescribe one. This logic is suitable for high-level reasoning with computer code, including tasks such as self-programming, objectoriented analysis, refactoring, systems integration, code reuse, and automated programming from sensor-acquired data. A strong theoretical foundation exists for the new logic. The inference derives laws of conservation from the permutation symmetry of the causal set, and calculates the corresponding conserved quantities. The association between symmetries and conservation laws is a fundamental and well-known law of nature and a general principle in modern theoretical Physics. The conserved quantities take the form of a nested hierarchy of invariant partitions of the given set. The logic associates elements of the set and binds them together to form the levels of the hierarchy. It is conjectured that the hierarchy corresponds to the invariant representations that the brain is known to generate. The hierarchies also represent fully object-oriented, self-generated code, that can be directly compiled and executed (when a compiler becomes available), or translated to a suitable programming language. The approach is constructivist because all entities are constructed bottom-up, with the fundamental principles of nature being at the bottom, and their existence is proved by construction. The new logic is mathematically introduced and later discussed in the context of transformations of algorithms and computer programs. We discuss what a full self-programming capability would really mean. We argue that self
Nonlinear helical MHD instability
Zueva, N.M.; Solov' ev, L.S.
1977-07-01
An examination is made of the boundary problem on the development of MHD instability in a toroidal plasma. Two types of local helical instability are noted - Alfven and thermal, and the corresponding criteria of instability are cited. An evaluation is made of the maximum attainable kinetic energy, limited by the degree to which the law of conservation is fulfilled. An examination is made of a precise solution to a kinematic problem on the helical evolution of a cylindrical magnetic configuration at a given velocity distribution in a plasma. A numerical computation of the development of MHD instability in a plasma cylinder by a computerized solution of MHD equations is made where the process's helical symmetry is conserved. The development of instability is of a resonance nature. The instability involves the entire cross section of the plasma and leads to an inside-out reversal of the magnetic surfaces when there is a maximum unstable equilibrium configuration in the nonlinear stage. The examined instability in the tore is apparently stabilized by a magnetic hole when certain limitations are placed on the distribution of flows in the plasma. 29 references, 8 figures.
Toth, G.; Daldorff, L. K. S.; Jia, X.; Gombosi, T. I.; Lapenta, G.
2014-12-01
We have recently developed a new modeling capability to embed theimplicit Particle-in-Cell (PIC) model iPIC3D into the BATS-R-USmagnetohydrodynamic model. The PIC domain can cover the regions wherekinetic effects are most important, such as reconnection sites. TheBATS-R-US code, on the other hand, can efficiently handle the rest ofthe computational domain where the MHD or Hall MHD description issufficient. As one of the very first applications of the MHD-EPICalgorithm (Daldorff et al. 2014, JCP, 268, 236) we simulate theinteraction between Jupiter's magnetospheric plasma with Ganymede'smagnetosphere, where the separation of kinetic and global scalesappears less severe than for the Earth's magnetosphere. Because theexternal Jovian magnetic field remains in an anti-parallel orientationwith respect to Ganymede's intrinsic magnetic field, magneticreconnection is believed to be the major process that couples the twomagnetospheres. As the PIC model is able to describe self-consistentlythe electron behavior, our coupled MHD-EPIC model is well suited forinvestigating the nature of magnetic reconnection in thisreconnection-driven mini-magnetosphere. We will compare the MHD-EPICsimulations with pure Hall MHD simulations and compare both modelresults with Galileo plasma and magnetic field measurements to assess therelative importance of ion and electron kinetics in controlling theconfiguration and dynamics of Ganymede's magnetosphere.
Interface design of VSOP'94 computer code for safety analysis
Natsir, Khairina; Yazid, Putranto Ilham; Andiwijayakusuma, D.; Wahanani, Nursinta Adi
2014-09-01
Today, most software applications, also in the nuclear field, come with a graphical user interface. VSOP'94 (Very Superior Old Program), was designed to simplify the process of performing reactor simulation. VSOP is a integrated code system to simulate the life history of a nuclear reactor that is devoted in education and research. One advantage of VSOP program is its ability to calculate the neutron spectrum estimation, fuel cycle, 2-D diffusion, resonance integral, estimation of reactors fuel costs, and integrated thermal hydraulics. VSOP also can be used to comparative studies and simulation of reactor safety. However, existing VSOP is a conventional program, which was developed using Fortran 65 and have several problems in using it, for example, it is only operated on Dec Alpha mainframe platforms and provide text-based output, difficult to use, especially in data preparation and interpretation of results. We develop a GUI-VSOP, which is an interface program to facilitate the preparation of data, run the VSOP code and read the results in a more user friendly way and useable on the Personal 'Computer (PC). Modifications include the development of interfaces on preprocessing, processing and postprocessing. GUI-based interface for preprocessing aims to provide a convenience way in preparing data. Processing interface is intended to provide convenience in configuring input files and libraries and do compiling VSOP code. Postprocessing interface designed to visualized the VSOP output in table and graphic forms. GUI-VSOP expected to be useful to simplify and speed up the process and analysis of safety aspects.
A computer code for analysis of severe accidents in LWRs
NONE
2001-07-01
The ICARE2 computer code, developed and validated since 1988 at IPSN (nuclear safety and protection institute), calculates in a mechanistic way the physical and chemical phenomena involved in the core degradation process during possible severe accidents in LWR's. The coupling between ICARE2 and the best-estimate thermal-hydraulics code CATHARE2 was completed at IPSN and led to the release of a first ICARE/CATHARE V1 version in 1999, followed by 2 successive revisions in 2000 and 2001. This documents gathers all the contributions presented at the first international ICARE/CATHARE users'club seminar that took place in November 2001. This seminar was characterized by a high quality and variety of the presentations, showing an increase of reactor applications and user needs in this area (2D/3D aspects, reflooding, corium slumping into the lower head,...). 2 sessions were organized. The first one was dedicated to the applications of ICARE2 V3mod1 against small-scale experiments such as PHEBUS FPT2 and FPT3 tests, PHEBUS AIC, QUENCH experiments, NRU-FLHT-5 test, ACRR-MP1 and DC1 experiments, CORA-PWR tests, and PBF-SFD1.4 test. The second session involved ICARE/CATHARE V1mod1 reactor applications and users'guidelines. Among reactor applications we found: code applicability to high burn-up fuel rods, simulation of the TMI-2 transient, simulation of a PWR-900 high pressure severe accident sequence, and the simulation of a VVER-1000 large break LOCA scenario. (A.C.)
Compute-and-Forward: Harnessing Interference through Structured Codes
Nazer, Bobak
2009-01-01
Interference is usually viewed as an obstacle to communication in wireless networks. This paper proposes a new strategy, compute-and-forward, that exploits interference to obtain significantly higher rates between users in a network. The key idea is that relays should decode linear functions of transmitted messages according to their observed channel coefficients rather than ignoring the interference as noise. After decoding these linear equations, the relays simply send them towards the destinations, which given enough equations, can recover their desired messages. The underlying codes are based on nested lattices whose algebraic structure ensures that integer combinations of codewords can be decoded reliably. Encoders map messages from a finite field to a lattice and decoders recover equations of lattice points which are then mapped back to equations over the finite field. This scheme is applicable even if the transmitters lack channel state information. Its potential is demonstrated through examples drawn ...
Benchmark Solutions for Computational Aeroacoustics (CAA) Code Validation
Scott, James R.
2004-01-01
NASA has conducted a series of Computational Aeroacoustics (CAA) Workshops on Benchmark Problems to develop a set of realistic CAA problems that can be used for code validation. In the Third (1999) and Fourth (2003) Workshops, the single airfoil gust response problem, with real geometry effects, was included as one of the benchmark problems. Respondents were asked to calculate the airfoil RMS pressure and far-field acoustic intensity for different airfoil geometries and a wide range of gust frequencies. This paper presents the validated that have been obtained to the benchmark problem, and in addition, compares them with classical flat plate results. It is seen that airfoil geometry has a strong effect on the airfoil unsteady pressure, and a significant effect on the far-field acoustic intensity. Those parts of the benchmark problem that have not yet been adequately solved are identified and presented as a challenge to the CAA research community.
Fire aerosol experiment and comparisons with computer code predictions
Gregory, W.S.; Nichols, B.D.; White, B.W.; Smith, P.R.; Leslie, I.H.; Corkran, J.R.
1988-01-01
Los Alamos National Laboratory, in cooperation with New Mexico State University, has carried on a series of tests to provide experimental data on fire-generated aerosol transport. These data will be used to verify the aerosol transport capabilities of the FIRAC computer code. FIRAC was developed by Los Alamos for the US Nuclear Regulatory Commission. It is intended to be used by safety analysts to evaluate the effects of hypothetical fires on nuclear plants. One of the most significant aspects of this analysis deals with smoke and radioactive material movement throughout the plant. The tests have been carried out using an industrial furnace that can generate gas temperatures to 300/degree/C. To date, we have used quartz aerosol with a median diameter of about 10 ..mu..m as the fire aerosol simulant. We also plan to use fire-generated aerosols of polystyrene and polymethyl methacrylate (PMMA). The test variables include two nominal gas flow rates (150 and 300 ft/sup 3//min) and three nominal gas temperatures (ambient, 150/degree/C, and 300/degree/C). The test results are presented in the form of plots of aerosol deposition vs length of duct. In addition, the mass of aerosol caught in a high-efficiency particulate air (HEPA) filter during the tests is reported. The tests are simulated with the FIRAC code, and the results are compared with the experimental data. 3 refs., 10 figs., 1 tab.
Application of the RESRAD computer code to VAMP scenario S
Gnanapragasam, E.K.; Yu, C.
1997-03-01
The RESRAD computer code developed at Argonne National Laboratory was among 11 models from 11 countries participating in the international Scenario S validation of radiological assessment models with Chernobyl fallout data from southern Finland. The validation test was conducted by the Multiple Pathways Assessment Working Group of the Validation of Environmental Model Predictions (VAMP) program coordinated by the International Atomic Energy Agency. RESRAD was enhanced to provide an output of contaminant concentrations in environmental media and in food products to compare with measured data from southern Finland. Probability distributions for inputs that were judged to be most uncertain were obtained from the literature and from information provided in the scenario description prepared by the Finnish Centre for Radiation and Nuclear Safety. The deterministic version of RESRAD was run repeatedly to generate probability distributions for the required predictions. These predictions were used later to verify the probabilistic RESRAD code. The RESRAD predictions of radionuclide concentrations are compared with measured concentrations in selected food products. The radiological doses predicted by RESRAD are also compared with those estimated by the Finnish Centre for Radiation and Nuclear Safety.
Computer Tensor Codes to Design the War Drive
Maccone, C.
To address problems in Breakthrough Propulsion Physics (BPP) and design the Warp Drive one needs sheer computing capabilities. This is because General Relativity (GR) and Quantum Field Theory (QFT) are so mathematically sophisticated that the amount of analytical calculations is prohibitive and one can hardly do all of them by hand. In this paper we make a comparative review of the main tensor calculus capabilities of the three most advanced and commercially available “symbolic manipulator” codes. We also point out that currently one faces such a variety of different conventions in tensor calculus that it is difficult or impossible to compare results obtained by different scholars in GR and QFT. Mathematical physicists, experimental physicists and engineers have each their own way of customizing tensors, especially by using different metric signatures, different metric determinant signs, different definitions of the basic Riemann and Ricci tensors, and by adopting different systems of physical units. This chaos greatly hampers progress toward the design of the Warp Drive. It is thus suggested that NASA would be a suitable organization to establish standards in symbolic tensor calculus and anyone working in BPP should adopt these standards. Alternatively other institutions, like CERN in Europe, might consider the challenge of starting the preliminary implementation of a Universal Tensor Code to design the Warp Drive.
Computer code for the atomistic simulation of lattice defects and dynamics. [COMENT code
Schiffgens, J.O.; Graves, N.J.; Oster, C.A.
1980-04-01
This document has been prepared to satisfy the need for a detailed, up-to-date description of a computer code that can be used to simulate phenomena on an atomistic level. COMENT was written in FORTRAN IV and COMPASS (CDC assembly language) to solve the classical equations of motion for a large number of atoms interacting according to a given force law, and to perform the desired ancillary analysis of the resulting data. COMENT is a dual-purpose intended to describe static defect configurations as well as the detailed motion of atoms in a crystal lattice. It can be used to simulate the effect of temperature, impurities, and pre-existing defects on radiation-induced defect production mechanisms, defect migration, and defect stability.
An MHD model of the earth's magnetosphere
Wu, C. C.
1985-01-01
It is pointed out that the earth's magnetosphere arises from the interaction of the solar wind with the earth's geomagnetic field. A global magnetohydrodynamics (MHD) model of the earth's magnetosphere has drawn much attention in recent years. In this model, MHD equations are used to describe the solar wind interaction with the magnetosphere. In the present paper, some numerical aspects of the model are considered. Attention is given to the ideal MHD equations, an equation of state for the plasma, the model as an initial- and boundary-value problem, the shock capturing technique, computational requirements and techniques for global MHD modeling, a three-dimensional mesh system employed in the global MHD model, and some computational results.
FEMHD: An adaptive finite element method for MHD and edge modelling
Strauss, H.R.
1995-07-01
This paper describes the code FEMHD, an adaptive finite element MHD code, which is applied in a number of different manners to model MHD behavior and edge plasma phenomena on a diverted tokamak. The code uses an unstructured triangular mesh in 2D and wedge shaped mesh elements in 3D. The code has been adapted to look at neutral and charged particle dynamics in the plasma scrape off region, and into a full MHD-particle code.
Acceleration of the OpenFOAM-based MHD solver using graphics processing units
He, Qingyun; Chen, Hongli, E-mail: hlchen1@ustc.edu.cn; Feng, Jingchao
2015-12-15
Highlights: • A 3D PISO-MHD was implemented on Kepler-class graphics processing units (GPUs) using CUDA technology. • A consistent and conservative scheme is used in the code which was validated by three basic benchmarks in a rectangular and round ducts. • Parallelized of CPU and GPU acceleration were compared relating to single core CPU in MHD problems and non-MHD problems. • Different preconditions for solving MHD solver were compared and the results showed that AMG method is better for calculations. - Abstract: The pressure-implicit with splitting of operators (PISO) magnetohydrodynamics MHD solver of the couple of Navier–Stokes equations and Maxwell equations was implemented on Kepler-class graphics processing units (GPUs) using the CUDA technology. The solver is developed on open source code OpenFOAM based on consistent and conservative scheme which is suitable for simulating MHD flow under strong magnetic field in fusion liquid metal blanket with structured or unstructured mesh. We verified the validity of the implementation on several standard cases including the benchmark I of Shercliff and Hunt's cases, benchmark II of fully developed circular pipe MHD flow cases and benchmark III of KIT experimental case. Computational performance of the GPU implementation was examined by comparing its double precision run times with those of essentially the same algorithms and meshes. The resulted showed that a GPU (GTX 770) can outperform a server-class 4-core, 8-thread CPU (Intel Core i7-4770k) by a factor of 2 at least.
Synchrotron radiation of self-collimating relativistic MHD jets
Porth, Oliver; Meliani, Zakaria; Vaidya, Bhargav
2011-01-01
The goal of this paper is to derive signatures of synchrotron radiation from state-of-the-art simulation models of collimating relativistic magnetohydrodynamic (MHD) jets featuring a large-scale helical magnetic field. We perform axisymmetric special relativistic MHD simulations of the jet acceleration region using the PLUTO code. The computational domain extends from the slow magnetosonic launching surface of the disk up to 6000^2 Schwarzschild radii allowing to reach highly relativistic Lorentz factors. The Poynting dominated disk wind develops into a jet with Lorentz factors of 8 and is collimated to 1 degree. In addition to the disk jet, we evolve a thermally driven spine jet, emanating from a hypothetical black hole corona. Solving the linearly polarized synchrotron radiation transport within the jet, we derive VLBI radio and (sub-) mm diagnostics such as core shift, polarization structure, intensity maps, spectra and Faraday rotation measure (RM), directly from the Stokes parameters. We also investigate...
Nabert, Christian; Othmer, Carsten; Glassmeier, Karl-Heinz
2017-05-01
The interaction of the solar wind with a planetary magnetic field causes electrical currents that modify the magnetic field distribution around the planet. We present an approach to estimating the planetary magnetic field from in situ spacecraft data using a magnetohydrodynamic (MHD) simulation approach. The method is developed with respect to the upcoming BepiColombo mission to planet Mercury aimed at determining the planet's magnetic field and its interior electrical conductivity distribution. In contrast to the widely used empirical models, global MHD simulations allow the calculation of the strongly time-dependent interaction process of the solar wind with the planet. As a first approach, we use a simple MHD simulation code that includes time-dependent solar wind and magnetic field parameters. The planetary parameters are estimated by minimizing the misfit of spacecraft data and simulation results with a gradient-based optimization. As the calculation of gradients with respect to many parameters is usually very time-consuming, we investigate the application of an adjoint MHD model. This adjoint MHD model is generated by an automatic differentiation tool to compute the gradients efficiently. The computational cost for determining the gradient with an adjoint approach is nearly independent of the number of parameters. Our method is validated by application to THEMIS (Time History of Events and Macroscale Interactions during Substorms) magnetosheath data to estimate Earth's dipole moment.
Implementation of a 3D mixing layer code on parallel computers
Roe, K.; Thakur, R.; Dang, T.; Bogucz, E. [Syracuse Univ., NY (United States)
1995-09-01
This paper summarizes our progress and experience in the development of a Computational-Fluid-Dynamics code on parallel computers to simulate three-dimensional spatially-developing mixing layers. In this initial study, the three-dimensional time-dependent Euler equations are solved using a finite-volume explicit time-marching algorithm. The code was first programmed in Fortran 77 for sequential computers. The code was then converted for use on parallel computers using the conventional message-passing technique, while we have not been able to compile the code with the present version of HPF compilers.
Ambipolar diffusion in low-mass star formation. I. General comparison with the ideal MHD case
Masson, Jacques; Hennebelle, Patrick; Vaytet, Neil; Commerçon, Benoit
2015-01-01
In this paper, we provide a more accurate description of the evolution of the magnetic flux redistribution during prestellar core collapse by including resistive terms in the magnetohydrodynamics (MHD) equations. We focus more particularly on the impact of ambipolar diffusion. We use the adaptive mesh refinement code RAMSES to carry out such calculations. The resistivities required to calculate the ambipolar diffusion terms were computed using a reduced chemical network of charged, neutral and grain species. The inclusion of ambipolar diffusion leads to the formation of a magnetic diffusion barrier in the vicinity of the core, preventing accumulation of magnetic flux in and around the core and amplification of the field above 0.1G. The mass and radius of the first Larson core remain similar between ideal and non-ideal MHD models. This diffusion plateau has crucial consequences on magnetic braking processes, allowing the formation of disk structures. Magnetically supported outflows launched in ideal MHD models...
3-D nonlinear evolution of MHD instabilities
Bateman, G.; Hicks, H. R.; Wooten, J. W.
1977-03-01
The nonlinear evolution of ideal MHD internal instabilities is investigated in straight cylindrical geometry by means of a 3-D initial-value computer code. These instabilities are characterized by pairs of velocity vortex cells rolling off each other and helically twisted down the plasma column. The cells persist until the poloidal velocity saturates at a few tenths of the Alfven velocity. The nonlinear phase is characterized by convection around these essentially fixed vortex cells. For example, the initially centrally peaked temperature profile is convected out and around to form an annulus of high temperature surrounding a small region of lower temperature. Weak, centrally localized instabilities do not alter the edge of the plasma. Strong, large-scale instabilities, resulting from a stronger longitudinal equilibrium current, drive the plasma against the wall. After three examples of instability are analyzed in detail, the numerical methods and their verification are discussed.
Assessment of the computer code COBRA/CFTL
Baxi, C. B.; Burhop, C. J.
1981-07-01
The COBRA/CFTL code has been developed by Oak Ridge National Laboratory (ORNL) for thermal-hydraulic analysis of simulated gas-cooled fast breeder reactor (GCFR) core assemblies to be tested in the core flow test loop (CFTL). The COBRA/CFTL code was obtained by modifying the General Atomic code COBRA*GCFR. This report discusses these modifications, compares the two code results for three cases which represent conditions from fully rough turbulent flow to laminar flow. Case 1 represented fully rough turbulent flow in the bundle. Cases 2 and 3 represented laminar and transition flow regimes. The required input for the COBRA/CFTL code, a sample problem input/output and the code listing are included in the Appendices.
Initial Studies of Validation of MHD Models for MST Reversed Field Pinch Plasmas
Jacobson, C. M.; Almagri, A. F.; Craig, D.; McCollam, K. J.; Reusch, J. A.; Sauppe, J. P.; Sovinec, C. R.; Triana, J. C.
2015-11-01
Quantitative validation of visco-resistive MHD models for RFP plasmas takes advantage of MST's advanced diagnostics. These plasmas are largely governed by MHD relaxation activity, so that a broad range of validation metrics can be evaluated. Previous nonlinear simulations using the visco-resistive MHD code DEBS at Lundquist number S = 4 ×106 produced equilibrium relaxation cycles in qualitative agreement with experiment, but magnetic fluctuation amplitudes b~ were at least twice as large as in experiment. The extended-MHD code NIMROD previously suggested that a two-fluid model may be necessary to produce b~ in agreement with experiment. For best comparisons with DEBS and to keep computational expense tractable, NIMROD is run in single-fluid mode at low S. These simulations are complemented by DEBS at higher S in cylindrical geometry, which will be used to examine b~ as a function of S. Experimental measurements are used with results from these simulations to evaluate validation metrics. Convergence tests of previous high S DEBS simulations are also discussed, along with benchmarking of DEBS and NIMROD with the SPECYL and PIXIE3D codes. Work supported by U.S. DOE and NSF.
Superimposed Code Theoretic Analysis of Deoxyribonucleic Acid (DNA) Codes and DNA Computing
2010-01-01
DNA Codes Based on Fibonacci Ensembles of DNA Sequences ”, 2008 IEEE Proceedings of International Symposium on Information Theory, pp. 2292 – 2296...2008, pp. 525-34. 28. A. Macula, et al., “Random Coding Bounds for DNA Codes Based on Fibonacci Ensembles of DNA Sequences ”, 2008 IEEE...component of this innovation is the combinatorial method of bio-memory design and detection that encodes item or process information as numerical sequences
User manual for PACTOLUS: a code for computing power costs.
Huber, H.D.; Bloomster, C.H.
1979-02-01
PACTOLUS is a computer code for calculating the cost of generating electricity. Through appropriate definition of the input data, PACTOLUS can calculate the cost of generating electricity from a wide variety of power plants, including nuclear, fossil, geothermal, solar, and other types of advanced energy systems. The purpose of PACTOLUS is to develop cash flows and calculate the unit busbar power cost (mills/kWh) over the entire life of a power plant. The cash flow information is calculated by two principal models: the Fuel Model and the Discounted Cash Flow Model. The Fuel Model is an engineering cost model which calculates the cash flow for the fuel cycle costs over the project lifetime based on input data defining the fuel material requirements, the unit costs of fuel materials and processes, the process lead and lag times, and the schedule of the capacity factor for the plant. For nuclear plants, the Fuel Model calculates the cash flow for the entire nuclear fuel cycle. For fossil plants, the Fuel Model calculates the cash flow for the fossil fuel purchases. The Discounted Cash Flow Model combines the fuel costs generated by the Fuel Model with input data on the capital costs, capital structure, licensing time, construction time, rates of return on capital, tax rates, operating costs, and depreciation method of the plant to calculate the cash flow for the entire lifetime of the project. The financial and tax structure for both investor-owned utilities and municipal utilities can be simulated through varying the rates of return on equity and debt, the debt-equity ratios, and tax rates. The Discounted Cash Flow Model uses the principal that the present worth of the revenues will be equal to the present worth of the expenses including the return on investment over the economic life of the project. This manual explains how to prepare the input data, execute cases, and interpret the output results. (RWR)
MMA, A Computer Code for Multi-Model Analysis
Eileen P. Poeter and Mary C. Hill
2007-08-20
This report documents the Multi-Model Analysis (MMA) computer code. MMA can be used to evaluate results from alternative models of a single system using the same set of observations for all models. As long as the observations, the observation weighting, and system being represented are the same, the models can differ in nearly any way imaginable. For example, they may include different processes, different simulation software, different temporal definitions (for example, steady-state and transient models could be considered), and so on. The multiple models need to be calibrated by nonlinear regression. Calibration of the individual models needs to be completed before application of MMA. MMA can be used to rank models and calculate posterior model probabilities. These can be used to (1) determine the relative importance of the characteristics embodied in the alternative models, (2) calculate model-averaged parameter estimates and predictions, and (3) quantify the uncertainty of parameter estimates and predictions in a way that integrates the variations represented by the alternative models. There is a lack of consensus on what model analysis methods are best, so MMA provides four default methods. Two are based on Kullback-Leibler information, and use the AIC (Akaike Information Criterion) or AICc (second-order-bias-corrected AIC) model discrimination criteria. The other two default methods are the BIC (Bayesian Information Criterion) and the KIC (Kashyap Information Criterion) model discrimination criteria. Use of the KIC criterion is equivalent to using the maximum-likelihood Bayesian model averaging (MLBMA) method. AIC, AICc, and BIC can be derived from Frequentist or Bayesian arguments. The default methods based on Kullback-Leibler information have a number of theoretical advantages, including that they tend to favor more complicated models as more data become available than do the other methods, which makes sense in many situations.
Reducing Computational Overhead of Network Coding with Intrinsic Information Conveying
Heide, Janus; Zhang, Qi; Pedersen, Morten V.
This paper investigated the possibility of intrinsic information conveying in network coding systems. The information is embedded into the coding vector by constructing the vector based on a set of predefined rules. This information can subsequently be retrieved by any receiver. The starting point...... to the overall energy consumption, which is particular problematic for mobile battery-driven devices. In RLNC coding is performed over a FF (Finite Field). We propose to divide this field into sub fields, and let each sub field signify some information or state. In order to embed the information correctly...... the coding operations must be performed in a particular way, which we introduce. Finally we evaluate the suggested system and find that the amount of coding can be significantly reduced both at nodes that recode and decode....
3-D field computation: The near-triumph of commerical codes
Turner, L.R.
1995-07-01
In recent years, more and more of those who design and analyze magnets and other devices are using commercial codes rather than developing their own. This paper considers the commercial codes and the features available with them. Other recent trends with 3-D field computation include parallel computation and visualization methods such as virtual reality systems.
ORNL ALICE: a statistical model computer code including fission competition. [In FORTRAN
Plasil, F.
1977-11-01
A listing of the computer code ORNL ALICE is given. This code is a modified version of computer codes ALICE and OVERLAID ALICE. It allows for higher excitation energies and for a greater number of evaporated particles than the earlier versions. The angular momentum removal option was made more general and more internally consistent. Certain roundoff errors are avoided by keeping a strict accounting of partial probabilities. Several output options were added.
Coupled simulation of kinetic pedestal growth and MHD ELM crash
Park, G [Courant Institute of Mathematical Sciences, New York University (United States); Cummings, J [California Institute of Technology (United States); Chang, C S [Courant Institute of Mathematical Sciences, New York University (United States); Podhorszki, N [Univ. California at Davis (United States); Klasky, S [ORNL (United States); Ku, S [Courant Institute of Mathematical Sciences, New York University (United States); Pankin, A [Lehigh Univ. (United States); Samtaney, R [Princeton Plasma Physics Laboratory (United States); Shoshani, A [LBNL (United States); Snyder, P [General Atomics (United States); Strauss, H [Courant Institute of Mathematical Sciences, New York University (United States); Sugiyama, L [MIT (United States)
2007-07-15
Edge pedestal height and the accompanying ELM crash are critical elements of ITER physics yet to be understood and predicted through high performance computing. An entirely self-consistent first principles simulation is being pursued as a long term research goal, and the plan is planned for completion in time for ITER operation. However, a proof-of-principle work has already been established using a computational tool that employs the best first principles physics available at the present time. A kinetic edge equilibrium code XGC0, which can simulate the neoclassically dominant pedestal growth from neutral ionization (using a phenomenological residual turbulence diffusion motion superposed upon the neoclassical particle motion) is coupled to an extended MHD code M3D, which can perform the nonlinear ELM crash. The stability boundary of the pedestal is checked by an ideal MHD linear peeling-ballooning code, which has been validated against many experimental data sets for the large scale (type I) ELMs onset boundary. The coupling workflow and scientific results to be enabled by it are described.
Comparison of computer codes for estimates of the symmetric coupled bunch instabilities growth times
Angal-Kalinin, Deepa
2002-01-01
The standard computer codes used for estimating the growth times of the symmetric coupled bunch instabilities are ZAP and BBI.The code Vlasov was earlier used for the LHC for the estimates of the coupled bunch instabilities growth time[1]. The results obtained by these three codes have been compared and the options under which their results can be compared are discussed. The differences in the input and the output for these three codes are given for a typical case.
Implementation of a 3-D nonlinear MHD calculation on the Intel hypercube
Drake, J.B.; Lawkins, W.F.; Carreras, B.A.; Hicks, H.R.
1987-08-01
As part of an exploratory study of the suitability of hypercube multiprocessors for scientific computations, the non-linear magnetohydrodynamics (MHD) code RSF was parallelized for use on an Intel iPSC hypercube. This report presents the numerical algorithm of RSF and the techniques used to obtain parallelism without sacrificing the numerical properties of the serial algorithm. Timing results are presented for a sample problem.
MMA, A Computer Code for Multi-Model Analysis
Poeter, Eileen P.; Hill, Mary C.
2007-01-01
This report documents the Multi-Model Analysis (MMA) computer code. MMA can be used to evaluate results from alternative models of a single system using the same set of observations for all models. As long as the observations, the observation weighting, and system being represented are the same, the models can differ in nearly any way imaginable. For example, they may include different processes, different simulation software, different temporal definitions (for example, steady-state and transient models could be considered), and so on. The multiple models need to be calibrated by nonlinear regression. Calibration of the individual models needs to be completed before application of MMA. MMA can be used to rank models and calculate posterior model probabilities. These can be used to (1) determine the relative importance of the characteristics embodied in the alternative models, (2) calculate model-averaged parameter estimates and predictions, and (3) quantify the uncertainty of parameter estimates and predictions in a way that integrates the variations represented by the alternative models. There is a lack of consensus on what model analysis methods are best, so MMA provides four default methods. Two are based on Kullback-Leibler information, and use the AIC (Akaike Information Criterion) or AICc (second-order-bias-corrected AIC) model discrimination criteria. The other two default methods are the BIC (Bayesian Information Criterion) and the KIC (Kashyap Information Criterion) model discrimination criteria. Use of the KIC criterion is equivalent to using the maximum-likelihood Bayesian model averaging (MLBMA) method. AIC, AICc, and BIC can be derived from Frequentist or Bayesian arguments. The default methods based on Kullback-Leibler information have a number of theoretical advantages, including that they tend to favor more complicated models as more data become available than do the other methods, which makes sense in many situations. Many applications of MMA will
Resistive MHD jet simulations with large resistivity
Cemeljic, Miljenko; Vlahakis, Nektarios; Tsinganos, Kanaris
2009-01-01
Axisymmetric resistive MHD simulations for radially self-similar initial conditions are performed, using the NIRVANA code. The magnetic diffusivity could occur in outflows above an accretion disk, being transferred from the underlying disk into the disk corona by MHD turbulence (anomalous turbulent diffusivity), or as a result of ambipolar diffusion in partially ionized flows. We introduce, in addition to the classical magnetic Reynolds number Rm, which measures the importance of resistive effects in the induction equation, a new number Rb, which measures the importance of the resistive effects in the energy equation. We find two distinct regimes of solutions in our simulations. One is the low-resistivity regime, in which results do not differ much from ideal-MHD solutions. In the high-resistivity regime, results seem to show some periodicity in time-evolution, and depart significantly from the ideal-MHD case. Whether this departure is caused by numerical or physical reasons is of considerable interest for nu...
Quantum error correcting codes and one-way quantum computing: Towards a quantum memory
Schlingemann, D
2003-01-01
For realizing a quantum memory we suggest to first encode quantum information via a quantum error correcting code and then concatenate combined decoding and re-encoding operations. This requires that the encoding and the decoding operation can be performed faster than the typical decoherence time of the underlying system. The computational model underlying the one-way quantum computer, which has been introduced by Hans Briegel and Robert Raussendorf, provides a suitable concept for a fast implementation of quantum error correcting codes. It is shown explicitly in this article is how encoding and decoding operations for stabilizer codes can be realized on a one-way quantum computer. This is based on the graph code representation for stabilizer codes, on the one hand, and the relation between cluster states and graph codes, on the other hand.
Application of computational fluid dynamics methods to improve thermal hydraulic code analysis
Sentell, Dennis Shannon, Jr.
A computational fluid dynamics code is used to model the primary natural circulation loop of a proposed small modular reactor for comparison to experimental data and best-estimate thermal-hydraulic code results. Recent advances in computational fluid dynamics code modeling capabilities make them attractive alternatives to the current conservative approach of coupled best-estimate thermal hydraulic codes and uncertainty evaluations. The results from a computational fluid dynamics analysis are benchmarked against the experimental test results of a 1:3 length, 1:254 volume, full pressure and full temperature scale small modular reactor during steady-state power operations and during a depressurization transient. A comparative evaluation of the experimental data, the thermal hydraulic code results and the computational fluid dynamics code results provides an opportunity to validate the best-estimate thermal hydraulic code's treatment of a natural circulation loop and provide insights into expanded use of the computational fluid dynamics code in future designs and operations. Additionally, a sensitivity analysis is conducted to determine those physical phenomena most impactful on operations of the proposed reactor's natural circulation loop. The combination of the comparative evaluation and sensitivity analysis provides the resources for increased confidence in model developments for natural circulation loops and provides for reliability improvements of the thermal hydraulic code.
VOOGD, J.A.
1999-04-19
An analysis of three software proposals is performed to recommend a computer code for immobilized low activity waste flow and transport modeling. The document uses criteria restablished in HNF-1839, ''Computer Code Selection Criteria for Flow and Transport Codes to be Used in Undisturbed Vadose Zone Calculation for TWRS Environmental Analyses'' as the basis for this analysis.
Quantum computation with topological codes from qubit to topological fault-tolerance
Fujii, Keisuke
2015-01-01
This book presents a self-consistent review of quantum computation with topological quantum codes. The book covers everything required to understand topological fault-tolerant quantum computation, ranging from the definition of the surface code to topological quantum error correction and topological fault-tolerant operations. The underlying basic concepts and powerful tools, such as universal quantum computation, quantum algorithms, stabilizer formalism, and measurement-based quantum computation, are also introduced in a self-consistent way. The interdisciplinary fields between quantum information and other fields of physics such as condensed matter physics and statistical physics are also explored in terms of the topological quantum codes. This book thus provides the first comprehensive description of the whole picture of topological quantum codes and quantum computation with them.
Two-Phase Flow in Geothermal Wells: Development and Uses of a Good Computer Code
Ortiz-Ramirez, Jaime
1983-06-01
A computer code is developed for vertical two-phase flow in geothermal wellbores. The two-phase correlations used were developed by Orkiszewski (1967) and others and are widely applicable in the oil and gas industry. The computer code is compared to the flowing survey measurements from wells in the East Mesa, Cerro Prieto, and Roosevelt Hot Springs geothermal fields with success. Well data from the Svartsengi field in Iceland are also used. Several applications of the computer code are considered. They range from reservoir analysis to wellbore deposition studies. It is considered that accurate and workable wellbore simulators have an important role to play in geothermal reservoir engineering.
Efficient Quantification of Uncertainties in Complex Computer Code Results Project
National Aeronautics and Space Administration — Propagation of parameter uncertainties through large computer models can be very resource intensive. Frameworks and tools for uncertainty quantification are...
Second Generation Integrated Composite Analyzer (ICAN) Computer Code
Murthy, Pappu L. N.; Ginty, Carol A.; Sanfeliz, Jose G.
1993-01-01
This manual updates the original 1986 NASA TP-2515, Integrated Composite Analyzer (ICAN) Users and Programmers Manual. The various enhancements and newly added features are described to enable the user to prepare the appropriate input data to run this updated version of the ICAN code. For reference, the micromechanics equations are provided in an appendix and should be compared to those in the original manual for modifications. A complete output for a sample case is also provided in a separate appendix. The input to the code includes constituent material properties, factors reflecting the fabrication process, and laminate configuration. The code performs micromechanics, macromechanics, and laminate analyses, including the hygrothermal response of polymer-matrix-based fiber composites. The output includes the various ply and composite properties, the composite structural response, and the composite stress analysis results with details on failure. The code is written in FORTRAN 77 and can be used efficiently as a self-contained package (or as a module) in complex structural analysis programs. The input-output format has changed considerably from the original version of ICAN and is described extensively through the use of a sample problem.
Code of Ethical Conduct for Computer-Using Educators: An ICCE Policy Statement.
Computing Teacher, 1987
1987-01-01
Prepared by the International Council for Computers in Education's Ethics and Equity Committee, this code of ethics for educators using computers covers nine main areas: curriculum issues, issues relating to computer access, privacy/confidentiality issues, teacher-related issues, student issues, the community, school organizational issues,…
Computational Participation: Understanding Coding as an Extension of Literacy Instruction
Burke, Quinn; O'Byrne, W. Ian; Kafai, Yasmin B.
2016-01-01
Understanding the computational concepts on which countless digital applications run offers learners the opportunity to no longer simply read such media but also become more discerning end users and potentially innovative "writers" of new media themselves. To think computationally--to solve problems, to design systems, and to process and…
Selection of a computer code for Hanford low-level waste engineered-system performance assessment
McGrail, B.P.; Mahoney, L.A.
1995-10-01
Planned performance assessments for the proposed disposal of low-level waste (LLW) glass produced from remediation of wastes stored in underground tanks at Hanford, Washington will require calculations of radionuclide release rates from the subsurface disposal facility. These calculations will be done with the aid of computer codes. Currently available computer codes were ranked in terms of the feature sets implemented in the code that match a set of physical, chemical, numerical, and functional capabilities needed to assess release rates from the engineered system. The needed capabilities were identified from an analysis of the important physical and chemical process expected to affect LLW glass corrosion and the mobility of radionuclides. The highest ranked computer code was found to be the ARES-CT code developed at PNL for the US Department of Energy for evaluation of and land disposal sites.
2006-09-01
Aerospace Applications, AIAA-Paper 96-2355, New Orleans, 1996 2. V.A.Bityurin, A.N.Bocharov, J.Lineberry, MHD Aerospace Applications, Invited Lecture ...Paper 2003- 4303, Orlando, FL 8. V.A.Bityurin, Prospective of MHD Interaction in Hypersonic and Propulsion Technologies, In: von Karman Series : Lectures ...Efforts in MHD AeoSpace Applications, In: von Karman Series : Lectures , Introduction of Magneto-Fluid Dynamics for AeroSpace Applications, von Karman
MHD and heat transfer benchmark problems for liquid metal flow in rectangular ducts
Sidorenkov, S.I. [D.V. Efremov Scientific Research Inst. of Electrophysical Apparatus, St. Petersburg (Russian Federation); Hua, T.Q. [Argonne National Lab., IL (United States); Araseki, H. [Central Research Inst. of Electric Power Industry, Tokoyo (Japan)
1994-12-31
Liquid metal cooling systems of a self-cooled blanket in a tokamak reactor will likely include channels of rectangular cross section where liquid metal is circulated in the presence of strong magnetic fields. MHD pressure drop, velocity distribution and heat transfer characteristics are important issues in the engineering design considerations. Computer codes for the reliable solution of three-dimensional MHD flow problems are needed for fusion relevant conditions. Argonne National Laboratory and The Efremov Institute have jointly defined several benchmark problems for code validation. The problems, described in this paper, are based on two series of rectangular duct experiments conducted at ANL; one of the series is a joint ANL/Efremov experiment. The geometries consist of variation of aspect ratio and wall thickness (thus wall conductance ratio). The transverse magnetic fields are uniform and nonuniform in the axial direction.
Computer codes used during upgrading activities at MINT TRIGA reactor
Mohammad Suhaimi Kassim; Adnan Bokhari; Mohd. Idris Taib [Malaysian Institute for Nuclear Technology Research, Kajang (Malaysia)
1999-10-01
MINT TRIGA Reactor is a 1-MW swimming pool nuclear research reactor commissioned in 1982. In 1993, a project was initiated to upgrade the thermal power to 2 MW. The IAEA assistance was sought to assist the various activities relevant to an upgrading exercise. For neutronics calculations, the IAEA has provided expert assistance to introduce the WIMS code, TRIGAP, and EXTERMINATOR2. For thermal-hydraulics calculations, PARET and RELAP5 were introduced. Shielding codes include ANISN and MERCURE. However, in the middle of 1997, MINT has decided to change the scope of the project to safety upgrading of the MINT Reactor. This paper describes some of the activities carried out during the upgrading process. (author)
Validation of the NCC Code for Staged Transverse Injection and Computations for a RBCC Combustor
Ajmani, Kumud; Liu, Nan-Suey
2005-01-01
The NCC code was validated for a case involving staged transverse injection into Mach 2 flow behind a rearward facing step. Comparisons with experimental data and with solutions from the FPVortex code was then used to perform computations to study fuel-air mixing for the combustor of a candidate rocket based combined cycle engine geometry. Comparisons with a one-dimensional analysis and a three-dimensional code (VULCAN) were performed to assess the qualitative and quantitative performance of the NCC solver.
Challenges of Computational Processing of Code-Switching
Çetinoğlu, Özlem; Schulz, Sarah; Vu, Ngoc Thang
2016-01-01
This paper addresses challenges of Natural Language Processing (NLP) on non-canonical multilingual data in which two or more languages are mixed. It refers to code-switching which has become more popular in our daily life and therefore obtains an increasing amount of attention from the research community. We report our experience that cov- ers not only core NLP tasks such as normalisation, language identification, language modelling, part-of-speech tagging and dependency parsing but also more...
POTRE: A computer code for the assessment of dose from ingestion
Hanusik, V.; Mitro, A.; Niedel, S.; Grosikova, B.; Uvirova, E.; Stranai, I. (Institute of Radioecology and Applied Nuclear Techniques, Kosice (Czechoslovakia))
1991-01-01
The paper describes the computer code PORET and the auxiliary database system which allow to assess the radiation exposure from ingestion of foodstuffs contaminated by radionuclides released from a nuclear facility during normal operation into the atmosphere. (orig.).
Speeding-up MADYMO 3D on serial and parallel computers using a portable coding environment
Tsiandikos, T.; Rooijackers, H.F.L.; Asperen, F.G.J. van; Lupker, H.A.
1996-01-01
This paper outlines the strategy and methodology used to create a portable coding environment for the commercial package MADYMO. The objective is to design a global data structure that efficiently utilises the memory and cache of computers, so that one source code can be used for serial, vector and
Moral, Cristian; de Antonio, Angelica; Ferre, Xavier; Lara, Graciela
2015-01-01
Introduction: In this article we propose a qualitative analysis tool--a coding system--that can support the formalisation of the information-seeking process in a specific field: research in computer science. Method: In order to elaborate the coding system, we have conducted a set of qualitative studies, more specifically a focus group and some…
Holbrook, M. Cay; MacCuspie, P. Ann
2010-01-01
Braille-reading mathematicians, scientists, and computer scientists were asked to examine the usability of the Unified English Braille Code (UEB) for technical materials. They had little knowledge of the code prior to the study. The research included two reading tasks, a short tutorial about UEB, and a focus group. The results indicated that the…
Metropol, a computer code for the simulation of transport of contaminants with groundwater
Sauter FJ; Hassanizadeh SM; Leijnse A; Glasbergen P; Slot AFM
1990-01-01
In this report a description is given of the computer code METROPOL. This code simulates the three dimensional flow of groundwater with varying density and the simultaneous transport of contaminants in low concentration and is based on the finite element method. The basic equations for groundwater
Comparison of different computer platforms for running the Versatile Advection Code
Toth, G.; Keppens, R.; Sloot, P.; Bubak, M.; Hertzberger, B.
1998-01-01
The Versatile Advection Code is a general tool for solving hydrodynamical and magnetohydrodynamical problems arising in astrophysics. We compare the performance of the code on different computer platforms, including work stations and vector and parallel supercomputers. Good parallel scaling can be a
Code and papers: computing publication patterns in the LHC era
CERN. Geneva
2012-01-01
Publications in scholarly journals establish the body of knowledge deriving from scientific research; they also play a fundamental role in the career path of scientists and in the evaluation criteria of funding agencies. This presentation reviews the evolution of computing-oriented publications in HEP following the start of operation of LHC. Quantitative analyses are illustrated, which document the production of scholarly papers on computing-related topics by HEP experiments and core tools projects (including distributed computing R&D), and the citations they receive. Several scientometric indicators are analyzed to characterize the role of computing in HEP literature. Distinctive features of scholarly publication production in the software-oriented and hardware-oriented experimental HEP communities are highlighted. Current patterns and trends are compared to the situation in previous generations' HEP experiments at LEP, Tevatron and B-factories. The results of this scientometric analysis document objec...
Proposed standards for peer-reviewed publication of computer code
Computer simulation models are mathematical abstractions of physical systems. In the area of natural resources and agriculture, these physical systems encompass selected interacting processes in plants, soils, animals, or watersheds. These models are scientific products and have become important i...
Joshua J. Cogliati; Abderrafi M. Ougouag
2006-10-01
A comprehensive, high fidelity model for pebble flow has been developed and embodied in the PEBBLES computer code. In this paper, a description of the physical artifacts included in the model is presented and some results from using the computer code for predicting the features of pebble flow and packing in a realistic pebble bed reactor design are shown. The sensitivity of models to various physical parameters is also discussed.
Computation of Grobner basis for systematic encoding of generalized quasi-cyclic codes
Van, Vo Tam; Mita, Seiichi
2008-01-01
Generalized quasi-cyclic (GQC) codes form a wide and useful class of linear codes that includes thoroughly quasi-cyclic codes, finite geometry (FG) low density parity check (LDPC) codes, and Hermitian codes. Although it is known that the systematic encoding of GQC codes is equivalent to the division algorithm in the theory of Grobner basis of modules, there has been no algorithm that computes Grobner basis for all types of GQC codes. In this paper, we propose two algorithms to compute Grobner basis for GQC codes from their parity check matrices: echelon canonical form algorithm and transpose algorithm. Both algorithms require sufficiently small number of finite-field operations with the order of the third power of code-length. Each algorithm has its own characteristic; the first algorithm is composed of elementary methods, and the second algorithm is based on a novel formula and is faster than the first one for high-rate codes. Moreover, we show that a serial-in serial-out encoder architecture for FG LDPC cod...
Windtalking Computers: Frequency Normalization, Binary Coding Systems and Encryption
Zirkind, Givon
2009-01-01
The goal of this paper is to discuss the application of known techniques, knowledge and technology in a novel way, to encrypt computer and non-computer data. To-date most computers use base 2 and most encryption systems use ciphering and/or an encryption algorithm, to convert data into a secret message. The method of having the computer "speak another secret language" as used in human military secret communications has never been imitated. The author presents the theory and several possible implementations of a method for computers for secret communications analogous to human beings using a secret language or; speaking multiple languages. The kind of encryption scheme proposed significantly increases the complexity of and the effort needed for, decryption. As every methodology has its drawbacks, so too, the data of the proposed system has its drawbacks. It is not as compressed as base 2 would be. However, this is manageable and acceptable, if the goal is very strong encryption: At least two methods and their ...
SENDIN and SENTINEL: two computer codes to assess the effects of nuclear data changes
Marable, J. H.; Drischler, J. D.; Weisbin, C. R.
1977-07-01
A description is given of the computer code SENTINEL, which provides a simple means for finding the effects on calculated reactor and shielding performance parameters due to proposed changes in the cross section data base. This code uses predetermined detailed sensitivity coefficients in SENPRO format, which is described in Appendix A. Knowledge of details of the particular reactor and/or shielding assemblies is not required of the user. Also described is the computer code SENDIN, which converts unformatted (binary) sensitivity files to card image form and vice versa. This is useful for transferring sensitivity files from one installation to another.
TPASS: a gamma-ray spectrum analysis and isotope identification computer code
Dickens, J.K.
1981-03-01
The gamma-ray spectral data-reduction and analysis computer code TPASS is described. This computer code is used to analyze complex Ge(Li) gamma-ray spectra to obtain peak areas corrected for detector efficiencies, from which are determined gamma-ray yields. These yields are compared with an isotope gamma-ray data file to determine the contributions to the observed spectrum from decay of specific radionuclides. A complete FORTRAN listing of the code and a complex test case are given.
Mathematical model and computer code for the analysis of advanced fast reactor dynamics
Schukin, N.V. (Moscow Engineering Physics Inst. (Russian Federation)); Korsun, A.S. (Moscow Engineering Physics Inst. (Russian Federation)); Vitruk, S.G. (Moscow Engineering Physics Inst. (Russian Federation)); Zimin, V.G. (Moscow Engineering Physics Inst. (Russian Federation)); Romanin, S.D. (Moscow Engineering Physics Inst. (Russian Federation))
1993-04-01
Efficient algorithms for mathematical modeling of 3-D neutron kinetics and thermal hydraulics are described. The model and appropriate computer code make it possible to analyze a variety of transient events ranging from normal operational states to catastrophic accident excursions. To verify the code, a number of calculations of different kind of transients was carried out. The results of the calculations show that the model and the computer code could be used for conceptual design of advanced liquid metal reactors. The detailed description of calculations of TOP WS accident is presented. (orig./DG)
Development of a system of computer codes for severe accident analyses and its applications
Chang, Soon Hong; Cheon, Moon Heon; Cho, Nam jin; No, Hui Cheon; Chang, Hyeon Seop; Moon, Sang Kee; Park, Seok Jeong; Chung, Jee Hwan [Korea Advanced Institute of Science and Technology, Taejon (Korea, Republic of)
1991-12-15
The objectives of this study is to develop a system of computer codes for postulated severe accident analyses in Nuclear Power Plants. This system of codes is necessary to conduct individual plant examination for domestic nuclear power plants. As a result of this study, one can conduct severe accident assessments more easily, and can extract the plant-specific vulnerabilities for severe accidents and at the same time the ideas for enhancing overall accident resistance. The scope and contents of this study are as follows : development of a system of computer codes for severe accident analyses, development of severe accident management strategy.
Proceedings of the conference on computer codes and the linear accelerator community
Cooper, R.K. (comp.)
1990-07-01
The conference whose proceedings you are reading was envisioned as the second in a series, the first having been held in San Diego in January 1988. The intended participants were those people who are actively involved in writing and applying computer codes for the solution of problems related to the design and construction of linear accelerators. The first conference reviewed many of the codes both extant and under development. This second conference provided an opportunity to update the status of those codes, and to provide a forum in which emerging new 3D codes could be described and discussed. The afternoon poster session on the second day of the conference provided an opportunity for extended discussion. All in all, this conference was felt to be quite a useful interchange of ideas and developments in the field of 3D calculations, parallel computation, higher-order optics calculations, and code documentation and maintenance for the linear accelerator community. A third conference is planned.
Exact Gap Computation for Code Coverage Metrics in ISO-C
Richter, Dirk; 10.4204/EPTCS.80.4
2012-01-01
Test generation and test data selection are difficult tasks for model based testing. Tests for a program can be meld to a test suite. A lot of research is done to quantify the quality and improve a test suite. Code coverage metrics estimate the quality of a test suite. This quality is fine, if the code coverage value is high or 100%. Unfortunately it might be impossible to achieve 100% code coverage because of dead code for example. There is a gap between the feasible and theoretical maximal possible code coverage value. Our review of the research indicates, none of current research is concerned with exact gap computation. This paper presents a framework to compute such gaps exactly in an ISO-C compatible semantic and similar languages. We describe an efficient approximation of the gap in all the other cases. Thus, a tester can decide if more tests might be able or necessary to achieve better coverage.
Visualization of elastic wavefields computed with a finite difference code
Larsen, S. [Lawrence Livermore National Lab., CA (United States); Harris, D.
1994-11-15
The authors have developed a finite difference elastic propagation model to simulate seismic wave propagation through geophysically complex regions. To facilitate debugging and to assist seismologists in interpreting the seismograms generated by the code, they have developed an X Windows interface that permits viewing of successive temporal snapshots of the (2D) wavefield as they are calculated. The authors present a brief video displaying the generation of seismic waves by an explosive source on a continent, which propagate to the edge of the continent then convert to two types of acoustic waves. This sample calculation was part of an effort to study the potential of offshore hydroacoustic systems to monitor seismic events occurring onshore.
Magnetohydrodynamic Simulation Code CANS+: Assessments and Applications
Matsumoto, Yosuke; Kudoh, Yuki; Kawashima, Tomohisa; Matsumoto, Jin; Takahashi, Hiroyuki R; Minoshima, Takashi; Zenitani, Seiji; Miyoshi, Takahiro; Matsumoto, Ryoji
2016-01-01
We present a new magnetohydrodynamic (MHD) simulation code with the aim of providing accurate numerical solutions to astrophysical phenomena where discontinuities, shock waves, and turbulence are inherently important. The code implements the HLLD approximate Riemann solver, the fifth-order-monotonicity-preserving interpolation scheme, and the hyperbolic divergence cleaning method for a magnetic field. This choice of schemes significantly improved numerical accuracy and stability, and saved computational costs in multidimensional problems. Numerical tests of one- and two-dimensional problems showed the advantages of using the high-order scheme by comparing with results from a standard second-order TVD scheme. The present code enabled us to explore long-term evolution of a three-dimensional global accretion disk, in which compressible MHD turbulence saturated at much higher levels via the magneto-rotational instability than that given by the second-order scheme owing to the adoption of the high-resolution, nume...
Introduction to error correcting codes in quantum computers
Salas, P J
2006-01-01
The goal of this paper is to review the theoretical basis for achieving a faithful quantum information transmission and processing in the presence of noise. Initially encoding and decoding, implementing gates and quantum error correction will be considered error free. Finally we will relax this non realistic assumption, introducing the quantum fault-tolerant concept. The existence of an error threshold permits to conclude that there is no physical law preventing a quantum computer from being built. An error model based on the depolarizing channel will be able to provide a simple estimation of the storage or memory computation error threshold: < 5.2 10-5. The encoding is made by means of the [[7,1,3
High-Performance Java Codes for Computational Fluid Dynamics
Riley, Christopher; Chatterjee, Siddhartha; Biswas, Rupak; Biegel, Bryan (Technical Monitor)
2001-01-01
The computational science community is reluctant to write large-scale computationally -intensive applications in Java due to concerns over Java's poor performance, despite the claimed software engineering advantages of its object-oriented features. Naive Java implementations of numerical algorithms can perform poorly compared to corresponding Fortran or C implementations. To achieve high performance, Java applications must be designed with good performance as a primary goal. This paper presents the object-oriented design and implementation of two real-world applications from the field of Computational Fluid Dynamics (CFD): a finite-volume fluid flow solver (LAURA, from NASA Langley Research Center), and an unstructured mesh adaptation algorithm (2D_TAG, from NASA Ames Research Center). This work builds on our previous experience with the design of high-performance numerical libraries in Java. We examine the performance of the applications using the currently available Java infrastructure and show that the Java version of the flow solver LAURA performs almost within a factor of 2 of the original procedural version. Our Java version of the mesh adaptation algorithm 2D_TAG performs within a factor of 1.5 of its original procedural version on certain platforms. Our results demonstrate that object-oriented software design principles are not necessarily inimical to high performance.
Compendium of computer codes for the safety analysis of fast breeder reactors
1977-10-01
The objective of the compendium is to provide the reader with a guide which briefly describes many of the computer codes used for liquid metal fast breeder reactor safety analyses, since it is for this system that most of the codes have been developed. The compendium is designed to address the following frequently asked questions from individuals in licensing and research and development activities: (1) What does the code do. (2) To what safety problems has it been applied. (3) What are the code's limitations. (4) What is being done to remove these limitations. (5) How does the code compare with experimental observations and other code predictions. (6) What reference documents are available.
Performance evaluation of moment-method codes on an Intel iPSC/860 hypercube computer
Klimkowski, K.; Ling, H. (Texas Univ., Austin (United States))
1993-09-01
An analytical evaluation is conducted of the performance of a moment-method code on a parallel computer, treating algorithmic complexity costs within the framework of matrix size and the 'subblock-size' matrix-partitioning parameter. A scaled-efficiencies analysis is conducted for the measured computation times of the matrix-fill operation and LU decomposition. 6 refs.
Computing the Feng-Rao distances for codes from order domains
Ruano Benito, Diego
2007-01-01
We compute the Feng–Rao distance of a code coming from an order domain with a simplicial value semigroup. The main tool is the Apéry set of a semigroup that can be computed using a Gröbner basis....
Safety and reliability in superconducting MHD magnets
Laverick, C.; Powell, J.; Hsieh, S.; Reich, M.; Botts, T.; Prodell, A.
1979-07-01
This compilation adapts studies on safety and reliability in fusion magnets to similar problems in superconducting MHD magnets. MHD base load magnet requirements have been identified from recent Francis Bitter National Laboratory reports and that of other contracts. Information relevant to this subject in recent base load magnet design reports for AVCO - Everett Research Laboratories and Magnetic Corporation of America is included together with some viewpoints from a BNL workshop on structural analysis needed for superconducting coils in magnetic fusion energy. A summary of design codes used in large bubble chamber magnet design is also included.
Development of MCNPX-ESUT computer code for simulation of neutron/gamma pulse height distribution
Abolfazl Hosseini, Seyed; Vosoughi, Naser; Zangian, Mehdi
2015-05-01
In this paper, the development of the MCNPX-ESUT (MCNPX-Energy Engineering of Sharif University of Technology) computer code for simulation of neutron/gamma pulse height distribution is reported. Since liquid organic scintillators like NE-213 are well suited and routinely used for spectrometry in mixed neutron/gamma fields, this type of detectors is selected for simulation in the present study. The proposed algorithm for simulation includes four main steps. The first step is the modeling of the neutron/gamma particle transport and their interactions with the materials in the environment and detector volume. In the second step, the number of scintillation photons due to charged particles such as electrons, alphas, protons and carbon nuclei in the scintillator material is calculated. In the third step, the transport of scintillation photons in the scintillator and lightguide is simulated. Finally, the resolution corresponding to the experiment is considered in the last step of the simulation. Unlike the similar computer codes like SCINFUL, NRESP7 and PHRESP, the developed computer code is applicable to both neutron and gamma sources. Hence, the discrimination of neutron and gamma in the mixed fields may be performed using the MCNPX-ESUT computer code. The main feature of MCNPX-ESUT computer code is that the neutron/gamma pulse height simulation may be performed without needing any sort of post processing. In the present study, the pulse height distributions due to a monoenergetic neutron/gamma source in NE-213 detector using MCNPX-ESUT computer code is simulated. The simulated neutron pulse height distributions are validated through comparing with experimental data (Gohil et al. Nuclear Instruments and Methods in Physics Research Section A: Accelerators, Spectrometers, Detectors and Associated Equipment, 664 (2012) 304-309.) and the results obtained from similar computer codes like SCINFUL, NRESP7 and Geant4. The simulated gamma pulse height distribution for a 137Cs
Kako, T.; Watanabe, T. [eds.
2000-06-01
This is the proceeding of 'study on numerical methods related to plasma confinement' held in National Institute for Fusion Science. In this workshop, theoretical and numerical analyses of possible plasma equilibria with their stability properties are presented. There are also various lectures on mathematical as well as numerical analyses related to the computational methods for fluid dynamics and plasma physics. Separate abstracts were presented for 13 of the papers in this report. The remaining 6 were considered outside the subject scope of INIS. (J.P.N.)
Robust Coding for Lossy Computing with Observation Costs
Ahmadi, Behzad
2011-01-01
An encoder wishes to minimize the bit rate necessary to guarantee that a decoder is able to calculate a symbol-wise function of a sequence available only at the encoder and a sequence that can be measured only at the decoder. This classical problem, first studied by Yamamoto, is addressed here by including two new aspects: (i) The decoder obtains noisy measurements of its sequence, where the quality of such measurements can be controlled via a cost-constrained "action" sequence, which is taken at the decoder or at the encoder; (ii) Measurement at the decoder may fail in a way that is unpredictable to the encoder, thus requiring robust encoding. The considered scenario generalizes known settings such as the Heegard-Berger-Kaspi and the "source coding with a vending machine" problems. The rate-distortion-cost function is derived in relevant special cases, along with general upper and lower bounds. Numerical examples are also worked out to obtain further insight into the optimal system design.
Automatic Parallelization Tool: Classification of Program Code for Parallel Computing
Mustafa Basthikodi
2016-04-01
Full Text Available Performance growth of single-core processors has come to a halt in the past decade, but was re-enabled by the introduction of parallelism in processors. Multicore frameworks along with Graphical Processing Units empowered to enhance parallelism broadly. Couples of compilers are updated to developing challenges forsynchronization and threading issues. Appropriate program and algorithm classifications will have advantage to a great extent to the group of software engineers to get opportunities for effective parallelization. In present work we investigated current species for classification of algorithms, in that related work on classification is discussed along with the comparison of issues that challenges the classification. The set of algorithms are chosen which matches the structure with different issues and perform given task. We have tested these algorithms utilizing existing automatic species extraction toolsalong with Bones compiler. We have added functionalities to existing tool, providing a more detailed characterization. The contributions of our work include support for pointer arithmetic, conditional and incremental statements, user defined types, constants and mathematical functions. With this, we can retain significant data which is not captured by original speciesof algorithms. We executed new theories into the device, empowering automatic characterization of program code.
The Uncertainty Test for the MAAP Computer Code
Park, S. H.; Song, Y. M.; Park, S. Y.; Ahn, K. I.; Kim, K. R.; Lee, Y. J. [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)
2008-10-15
After the Three Mile Island Unit 2 (TMI-2) and Chernobyl accidents, safety issues for a severe accident are treated in various aspects. Major issues in our research part include a level 2 PSA. The difficulty in expanding the level 2 PSA as a risk information activity is the uncertainty. In former days, it attached a weight to improve the quality in a internal accident PSA, but the effort is insufficient for decrease the phenomenon uncertainty in the level 2 PSA. In our country, the uncertainty degree is high in the case of a level 2 PSA model, and it is necessary to secure a model to decrease the uncertainty. We have not yet experienced the uncertainty assessment technology, the assessment system itself depends on advanced nations. In advanced nations, the severe accident simulator is implemented in the hardware level. But in our case, basic function in a software level can be implemented. In these circumstance at home and abroad, similar instances are surveyed such as UQM and MELCOR. Referred to these instances, SAUNA (Severe Accident UNcertainty Analysis) system is being developed in our project to assess and decrease the uncertainty in a level 2 PSA. It selects the MAAP code to analyze the uncertainty in a severe accident.
Development of a model and computer code to describe solar grade silicon production processes
Gould, R. K.; Srivastava, R.
1979-01-01
Two computer codes were developed for describing flow reactors in which high purity, solar grade silicon is produced via reduction of gaseous silicon halides. The first is the CHEMPART code, an axisymmetric, marching code which treats two phase flows with models describing detailed gas-phase chemical kinetics, particle formation, and particle growth. It can be used to described flow reactors in which reactants, mix, react, and form a particulate phase. Detailed radial gas-phase composition, temperature, velocity, and particle size distribution profiles are computed. Also, deposition of heat, momentum, and mass (either particulate or vapor) on reactor walls is described. The second code is a modified version of the GENMIX boundary layer code which is used to compute rates of heat, momentum, and mass transfer to the reactor walls. This code lacks the detailed chemical kinetics and particle handling features of the CHEMPART code but has the virtue of running much more rapidly than CHEMPART, while treating the phenomena occurring in the boundary layer in more detail.
Validation of physics and thermalhydraulic computer codes for advanced Candu reactor applications
Wren, D.J.; Popov, N.; Snell, V.G. [Atomic Energy of Canada Ltd, (Canada)
2004-07-01
Atomic Energy of Canada Ltd. (AECL) is developing an Advanced Candu Reactor (ACR) that is an evolutionary advancement of the currently operating Candu 6 reactors. The ACR is being designed to produce electrical power for a capital cost and at a unit-energy cost significantly less than that of the current reactor designs. The ACR retains the modular Candu concept of horizontal fuel channels surrounded by a heavy water moderator. However, ACR uses slightly enriched uranium fuel compared to the natural uranium used in Candu 6. This achieves the twin goals of improved economics (via large reductions in the heavy water moderator volume and replacement of the heavy water coolant with light water coolant) and improved safety. AECL has developed and implemented a software quality assurance program to ensure that its analytical, scientific and design computer codes meet the required standards for software used in safety analyses. Since the basic design of the ACR is equivalent to that of the Candu 6, most of the key phenomena associated with the safety analyses of ACR are common, and the Candu industry standard tool-set of safety analysis codes can be applied to the analysis of the ACR. A systematic assessment of computer code applicability addressing the unique features of the ACR design was performed covering the important aspects of the computer code structure, models, constitutive correlations, and validation database. Arising from this assessment, limited additional requirements for code modifications and extensions to the validation databases have been identified. This paper provides an outline of the AECL software quality assurance program process for the validation of computer codes used to perform physics and thermal-hydraulics safety analyses of the ACR. It describes the additional validation work that has been identified for these codes and the planned, and ongoing, experimental programs to extend the code validation as required to address specific ACR design
Issues in computational fluid dynamics code verification and validation
Oberkampf, W.L.; Blottner, F.G.
1997-09-01
A broad range of mathematical modeling errors of fluid flow physics and numerical approximation errors are addressed in computational fluid dynamics (CFD). It is strongly believed that if CFD is to have a major impact on the design of engineering hardware and flight systems, the level of confidence in complex simulations must substantially improve. To better understand the present limitations of CFD simulations, a wide variety of physical modeling, discretization, and solution errors are identified and discussed. Here, discretization and solution errors refer to all errors caused by conversion of the original partial differential, or integral, conservation equations representing the physical process, to algebraic equations and their solution on a computer. The impact of boundary conditions on the solution of the partial differential equations and their discrete representation will also be discussed. Throughout the article, clear distinctions are made between the analytical mathematical models of fluid dynamics and the numerical models. Lax`s Equivalence Theorem and its frailties in practical CFD solutions are pointed out. Distinctions are also made between the existence and uniqueness of solutions to the partial differential equations as opposed to the discrete equations. Two techniques are briefly discussed for the detection and quantification of certain types of discretization and grid resolution errors.
Fully Parallel MHD Stability Analysis Tool
Svidzinski, Vladimir; Galkin, Sergei; Kim, Jin-Soo; Liu, Yueqiang
2015-11-01
Progress on full parallelization of the plasma stability code MARS will be reported. MARS calculates eigenmodes in 2D axisymmetric toroidal equilibria in MHD-kinetic plasma models. It is a powerful tool for studying MHD and MHD-kinetic instabilities and it is widely used by fusion community. Parallel version of MARS is intended for simulations on local parallel clusters. It will be an efficient tool for simulation of MHD instabilities with low, intermediate and high toroidal mode numbers within both fluid and kinetic plasma models, already implemented in MARS. Parallelization of the code includes parallelization of the construction of the matrix for the eigenvalue problem and parallelization of the inverse iterations algorithm, implemented in MARS for the solution of the formulated eigenvalue problem. Construction of the matrix is parallelized by distributing the load among processors assigned to different magnetic surfaces. Parallelization of the solution of the eigenvalue problem is made by repeating steps of the present MARS algorithm using parallel libraries and procedures. Results of MARS parallelization and of the development of a new fix boundary equilibrium code adapted for MARS input will be reported. Work is supported by the U.S. DOE SBIR program.
Computer code simulations of explosions in flow networks and comparison with experiments
Gregory, W. S.; Nichols, B. D.; Moore, J. A.; Smith, P. R.; Steinke, R. G.; Idzorek, R. D.
1987-10-01
A program of experimental testing and computer code development for predicting the effects of explosions in air-cleaning systems is being carried out for the Department of Energy. This work is a combined effort by the Los Alamos National Laboratory and New Mexico State University (NMSU). Los Alamos has the lead responsibility in the project and develops the computer codes; NMSU performs the experimental testing. The emphasis in the program is on obtaining experimental data to verify the analytical work. The primary benefit of this work will be the development of a verified computer code that safety analysts can use to analyze the effects of hypothetical explosions in nuclear plant air cleaning systems. The experimental data show the combined effects of explosions in air-cleaning systems that contain all of the important air-cleaning elements (blowers, dampers, filters, ductwork, and cells). A small experimental set-up consisting of multiple rooms, ductwork, a damper, a filter, and a blower was constructed. Explosions were simulated with a shock tube, hydrogen/air-filled gas balloons, and blasting caps. Analytical predictions were made using the EVENT84 and NF85 computer codes. The EVENT84 code predictions were in good agreement with the effects of the hydrogen/air explosions, but they did not model the blasting cap explosions adequately. NF85 predicted shock entrance to and within the experimental set-up very well. The NF85 code was not used to model the hydrogen/air or blasting cap explosions.
Algorithms and computer codes for atomic and molecular quantum scattering theory
Thomas, L. (ed.)
1979-01-01
This workshop has succeeded in bringing up 11 different coupled equation codes on the NRCC computer, testing them against a set of 24 different test problems and making them available to the user community. These codes span a wide variety of methodologies, and factors of up to 300 were observed in the spread of computer times on specific problems. A very effective method was devised for examining the performance of the individual codes in the different regions of the integration range. Many of the strengths and weaknesses of the codes have been identified. Based on these observations, a hybrid code has been developed which is significantly superior to any single code tested. Thus, not only have the original goals been fully met, the workshop has resulted directly in an advancement of the field. All of the computer programs except VIVS are available upon request from the NRCC. Since an improved version of VIVS is contained in the hybrid program, VIVAS, it was not made available for distribution. The individual program LOGD is, however, available. In addition, programs which compute the potential energy matrices of the test problems are also available. The software library names for Tests 1, 2 and 4 are HEH2, LICO, and EN2, respectively.
Carbajo, Juan (Oak Ridge National Laboratory, Oak Ridge, TN); Jeong, Hae-Yong (Korea Atomic Energy Research Institute, Daejeon, Korea); Wigeland, Roald (Idaho National Laboratory, Idaho Falls, ID); Corradini, Michael (University of Wisconsin, Madison, WI); Schmidt, Rodney Cannon; Thomas, Justin (Argonne National Laboratory, Argonne, IL); Wei, Tom (Argonne National Laboratory, Argonne, IL); Sofu, Tanju (Argonne National Laboratory, Argonne, IL); Ludewig, Hans (Brookhaven National Laboratory, Upton, NY); Tobita, Yoshiharu (Japan Atomic Energy Agency, Ibaraki-ken, Japan); Ohshima, Hiroyuki (Japan Atomic Energy Agency, Ibaraki-ken, Japan); Serre, Frederic (Centre d' %C3%94etudes nucl%C3%94eaires de Cadarache %3CU%2B2013%3E CEA, France)
2011-06-01
This report summarizes the results of an expert-opinion elicitation activity designed to qualitatively assess the status and capabilities of currently available computer codes and models for accident analysis and reactor safety calculations of advanced sodium fast reactors, and identify important gaps. The twelve-member panel consisted of representatives from five U.S. National Laboratories (SNL, ANL, INL, ORNL, and BNL), the University of Wisconsin, the KAERI, the JAEA, and the CEA. The major portion of this elicitation activity occurred during a two-day meeting held on Aug. 10-11, 2010 at Argonne National Laboratory. There were two primary objectives of this work: (1) Identify computer codes currently available for SFR accident analysis and reactor safety calculations; and (2) Assess the status and capability of current US computer codes to adequately model the required accident scenarios and associated phenomena, and identify important gaps. During the review, panel members identified over 60 computer codes that are currently available in the international community to perform different aspects of SFR safety analysis for various event scenarios and accident categories. A brief description of each of these codes together with references (when available) is provided. An adaptation of the Predictive Capability Maturity Model (PCMM) for computational modeling and simulation is described for use in this work. The panel's assessment of the available US codes is presented in the form of nine tables, organized into groups of three for each of three risk categories considered: anticipated operational occurrences (AOOs), design basis accidents (DBA), and beyond design basis accidents (BDBA). A set of summary conclusions are drawn from the results obtained. At the highest level, the panel judged that current US code capabilities are adequate for licensing given reasonable margins, but expressed concern that US code development activities had stagnated and that the
无
2010-01-01
The eukaryotic genome contains varying numbers of non-coding RNA(ncRNA) genes."Computational RNomics" takes a multidisciplinary approach,like information science,to resolve the structure and function of ncRNAs.Here,we review the main issues in "Computational RNomics" of data storage and management,ncRNA gene identification and characterization,ncRNA target identification and functional prediction,and we summarize the main methods and current content of "computational RNomics".
[Vascular assessment in stroke codes: role of computed tomography angiography].
Mendigaña Ramos, M; Cabada Giadas, T
2015-01-01
Advances in imaging studies for acute ischemic stroke are largely due to the development of new efficacious treatments carried out in the acute phase. Together with computed tomography (CT) perfusion studies, CT angiography facilitates the selection of patients who are likely to benefit from appropriate early treatment. CT angiography plays an important role in the workup for acute ischemic stroke because it makes it possible to confirm vascular occlusion, assess the collateral circulation, and obtain an arterial map that is very useful for planning endovascular treatment. In this review about CT angiography, we discuss the main technical characteristics, emphasizing the usefulness of the technique in making the right diagnosis and improving treatment strategies. Copyright © 2012 SERAM. Published by Elsevier España, S.L.U. All rights reserved.
Symbolic coding for noninvertible systems: uniform approximation and numerical computation
Beyn, Wolf-Jürgen; Hüls, Thorsten; Schenke, Andre
2016-11-01
It is well known that the homoclinic theorem, which conjugates a map near a transversal homoclinic orbit to a Bernoulli subshift, extends from invertible to specific noninvertible dynamical systems. In this paper, we provide a unifying approach that combines such a result with a fully discrete analog of the conjugacy for finite but sufficiently long orbit segments. The underlying idea is to solve appropriate discrete boundary value problems in both cases, and to use the theory of exponential dichotomies to control the errors. This leads to a numerical approach that allows us to compute the conjugacy to any prescribed accuracy. The method is demonstrated for several examples where invertibility of the map fails in different ways.
Benchmark Problems Used to Assess Computational Aeroacoustics Codes
Dahl, Milo D.; Envia, Edmane
2005-01-01
The field of computational aeroacoustics (CAA) encompasses numerical techniques for calculating all aspects of sound generation and propagation in air directly from fundamental governing equations. Aeroacoustic problems typically involve flow-generated noise, with and without the presence of a solid surface, and the propagation of the sound to a receiver far away from the noise source. It is a challenge to obtain accurate numerical solutions to these problems. The NASA Glenn Research Center has been at the forefront in developing and promoting the development of CAA techniques and methodologies for computing the noise generated by aircraft propulsion systems. To assess the technological advancement of CAA, Glenn, in cooperation with the Ohio Aerospace Institute and the AeroAcoustics Research Consortium, organized and hosted the Fourth CAA Workshop on Benchmark Problems. Participants from industry and academia from both the United States and abroad joined to present and discuss solutions to benchmark problems. These demonstrated technical progress ranging from the basic challenges to accurate CAA calculations to the solution of CAA problems of increasing complexity and difficulty. The results are documented in the proceedings of the workshop. Problems were solved in five categories. In three of the five categories, exact solutions were available for comparison with CAA results. A fourth category of problems representing sound generation from either a single airfoil or a blade row interacting with a gust (i.e., problems relevant to fan noise) had approximate analytical or completely numerical solutions. The fifth category of problems involved sound generation in a viscous flow. In this case, the CAA results were compared with experimental data.
Scherer, W.; Brockmann, H.; Haas, K. A.; Rütten, H. J.
2005-01-01
V.S.O.P. is a computer code system for the comprehensive numerical simulation of the physics of thermal reactors. It implies the setup of the reactor and of the fuel element, processing of cross sections, neutron spectrum evaluation, neutron diffusion calculation in two or three dimensions, fuel burnup, fuel shuffling, reactor control, thermal hydraulics and fuel cycle costs. The thermal hydraulics part (steady state and time-dependent) is restricted to HTRs and to two spatial dimensions. The...
Bordy, J.M.; Kodeli, I.; Menard, St.; Bouchet, J.L.; Renard, F.; Martin, E.; Blazy, L.; Voros, S.; Bochud, F.; Laedermann, J.P.; Beaugelin, K.; Makovicka, L.; Quiot, A.; Vermeersch, F.; Roche, H.; Perrin, M.C.; Laye, F.; Bardies, M.; Struelens, L.; Vanhavere, F.; Gschwind, R.; Fernandez, F.; Quesne, B.; Fritsch, P.; Lamart, St.; Crovisier, Ph.; Leservot, A.; Antoni, R.; Huet, Ch.; Thiam, Ch.; Donadille, L.; Monfort, M.; Diop, Ch.; Ricard, M
2006-07-01
The purpose of this conference was to describe the present state of computer codes dedicated to radiation transport or radiation source assessment or dosimetry. The presentations have been parted into 2 sessions: 1) methodology and 2) uses in industrial or medical or research domains. It appears that 2 different calculation strategies are prevailing, both are based on preliminary Monte-Carlo calculations with data storage. First, quick simulations made from a database of particle histories built though a previous Monte-Carlo simulation and secondly, a neuronal approach involving a learning platform generated through a previous Monte-Carlo simulation. This document gathers the slides of the presentations.
MHD control in burning plasmas MHD control in burning plasmas
Donné, Tony; Liang, Yunfeng
2012-07-01
in the field of burn control is to find the proper balance between desired and detrimental effects of the various MHD modes and to develop the methods and tools for active feedback control of MHD modes in burning plasmas. Therefore, it is necessary to understand the dynamics of the system, in this case the mutual interactions between the fast alpha particles and the MHD instabilities. Since burning plasmas do not yet exist, the relevant experimental work until ITER comes into full operation needs to be largely based on alpha-particle simulation experiments in which the alpha particles are accelerated to high energies by means of special heating techniques. The precise conditions of a burning plasma can be only partly mimicked in present tokamaks. Hence, also a detailed computational modelling effort is needed, in order to understand the impact of findings in present machines for those of the future. In 2011 two dedicated workshops were devoted to MHD control. Firstly, there was a workshop on Control of Burning Plasmas that took place from 21-25 March 2011 at the Lorentz Centre in Leiden, The Netherlands. Secondly, the 480th Wilhelm and Else Heraeus Seminar that took place from 16-18 June in Bad Honnef, Germany was devoted to Active Control of Instabilities in Hot Plasmas. This special issue presents a collection of papers that have been presented at the two workshops, along with a few papers that are the result of an open call to contribute to this special issue.
Marchand, Pierre; Chabrier, Gilles; Hennebelle, Patrick; Commerçon, Benoit; Vaytet, Neil
2016-01-01
We develop a detailed chemical network relevant to the conditions characteristic of prestellar core collapse. We solve the system of time-dependent differential equations to calculate the equilibrium abundances of molecules and dust grains, with a size distribution given by size-bins for these latter. These abundances are used to compute the different non-ideal magneto-hydrodynamics resistivities (ambipolar, Ohmic and Hall), needed to carry out simulations of protostellar collapse. For the first time in this context, we take into account the evaporation of the grains, the thermal ionisation of Potassium, Sodium and Hydrogen at high temperature, and the thermionic emission of grains in the chemical network, and we explore the impact of various cosmic ray ionisation rates. All these processes significantly affect the non-ideal magneto-hydrodynamics resistivities, which will modify the dynamics of the collapse. Ambipolar diffusion and Hall effect dominate at low densities, up to n_H = 10^12 cm^-3, after which Oh...
Hoffman, F. O.; Miller, C. W.; Shaeffer, D. L.; Garten, Jr., C. T.; Shor, R. W.; Ensminger, J. T.
1977-04-01
The objective of this paper is to present a compilation of computer codes for the assessment of accidental or routine releases of radioactivity to the environment from nuclear power facilities. The capabilities of 83 computer codes in the areas of environmental transport and radiation dosimetry are summarized in tabular form. This preliminary analysis clearly indicates that the initial efforts in assessment methodology development have concentrated on atmospheric dispersion, external dosimetry, and internal dosimetry via inhalation. The incorporation of terrestrial and aquatic food chain pathways has been a more recent development and reflects the current requirements of environmental legislation and the needs of regulatory agencies. The characteristics of the conceptual models employed by these codes are reviewed. The appendixes include abstracts of the codes and indexes by author, key words, publication description, and title.
Compendium of computer codes for the researcher in magnetic fusion energy
Porter, G.D. (ed.)
1989-03-10
This is a compendium of computer codes, which are available to the fusion researcher. It is intended to be a document that permits a quick evaluation of the tools available to the experimenter who wants to both analyze his data, and compare the results of his analysis with the predictions of available theories. This document will be updated frequently to maintain its usefulness. I would appreciate receiving further information about codes not included here from anyone who has used them. The information required includes a brief description of the code (including any special features), a bibliography of the documentation available for the code and/or the underlying physics, a list of people to contact for help in running the code, instructions on how to access the code, and a description of the output from the code. Wherever possible, the code contacts should include people from each of the fusion facilities so that the novice can talk to someone ''down the hall'' when he first tries to use a code. I would also appreciate any comments about possible additions and improvements in the index. I encourage any additional criticism of this document. 137 refs.
Mann, F.M.
1998-01-26
The Tank Waste Remediation System (TWRS) is responsible for the safe storage, retrieval, and disposal of waste currently being held in 177 underground tanks at the Hanford Site. In order to successfully carry out its mission, TWRS must perform environmental analyses describing the consequences of tank contents leaking from tanks and associated facilities during the storage, retrieval, or closure periods and immobilized low-activity tank waste contaminants leaving disposal facilities. Because of the large size of the facilities and the great depth of the dry zone (known as the vadose zone) underneath the facilities, sophisticated computer codes are needed to model the transport of the tank contents or contaminants. This document presents the code selection criteria for those vadose zone analyses (a subset of the above analyses) where the hydraulic properties of the vadose zone are constant in time the geochemical behavior of the contaminant-soil interaction can be described by simple models, and the geologic or engineered structures are complicated enough to require a two-or three dimensional model. Thus, simple analyses would not need to use the fairly sophisticated codes which would meet the selection criteria in this document. Similarly, those analyses which involve complex chemical modeling (such as those analyses involving large tank leaks or those analyses involving the modeling of contaminant release from glass waste forms) are excluded. The analyses covered here are those where the movement of contaminants can be relatively simply calculated from the moisture flow. These code selection criteria are based on the information from the low-level waste programs of the US Department of Energy (DOE) and of the US Nuclear Regulatory Commission as well as experience gained in the DOE Complex in applying these criteria. Appendix table A-1 provides a comparison between the criteria in these documents and those used here. This document does not define the models (that
POPCYCLE: a computer code for calculating nuclear and fossil plant levelized life-cycle power costs
Hardie, R.W.
1982-02-01
POPCYCLE, a computer code designed to calculate levelized life-cycle power costs for nuclear and fossil electrical generating plants is described. Included are (1) derivations of the equations and a discussion of the methodology used by POPCYCLE, (2) a description of the input required by the code, (3) a listing of the input for a sample case, and (4) the output for a sample case.
Fault-tolerant quantum computation with asymmetric Bacon-Shor codes
Brooks, Peter; Preskill, John
2013-03-01
We develop a scheme for fault-tolerant quantum computation based on asymmetric Bacon-Shor codes, which works effectively against highly biased noise dominated by dephasing. We find the optimal Bacon-Shor block size as a function of the noise strength and the noise bias, and estimate the logical error rate and overhead cost achieved by this optimal code. Our fault-tolerant gadgets, based on gate teleportation, are well suited for hardware platforms with geometrically local gates in two dimensions.
HIFI: a computer code for projectile fragmentation accompanied by incomplete fusion
Wu, J.R.
1980-07-01
A brief summary of a model proposed to describe projectile fragmentation accompanied by incomplete fusion and the instructions for the use of the computer code HIFI are given. The code HIFI calculates single inclusive spectra, coincident spectra and excitation functions resulting from particle-induced reactions. It is a multipurpose program which can calculate any type of coincident spectra as long as the reaction is assumed to take place in two steps.
SAMDIST: A Computer Code for Calculating Statistical Distributions for R-Matrix Resonance Parameters
Leal, L.C.
1995-01-01
The: SAMDIST computer code has been developed to calculate distribution of resonance parameters of the Reich-Moore R-matrix type. The program assumes the parameters are in the format compatible with that of the multilevel R-matrix code SAMMY. SAMDIST calculates the energy-level spacing distribution, the resonance width distribution, and the long-range correlation of the energy levels. Results of these calculations are presented in both graphic and tabular forms.
SAMDIST A Computer Code for Calculating Statistical Distributions for R-Matrix Resonance Parameters
Leal, L C
1995-01-01
The: SAMDIST computer code has been developed to calculate distribution of resonance parameters of the Reich-Moore R-matrix type. The program assumes the parameters are in the format compatible with that of the multilevel R-matrix code SAMMY. SAMDIST calculates the energy-level spacing distribution, the resonance width distribution, and the long-range correlation of the energy levels. Results of these calculations are presented in both graphic and tabular forms.
The development of an intelligent interface to a computational fluid dynamics flow-solver code
Williams, Anthony D.
1988-01-01
Researchers at NASA Lewis are currently developing an 'intelligent' interface to aid in the development and use of large, computational fluid dynamics flow-solver codes for studying the internal fluid behavior of aerospace propulsion systems. This paper discusses the requirements, design, and implementation of an intelligent interface to Proteus, a general purpose, three-dimensional, Navier-Stokes flow solver. The interface is called PROTAIS to denote its introduction of artificial intelligence (AI) concepts to the Proteus code.
ANL/HTP: a computer code for the simulation of heat pipe operation
McLennan, G.A.
1983-11-01
ANL/HTP is a computer code for the simulation of heat pipe operation, to predict heat pipe performance and temperature distributions during steady state operation. Source and sink temperatures and heat transfer coefficients can be set as input boundary conditions, and varied for parametric studies. Five code options are included to calculate performance for fixed operating conditions, or to vary any one of the four boundary conditions to determine the heat pipe limited performance. The performance limits included are viscous, sonic, entrainment capillary, and boiling, using the best available theories to model these effects. The code has built-in models for a number of wick configurations - open grooves, screen-covered grooves, screen-wrap, and arteries, with provision for expansion. The current version of the code includes the thermophysical properties of sodium as the working fluid in an expandable subroutine. The code-calculated performance agrees quite well with measured experiment data.
LEADS-DC: A computer code for intense dc beam nonlinear transport simulation
无
2011-01-01
An intense dc beam nonlinear transport code has been developed. The code is written in Visual FORTRAN 6.6 and has ~13000 lines. The particle distribution in the transverse cross section is uniform or Gaussian. The space charge forces are calculated by the PIC (particle in cell) scheme, and the effects of the applied fields on the particle motion are calculated with the Lie algebraic method through the third order approximation. Obviously,the solutions to the equations of particle motion are self-consistent. The results obtained from the theoretical analysis have been put in the computer code. Many optical beam elements are contained in the code. So, the code can simulate the intense dc particle motions in the beam transport lines, high voltage dc accelerators and ion implanters.
Benyo, Theresa Louise
Historically, the National Aeronautics and Space Administration (NASA) has used rocket-powered vehicles as launch vehicles for access to space. A familiar example is the Space Shuttle launch system. These vehicles carry both fuel and oxidizer onboard. If an external oxidizer (such as the Earth's atmosphere) is utilized, the need to carry an onboard oxidizer is eliminated, and future launch vehicles could carry a larger payload into orbit at a fraction of the total fuel expenditure. For this reason, NASA is currently researching the use of air-breathing engines to power the first stage of two-stage-to-orbit hypersonic launch systems. Removing the need to carry an onboard oxidizer leads also to reductions in total vehicle weight at liftoff. This in turn reduces the total mass of propellant required, and thus decreases the cost of carrying a specific payload into orbit or beyond. However, achieving hypersonic flight with air-breathing jet engines has several technical challenges. These challenges, such as the mode transition from supersonic to hypersonic engine operation, are under study in NASA's Fundamental Aeronautics Program. One propulsion concept that is being explored is a magnetohydrodynamic (MHD) energy- bypass generator coupled with an off-the-shelf turbojet/turbofan. It is anticipated that this engine will be capable of operation from takeoff to Mach 7 in a single flowpath without mode transition. The MHD energy bypass consists of an MHD generator placed directly upstream of the engine, and converts a portion of the enthalpy of the inlet flow through the engine into electrical current. This reduction in flow enthalpy corresponds to a reduced Mach number at the turbojet inlet so that the engine stays within its design constraints. Furthermore, the generated electrical current may then be used to power aircraft systems or an MHD accelerator positioned downstream of the turbojet. The MHD accelerator operates in reverse of the MHD generator, re-accelerating the
MARG2D code. 1. Eigenvalue problem for two dimensional Newcomb equation
Tokuda, Shinji [Japan Atomic Energy Research Inst., Naka, Ibaraki (Japan). Naka Fusion Research Establishment; Watanabe, Tomoko
1997-10-01
A new method and a code MARG2D have been developed to solve the 2-dimensional Newcomb equation which plays an important role in the magnetohydrodynamic (MHD) stability analysis in an axisymmetric toroidal plasma such as a tokamak. In the present formulation, an eigenvalue problem is posed for the 2-D Newcomb equation, where the weight function (the kinetic energy integral) and the boundary conditions at rational surfaces are chosen so that an eigenfunction correctly behaves as the linear combination of the small solution and the analytical solutions around each of the rational surfaces. Thus, the difficulty on solving the 2-D Newcomb equation has been resolved. By using the MARG2D code, the ideal MHD marginally stable state can be identified for a 2-D toroidal plasma. The code is indispensable on computing the outer-region matching data necessary for the resistive MHD stability analysis. Benchmark with ERATOJ, an ideal MHD stability code, has been carried out and the MARG2D code demonstrates that it indeed identifies both stable and marginally stable states against ideal MHD motion. (author)
SCALE: A modular code system for performing standardized computer analyses for licensing evaluation
NONE
1997-03-01
This Manual represents Revision 5 of the user documentation for the modular code system referred to as SCALE. The history of the SCALE code system dates back to 1969 when the current Computational Physics and Engineering Division at Oak Ridge National Laboratory (ORNL) began providing the transportation package certification staff at the U.S. Atomic Energy Commission with computational support in the use of the new KENO code for performing criticality safety assessments with the statistical Monte Carlo method. From 1969 to 1976 the certification staff relied on the ORNL staff to assist them in the correct use of codes and data for criticality, shielding, and heat transfer analyses of transportation packages. However, the certification staff learned that, with only occasional use of the codes, it was difficult to become proficient in performing the calculations often needed for an independent safety review. Thus, shortly after the move of the certification staff to the U.S. Nuclear Regulatory Commission (NRC), the NRC staff proposed the development of an easy-to-use analysis system that provided the technical capabilities of the individual modules with which they were familiar. With this proposal, the concept of the Standardized Computer Analyses for Licensing Evaluation (SCALE) code system was born. This manual covers an array of modules written for the SCALE package, consisting of drivers, system libraries, cross section and materials properties libraries, input/output routines, storage modules, and help files.
ASHMET: a computer code for estimating insolation incident on tilted surfaces
Elkin, R.F.; Toelle, R.G.
1980-05-01
A computer code, ASHMET, has been developed by MSFC to estimate the amount of solar insolation incident on the surfaces of solar collectors. Both tracking and fixed-position collectors have been included. Climatological data for 248 US locations are built into the code. This report describes the methodology of the code, and its input and output. The basic methodology used by ASHMET is the ASHRAE clear-day insolation relationships modified by a clearness index derived from SOLMET-measured solar radiation data to a horizontal surface.
Schnack, Dalton D.
In this lecture we will examine some simple examples of MHD equilibrium configurations. These will all be in cylindrical geometry. They form the basis for more complicated equilibrium states in toroidal geometry.
Solution of 3-dimensional time-dependent viscous flows. Part 2: Development of the computer code
Weinberg, B. C.; Mcdonald, H.
1980-01-01
There is considerable interest in developing a numerical scheme for solving the time dependent viscous compressible three dimensional flow equations to aid in the design of helicopter rotors. The development of a computer code to solve a three dimensional unsteady approximate form of the Navier-Stokes equations employing a linearized block emplicit technique in conjunction with a QR operator scheme is described. Results of calculations of several Cartesian test cases are presented. The computer code can be applied to more complex flow fields such as these encountered on rotating airfoils.
Aeschliman, D. P.; Oberkampf, W. L.; Blottner, F. G.
Verification, calibration, and validation (VCV) of Computational Fluid Dynamics (CFD) codes is an essential element of the code development process. The exact manner in which code VCV activities are planned and conducted, however, is critically important. It is suggested that the way in which code validation, in particular, is often conducted--by comparison to published experimental data obtained for other purposes--is in general difficult and unsatisfactory, and that a different approach is required. This paper describes a proposed methodology for CFD code VCV that meets the technical requirements and is philosophically consistent with code development needs. The proposed methodology stresses teamwork and cooperation between code developers and experimentalists throughout the VCV process, and takes advantage of certain synergisms between CFD and experiment. A novel approach to uncertainty analysis is described which can both distinguish between and quantify various types of experimental error, and whose attributes are used to help define an appropriate experimental design for code VCV experiments. The methodology is demonstrated with an example of laminar, hypersonic, near perfect gas, 3-dimensional flow over a sliced sphere/cone of varying geometrical complexity.
NADAC and MERGE: computer codes for processing neutron activation analysis data
Heft, R.E.; Martin, W.E.
1977-05-19
Absolute disintegration rates of specific radioactive products induced by neutron irradition of a sample are determined by spectrometric analysis of gamma-ray emissions. Nuclide identification and quantification is carried out by a complex computer code GAMANAL (described elsewhere). The output of GAMANAL is processed by NADAC, a computer code that converts the data on observed distintegration rates to data on the elemental composition of the original sample. Computations by NADAC are on an absolute basis in that stored nuclear parameters are used rather than the difference between the observed disintegration rate and the rate obtained by concurrent irradiation of elemental standards. The NADAC code provides for the computation of complex cases including those involving interrupted irradiations, parent and daughter decay situations where the daughter may also be produced independently, nuclides with very short half-lives compared to counting interval, and those involving interference by competing neutron-induced reactions. The NADAC output consists of a printed report, which summarizes analytical results, and a card-image file, which can be used as input to another computer code MERGE. The purpose of MERGE is to combine the results of multiple analyses and produce a single final answer, based on all available information, for each element found.
Baes, C.F. III; Sharp, R.D.; Sjoreen, A.L.; Hermann, O.W.
1984-11-01
TERRA is a computer code which calculates concentrations of radionuclides and ingrowing daughters in surface and root-zone soil, produce and feed, beef, and milk from a given deposition rate at any location in the conterminous United States. The code is fully integrated with seven other computer codes which together comprise a Computerized Radiological Risk Investigation System, CRRIS. Output from either the long range (> 100 km) atmospheric dispersion code RETADD-II or the short range (<80 km) atmospheric dispersion code ANEMOS, in the form of radionuclide air concentrations and ground deposition rates by downwind location, serves as input to TERRA. User-defined deposition rates and air concentrations may also be provided as input to TERRA through use of the PRIMUS computer code. The environmental concentrations of radionuclides predicted by TERRA serve as input to the ANDROS computer code which calculates population and individual intakes, exposures, doses, and risks. TERRA incorporates models to calculate uptake from soil and atmospheric deposition on four groups of produce for human consumption and four groups of livestock feeds. During the environmental transport simulation, intermediate calculations of interception fraction for leafy vegetables, produce directly exposed to atmospherically depositing material, pasture, hay, and silage are made based on location-specific estimates of standing crop biomass. Pasture productivity is estimated by a model which considers the number and types of cattle and sheep, pasture area, and annual production of other forages (hay and silage) at a given location. Calculations are made of the fraction of grain imported from outside the assessment area. TERRA output includes the above calculations and estimated radionuclide concentrations in plant produce, milk, and a beef composite by location.
A Novel High-Order, Entropy Stable, 3D AMR MHD Solver with Guaranteed Positive Pressure
Derigs, Dominik; Gassner, Gregor J; Walch, Stefanie
2016-01-01
We describe a high-order numerical magnetohydrodynamics (MHD) solver built upon a novel non-linear entropy stable numerical flux function that supports eight travelling wave solutions. By construction the solver conserves mass, momentum, and energy and is entropy stable. The method is designed to treat the divergence-free constraint on the magnetic field in a similar fashion to a hyperbolic divergence cleaning technique. The solver described herein is especially well-suited for flows involving strong discontinuities. Furthermore, we present a new formulation to guarantee positivity of the pressure. We present the underlying theory and implementation of the new solver into the multi-physics, multi-scale adaptive mesh refinement (AMR) simulation code $\\texttt{FLASH}$ (http://flash.uchicago.edu). The accuracy, robustness and computational efficiency is demonstrated with a number of tests, including comparisons to available MHD implementations in $\\texttt{FLASH}$.
Lilley, D. G.; Rhode, D. L.
1982-01-01
A primitive pressure-velocity variable finite difference computer code was developed to predict swirling recirculating inert turbulent flows in axisymmetric combustors in general, and for application to a specific idealized combustion chamber with sudden or gradual expansion. The technique involves a staggered grid system for axial and radial velocities, a line relaxation procedure for efficient solution of the equations, a two-equation k-epsilon turbulence model, a stairstep boundary representation of the expansion flow, and realistic accommodation of swirl effects. A user's manual, dealing with the computational problem, showing how the mathematical basis and computational scheme may be translated into a computer program is presented. A flow chart, FORTRAN IV listing, notes about various subroutines and a user's guide are supplied as an aid to prospective users of the code.
Feasibility of MHD submarine propulsion
Doss, E.D. (ed.) (Argonne National Lab., IL (United States)); Sikes, W.C. (ed.) (Newport News Shipbuilding and Dry Dock Co., VA (United States))
1992-09-01
This report describes the work performed during Phase 1 and Phase 2 of the collaborative research program established between Argonne National Laboratory (ANL) and Newport News Shipbuilding and Dry Dock Company (NNS). Phase I of the program focused on the development of computer models for Magnetohydrodynamic (MHD) propulsion. Phase 2 focused on the experimental validation of the thruster performance models and the identification, through testing, of any phenomena which may impact the attractiveness of this propulsion system for shipboard applications. The report discusses in detail the work performed in Phase 2 of the program. In Phase 2, a two Tesla test facility was designed, built, and operated. The facility test loop, its components, and their design are presented. The test matrix and its rationale are discussed. Representative experimental results of the test program are presented, and are compared to computer model predictions. In general, the results of the tests and their comparison with the predictions indicate that thephenomena affecting the performance of MHD seawater thrusters are well understood and can be accurately predicted with the developed thruster computer models.
Just-in-Time Compilation-Inspired Methodology for Parallelization of Compute Intensive Java Code
GHULAM MUSTAFA
2017-01-01
Full Text Available Compute intensive programs generally consume significant fraction of execution time in a small amount of repetitive code. Such repetitive code is commonly known as hotspot code. We observed that compute intensive hotspots often possess exploitable loop level parallelism. A JIT (Just-in-Time compiler profiles a running program to identify its hotspots. Hotspots are then translated into native code, for efficient execution. Using similar approach, we propose a methodology to identify hotspots and exploit their parallelization potential on multicore systems. Proposed methodology selects and parallelizes each DOALL loop that is either contained in a hotspot method or calls a hotspot method. The methodology could be integrated in front-end of a JIT compiler to parallelize sequential code, just before native translation. However, compilation to native code is out of scope of this work. As a case study, we analyze eighteen JGF (Java Grande Forum benchmarks to determine parallelization potential of hotspots. Eight benchmarks demonstrate a speedup of up to 7.6x on an 8-core system
Development of a computer code for thermal hydraulics of reactors (THOR). [BWR and PWR
Wulff, W
1975-01-01
The purpose of the advanced code development work is to construct a computer code for the prediction of thermohydraulic transients in water-cooled nuclear reactor systems. The fundamental formulation of fluid dynamics is to be based on the one-dimensional drift flux model for non-homogeneous, non-equilibrium flows of two-phase mixtures. Particular emphasis is placed on component modeling, automatic prediction of initial steady state conditions, inclusion of one-dimensional transient neutron kinetics, freedom in the selection of computed spatial detail, development of reliable constitutive descriptions, and modular code structure. Numerical solution schemes have been implemented to integrate simultaneously the one-dimensional transient drift flux equations. The lumped-parameter modeling analyses of thermohydraulic transients in the reactor core and in the pressurizer have been completed. The code development for the prediction of the initial steady state has been completed with preliminary representation of individual reactor system components. A program has been developed to predict critical flow expanding from a dead-ended pipe; the computed results have been compared and found in good agreement with idealized flow solutions. Transport properties for liquid water and water vapor have been coded and verified.
PIC codes for plasma accelerators on emerging computer architectures (GPUS, Multicore/Manycore CPUS)
Vincenti, Henri
2016-03-01
The advent of exascale computers will enable 3D simulations of a new laser-plasma interaction regimes that were previously out of reach of current Petasale computers. However, the paradigm used to write current PIC codes will have to change in order to fully exploit the potentialities of these new computing architectures. Indeed, achieving Exascale computing facilities in the next decade will be a great challenge in terms of energy consumption and will imply hardware developments directly impacting our way of implementing PIC codes. As data movement (from die to network) is by far the most energy consuming part of an algorithm future computers will tend to increase memory locality at the hardware level and reduce energy consumption related to data movement by using more and more cores on each compute nodes (''fat nodes'') that will have a reduced clock speed to allow for efficient cooling. To compensate for frequency decrease, CPU machine vendors are making use of long SIMD instruction registers that are able to process multiple data with one arithmetic operator in one clock cycle. SIMD register length is expected to double every four years. GPU's also have a reduced clock speed per core and can process Multiple Instructions on Multiple Datas (MIMD). At the software level Particle-In-Cell (PIC) codes will thus have to achieve both good memory locality and vectorization (for Multicore/Manycore CPU) to fully take advantage of these upcoming architectures. In this talk, we present the portable solutions we implemented in our high performance skeleton PIC code PICSAR to both achieve good memory locality and cache reuse as well as good vectorization on SIMD architectures. We also present the portable solutions used to parallelize the Pseudo-sepctral quasi-cylindrical code FBPIC on GPUs using the Numba python compiler.
Nonlinear evolution of parallel propagating Alfven waves: Vlasov - MHD simulation
Nariyuki, Y; Kumashiro, T; Hada, T
2009-01-01
Nonlinear evolution of circularly polarized Alfv\\'en waves are discussed by using the recently developed Vlasov-MHD code, which is a generalized Landau-fluid model. The numerical results indicate that as far as the nonlinearity in the system is not so large, the Vlasov-MHD model can validly solve time evolution of the Alfv\\'enic turbulence both in the linear and nonlinear stages. The present Vlasov-MHD model is proper to discuss the solar coronal heating and solar wind acceleration by Alfve\\'n waves propagating from the photosphere.
TP Clement
1999-06-24
RT3DV1 (Reactive Transport in 3-Dimensions) is computer code that solves the coupled partial differential equations that describe reactive-flow and transport of multiple mobile and/or immobile species in three-dimensional saturated groundwater systems. RT3D is a generalized multi-species version of the US Environmental Protection Agency (EPA) transport code, MT3D (Zheng, 1990). The current version of RT3D uses the advection and dispersion solvers from the DOD-1.5 (1997) version of MT3D. As with MT3D, RT3D also requires the groundwater flow code MODFLOW for computing spatial and temporal variations in groundwater head distribution. The RT3D code was originally developed to support the contaminant transport modeling efforts at natural attenuation demonstration sites. As a research tool, RT3D has also been used to model several laboratory and pilot-scale active bioremediation experiments. The performance of RT3D has been validated by comparing the code results against various numerical and analytical solutions. The code is currently being used to model field-scale natural attenuation at multiple sites. The RT3D code is unique in that it includes an implicit reaction solver that makes the code sufficiently flexible for simulating various types of chemical and microbial reaction kinetics. RT3D V1.0 supports seven pre-programmed reaction modules that can be used to simulate different types of reactive contaminants including benzene-toluene-xylene mixtures (BTEX), and chlorinated solvents such as tetrachloroethene (PCE) and trichloroethene (TCE). In addition, RT3D has a user-defined reaction option that can be used to simulate any other types of user-specified reactive transport systems. This report describes the mathematical details of the RT3D computer code and its input/output data structure. It is assumed that the user is familiar with the basics of groundwater flow and contaminant transport mechanics. In addition, RT3D users are expected to have some experience in
Hayek, W; Carlsson, M; Trampedach, R; Collet, R; Gudiksen, B V; Hansteen, V H; Leenaarts, J
2010-01-01
We present the implementation of a radiative transfer solver with coherent scattering in the new BIFROST code for radiative magneto-hydrodynamical (MHD) simulations of stellar surface convection. The code is fully parallelized using MPI domain decomposition, which allows for large grid sizes and improved resolution of hydrodynamical structures. We apply the code to simulate the surface granulation in a solar-type star, ignoring magnetic fields, and investigate the importance of coherent scattering for the atmospheric structure. A scattering term is added to the radiative transfer equation, requiring an iterative computation of the radiation field. We use a short-characteristics-based Gauss-Seidel acceleration scheme to compute radiative flux divergences for the energy equation. The effects of coherent scattering are tested by comparing the temperature stratification of three 3D time-dependent hydrodynamical atmosphere models of a solar-type star: without scattering, with continuum scattering only, and with bo...
Physical implementation of a Majorana fermion surface code for fault-tolerant quantum computation
Vijay, Sagar; Fu, Liang
2016-12-01
We propose a physical realization of a commuting Hamiltonian of interacting Majorana fermions realizing Z 2 topological order, using an array of Josephson-coupled topological superconductor islands. The required multi-body interaction Hamiltonian is naturally generated by a combination of charging energy induced quantum phase-slips on the superconducting islands and electron tunneling between islands. Our setup improves on a recent proposal for implementing a Majorana fermion surface code (Vijay et al 2015 Phys. Rev. X 5 041038), a ‘hybrid’ approach to fault-tolerant quantum computation that combines (1) the engineering of a stabilizer Hamiltonian with a topologically ordered ground state with (2) projective stabilizer measurements to implement error correction and a universal set of logical gates. Our hybrid strategy has advantages over the traditional surface code architecture in error suppression and single-step stabilizer measurements, and is widely applicable to implementing stabilizer codes for quantum computation.
Modeling of BWR core meltdown accidents - for application in the MELRPI. MOD2 computer code
Koh, B R; Kim, S H; Taleyarkhan, R P; Podowski, M Z; Lahey, Jr, R T
1985-04-01
This report summarizes improvements and modifications made in the MELRPI computer code. A major difference between this new, updated version of the code, called MELRPI.MOD2, and the one reported previously, concerns the inclusion of a model for the BWR emergency core cooling systems (ECCS). This model and its computer implementation, the ECCRPI subroutine, account for various emergency injection modes, for both intact and rubblized geometries. Other changes to MELRPI deal with an improved model for canister wall oxidation, rubble bed modeling, and numerical integration of system equations. A complete documentation of the entire MELRPI.MOD2 code is also given, including an input guide, list of subroutines, sample input/output and program listing.
Computational approaches towards understanding human long non-coding RNA biology.
Jalali, Saakshi; Kapoor, Shruti; Sivadas, Ambily; Bhartiya, Deeksha; Scaria, Vinod
2015-07-15
Long non-coding RNAs (lncRNAs) form the largest class of non-protein coding genes in the human genome. While a small subset of well-characterized lncRNAs has demonstrated their significant role in diverse biological functions like chromatin modifications, post-transcriptional regulation, imprinting etc., the functional significance of a vast majority of them still remains an enigma. Increasing evidence of the implications of lncRNAs in various diseases including cancer and major developmental processes has further enhanced the need to gain mechanistic insights into the lncRNA functions. Here, we present a comprehensive review of the various computational approaches and tools available for the identification and annotation of long non-coding RNAs. We also discuss a conceptual roadmap to systematically explore the functional properties of the lncRNAs using computational approaches.
Algorithms and computer codes for atomic and molecular quantum scattering theory. Volume I
Thomas, L. (ed.)
1979-01-01
The goals of this workshop are to identify which of the existing computer codes for solving the coupled equations of quantum molecular scattering theory perform most efficiently on a variety of test problems, and to make tested versions of those codes available to the chemistry community through the NRCC software library. To this end, many of the most active developers and users of these codes have been invited to discuss the methods and to solve a set of test problems using the LBL computers. The first volume of this workshop report is a collection of the manuscripts of the talks that were presented at the first meeting held at the Argonne National Laboratory, Argonne, Illinois June 25-27, 1979. It is hoped that this will serve as an up-to-date reference to the most popular methods with their latest refinements and implementations.
Tight bounds on computing error-correcting codes by bounded-depth circuits with arbitrary gates
Gál, Anna; Hansen, Kristoffer Arnsfelt; Koucký, Michal;
2011-01-01
We bound the minimum number w of wires needed to compute any (asymptotically good) error-correcting code C:01(n)01n with minimum distance (n), using unbounded fan-in circuits of depth d with arbitrary gates. Our main results are: (1) If d=2 then w=(n(lognloglogn)2) . (2) If d=3 then w=(nlglgn). (3...
Application of Multiple Description Coding for Adaptive QoS Mechanism for Mobile Cloud Computing
Ilan Sadeh
2014-02-01
Full Text Available Multimedia transmission over cloud infrastructure is a hot research topic worldwide. It is very strongly related to video streaming, VoIP, mobile networks, and computer networks. The goal is a reliable integration of telephony, video and audio transmission, computing and broadband transmission based on cloud computing. One right approach to pave the way for mobile multimedia and cloud computing is Multiple Description Coding (MDC, i.e. the solution would be: TCP/IP and similar protocols to be used for transmission of text files, and Multiple Description Coding “Send and Forget” algorithm to be used as transmission method for Multimedia over the cloud. Multiple Description Coding would improve the Quality of Service and would provide new service of rate adaptive streaming. This paper presents a new approach for improving the quality of multimedia and other services in the cloud, by using Multiple Description Coding (MDC. Firsty MDC Send and Forget Algorithm is compared with the existing protocols such as TCP/IP, UDP, RTP, etc. Then the Achievable Rate Region for MDC system is evaluated. Finally, a new subset of Quality of Service that considers the blocking in multi-terminal multimedia network and fidelity losses is considered.
Ivanov, Anisoara; Neacsu, Andrei
2011-01-01
This study describes the possibility and advantages of utilizing simple computer codes to complement the teaching techniques for high school physics. The authors have begun working on a collection of open source programs which allow students to compare the results and graphics from classroom exercises with the correct solutions and further more to…
Methods, algorithms and computer codes for calculation of electron-impact excitation parameters
Bogdanovich, P; Stonys, D
2015-01-01
We describe the computer codes, developed at Vilnius University, for the calculation of electron-impact excitation cross sections, collision strengths, and excitation rates in the plane-wave Born approximation. These codes utilize the multireference atomic wavefunctions which are also adopted to calculate radiative transition parameters of complex many-electron ions. This leads to consistent data sets suitable in plasma modelling codes. Two versions of electron scattering codes are considered in the present work, both of them employing configuration interaction method for inclusion of correlation effects and Breit-Pauli approximation to account for relativistic effects. These versions differ only by one-electron radial orbitals, where the first one employs the non-relativistic numerical radial orbitals, while another version uses the quasirelativistic radial orbitals. The accuracy of produced results is assessed by comparing radiative transition and electron-impact excitation data for neutral hydrogen, helium...
Computer code to interchange CDS and wave-drag geometry formats
Johnson, V. S.; Turnock, D. L.
1986-01-01
A computer program has been developed on the PRIME minicomputer to provide an interface for the passage of aircraft configuration geometry data between the Rockwell Configuration Development System (CDS) and a wireframe geometry format used by aerodynamic design and analysis codes. The interface program allows aircraft geometry which has been developed in CDS to be directly converted to the wireframe geometry format for analysis. Geometry which has been modified in the analysis codes can be transformed back to a CDS geometry file and examined for physical viability. Previously created wireframe geometry files may also be converted into CDS geometry files. The program provides a useful link between a geometry creation and manipulation code and analysis codes by providing rapid and accurate geometry conversion.
Users manual for CAFE-3D : a computational fluid dynamics fire code.
Khalil, Imane; Lopez, Carlos; Suo-Anttila, Ahti Jorma (Alion Science and Technology, Albuquerque, NM)
2005-03-01
The Container Analysis Fire Environment (CAFE) computer code has been developed to model all relevant fire physics for predicting the thermal response of massive objects engulfed in large fires. It provides realistic fire thermal boundary conditions for use in design of radioactive material packages and in risk-based transportation studies. The CAFE code can be coupled to commercial finite-element codes such as MSC PATRAN/THERMAL and ANSYS. This coupled system of codes can be used to determine the internal thermal response of finite element models of packages to a range of fire environments. This document is a user manual describing how to use the three-dimensional version of CAFE, as well as a description of CAFE input and output parameters. Since this is a user manual, only a brief theoretical description of the equations and physical models is included.
TEMP: a computer code to calculate fuel pin temperatures during a transient. [LMFBR
Bard, F E; Christensen, B Y; Gneiting, B C
1980-04-01
The computer code TEMP calculates fuel pin temperatures during a transient. It was developed to accommodate temperature calculations in any system of axi-symmetric concentric cylinders. When used to calculate fuel pin temperatures, the code will handle a fuel pin as simple as a solid cylinder or as complex as a central void surrounded by fuel that is broken into three regions by two circumferential cracks. Any fuel situation between these two extremes can be analyzed along with additional cladding, heat sink, coolant or capsule regions surrounding the fuel. The one-region version of the code accurately calculates the solution to two problems having closed-form solutions. The code uses an implicit method, an explicit method and a Crank-Nicolson (implicit-explicit) method.
NASCRAC - A computer code for fracture mechanics analysis of crack growth
Harris, D. O.; Eason, E. D.; Thomas, J. M.; Bianca, C. J.; Salter, L. D.
1987-01-01
NASCRAC - a computer code for fracture mechanics analysis of crack growth - is described in this paper. The need for such a code is increasing as requirements grow for high reliability and low weight in aerospace components. The code is comprehensive and versatile, as well as user friendly. The major purpose of the code is calculation of fatigue, corrosion fatigue, or stress corrosion crack growth, and a variety of crack growth relations can be selected by the user. Additionally, crack retardation models are included. A very wide variety of stress intensity factor solutions are contained in the code, and extensive use is made of influence functions. This allows complex stress gradients in three-dimensional crack problems to be treated easily and economically. In cases where previous stress intensity factor solutions are not adequate, new influence functions can be calculated by the code. Additional features include incorporation of J-integral solutions from the literature and a capability for estimating elastic-plastic stress redistribution from the results of a corresponding elastic analysis. An example problem is presented which shows typical outputs from the code.
A proposed framework for computational fluid dynamics code calibration/validation
Oberkampf, W.L.
1993-12-31
The paper reviews the terminology and methodology that have been introduced during the last several years for building confidence n the predictions from Computational Fluid Dynamics (CID) codes. Code validation terminology developed for nuclear reactor analyses and aerospace applications is reviewed and evaluated. Currently used terminology such as ``calibrated code,`` ``validated code,`` and a ``validation experiment`` is discussed along with the shortcomings and criticisms of these terms. A new framework is proposed for building confidence in CFD code predictions that overcomes some of the difficulties of past procedures and delineates the causes of uncertainty in CFD predictions. Building on previous work, new definitions of code verification and calibration are proposed. These definitions provide more specific requirements for the knowledge level of the flow physics involved and the solution accuracy of the given partial differential equations. As part of the proposed framework, categories are also proposed for flow physics research, flow modeling research, and the application of numerical predictions. The contributions of physical experiments, analytical solutions, and other numerical solutions are discussed, showing that each should be designed to achieve a distinctively separate purpose in building confidence in accuracy of CFD predictions. A number of examples are given for each approach to suggest methods for obtaining the highest value for CFD code quality assurance.
Simulation of wave interactions with MHD
Batchelor, D; Bernholdt, D; Berry, L; Elwasif, W; Jaeger, E; Keyes, D; Klasky, S [Oak Ridge National Laboratory, Oak Ridge, TN 37331 (United States); Alba, C; Choi, M [General Atomics, San Diego, CA 92186 (United States); Bateman, G [Lehigh University, Bethlehem, PA 18015 (United States); Bonoli, P [Plasma Science and Fusion Center, MTT, Cambridge, MA 02139 (United States); Bramley, R [Indiana University, Bloomington, IN 47405 (United States); Breslau, J; Chance, M; Chen, J; Fu, G; Jardin, S [Princeton Plasma Physics Laboratory, Princeton, NJ 08543 (United States); Harvey, R [CompX, Del Mar, CA 92014 (United States); Jenkins, T [University of Wisconsin, Madison, WI 53706 (United States); Kruger, S [Tech-X, Boulder, CO 80303 (United States)], E-mail: batchelordb@ornl.gov (and others)
2008-07-15
The broad scientific objectives of the SWIM (Simulation 01 Wave Interaction with MHD) project are twofold: (1) improve our understanding of interactions that both radio frequency (RF) wave and particle sources have on extended-MHD phenomena, and to substantially improve our capability for predicting and optimizing the performance of burning plasmas in devices such as ITER: and (2) develop an integrated computational system for treating multiphysics phenomena with the required flexibility and extensibility to serve as a prototype for the Fusion Simulation Project. The Integrated Plasma Simulator (IPS) has been implemented. Presented here are initial physics results on RP effects on MHD instabilities in tokamaks as well as simulation results for tokamak discharge evolution using the IPS.
Simulation of wave interactions with MHD
Batchelor, Donald B [ORNL; Abla, G [General Atomics, San Diego; Bateman, Glenn [Lehigh University, Bethlehem, PA; Bernholdt, David E [ORNL; Berry, Lee A [ORNL; Bonoli, P. [Massachusetts Institute of Technology (MIT); Bramley, R [Indiana University; Breslau, J. [Princeton Plasma Physics Laboratory (PPPL); Chance, M. [Princeton Plasma Physics Laboratory (PPPL); Chen, J. [Princeton Plasma Physics Laboratory (PPPL); Choi, M. [General Atomics; Elwasif, Wael R [ORNL; Fu, GuoYong [Princeton Plasma Physics Laboratory (PPPL); Harvey, R. W. [CompX, Del Mar, CA; Jaeger, Erwin Frederick [ORNL; Jardin, S. C. [Princeton Plasma Physics Laboratory (PPPL); Jenkins, T [University of Wisconsin; Keyes, David E [Columbia University; Klasky, Scott A [ORNL; Kruger, Scott [Tech-X Corporation; Ku, Long-Poe [Princeton Plasma Physics Laboratory (PPPL); Lynch, Vickie E [ORNL; McCune, Douglas [Princeton Plasma Physics Laboratory (PPPL); Ramos, J. [Massachusetts Institute of Technology (MIT); Schissel, D. [General Atomics; Schnack, [University of Wisconsin; Wright, J. [Massachusetts Institute of Technology (MIT)
2008-07-01
The broad scientific objectives of the SWIM (Simulation of Wave Interaction with MHD) project are twofold: (1) improve our understanding of interactions that both radio frequency (RF) wave and particle sources have on extended-MHD phenomena, and to substantially improve our capability for predicting and optimizing the performance of burning plasmas in devices such as ITER: and (2) develop an integrated computational system for treating multiphysics phenomena with the required flexibility and extensibility to serve as a prototype for the Fusion Simulation Project. The Integrated Plasma Simulator (IPS) has been implemented. Presented here are initial physics results on RF effects on MHD instabilities in tokamaks as well as simulation results for tokamak discharge evolution using the IPS.
Physical Model Development and Benchmarking for MHD Flows in Blanket Design
Ramakanth Munipalli; P.-Y.Huang; C.Chandler; C.Rowell; M.-J.Ni; N.Morley; S.Smolentsev; M.Abdou
2008-06-05
An advanced simulation environment to model incompressible MHD flows relevant to blanket conditions in fusion reactors has been developed at HyPerComp in research collaboration with TEXCEL. The goals of this phase-II project are two-fold: The first is the incorporation of crucial physical phenomena such as induced magnetic field modeling, and extending the capabilities beyond fluid flow prediction to model heat transfer with natural convection and mass transfer including tritium transport and permeation. The second is the design of a sequence of benchmark tests to establish code competence for several classes of physical phenomena in isolation as well as in select (termed here as “canonical”,) combinations. No previous attempts to develop such a comprehensive MHD modeling capability exist in the literature, and this study represents essentially uncharted territory. During the course of this Phase-II project, a significant breakthrough was achieved in modeling liquid metal flows at high Hartmann numbers. We developed a unique mathematical technique to accurately compute the fluid flow in complex geometries at extremely high Hartmann numbers (10,000 and greater), thus extending the state of the art of liquid metal MHD modeling relevant to fusion reactors at the present time. These developments have been published in noted international journals. A sequence of theoretical and experimental results was used to verify and validate the results obtained. The code was applied to a complete DCLL module simulation study with promising results.
An Object-Oriented Computer Code for Aircraft Engine Weight Estimation
Tong, Michael T.; Naylor, Bret A.
2009-01-01
Reliable engine-weight estimation at the conceptual design stage is critical to the development of new aircraft engines. It helps to identify the best engine concept amongst several candidates. At NASA Glenn Research Center (GRC), the Weight Analysis of Turbine Engines (WATE) computer code, originally developed by Boeing Aircraft, has been used to estimate the engine weight of various conceptual engine designs. The code, written in FORTRAN, was originally developed for NASA in 1979. Since then, substantial improvements have been made to the code to improve the weight calculations for most of the engine components. Most recently, to improve the maintainability and extensibility of WATE, the FORTRAN code has been converted into an object-oriented version. The conversion was done within the NASA's NPSS (Numerical Propulsion System Simulation) framework. This enables WATE to interact seamlessly with the thermodynamic cycle model which provides component flow data such as airflows, temperatures, and pressures, etc., that are required for sizing the components and weight calculations. The tighter integration between the NPSS and WATE would greatly enhance system-level analysis and optimization capabilities. It also would facilitate the enhancement of the WATE code for next-generation aircraft and space propulsion systems. In this paper, the architecture of the object-oriented WATE code (or WATE++) is described. Both the FORTRAN and object-oriented versions of the code are employed to compute the dimensions and weight of a 300-passenger aircraft engine (GE90 class). Both versions of the code produce essentially identical results as should be the case.
Multiple frequencies sequential coding for SSVEP-based brain-computer interface.
Yangsong Zhang
Full Text Available BACKGROUND: Steady-state visual evoked potential (SSVEP-based brain-computer interface (BCI has become one of the most promising modalities for a practical noninvasive BCI system. Owing to both the limitation of refresh rate of liquid crystal display (LCD or cathode ray tube (CRT monitor, and the specific physiological response property that only a very small number of stimuli at certain frequencies could evoke strong SSVEPs, the available frequencies for SSVEP stimuli are limited. Therefore, it may not be enough to code multiple targets with the traditional frequencies coding protocols, which poses a big challenge for the design of a practical SSVEP-based BCI. This study aimed to provide an innovative coding method to tackle this problem. METHODOLOGY/PRINCIPAL FINDINGS: In this study, we present a novel protocol termed multiple frequencies sequential coding (MFSC for SSVEP-based BCI. In MFSC, multiple frequencies are sequentially used in each cycle to code the targets. To fulfill the sequential coding, each cycle is divided into several coding epochs, and during each epoch, certain frequency is used. Obviously, different frequencies or the same frequency can be presented in the coding epochs, and the different epoch sequence corresponds to the different targets. To show the feasibility of MFSC, we used two frequencies to realize four targets and carried on an offline experiment. The current study shows that: 1 MFSC is feasible and efficient; 2 the performance of SSVEP-based BCI based on MFSC can be comparable to some existed systems. CONCLUSIONS/SIGNIFICANCE: The proposed protocol could potentially implement much more targets with the limited available frequencies compared with the traditional frequencies coding protocol. The efficiency of the new protocol was confirmed by real data experiment. We propose that the SSVEP-based BCI under MFSC might be a promising choice in the future.
Alexakis, A.
2009-04-01
Most astrophysical and planetary systems e.g., solar convection and stellar winds, are in a turbulent state and coupled to magnetic fields. Understanding and quantifying the statistical properties of magneto-hydro-dynamic (MHD) turbulence is crucial to explain the involved physical processes. Although the phenomenological theory of hydro-dynamic (HD) turbulence has been verified up to small corrections, a similar statement cannot be made for MHD turbulence. Since the phenomenological description of Hydrodynamic turbulence by Kolmogorov in 1941 there have been many attempts to derive a similar description for turbulence in conducting fluids (i.e Magneto-Hydrodynamic turbulence). However such a description is going to be based inevitably on strong assumptions (typically borrowed from hydrodynamics) that do not however necessarily apply to the MHD case. In this talk I will discuss some of the properties and differences of the energy and helicity cascades in turbulent MHD and HD flows. The investigation is going to be based on the analysis of direct numerical simulations. The cascades in MHD turbulence appear to be a more non-local process (in scale space) than in Hydrodynamics. Some implications of these results to turbulent modeling will be discussed
Zhao, Shengmei; Wang, Le; Liang, Wenqiang; Cheng, Weiwen; Gong, Longyan
2015-10-01
In this paper, we propose a high performance optical encryption (OE) scheme based on computational ghost imaging (GI) with QR code and compressive sensing (CS) technique, named QR-CGI-OE scheme. N random phase screens, generated by Alice, is a secret key and be shared with its authorized user, Bob. The information is first encoded by Alice with QR code, and the QR-coded image is then encrypted with the aid of computational ghost imaging optical system. Here, measurement results from the GI optical system's bucket detector are the encrypted information and be transmitted to Bob. With the key, Bob decrypts the encrypted information to obtain the QR-coded image with GI and CS techniques, and further recovers the information by QR decoding. The experimental and numerical simulated results show that the authorized users can recover completely the original image, whereas the eavesdroppers can not acquire any information about the image even the eavesdropping ratio (ER) is up to 60% at the given measurement times. For the proposed scheme, the number of bits sent from Alice to Bob are reduced considerably and the robustness is enhanced significantly. Meantime, the measurement times in GI system is reduced and the quality of the reconstructed QR-coded image is improved.
MOLOCH computer code for molecular-dynamics simulation of processes in condensed matter
Derbenev I.V.
2011-01-01
Full Text Available Theoretical and experimental investigation into properties of condensed matter is one of the mainstreams in RFNC-VNIITF scientific activity. The method of molecular dynamics (MD is an innovative method of theoretical materials science. Modern supercomputers allow the direct simulation of collective effects in multibillion atom sample, making it possible to model physical processes on the atomistic level, including material response to dynamic load, radiation damage, influence of defects and alloying additions upon material mechanical properties, or aging of actinides. During past ten years, the computer code MOLOCH has been developed at RFNC-VNIITF. It is a parallel code suitable for massive parallel computing. Modern programming techniques were used to make the code almost 100% efficient. Practically all instruments required for modelling were implemented in the code: a potential builder for different materials, simulation of physical processes in arbitrary 3D geometry, and calculated data processing. A set of tests was developed to analyse algorithms efficiency. It can be used to compare codes with different MD implementation between each other.
MHD Simulation of the Inner-Heliospheric Magnetic Field
Wiengarten, T; Fichtner, H; Cameron, R; Jiang, J; Kissmann, R; Scherer, K; 10.1029/2012JA018089
2013-01-01
Maps of the radial magnetic field at a heliocentric distance of ten solar radii are used as boundary conditions in the MHD code CRONOS to simulate a 3D inner-heliospheric solar wind emanating from the rotating Sun out to 1 AU. The input data for the magnetic field are the result of solar surface flux transport modelling using observational data of sunspot groups coupled with a current sheet source surface model. Amongst several advancements, this allows for higher angular resolution than that of comparable observational data from synoptic magnetograms. The required initial conditions for the other MHD quantities are obtained following an empirical approach using an inverse relation between flux tube expansion and radial solar wind speed. The computations are performed for representative solar minimum and maximum conditions, and the corresponding state of the solar wind up to the Earths orbit is obtained. After a successful comparison of the latter with observational data, they can be used to drive outer-helio...
Daniel Litinski
2017-09-01
Full Text Available We present a scalable architecture for fault-tolerant topological quantum computation using networks of voltage-controlled Majorana Cooper pair boxes and topological color codes for error correction. Color codes have a set of transversal gates which coincides with the set of topologically protected gates in Majorana-based systems, namely, the Clifford gates. In this way, we establish color codes as providing a natural setting in which advantages offered by topological hardware can be combined with those arising from topological error-correcting software for full-fledged fault-tolerant quantum computing. We provide a complete description of our architecture, including the underlying physical ingredients. We start by showing that in topological superconductor networks, hexagonal cells can be employed to serve as physical qubits for universal quantum computation, and we present protocols for realizing topologically protected Clifford gates. These hexagonal-cell qubits allow for a direct implementation of open-boundary color codes with ancilla-free syndrome read-out and logical T gates via magic-state distillation. For concreteness, we describe how the necessary operations can be implemented using networks of Majorana Cooper pair boxes, and we give a feasibility estimate for error correction in this architecture. Our approach is motivated by nanowire-based networks of topological superconductors, but it could also be realized in alternative settings such as quantum-Hall–superconductor hybrids.
Once-through CANDU reactor models for the ORIGEN2 computer code
Croff, A.G.; Bjerke, M.A.
1980-11-01
Reactor physics calculations have led to the development of two CANDU reactor models for the ORIGEN2 computer code. The model CANDUs are based on (1) the existing once-through fuel cycle with feed comprised of natural uranium and (2) a projected slightly enriched (1.2 wt % /sup 235/U) fuel cycle. The reactor models are based on cross sections taken directly from the reactor physics codes. Descriptions of the reactor models, as well as values for the ORIGEN2 flux parameters THERM, RES, and FAST, are given.
Experimental assessment of computer codes used for safety analysis of integral reactors
Falkov, A.A.; Kuul, V.S.; Samoilov, O.B. [OKB Mechanical Engineering, Nizhny Novgorod (Russian Federation)
1995-09-01
Peculiarities of integral reactor thermohydraulics in accidents are associated with presence of noncondensable gas in built-in pressurizer, absence of pumped ECCS, use of guard vessel for LOCAs localisation and passive RHRS through in-reactor HX`s. These features defined the main trends in experimental investigations and verification efforts for computer codes applied. The paper reviews briefly the performed experimental investigation of thermohydraulics of AST-500, VPBER600-type integral reactors. The characteristic of UROVEN/MB-3 code for LOCAs analysis in integral reactors and results of its verification are given. The assessment of RELAP5/mod3 applicability for accident analysis in integral reactor is presented.
The MELTSPREAD-1 computer code for the analysis of transient spreading in containments
Farmer, M.T.; Sienicki, J.J.; Spencer, B.W.
1990-01-01
A one-dimensional, multicell, Eulerian finite difference computer code (MELTSPREAD-1) has been developed to provide an improved prediction of the gravity driven spreading and thermal interactions of molten corium flowing over a concrete or steel surface. In this paper, the modeling incorporated into the code is described and the spreading models are benchmarked against a simple dam break'' problem as well as water simulant spreading data obtained in a scaled apparatus of the Mk I containment. Results are also presented for a scoping calculation of the spreading behavior and shell thermal response in the full scale Mk I system following vessel meltthrough. 24 refs., 15 figs.
Strenge, D.L.; Peloquin, R.A.
1981-04-01
The computer code HADOC (Hanford Acute Dose Calculations) is described and instructions for its use are presented. The code calculates external dose from air submersion and inhalation doses following acute radionuclide releases. Atmospheric dispersion is calculated using the Hanford model with options to determine maximum conditions. Building wake effects and terrain variation may also be considered. Doses are calculated using dose conversion factor supplied in a data library. Doses are reported for one and fifty year dose commitment periods for the maximum individual and the regional population (within 50 miles). The fractional contribution to dose by radionuclide and exposure mode are also printed if requested.
V.S.O.P. (99/05) computer code system
Ruetten, H.J.; Haas, K.A.; Brockmann, H.; Scherer, W.
2005-11-01
V.S.O.P. is a computer code system for the comprehensive numerical simulation of the physics of thermal reactors. It implies the setup of the reactor and of the fuel element, processing of cross sections, neutron spectrum evaluation, neutron diffusion calculation in two or three dimensions, fuel burnup, fuel shuffling, reactor control, thermal hydraulics and fuel cycle costs. The thermal hydraulics part (steady state and time-dependent) is restricted to HTRs and to two spatial dimensions. The code can simulate the reactor operation from the initial core towards the equilibrium core. V.S.O.P.(99 / 05) represents the further development of V.S.O.P. (99). Compared to its precursor, the code system has been improved in many details. Major improvements and extensions have been included concerning the neutron spectrum calculation, the 3-d neutron diffusion options, and the thermal hydraulic section with respect to 'multi-pass'-fuelled pebblebed cores. This latest code version was developed and tested under the WINDOWS-XP - operating system. The storage requirement for the executables and the basic libraries associated with the code amounts to about 15 MB. Another 5 MB are required - if desired - for storage of the source code ({approx}65000 Fortran statements). (orig.)
V.S.O.P. (99/05) computer code system
Ruetten, H.J.; Haas, K.A.; Brockmann, H.; Scherer, W.
2005-11-01
V.S.O.P. is a computer code system for the comprehensive numerical simulation of the physics of thermal reactors. It implies the setup of the reactor and of the fuel element, processing of cross sections, neutron spectrum evaluation, neutron diffusion calculation in two or three dimensions, fuel burnup, fuel shuffling, reactor control, thermal hydraulics and fuel cycle costs. The thermal hydraulics part (steady state and time-dependent) is restricted to HTRs and to two spatial dimensions. The code can simulate the reactor operation from the initial core towards the equilibrium core. V.S.O.P.(99 / 05) represents the further development of V.S.O.P. (99). Compared to its precursor, the code system has been improved in many details. Major improvements and extensions have been included concerning the neutron spectrum calculation, the 3-d neutron diffusion options, and the thermal hydraulic section with respect to 'multi-pass'-fuelled pebblebed cores. This latest code version was developed and tested under the WINDOWS-XP - operating system. The storage requirement for the executables and the basic libraries associated with the code amounts to about 15 MB. Another 5 MB are required - if desired - for storage of the source code ({approx}65000 Fortran statements). (orig.)
Rutishauser, David
2006-01-01
The motivation for this work comes from an observation that amidst the push for Massively Parallel (MP) solutions to high-end computing problems such as numerical physical simulations, large amounts of legacy code exist that are highly optimized for vector supercomputers. Because re-hosting legacy code often requires a complete re-write of the original code, which can be a very long and expensive effort, this work examines the potential to exploit reconfigurable computing machines in place of a vector supercomputer to implement an essentially unmodified legacy source code. Custom and reconfigurable computing resources could be used to emulate an original application's target platform to the extent required to achieve high performance. To arrive at an architecture that delivers the desired performance subject to limited resources involves solving a multi-variable optimization problem with constraints. Prior research in the area of reconfigurable computing has demonstrated that designing an optimum hardware implementation of a given application under hardware resource constraints is an NP-complete problem. The premise of the approach is that the general issue of applying reconfigurable computing resources to the implementation of an application, maximizing the performance of the computation subject to physical resource constraints, can be made a tractable problem by assuming a computational paradigm, such as vector processing. This research contributes a formulation of the problem and a methodology to design a reconfigurable vector processing implementation of a given application that satisfies a performance metric. A generic, parametric, architectural framework for vector processing implemented in reconfigurable logic is developed as a target for a scheduling/mapping algorithm that maps an input computation to a given instance of the architecture. This algorithm is integrated with an optimization framework to arrive at a specification of the architecture parameters
Realistic radiative MHD simulation of a solar flare
Rempel, Matthias D.; Cheung, Mark; Chintzoglou, Georgios; Chen, Feng; Testa, Paola; Martinez-Sykora, Juan; Sainz Dalda, Alberto; DeRosa, Marc L.; Viktorovna Malanushenko, Anna; Hansteen, Viggo H.; De Pontieu, Bart; Carlsson, Mats; Gudiksen, Boris; McIntosh, Scott W.
2017-08-01
We present a recently developed version of the MURaM radiative MHD code that includes coronal physics in terms of optically thin radiative loss and field aligned heat conduction. The code employs the "Boris correction" (semi-relativistic MHD with a reduced speed of light) and a hyperbolic treatment of heat conduction, which allow for efficient simulations of the photosphere/corona system by avoiding the severe time-step constraints arising from Alfven wave propagation and heat conduction. We demonstrate that this approach can be used even in dynamic phases such as a flare. We consider a setup in which a flare is triggered by flux emergence into a pre-existing bipolar active region. After the coronal energy release, efficient transport of energy along field lines leads to the formation of flare ribbons within seconds. In the flare ribbons we find downflows for temperatures lower than ~5 MK and upflows at higher temperatures. The resulting soft X-ray emission shows a fast rise and slow decay, reaching a peak corresponding to a mid C-class flare. The post reconnection energy release in the corona leads to average particle energies reaching 50 keV (500 MK under the assumption of a thermal plasma). We show that hard X-ray emission from the corona computed under the assumption of thermal bremsstrahlung can produce a power-law spectrum due to the multi-thermal nature of the plasma. The electron energy flux into the flare ribbons (classic heat conduction with free streaming limit) is highly inhomogeneous and reaches peak values of about 3x1011 erg/cm2/s in a small fraction of the ribbons, indicating regions that could potentially produce hard X-ray footpoint sources. We demonstrate that these findings are robust by comparing simulations computed with different values of the saturation heat flux as well as the "reduced speed of light".
Reginatto, M.; Goldhagen, P.
1998-06-01
The problem of analyzing data from a multisphere neutron spectrometer to infer the energy spectrum of the incident neutrons is discussed. The main features of the code MAXED, a computer program developed to apply the maximum entropy principle to the deconvolution (unfolding) of multisphere neutron spectrometer data, are described, and the use of the code is illustrated with an example. A user`s guide for the code MAXED is included in an appendix. The code is available from the authors upon request.
The MELTSPREAD-1 computer code for the analysis of transient spreading in containments
Farmer, M.T.; Sienicki, J.J.; Spencer, B.W.
1990-01-01
Transient spreading of molten core materials is important in the assessment of severe-accident sequences for Mk-I boiling water reactors (BWRs). Of interest is whether core materials are able to spread over the pedestal and drywell floors to contact the containment shell and cause thermally induced shell failure, or whether heat transfer to underlying concrete and overlying water will freeze the melt short of the shell. The development of a computational capability for the assessment of this problem was initiated by Sienicki et al. in the form of MELTSPREAD-O code. Development is continuing in the form of the MELTSPREAD-1 code, which contains new models for phenomena that were ignored in the earlier code. This paper summarizes these new models, provides benchmarking calculations of the relocation model against an analytical solution as well as simulant spreading data, and summarizes the results of a scoping calculation for the full Mk-I system.
Computer code simulations of the formation of Meteor Crater, Arizona - Calculations MC-1 and MC-2
Roddy, D. J.; Schuster, S. H.; Kreyenhagen, K. N.; Orphal, D. L.
1980-01-01
It has been widely accepted that hypervelocity impact processes play a major role in the evolution of the terrestrial planets and satellites. In connection with the development of quantitative methods for the description of impact cratering, it was found that the results provided by two-dimensional finite difference, computer codes is greatly improved when initial impact conditions can be defined and when the numerical results can be tested against field and laboratory data. In order to address this problem, a numerical code study of the formation of Meteor (Barringer) Crater, Arizona, has been undertaken. A description is presented of the major results from the first two code calculations, MC-1 and MC-2, that have been completed for Meteor Crater. Both calculations used an iron meteorite with a kinetic energy of 3.8 Megatons. Calculation MC-1 had an impact velocity of 25 km/sec and MC-2 had an impact velocity of 15 km/sec.
WOLF: a computer code package for the calculation of ion beam trajectories
Vogel, D.L.
1985-10-01
The WOLF code solves POISSON'S equation within a user-defined problem boundary of arbitrary shape. The code is compatible with ANSI FORTRAN and uses a two-dimensional Cartesian coordinate geometry represented on a triangular lattice. The vacuum electric fields and equipotential lines are calculated for the input problem. The use may then introduce a series of emitters from which particles of different charge-to-mass ratios and initial energies can originate. These non-relativistic particles will then be traced by WOLF through the user-defined region. Effects of ion and electron space charge are included in the calculation. A subprogram PISA forms part of this code and enables optimization of various aspects of the problem. The WOLF package also allows detailed graphics analysis of the computed results to be performed.
HYDRA-II: A hydrothermal analysis computer code: Volume 3, Verification/validation assessments
McCann, R.A.; Lowery, P.S.
1987-10-01
HYDRA-II is a hydrothermal computer code capable of three-dimensional analysis of coupled conduction, convection, and thermal radiation problems. This code is especially appropriate for simulating the steady-state performance of spent fuel storage systems. The code has been evaluated for this application for the US Department of Energy's Commercial Spent Fuel Management Program. HYDRA-II provides a finite difference solution in cartesian coordinates to the equations governing the conservation of mass, momentum, and energy. A cylindrical coordinate system may also be used to enclose the cartesian coordinate system. This exterior coordinate system is useful for modeling cylindrical cask bodies. The difference equations for conservation of momentum are enhanced by the incorporation of directional porosities and permeabilities that aid in modeling solid structures whose dimensions may be smaller than the computational mesh. The equation for conservation of energy permits modeling of orthotropic physical properties and film resistances. Several automated procedures are available to model radiation transfer within enclosures and from fuel rod to fuel rod. The documentation of HYDRA-II is presented in three separate volumes. Volume I - Equations and Numerics describes the basic differential equations, illustrates how the difference equations are formulated, and gives the solution procedures employed. Volume II - User's Manual contains code flow charts, discusses the code structure, provides detailed instructions for preparing an input file, and illustrates the operation of the code by means of a model problem. This volume, Volume III - Verification/Validation Assessments, provides a comparison between the analytical solution and the numerical simulation for problems with a known solution. This volume also documents comparisons between the results of simulations of single- and multiassembly storage systems and actual experimental data. 11 refs., 55 figs., 13 tabs.
HYDRA-II: A hydrothermal analysis computer code: Volume 2, User's manual
McCann, R.A.; Lowery, P.S.; Lessor, D.L.
1987-09-01
HYDRA-II is a hydrothermal computer code capable of three-dimensional analysis of coupled conduction, convection, and thermal radiation problems. This code is especially appropriate for simulating the steady-state performance of spent fuel storage systems. The code has been evaluated for this application for the US Department of Energy's Commercial Spent Fuel Management Program. HYDRA-II provides a finite-difference solution in cartesian coordinates to the equations governing the conservation of mass, momentum, and energy. A cylindrical coordinate system may also be used to enclose the cartesian coordinate system. This exterior coordinate system is useful for modeling cylindrical cask bodies. The difference equations for conservation of momentum incorporate directional porosities and permeabilities that are available to model solid structures whose dimensions may be smaller than the computational mesh. The equation for conservation of energy permits modeling of orthotropic physical properties and film resistances. Several automated methods are available to model radiation transfer within enclosures and from fuel rod to fuel rod. The documentation of HYDRA-II is presented in three separate volumes. Volume 1 - Equations and Numerics describes the basic differential equations, illustrates how the difference equations are formulated, and gives the solution procedures employed. This volume, Volume 2 - User's Manual, contains code flow charts, discusses the code structure, provides detailed instructions for preparing an input file, and illustrates the operation of the code by means of a sample problem. The final volume, Volume 3 - Verification/Validation Assessments, provides a comparison between the analytical solution and the numerical simulation for problems with a known solution. 6 refs.
Zhang, Shuai; Morita, Koji; Shirakawa, Noriyuki; Yamamoto, Yuichi
The COMPASS code is designed based on the moving particle semi-implicit method to simulate various complex mesoscale phenomena relevant to core disruptive accidents of sodium-cooled fast reactors. In this study, a computational framework for fluid-solid mixture flow simulations was developed for the COMPASS code. The passively moving solid model was used to simulate hydrodynamic interactions between fluid and solids. Mechanical interactions between solids were modeled by the distinct element method. A multi-time-step algorithm was introduced to couple these two calculations. The proposed computational framework for fluid-solid mixture flow simulations was verified by the comparison between experimental and numerical studies on the water-dam break with multiple solid rods.
Automatic code generation in SPARK: Applications of computer algebra and compiler-compilers
Nataf, J.M.; Winkelmann, F.
1992-09-01
We show how computer algebra and compiler-compilers are used for automatic code generation in the Simulation Problem Analysis and Research Kernel (SPARK), an object oriented environment for modeling complex physical systems that can be described by differential-algebraic equations. After a brief overview of SPARK, we describe the use of computer algebra in SPARK`s symbolic interface, which generates solution code for equations that are entered in symbolic form. We also describe how the Lex/Yacc compiler-compiler is used to achieve important extensions to the SPARK simulation language, including parametrized macro objects and steady-state resetting of a dynamic simulation. The application of these methods to solving the partial differential equations for two-dimensional heat flow is illustrated.
Automatic code generation in SPARK: Applications of computer algebra and compiler-compilers
Nataf, J.M.; Winkelmann, F.
1992-09-01
We show how computer algebra and compiler-compilers are used for automatic code generation in the Simulation Problem Analysis and Research Kernel (SPARK), an object oriented environment for modeling complex physical systems that can be described by differential-algebraic equations. After a brief overview of SPARK, we describe the use of computer algebra in SPARK's symbolic interface, which generates solution code for equations that are entered in symbolic form. We also describe how the Lex/Yacc compiler-compiler is used to achieve important extensions to the SPARK simulation language, including parametrized macro objects and steady-state resetting of a dynamic simulation. The application of these methods to solving the partial differential equations for two-dimensional heat flow is illustrated.
Computing element evolution towards Exascale and its impact on legacy simulation codes
Colin de Verdiere, Guillaume J.L. [CEA, DAM, DIF, Arpajon (France)
2015-12-15
In the light of the current race towards the Exascale, this article highlights the main features of the forthcoming computing elements that will be at the core of next generations of supercomputers. The market analysis, underlying this work, shows that computers are facing a major evolution in terms of architecture. As a consequence, it is important to understand the impacts of those evolutions on legacy codes or programming methods. The problems of dissipated power and memory access are discussed and will lead to a vision of what should be an exascale system. To survive, programming languages had to respond to the hardware evolutions either by evolving or with the creation of new ones. From the previous elements, we elaborate why vectorization, multithreading, data locality awareness and hybrid programming will be the key to reach the exascale, implying that it is time to start rewriting codes. (orig.)
Chen, Y. S.
1986-03-01
In this report, a numerical method for solving the equations of motion of three-dimensional incompressible flows in nonorthogonal body-fitted coordinate (BFC) systems has been developed. The equations of motion are transformed to a generalized curvilinear coordinate system from which the transformed equations are discretized using finite difference approximations in the transformed domain. The hybrid scheme is used to approximate the convection terms in the governing equations. Solutions of the finite difference equations are obtained iteratively by using a pressure-velocity correction algorithm (SIMPLE-C). Numerical examples of two- and three-dimensional, laminar and turbulent flow problems are employed to evaluate the accuracy and efficiency of the present computer code. The user's guide and computer program listing of the present code are also included.
Abstracts of digital computer code packages assembled by the Radiation Shielding Information Center
Carter, B.J.; Maskewitz, B.F.
1985-04-01
This publication, ORNL/RSIC-13, Volumes I to III Revised, has resulted from an internal audit of the first 168 packages of computing technology in the Computer Codes Collection (CCC) of the Radiation Shielding Information Center (RSIC). It replaces the earlier three documents published as single volumes between 1966 to 1972. A significant number of the early code packages were considered to be obsolete and were removed from the collection in the audit process and the CCC numbers were not reassigned. Others not currently being used by the nuclear R and D community were retained in the collection to preserve technology not replaced by newer methods, or were considered of potential value for reference purposes. Much of the early technology, however, has improved through developer/RSIC/user interaction and continues at the forefront of the advancing state-of-the-art.
Improvement of Level-1 PSA computer code package -A study for nuclear safety improvement-
Park, Chang Kyu; Kim, Tae Woon; Ha, Jae Joo; Han, Sang Hoon; Cho, Yeong Kyun; Jeong, Won Dae; Jang, Seung Cheol; Choi, Young; Seong, Tae Yong; Kang, Dae Il; Hwang, Mi Jeong; Choi, Seon Yeong; An, Kwang Il [Korea Atomic Energy Res. Inst., Taejon (Korea, Republic of)
1994-07-01
This year is the second year of the Government-sponsored Mid- and Long-Term Nuclear Power Technology Development Project. The scope of this subproject titled on `The Improvement of Level-1 PSA Computer Codes` is divided into three main activities : (1) Methodology development on the under-developed fields such as risk assessment technology for plant shutdown and external events, (2) Computer code package development for Level-1 PSA, (3) Applications of new technologies to reactor safety assessment. At first, in the area of PSA methodology development, foreign PSA reports on shutdown and external events have been reviewed and various PSA methodologies have been compared. Level-1 PSA code KIRAP and CCF analysis code COCOA are converted from KOS to Windows. Human reliability database has been also established in this year. In the area of new technology applications, fuzzy set theory and entropy theory are used to estimate component life and to develop a new measure of uncertainty importance. Finally, in the field of application study of PSA technique to reactor regulation, a strategic study to develop a dynamic risk management tool PEPSI and the determination of inspection and test priority of motor operated valves based on risk importance worths have been studied. (Author).
[Series: Medical Applications of the PHITS Code (2): Acceleration by Parallel Computing].
Furuta, Takuya; Sato, Tatsuhiko
2015-01-01
Time-consuming Monte Carlo dose calculation becomes feasible owing to the development of computer technology. However, the recent development is due to emergence of the multi-core high performance computers. Therefore, parallel computing becomes a key to achieve good performance of software programs. A Monte Carlo simulation code PHITS contains two parallel computing functions, the distributed-memory parallelization using protocols of message passing interface (MPI) and the shared-memory parallelization using open multi-processing (OpenMP) directives. Users can choose the two functions according to their needs. This paper gives the explanation of the two functions with their advantages and disadvantages. Some test applications are also provided to show their performance using a typical multi-core high performance workstation.
PREMOR: a point reactor exposure model computer code for survey analysis of power plant performance
Vondy, D.R.
1979-10-01
The PREMOR computer code was written to exploit a simple, two-group point nuclear reactor power plant model for survey analysis. Up to thirteen actinides, fourteen fission products, and one lumped absorber nuclide density are followed over a reactor history. Successive feed batches are accounted for with provision for from one to twenty batches resident. The effect of exposure of each of the batches to the same neutron flux is determined.
Assessment of uncertainties of the models used in thermal-hydraulic computer codes
Gricay, A. S.; Migrov, Yu. A.
2015-09-01
The article deals with matters concerned with the problem of determining the statistical characteristics of variable parameters (the variation range and distribution law) in analyzing the uncertainty and sensitivity of calculation results to uncertainty in input data. A comparative analysis of modern approaches to uncertainty in input data is presented. The need to develop an alternative method for estimating the uncertainty of model parameters used in thermal-hydraulic computer codes, in particular, in the closing correlations of the loop thermal hydraulics block, is shown. Such a method shall feature the minimal degree of subjectivism and must be based on objective quantitative assessment criteria. The method includes three sequential stages: selecting experimental data satisfying the specified criteria, identifying the key closing correlation using a sensitivity analysis, and carrying out case calculations followed by statistical processing of the results. By using the method, one can estimate the uncertainty range of a variable parameter and establish its distribution law in the above-mentioned range provided that the experimental information is sufficiently representative. Practical application of the method is demonstrated taking as an example the problem of estimating the uncertainty of a parameter appearing in the model describing transition to post-burnout heat transfer that is used in the thermal-hydraulic computer code KORSAR. The performed study revealed the need to narrow the previously established uncertainty range of this parameter and to replace the uniform distribution law in the above-mentioned range by the Gaussian distribution law. The proposed method can be applied to different thermal-hydraulic computer codes. In some cases, application of the method can make it possible to achieve a smaller degree of conservatism in the expert estimates of uncertainties pertinent to the model parameters used in computer codes.
Tight bounds on computing error-correcting codes by bounded-depth circuits with arbitrary gates
Gal, A.; Hansen, Kristoffer Arnsfelt; Koucky, Michal
2013-01-01
We bound the minimum number w of wires needed to compute any (asymptotically good) error-correcting code C:{0,1}Ω(n)→{0,1}n with minimum distance Ω(n), using unbounded fan-in circuits of depth d with arbitrary gates. Our main results are: 1) if d=2, then w=Θ(n (lgn/lglgn)2); 2) if d=3, then w...
Chia-Chang Hu
2005-01-01
A novel space-time adaptive near-far robust code-synchronization array detector for asynchronous DS-CDMA systems is developed in this paper. There are the same basic requirements that are needed by the conventional matched filter of an asynchronous DS-CDMA system. For the real-time applicability, a computationally efficient architecture of the proposed detector is developed that is based on the concept of the multistage Wiener filter (MWF) of Goldstein and Reed. This multistage technique resu...
Method for computing self-consistent solution in a gun code
Nelson, Eric M
2014-09-23
Complex gun code computations can be made to converge more quickly based on a selection of one or more relaxation parameters. An eigenvalue analysis is applied to error residuals to identify two error eigenvalues that are associated with respective error residuals. Relaxation values can be selected based on these eigenvalues so that error residuals associated with each can be alternately reduced in successive iterations. In some examples, relaxation values that would be unstable if used alone can be used.
Tight bounds on computing error-correcting codes by bounded-depth circuits with arbitrary gates
Gál, Anna; Hansen, Kristoffer Arnsfelt; Koucký, Michal;
2012-01-01
We bound the minimum number w of wires needed to compute any (asymptotically good) error-correcting code C:{0,1}Ω(n) -> {0,1}n with minimum distance Ω(n), using unbounded fan-in circuits of depth d with arbitrary gates. Our main results are: (1) If d=2 then w = Θ(n ({log n/ log log n})2). (2) If d...
Johansen, Peter Meincke
1996-01-01
New uniform closed-form expressions for physical theory of diffraction equivalent edge currents are derived for truncated incremental wedge strips. In contrast to previously reported expressions, the new expressions are well-behaved for all directions of incidence and observation and take a finit...... value for zero strip length. Consequently, the new equivalent edge currents are, to the knowledge of the author, the first that are well-suited for implementation in general computer codes...
Walowit, Jed A.
1994-01-01
A viewgraph presentation is made showing the capabilities of the computer code SPIRALI. Overall capabilities of SPIRALI include: computes rotor dynamic coefficients, flow, and power loss for cylindrical and face seals; treats turbulent, laminar, Couette, and Poiseuille dominated flows; fluid inertia effects are included; rotor dynamic coefficients in three (face) or four (cylindrical) degrees of freedom; includes effects of spiral grooves; user definable transverse film geometry including circular steps and grooves; independent user definable friction factor models for rotor and stator; and user definable loss coefficients for sudden expansions and contractions.
3D MHD Simulations of Tokamak Disruptions
Woodruff, Simon; Stuber, James
2014-10-01
Two disruption scenarios are modeled numerically by use of the CORSICA 2D equilibrium and NIMROD 3D MHD codes. The work follows the simulations of pressure-driven modes in DIII-D and VDEs in ITER. The aim of the work is to provide starting points for simulation of tokamak disruption mitigation techniques currently in the CDR phase for ITER. Pressure-driven instability growth rates previously observed in simulations of DIIID are verified; Halo and Hiro currents produced during vertical displacements are observed in simulations of ITER with implementation of resistive walls in NIMROD. We discuss plans to exercise new code capabilities and validation.
Multilevel Coding Schemes for Compute-and-Forward with Flexible Decoding
Hern, Brett
2011-01-01
We consider the design of coding schemes for the wireless two-way relaying channel when there is no channel state information at the transmitter. In the spirit of the compute and forward paradigm, we present a multilevel coding scheme that permits computation (or, decoding) of a class of functions at the relay. The function to be computed (or, decoded) is then chosen depending on the channel realization. We define such a class of functions which can be decoded at the relay using the proposed coding scheme and derive rates that are universally achievable over a set of channel gains when this class of functions is used at the relay. We develop our framework with general modulation formats in mind, but numerical results are presented for the case where each node transmits using the QPSK constellation. Numerical results with QPSK show that the flexibility afforded by our proposed scheme results in substantially higher rates than those achievable by always using a fixed function or by adapting the function at the ...
MHD thrust vectoring of a rocket engine
Labaune, Julien; Packan, Denis; Tholin, Fabien; Chemartin, Laurent; Stillace, Thierry; Masson, Frederic
2016-09-01
In this work, the possibility to use MagnetoHydroDynamics (MHD) to vectorize the thrust of a solid propellant rocket engine exhaust is investigated. Using a magnetic field for vectoring offers a mass gain and a reusability advantage compared to standard gimbaled, elastomer-joint systems. Analytical and numerical models were used to evaluate the flow deviation with a 1 Tesla magnetic field inside the nozzle. The fluid flow in the resistive MHD approximation is calculated using the KRONOS code from ONERA, coupling the hypersonic CFD platform CEDRE and the electrical code SATURNE from EDF. A critical parameter of these simulations is the electrical conductivity, which was evaluated using a set of equilibrium calculations with 25 species. Two models were used: local thermodynamic equilibrium and frozen flow. In both cases, chlorine captures a large fraction of free electrons, limiting the electrical conductivity to a value inadequate for thrust vectoring applications. However, when using chlorine-free propergols with 1% in mass of alkali, an MHD thrust vectoring of several degrees was obtained.
Multiphase integral reacting flow computer code (ICOMFLO): User`s guide
Chang, S.L.; Lottes, S.A.; Petrick, M.
1997-11-01
A copyrighted computational fluid dynamics computer code, ICOMFLO, has been developed for the simulation of multiphase reacting flows. The code solves conservation equations for gaseous species and droplets (or solid particles) of various sizes. General conservation laws, expressed by elliptic type partial differential equations, are used in conjunction with rate equations governing the mass, momentum, enthalpy, species, turbulent kinetic energy, and turbulent dissipation. Associated phenomenological submodels of the code include integral combustion, two parameter turbulence, particle evaporation, and interfacial submodels. A newly developed integral combustion submodel replacing an Arrhenius type differential reaction submodel has been implemented to improve numerical convergence and enhance numerical stability. A two parameter turbulence submodel is modified for both gas and solid phases. An evaporation submodel treats not only droplet evaporation but size dispersion. Interfacial submodels use correlations to model interfacial momentum and energy transfer. The ICOMFLO code solves the governing equations in three steps. First, a staggered grid system is constructed in the flow domain. The staggered grid system defines gas velocity components on the surfaces of a control volume, while the other flow properties are defined at the volume center. A blocked cell technique is used to handle complex geometry. Then, the partial differential equations are integrated over each control volume and transformed into discrete difference equations. Finally, the difference equations are solved iteratively by using a modified SIMPLER algorithm. The results of the solution include gas flow properties (pressure, temperature, density, species concentration, velocity, and turbulence parameters) and particle flow properties (number density, temperature, velocity, and void fraction). The code has been used in many engineering applications, such as coal-fired combustors, air
Agarwal, Sapan; Quach, Tu-Thach; Parekh, Ojas; Hsia, Alexander H.; DeBenedictis, Erik P.; James, Conrad D.; Marinella, Matthew J.; Aimone, James B.
2016-01-01
The exponential increase in data over the last decade presents a significant challenge to analytics efforts that seek to process and interpret such data for various applications. Neural-inspired computing approaches are being developed in order to leverage the computational properties of the analog, low-power data processing observed in biological systems. Analog resistive memory crossbars can perform a parallel read or a vector-matrix multiplication as well as a parallel write or a rank-1 update with high computational efficiency. For an N × N crossbar, these two kernels can be O(N) more energy efficient than a conventional digital memory-based architecture. If the read operation is noise limited, the energy to read a column can be independent of the crossbar size (O(1)). These two kernels form the basis of many neuromorphic algorithms such as image, text, and speech recognition. For instance, these kernels can be applied to a neural sparse coding algorithm to give an O(N) reduction in energy for the entire algorithm when run with finite precision. Sparse coding is a rich problem with a host of applications including computer vision, object tracking, and more generally unsupervised learning. PMID:26778946
Sapan eAgarwal
2016-01-01
Full Text Available The exponential increase in data over the last decade presents a significant challenge to analytics efforts that seek to process and interpret such data for various applications. Neural-inspired computing approaches are being developed in order to leverage the computational advantages of the analog, low-power data processing observed in biological systems. Analog resistive memory crossbars can perform a parallel read or a vector-matrix multiplication as well as a parallel write or a rank-1 update with high computational efficiency. For an NxN crossbar, these two kernels are at a minimum O(N more energy efficient than a digital memory-based architecture. If the read operation is noise limited, the energy to read a column can be independent of the crossbar size (O(1. These two kernels form the basis of many neuromorphic algorithms such as image, text, and speech recognition. For instance, these kernels can be applied to a neural sparse coding algorithm to give an O(N reduction in energy for the entire algorithm. Sparse coding is a rich problem with a host of applications including computer vision, object tracking, and more generally unsupervised learning.
Error threshold in topological quantum-computing models with color codes
Katzgraber, Helmut; Bombin, Hector; Martin-Delgado, Miguel A.
2009-03-01
Dealing with errors in quantum computing systems is possibly one of the hardest tasks when attempting to realize physical devices. By encoding the qubits in topological properties of a system, an inherent protection of the quantum states can be achieved. Traditional topologically-protected approaches are based on the braiding of quasiparticles. Recently, a braid-less implementation using brane-net condensates in 3-colexes has been proposed. In 2D it allows the transversal implementation of the whole Clifford group of quantum gates. In this work, we compute the error threshold for this topologically-protected quantum computing system in 2D, by means of mapping its error correction process onto a random 3-body Ising model on a triangular lattice. Errors manifest themselves as random perturbation of the plaquette interaction terms thus introducing frustration. Our results from Monte Carlo simulations suggest that these topological color codes are similarly robust to perturbations as the toric codes. Furthermore, they provide more computational capabilities and the possibility of having more qubits encoded in the quantum memory.
Farmer, M.T.; Sienicki, J.J.; Spencer, B.W.; Chu, C.C.
1992-01-01
A transient, one dimensional, finite difference computer code (MELTSPREAD-1) has been developed to predict spreading behavior of high temperature melts flowing over concrete and/or steel surfaces submerged in water, or without the effects of water if the surface is initially dry. This paper provides a summary overview of models and correlations currently implemented in the code, code validation activities completed thus far, LWR spreading-related safety issues for which the code has been applied, and the status of documentation for the code.
Farmer, M.T.; Sienicki, J.J.; Spencer, B.W.; Chu, C.C.
1992-04-01
A transient, one dimensional, finite difference computer code (MELTSPREAD-1) has been developed to predict spreading behavior of high temperature melts flowing over concrete and/or steel surfaces submerged in water, or without the effects of water if the surface is initially dry. This paper provides a summary overview of models and correlations currently implemented in the code, code validation activities completed thus far, LWR spreading-related safety issues for which the code has been applied, and the status of documentation for the code.
V.S.O.P. (99/09) Computer Code System for Reactor Physics and Fuel Cycle Simulation; Version 2009
Rütten, H.-J.; Haas, K. A.; Brockmann, H.; Ohlig, U.; Pohl, C.; Scherer, W.
2010-01-01
V.S.O.P.(99/ 09) represents the further development of V.S.O.P.(99/ 05). Compared to its precursor, the code system has been improved again in many details. The main motivation for this new code version was to update the basic nuclear libraries used by the code system. Thus, all cross section libraries involved in the code have now been based on ENDF/B-VII. V.S.O.P. is a computer code system for the comprehensive numerical simulation of the physics of thermal reactors. It implies the setup of...
Automatic Generation of OpenMP Directives and Its Application to Computational Fluid Dynamics Codes
Yan, Jerry; Jin, Haoqiang; Frumkin, Michael; Yan, Jerry (Technical Monitor)
2000-01-01
The shared-memory programming model is a very effective way to achieve parallelism on shared memory parallel computers. As great progress was made in hardware and software technologies, performance of parallel programs with compiler directives has demonstrated large improvement. The introduction of OpenMP directives, the industrial standard for shared-memory programming, has minimized the issue of portability. In this study, we have extended CAPTools, a computer-aided parallelization toolkit, to automatically generate OpenMP-based parallel programs with nominal user assistance. We outline techniques used in the implementation of the tool and discuss the application of this tool on the NAS Parallel Benchmarks and several computational fluid dynamics codes. This work demonstrates the great potential of using the tool to quickly port parallel programs and also achieve good performance that exceeds some of the commercial tools.
Louis, J.F.
1978-03-01
Research and development in open-cycle coal-fired MHD power generation is described. The scope and objectives of the MIT program are: (1) establish chemical, thermal, and electrical data to guide materials selection, develop improved detail designs, and support performance analyses of MHD electrode modules and insulator materials; (2) parametrically investigate selected electrode properties of critical design importance in chemical, thermal, and electrical environments simulating a coal-fired MHD generator; (3) develop combustion data pertinent to the design of MHD combustors; (4) establish techniques for the analysis and understanding of critical MHD phenomena which have an important bearing on MHD generator performance; such phenomena include inter-electrode breakdown, time-dependent behavior, effective plasma properties and plasma inhomogeneities; (5) establish the operating characteristics of an MHD disk generator; (6) continue work on computer techniques for modeling and for design and cost analysis of MHD components and the overall system; (7) integrate the engineering data and design criteria, as applicable, which are developed in the listed tasks into a model of the MHD channel; (8) participate in the US/USSR Cooperative Program in MHD Power Generation; and (9) participate in technical support of the DOE MHD Project Office. Progress in each of these areas is reported. (WHK)
On the Computational Complexity of Sphere Decoder for Lattice Space-Time Coded MIMO Channel
Abediseid, Walid
2011-01-01
The exact complexity analysis of the basic sphere decoder for general space-time codes applied to multi-input multi-output (MIMO) wireless channel is known to be difficult. In this work, we shed the light on the computational complexity of sphere decoding for the quasi-static, LAttice Space-Time (LAST) coded MIMO channel. Specifically, we derive the asymptotic tail distribution of the decoder's computational complexity in the high signal-to-noise ratio (SNR) regime. For the uncoded $M\\times N$ MIMO channel (e.g., V-BLAST), the analysis in [6] revealed that the tail distribution of such a decoder is of a Pareto-type with tail exponent that is equivalent to $N-M+1$. In our analysis, we show that the tail exponent of the sphere decoder's complexity distribution is equivalent to the diversity-multiplexing tradeoff achieved by LAST coding and lattice decoding schemes. This leads to extend the channel's tradeoff to include the decoding complexity. Moreover, we show analytically how minimum-mean square-error decisio...
Chung, Chang Hyun; You, Young Woo; Huh, Chang Wook; Kim, Ju Yeul; Kim Do Hyung; Kim, Yoon Ik; Yang, Hui Chang [Seoul National University, Seoul (Korea, Republic of); Jae, Moo Sung [Hansung University, Seoul (Korea, Republic of)
1997-07-01
The objective of this study is to develop the appropriate procedure that can evaluate the human error in LP/S(lower power/shutdown) and the computer code that calculate the human error probabilities(HEPs) using this framework. The assessment of applicability of the typical HRA methodologies to LP/S is conducted and a new HRA procedure, SEPLOT (Systematic Evaluation Procedure for LP/S Operation Tasks) which presents the characteristics of LP/S is developed by selection and categorization of human actions by reviewing present studies. This procedure is applied to evaluate the LOOP(Loss of Off-site Power) sequence and the HEPs obtained by using SEPLOT are used to quantitative evaluation of the core uncovery frequency. In this evaluation one of the dynamic reliability computer codes, DYLAM-3 which has the advantages against the ET/FT is used. The SEPLOT developed in this study can give the basis and arrangement as to the human error evaluation technique. And this procedure can make it possible to assess the dynamic aspects of accidents leading to core uncovery applying the HEPs obtained by using the SEPLOT as input data to DYLAM-3 code, Eventually, it is expected that the results of this study will contribute to improve safety in LP/S and reduce uncertainties in risk. 57 refs. 17 tabs., 33 figs. (author)
Research on the improvement of nuclear safety -Improvement of level 1 PSA computer code package-
Park, Chang Kyoo; Kim, Tae Woon; Kim, Kil Yoo; Han, Sang Hoon; Jung, Won Dae; Jang, Seung Chul; Yang, Joon Un; Choi, Yung; Sung, Tae Yong; Son, Yung Suk; Park, Won Suk; Jung, Kwang Sub; Kang Dae Il; Park, Jin Heui; Hwang, Mi Jung; Hah, Jae Joo [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)
1995-07-01
This year is the third year of the Government-sponsored mid- and long-term nuclear power technology development project. The scope of this sub project titled on `The improvement of level-1 PSA computer codes` is divided into three main activities : (1) Methodology development on the underdeveloped fields such as risk assessment technology for plant shutdown and low power situations, (2) Computer code package development for level-1 PSA, (3) Applications of new technologies to reactor safety assessment. At first, in this area of shutdown risk assessment technology development, plant outage experiences of domestic plants are reviewed and plant operating states (POS) are decided. A sample core damage frequency is estimated for over draining event in RCS low water inventory i.e. mid-loop operation. Human reliability analysis and thermal hydraulic support analysis are identified to be needed to reduce uncertainty. Two design improvement alternatives are evaluated using PSA technique for mid-loop operation situation: one is use of containment spray system as backup of shutdown cooling system and the other is installation of two independent level indication system. Procedure change is identified more preferable option to hardware modification in the core damage frequency point of view. Next, level-1 PSA code KIRAP is converted to PC-windows environment. For the improvement of efficiency in performing PSA, the fast cutest generation algorithm and an analytical technique for handling logical loop in fault tree modeling are developed. 48 figs, 15 tabs, 59 refs. (Author).
Calculations of reactor-accident consequences, Version 2. CRAC2: computer code user's guide
Ritchie, L.T.; Johnson, J.D.; Blond, R.M.
1983-02-01
The CRAC2 computer code is a revision of the Calculation of Reactor Accident Consequences computer code, CRAC, developed for the Reactor Safety Study. The CRAC2 computer code incorporates significant modeling improvements in the areas of weather sequence sampling and emergency response, and refinements to the plume rise, atmospheric dispersion, and wet deposition models. New output capabilities have also been added. This guide is to facilitate the informed and intelligent use of CRAC2. It includes descriptions of the input data, the output results, the file structures, control information, and five sample problems.
Computer simulation of an advanced combustor for clean coal technology
Chang, S.L.; Lottes, S.A.
1992-01-01
Magnetohydrodynamic (MHD) power generation is a clean coal technology because of its higher thermal efficiency and lower pollutant emission. Argonne National Laboratory used a comprehensive integral combustion computer code to aid the development of a TRW's second stage combustor for MHD power generation. The integral combustion code is a computer code for two-phase, two-dimensional, steady state, turbulent, and reacting flows, based on mass, momentum, and energy conservation laws for multiple gas species and solid particles of variable sizes. In the MHD second stage combustor, opposed jets of oxidizer are injected into a confined cross-stream coal gas flow laden with seed particles. The performance of the downstream MHD power generation channel depends mainly on the degree and the uniformity of gas ionization, which, in turn, depends on the uniformity of temperature and seed vapor distributions leaving the combustor. The simulation provides in-depth information of flow, combustion, and heat transfer patterns in the combustor, which is used to predict ranges of combustor operating conditions for optimum performance of the MHD system.
Computer simulation of an advanced combustor for clean coal technology
Chang, S.L.; Lottes, S.A.
1992-09-01
Magnetohydrodynamic (MHD) power generation is a clean coal technology because of its higher thermal efficiency and lower pollutant emission. Argonne National Laboratory used a comprehensive integral combustion computer code to aid the development of a TRW`s second stage combustor for MHD power generation. The integral combustion code is a computer code for two-phase, two-dimensional, steady state, turbulent, and reacting flows, based on mass, momentum, and energy conservation laws for multiple gas species and solid particles of variable sizes. In the MHD second stage combustor, opposed jets of oxidizer are injected into a confined cross-stream coal gas flow laden with seed particles. The performance of the downstream MHD power generation channel depends mainly on the degree and the uniformity of gas ionization, which, in turn, depends on the uniformity of temperature and seed vapor distributions leaving the combustor. The simulation provides in-depth information of flow, combustion, and heat transfer patterns in the combustor, which is used to predict ranges of combustor operating conditions for optimum performance of the MHD system.
SEACC: the systems engineering and analysis computer code for small wind systems
Tu, P.K.C.; Kertesz, V.
1983-03-01
The systems engineering and analysis (SEA) computer program (code) evaluates complete horizontal-axis SWECS performance. Rotor power output as a function of wind speed and energy production at various wind regions are predicted by the code. Efficiencies of components such as gearbox, electric generators, rectifiers, electronic inverters, and batteries can be included in the evaluation process to reflect the complete system performance. Parametric studies can be carried out for blade design characteristics such as airfoil series, taper rate, twist degrees and pitch setting; and for geometry such as rotor radius, hub radius, number of blades, coning angle, rotor rpm, etc. Design tradeoffs can also be performed to optimize system configurations for constant rpm, constant tip speed ratio and rpm-specific rotors. SWECS energy supply as compared to the load demand for each hour of the day and during each session of the year can be assessed by the code if the diurnal wind and load distributions are known. Also available during each run of the code is blade aerodynamic loading information.
A fully parallel, high precision, N-body code running on hybrid computing platforms
Capuzzo-Dolcetta, R; Punzo, D
2012-01-01
We present a new implementation of the numerical integration of the classical, gravitational, N-body problem based on a high order Hermite's integration scheme with block time steps, with a direct evaluation of the particle-particle forces. The main innovation of this code (called HiGPUs) is its full parallelization, exploiting both OpenMP and MPI in the use of the multicore Central Processing Units as well as either Compute Unified Device Architecture (CUDA) or OpenCL for the hosted Graphic Processing Units. We tested both performance and accuracy of the code using up to 256 GPUs in the supercomputer IBM iDataPlex DX360M3 Linux Infiniband Cluster provided by the italian supercomputing consortium CINECA, for values of N up to 8 millions. We were able to follow the evolution of a system of 8 million bodies for few crossing times, task previously unreached by direct summation codes. The code is freely available to the scientific community.
Development of a space radiation Monte Carlo computer simulation based on the FLUKA and ROOT codes
Pinsky, L; Ferrari, A; Sala, P; Carminati, F; Brun, R
2001-01-01
This NASA funded project is proceeding to develop a Monte Carlo-based computer simulation of the radiation environment in space. With actual funding only initially in place at the end of May 2000, the study is still in the early stage of development. The general tasks have been identified and personnel have been selected. The code to be assembled will be based upon two major existing software packages. The radiation transport simulation will be accomplished by updating the FLUKA Monte Carlo program, and the user interface will employ the ROOT software being developed at CERN. The end-product will be a Monte Carlo-based code which will complement the existing analytic codes such as BRYNTRN/HZETRN presently used by NASA to evaluate the effects of radiation shielding in space. The planned code will possess the ability to evaluate the radiation environment for spacecraft and habitats in Earth orbit, in interplanetary space, on the lunar surface, or on a planetary surface such as Mars. Furthermore, it will be usef...
The Proteus Navier-Stokes code. [two and three dimensional computational fluid dynamics
Towne, Charles E.; Schwab, John R.
1992-01-01
An effort is currently underway at NASA Lewis to develop two and three dimensional Navier-Stokes codes, called Proteus, for aerospace propulsion applications. Proteus solves the Reynolds-averaged, unsteady, compressible Navier-Stokes equations in strong conservation law form. Turbulence is modeled using a Baldwin-Lomax based algebraic eddy viscosity model. In addition, options are available to solve thin layer or Euler equations, and to eliminate the energy equation by assuming constant stagnation enthalpy. An extensive series of validation cases have been run, primarily using the two dimensional planar/axisymmetric version of the code. Several flows were computed that have exact solution such as: fully developed channel and pipe flow; Couette flow with and without pressure gradients; unsteady Couette flow formation; flow near a suddenly accelerated flat plate; flow between concentric rotating cylinders; and flow near a rotating disk. The two dimensional version of the Proteus code has been released, and the three dimensional code is scheduled for release in late 1991.
A 3rd Order WENO GLM-MHD Scheme for Magnetic Reconnection
FENG Xueshang; ZHOU Yufen; HU Yanqi
2006-01-01
A new numerical scheme of 3rd order Weighted Essentially Non-Oscillatory (WENO)type for 2.5D mixed GLM-MHD in Cartesian coordinates is proposed. The MHD equations are modified by combining the arguments as by Dellar and Dedner et al to couple the divergence constraint with the evolution equations using a Generalized Lagrange Multiplier (GLM). Moreover, the magnetohydrodynamic part of the GLM-MHD system is still in conservation form. Meanwhile, this method is very easy to add to an existing code since the underlying MHD solver does not have to be modified. To show the validation and capacity of its application to MHD problem modelling,interaction between a magnetosonic shock and a denser cloud and magnetic reconnection problems are used to verify this new MHD code. The numerical tests for 2D Orszag and Tang's MHD vortex,interaction between a magnetosonic shock and a denser cloud and magnetic reconnection problems show that the third order WENO MHD solvers are robust and yield reliable results by the new mixed GLM or the mixed EGLM correction here even if it can not be shown that how the divergence errors are transported as well as damped as done for one dimensional ideal MHD by Dedner et al.
Parallel Computing Characteristics of CUPID code under MPI and Hybrid environment
Lee, Jae Ryong; Yoon, Han Young [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of); Jeon, Byoung Jin; Choi, Hyoung Gwon [Seoul National Univ. of Science and Technology, Seoul (Korea, Republic of)
2014-05-15
In this paper, a characteristic of parallel algorithm is presented for solving an elliptic type equation of CUPID via domain decomposition method using the MPI and the parallel performance is estimated in terms of a scalability which shows the speedup ratio. In addition, the time-consuming pattern of major subroutines is studied. Two different grid systems are taken into account: 40,000 meshes for coarse system and 320,000 meshes for fine system. Since the matrix of the CUPID code differs according to whether the flow is single-phase or two-phase, the effect of matrix shape is evaluated. Finally, the effect of the preconditioner for matrix solver is also investigated. Finally, the hybrid (OpenMP+MPI) parallel algorithm is introduced and discussed in detail for solving pressure solver. Component-scale thermal-hydraulics code, CUPID has been developed for two-phase flow analysis, which adopts a three-dimensional, transient, three-field model, and parallelized to fulfill a recent demand for long-transient and highly resolved multi-phase flow behavior. In this study, the parallel performance of the CUPID code was investigated in terms of scalability. The CUPID code was parallelized with domain decomposition method. The MPI library was adopted to communicate the information at the neighboring domain. For managing the sparse matrix effectively, the CSR storage format is used. To take into account the characteristics of the pressure matrix which turns to be asymmetric for two-phase flow, both single-phase and two-phase calculations were run. In addition, the effect of the matrix size and preconditioning was also investigated. The fine mesh calculation shows better scalability than the coarse mesh because the number of coarse mesh does not need to decompose the computational domain excessively. The fine mesh can be present good scalability when dividing geometry with considering the ratio between computation and communication time. For a given mesh, single-phase flow
Assessment of computer codes for VVER-440/213-type nuclear power plants
Szabados, L.; Ezsol, Gy.; Perneczky [Atomic Energy Research Institute, Budapest (Hungary)
1995-09-01
Nuclear power plant of VVER-440/213 designed by the former USSR have a number of special features. As a consequence of these features the transient behaviour of such a reactor system should be different from the PWR system behaviour. To study the transient behaviour of the Hungarian Paks Nuclear Power Plant of VVER-440/213-type both analytical and experimental activities have been performed. The experimental basis of the research in the PMK-2 integral-type test facility , which is a scaled down model of the plant. Experiments performed on this facility have been used to assess thermal-hydraulic system codes. Four tests were selected for {open_quotes}Standard Problem Exercises{close_quotes} of the International Atomic Energy Agency. Results of the 4th Exercise, of high international interest, are presented in the paper, focusing on the essential findings of the assessment of computer codes.
Revised uranium--plutonium cycle PWR and BWR models for the ORIGEN computer code
Croff, A. G.; Bjerke, M. A.; Morrison, G. W.; Petrie, L. M.
1978-09-01
Reactor physics calculations and literature searches have been conducted, leading to the creation of revised enriched-uranium and enriched-uranium/mixed-oxide-fueled PWR and BWR reactor models for the ORIGEN computer code. These ORIGEN reactor models are based on cross sections that have been taken directly from the reactor physics codes and eliminate the need to make adjustments in uncorrected cross sections in order to obtain correct depletion results. Revised values of the ORIGEN flux parameters THERM, RES, and FAST were calculated along with new parameters related to the activation of fuel-assembly structural materials not located in the active fuel zone. Recommended fuel and structural material masses and compositions are presented. A summary of the new ORIGEN reactor models is given.
A general panel sizing computer code and its application to composite structural panels
Anderson, M. S.; Stroud, W. J.
1978-01-01
A computer code for obtaining the dimensions of optimum (least mass) stiffened composite structural panels is described. The procedure, which is based on nonlinear mathematical programming and a rigorous buckling analysis, is applicable to general cross sections under general loading conditions causing buckling. A simplified method of accounting for bow-type imperfections is also included. Design studies in the form of structural efficiency charts for axial compression loading are made with the code for blade and hat stiffened panels. The effects on panel mass of imperfections, material strength limitations, and panel stiffness requirements are also examined. Comparisons with previously published experimental data show that accounting for imperfections improves correlation between theory and experiment.
Development of system of computer codes for severe accident analysis and its applications
Jang, H. S.; Jeon, M. H.; Cho, N. J. and others [Korea Advanced Institute of Science and Technology, Taejon (Korea, Republic of)
1992-01-15
The objectives of this study is to develop a system of computer codes for postulated severe accident analyses in nuclear power plants. This system of codes is necessary to conduct Individual Plant Examination for domestic nuclear power plants. As a result of this study, one can conduct severe accident assessments more easily, and can extract the plant-specific vulnerabilities for severe accidents and at the same time the ideas for enhancing overall accident-resistance. Severe accident can be mitigated by the proper accident management strategies. Some operator action for mitigation can lead to more disastrous result and thus uncertain severe accident phenomena must be well recognized. There must be further research for development of severe accident management strategies utilizing existing plant resources as well as new design concepts.
Development of a computer code for dynamic analysis of the primary circuit of advanced reactors
Rocha, Jussie Soares da; Lira, Carlos A.B.O.; Magalhaes, Mardson A. de Sa, E-mail: cabol@ufpe.b [Universidade Federal de Pernambuco (DEN/UFPE), Recife, PE (Brazil). Dept. de Energia Nuclear
2011-07-01
Currently, advanced reactors are being developed, seeking for enhanced safety, better performance and low environmental impacts. Reactor designs must follow several steps and numerous tests before a conceptual project could be certified. In this sense, computational tools become indispensable in the preparation of such projects. Thus, this study aimed at the development of a computational tool for thermal-hydraulic analysis by coupling two computer codes to evaluate the influence of transients caused by pressure variations and flow surges in the region of the primary circuit of IRIS reactor between the core and the pressurizer. For the simulation, it was used a situation of 'insurge', characterized by the entry of water in the pressurizer, due to the expansion of the refrigerant in the primary circuit. This expansion was represented by a pressure disturbance in step form, through the block 'step' of SIMULINK, thus enabling the transient startup. The results showed that the dynamic tool, obtained through the coupling of the codes, generated very satisfactory responses within model limitations, preserving the most important phenomena in the process. (author)
Müller, C.; Hughes, E. D.; Niederauer, G. F.; Wilkening, H.; Travis, J. R.; Spore, J. W.; Royl, P.; Baumann, W.
1998-10-01
Los Alamos National Laboratory (LANL) and Forschungszentrum Karlsruhe (FzK) are developing GASFLOW, a three-dimensional (3D) fluid dynamics field code as a best- estimate tool to characterize local phenomena within a flow field. Examples of 3D phenomena include circulation patterns; flow stratification; hydrogen distribution mixing and stratification; combustion and flame propagation; effects of noncondensable gas distribution on local condensation and evaporation; and aerosol entrainment, transport, and deposition. An analysis with GASFLOW will result in a prediction of the gas composition and discrete particle distribution in space and time throughout the facility and the resulting pressure and temperature loadings on the walls and internal structures with or without combustion. A major application of GASFLOW is for predicting the transport, mixing, and combustion of hydrogen and other gases in nuclear reactor containment and other facilities. It has been applied to situations involving transporting and distributing combustible gas mixtures. It has been used to study gas dynamic behavior in low-speed, buoyancy-driven flows, as well as sonic flows or diffusion dominated flows; and during chemically reacting flows, including deflagrations. The effects of controlling such mixtures by safety systems can be analyzed. The code version described in this manual is designated GASFLOW 2.1, which combines previous versions of the United States Nuclear Regulatory Commission code HMS (for Hydrogen Mixing Studies) and the Department of Energy and FzK versions of GASFLOW. The code was written in standard Fortran 90. This manual comprises three volumes. Volume I describes the governing physical equations and computational model. Volume II describes how to use the code to set up a model geometry, specify gas species and material properties, define initial and boundary conditions, and specify different outputs, especially graphical displays. Sample problems are included. Volume
Mohebbi, Ali; Engelsholm, Signe K.D.; Puthusserypady, Sadasivan
2015-01-01
In this pilot study, a novel and minimalistic Brain Computer Interface (BCI) based wheelchair control application was developed. The system was based on pseudorandom code modulated Visual Evoked Potentials (c-VEPs). The visual stimuli in the scheme were generated based on the Gold code...
Sessarego, Matias; Ramos García, Néstor; Sørensen, Jens Nørkær
2017-01-01
Aerodynamic and structural dynamic performance analysis of modern wind turbines are routinely estimated in the wind energy field using computational tools known as aeroelastic codes. Most aeroelastic codes use the blade element momentum (BEM) technique to model the rotor aerodynamics and a modal...
Prediction of detonation and JWL eos parameters of energetic materials using EXPLO5 computer code
Peter, Xolani
2016-09-01
Full Text Available (Cowperthwaite and Zwisler, 1976), CHEETAH (Fried, 1996), EXPLO5(Sućeska , 2001), BARUT-X (Cengiz et al., 2007). These computer codes describe the detonation on the basis of the solution of Euler’s hydrodynamic equation based on the description of an equation... of detonation products equation of state from cylinder test: Analytical model and numerical analysis. Thermal Science, 19(1), pp. 35-48. Fried, L.E., 1996. CHEETAH 1.39 user’s manual. Lawrence Livermore National Laboratory. Göbel, M., 2009. Energetic...
On the application of computational fluid dynamics codes for liquefied natural gas dispersion.
Luketa-Hanlin, Anay; Koopman, Ronald P; Ermak, Donald L
2007-02-20
Computational fluid dynamics (CFD) codes are increasingly being used in the liquefied natural gas (LNG) industry to predict natural gas dispersion distances. This paper addresses several issues regarding the use of CFD for LNG dispersion such as specification of the domain, grid, boundary and initial conditions. A description of the k-epsilon model is presented, along with modifications required for atmospheric flows. Validation issues pertaining to the experimental data from the Burro, Coyote, and Falcon series of LNG dispersion experiments are also discussed. A description of the atmosphere is provided as well as discussion on the inclusion of the Coriolis force to model very large LNG spills.
Fletcher, C. D.
The capability to perform thermal-hydraulic analyses of a space reactor using the ATHENA computer code is demonstrated. The fast reactor, liquid-lithium coolant loops, and lithium-filled heat pipes of the preliminary General electric SP-100 design were modeled with ATHENA. Two demonstration transient calculations were performed simulating accident conditions. Calculated results are available for display using the Nuclear Plant Analyzer color graphics analysis tool in addition to traditional plots. ATHENA-calculated results appear reasonable, both for steady state full power conditions, and for the two transients. This analysis represents the first known transient thermal-hydraulic simulation using an integral space reactor system model incorporating heat pipes.
Capabilities of the ATHENA computer code for modeling the SP-100 space reactor concept
Fletcher, C. D.
1985-09-01
The capability to perform thermal-hydraulic analyses of an SP-100 space reactor was demonstrated using the ATHENA computer code. The preliminary General Electric SP-100 design was modeled using Athena. The model simulates the fast reactor, liquid-lithium coolant loops, and lithium-filled heat pipes of this design. Two ATHENA demonstration calculations were performed simulating accident scenarios. A mask for the SP-100 model and an interface with the Nuclear Plant Analyzer (NPA) were developed, allowing a graphic display of the calculated results on the NPA.
Discrete logarithm computations over finite fields using Reed-Solomon codes
Augot, Daniel; Morain, François
2012-01-01
Cheng and Wan have related the decoding of Reed-Solomon codes to the computation of discrete logarithms over finite fields, with the aim of proving the hardness of their decoding. In this work, we experiment with solving the discrete logarithm over GF(q^h) using Reed-Solomon decoding. For fixed h and q going to infinity, we introduce an algorithm (RSDL) needing O~(h! q^2) operations over GF(q), operating on a q x q matrix with (h+2) q non-zero coefficients. We give faster variants including a...
Resin Matrix/Fiber Reinforced Composite Material, Ⅱ: Method of Solution and Computer Code
Li Chensha(李辰砂); Jiao Caishan; Liu Ying; Wang Zhengping; Wang Hongjie; Cao Maosheng
2003-01-01
According to a mathematical model which describes the curing process of composites constructed from continuous fiber-reinforced, thermosetting resin matrix prepreg materials, and the consolidation of the composites, the solution method to the model is made and a computer code is developed, which for flat-plate composites cured by a specified cure cycle, provides the variation of temperature distribution, the cure reaction process in the resin, the resin flow and fibers stress inside the composite, the void variation and the residual stress distribution.
Fuel burnup analysis for Thai research reactor by using MCNPX computer code
Sangkaew, S.; Angwongtrakool, T.; Srimok, B.
2017-06-01
This paper presents the fuel burnup analysis of the Thai research reactor (TRR-1/M1), TRIGA Mark-III, operated by Thailand Institute of Nuclear Technology (TINT) in Bangkok, Thailand. The modelling software used in this analysis is MCNPX (MCNP eXtended) version 2.6.0, a Fortran90 Monte Carlo radiation transport computer code. The analysis results will cover the core excess reactivity, neutron fluxes at the irradiation positions and neutron detector tubes, power distribution, fuel burnup, and fission products based on fuel cycle of first reactor core arrangement.
Reznik, A. L.; Tuzikov, A. V.; Solov'ev, A. A.; Torgov, A. V.
2016-11-01
Original codes and combinatorial-geometrical computational schemes are presented, which are developed and applied for finding exact analytical formulas that describe the probability of errorless readout of random point images recorded by a scanning aperture with a limited number of threshold levels. Combinatorial problems encountered in the course of the study and associated with the new generalization of Catalan numbers are formulated and solved. An attempt is made to find the explicit analytical form of these numbers, which is, on the one hand, a necessary stage of solving the basic research problem and, on the other hand, an independent self-consistent problem.
Apparatus, Method, and Computer Program for a Resolution-Enhanced Pseudo-Noise Code Technique
Li, Steven X. (Inventor)
2015-01-01
An apparatus, method, and computer program for a resolution enhanced pseudo-noise coding technique for 3D imaging is provided. In one embodiment, a pattern generator may generate a plurality of unique patterns for a return to zero signal. A plurality of laser diodes may be configured such that each laser diode transmits the return to zero signal to an object. Each of the return to zero signal includes one unique pattern from the plurality of unique patterns to distinguish each of the transmitted return to zero signals from one another.
无
2001-01-01
The fundamental algorithm of light beam propagation in high powerlaser system is investigated and the corresponding computational codes are given. It is shown that the number of modulation ring due to the diffraction is related to the size of the pinhole in spatial filter (in terms of the times of diffraction limitation, i.e. TDL) and the Fresnel number of the laser system; for the complex laser system with multi-spatial filters and free space, the system can be investigated by the reciprocal rule of operators.
Modeling of field lysimeter release data using the computer code dust
Sullivan, T.M.; Fitzgerald, I.T. (Brookhaven National Lab., Upton, NY (United States)); McConnell, J.W.; Rogers, R.D. (Idaho National Engineering Lab., Idaho Falls, ID (United States))
1993-01-01
In this study, it was attempted to match the experimentally measured mass release data collected over a period of seven years by investigators from Idaho National Engineering Laboratory from the lysimeters at Oak Ridge National Laboratory and Argonne National Laboratory using the computer code DUST. The influence of the dispersion coefficient and distribution coefficient on mass release was investigated. Both were found to significantly influence mass release over the seven year period. It is recommended that these parameters be measured on a site specific basis to enhance the understanding of the system.
Modeling of field lysimeter release data using the computer code dust
Sullivan, T.M.; Fitzgerald, I.T. [Brookhaven National Lab., Upton, NY (United States); McConnell, J.W.; Rogers, R.D. [Idaho National Engineering Lab., Idaho Falls, ID (United States)
1993-03-01
In this study, it was attempted to match the experimentally measured mass release data collected over a period of seven years by investigators from Idaho National Engineering Laboratory from the lysimeters at Oak Ridge National Laboratory and Argonne National Laboratory using the computer code DUST. The influence of the dispersion coefficient and distribution coefficient on mass release was investigated. Both were found to significantly influence mass release over the seven year period. It is recommended that these parameters be measured on a site specific basis to enhance the understanding of the system.
Nichols, B.D.; Mueller, C.; Necker, G.A.; Travis, J.R.; Spore, J.W.; Lam, K.L.; Royl, P.; Redlinger, R.; Wilson, T.L.
1998-10-01
Los Alamos National Laboratory (LANL) and Forschungszentrum Karlsruhe (FzK) are developing GASFLOW, a three-dimensional (3D) fluid dynamics field code as a best-estimate tool to characterize local phenomena within a flow field. Examples of 3D phenomena include circulation patterns; flow stratification; hydrogen distribution mixing and stratification; combustion and flame propagation; effects of noncondensable gas distribution on local condensation and evaporation; and aerosol entrainment, transport, and deposition. An analysis with GASFLOW will result in a prediction of the gas composition and discrete particle distribution in space and time throughout the facility and the resulting pressure and temperature loadings on the walls and internal structures with or without combustion. A major application of GASFLOW is for predicting the transport, mixing, and combustion of hydrogen and other gases in nuclear reactor containments and other facilities. It has been applied to situations involving transporting and distributing combustible gas mixtures. It has been used to study gas dynamic behavior (1) in low-speed, buoyancy-driven flows, as well as sonic flows or diffusion dominated flows; and (2) during chemically reacting flows, including deflagrations. The effects of controlling such mixtures by safety systems can be analyzed. The code version described in this manual is designated GASFLOW 2.1, which combines previous versions of the United States Nuclear Regulatory Commission code HMS (for Hydrogen Mixing Studies) and the Department of Energy and FzK versions of GASFLOW. The code was written in standard Fortran 90. This manual comprises three volumes. Volume I describes the governing physical equations and computational model. Volume II describes how to use the code to set up a model geometry, specify gas species and material properties, define initial and boundary conditions, and specify different outputs, especially graphical displays. Sample problems are included
Digital Poetry: A Narrow Relation between Poetics and the Codes of the Computational Logic
Laurentiz, Silvia
The project "Percorrendo Escrituras" (Walking Through Writings Project) has been developed at ECA-USP Fine Arts Department. Summarizing, it intends to study different structures of digital information that share the same universe and are generators of a new aesthetics condition. The aim is to search which are the expressive possibilities of the computer among the algorithm functions and other of its specific properties. It is a practical, theoretical and interdisciplinary project where the study of programming evolutionary language, logic and mathematics take us to poetic experimentations. The focus of this research is the digital poetry, and it comes from poetics of permutation combinations and culminates with dynamic and complex systems, autonomous, multi-user and interactive, through agents generation derivations, filtration and emergent standards. This lecture will present artworks that use some mechanisms introduced by cybernetics and the notion of system in digital poetry that demonstrate the narrow relationship between poetics and the codes of computational logic.
Blanket-relevant liquid metal MHD channel flows: Data base and optimization simulation development
Evtushenko, I.A.; Kirillov, I.R.; Sidorenkov, S.I. [D.V. Efremov Inst. of Electrophysical Apparatus, St Petersburg (Russian Federation)
1995-12-31
The problems of generalization and integration of test, theoretical and design data relevant to liquid metal (LM) blanket are discussed in present work. First results on MHD data base and LM blanket optimization codes are presented.
An implementation of a tree code on a SIMD, parallel computer
Olson, Kevin M.; Dorband, John E.
1994-01-01
We describe a fast tree algorithm for gravitational N-body simulation on SIMD parallel computers. The tree construction uses fast, parallel sorts. The sorted lists are recursively divided along their x, y and z coordinates. This data structure is a completely balanced tree (i.e., each particle is paired with exactly one other particle) and maintains good spatial locality. An implementation of this tree-building algorithm on a 16k processor Maspar MP-1 performs well and constitutes only a small fraction (approximately 15%) of the entire cycle of finding the accelerations. Each node in the tree is treated as a monopole. The tree search and the summation of accelerations also perform well. During the tree search, node data that is needed from another processor is simply fetched. Roughly 55% of the tree search time is spent in communications between processors. We apply the code to two problems of astrophysical interest. The first is a simulation of the close passage of two gravitationally, interacting, disk galaxies using 65,636 particles. We also simulate the formation of structure in an expanding, model universe using 1,048,576 particles. Our code attains speeds comparable to one head of a Cray Y-MP, so single instruction, multiple data (SIMD) type computers can be used for these simulations. The cost/performance ratio for SIMD machines like the Maspar MP-1 make them an extremely attractive alternative to either vector processors or large multiple instruction, multiple data (MIMD) type parallel computers. With further optimizations (e.g., more careful load balancing), speeds in excess of today's vector processing computers should be possible.
Wavelet subband coding of computer simulation output using the A++ array class library
Bradley, J.N.; Brislawn, C.M.; Quinlan, D.J.; Zhang, H.D. [Los Alamos National Lab., NM (United States); Nuri, V. [Washington State Univ., Pullman, WA (United States). School of EECS
1995-07-01
The goal of the project is to produce utility software for off-line compression of existing data and library code that can be called from a simulation program for on-line compression of data dumps as the simulation proceeds. Naturally, we would like the amount of CPU time required by the compression algorithm to be small in comparison to the requirements of typical simulation codes. We also want the algorithm to accomodate a wide variety of smooth, multidimensional data types. For these reasons, the subband vector quantization (VQ) approach employed in has been replaced by a scalar quantization (SQ) strategy using a bank of almost-uniform scalar subband quantizers in a scheme similar to that used in the FBI fingerprint image compression standard. This eliminates the considerable computational burdens of training VQ codebooks for each new type of data and performing nearest-vector searches to encode the data. The comparison of subband VQ and SQ algorithms in indicated that, in practice, there is relatively little additional gain from using vector as opposed to scalar quantization on DWT subbands, even when the source imagery is from a very homogeneous population, and our subjective experience with synthetic computer-generated data supports this stance. It appears that a careful study is needed of the tradeoffs involved in selecting scalar vs. vector subband quantization, but such an analysis is beyond the scope of this paper. Our present work is focused on the problem of generating wavelet transform/scalar quantization (WSQ) implementations that can be ported easily between different hardware environments. This is an extremely important consideration given the great profusion of different high-performance computing architectures available, the high cost associated with learning how to map algorithms effectively onto a new architecture, and the rapid rate of evolution in the world of high-performance computing.
Computer simulation of Angra-2 PWR nuclear reactor core using MCNPX code
Medeiros, Marcos P.C. de; Rebello, Wilson F., E-mail: eng.cavaliere@ime.eb.br, E-mail: rebello@ime.eb.br [Instituto Militar de Engenharia - Secao de Engenharia Nuclear, Rio de Janeiro, RJ (Brazil); Oliveira, Claudio L. [Universidade Gama Filho, Departamento de Matematica, Rio de Janeiro, RJ (Brazil); Vellozo, Sergio O., E-mail: vellozo@cbpf.br [Centro Tecnologico do Exercito. Divisao de Defesa Quimica, Biologica e Nuclear, Rio de Janeiro, RJ (Brazil); Silva, Ademir X. da, E-mail: ademir@nuclear.ufrj.br [Coordenacao dos Programas de Pos Gaduacao de Engenharia (COPPE/UFRJ), Rio de Janeiro, RJ (Brazil)
2011-07-01
In this work the MCNPX (Monte Carlo N-Particle Transport Code) code was used to develop a computerized model of the core of Angra 2 PWR (Pressurized Water Reactor) nuclear reactor. The model was created without any kind of homogenization, but using real geometric information and material composition of that reactor, obtained from the FSAR (Final Safety Analysis Report). The model is still being improved and the version presented in this work is validated by comparing values calculated by MCNPX with results calculated by others means and presented on FSAR. This paper shows the results already obtained to K{sub eff} and K{infinity}, general parameters of the core, considering the reactor operating under stationary conditions of initial testing and operation. Other stationary operation conditions have been simulated and, in all tested cases, there was a close agreement between values calculated computationally through this model and data presented on the FSAR, which were obtained by other codes. This model is expected to become a valuable tool for many future applications. (author)
Development of computer code models for analysis of subassembly voiding in the LMFBR
Hinkle, W [ed.
1979-12-01
The research program discussed in this report was started in FY1979 under the combined sponsorship of the US Department of Energy (DOE), General Electric (GE) and Hanford Engineering Development Laboratory (HEDL). The objective of the program is to develop multi-dimensional computer codes which can be used for the analysis of subassembly voiding incoherence under postulated accident conditions in the LMFBR. Two codes are being developed in parallel. The first will use a two fluid (6 equation) model which is more difficult to develop but has the potential for providing a code with the utmost in flexibility and physical consistency for use in the long term. The other will use a mixture (< 6 equation) model which is less general but may be more amenable to interpretation and use of experimental data and therefore, easier to develop for use in the near term. To assure that the models developed are not design dependent, geometries and transient conditions typical of both foreign and US designs are being considered.
Sodium combustion computer code ASSCOPS version 2.0 user`s manual
Ishikawa, Hiroyasu; Futagami, Satoshi; Ohno, Shuji; Seino, Hiroshi; Miyake, Osamu [Power Reactor and Nuclear Fuel Development Corp., Oarai, Ibaraki (Japan). Oarai Engineering Center
1997-12-01
ASSCOPS (Analysis of Simultaneous Sodium Combustion in Pool and Spray) has been developed for analyses of thermal consequences of sodium leak and fire accidents in LMFBRs. This report presents a description of the computational models, input, and output as the user`s manual of ASSCOPS version 2.0. ASSCOPS is an integrated code based on the sodium pool fire code SOFIRE II developed by the Atomics International Division of Rockwell International, and the sodium spray fire code SPRAY developed by the Hanford Engineering Development Laboratory in the U.S. The experimental studies conducted at PNC have been reflected in the ASSCOPS improvement. The users of ASSCOPS need to specify the sodium leak conditions (leak flow rate and temperature, etc.), the cell geometries (volume and structure surface area and thickness, etc.), and the atmospheric initial conditions, such as gas temperature, pressure, and gas composition. ASSCOPS calculates the time histories of atmospheric pressure and temperature changes along with those of the structural temperatures. (author)
Hierarchical surface code for network quantum computing with modules of arbitrary size
Li, Ying; Benjamin, Simon C.
2016-10-01
The network paradigm for quantum computing involves interconnecting many modules to form a scalable machine. Typically it is assumed that the links between modules are prone to noise while operations within modules have a significantly higher fidelity. To optimize fault tolerance in such architectures we introduce a hierarchical generalization of the surface code: a small "patch" of the code exists within each module and constitutes a single effective qubit of the logic-level surface code. Errors primarily occur in a two-dimensional subspace, i.e., patch perimeters extruded over time, and the resulting noise threshold for intermodule links can exceed ˜10 % even in the absence of purification. Increasing the number of qubits within each module decreases the number of qubits necessary for encoding a logical qubit. But this advantage is relatively modest, and broadly speaking, a "fine-grained" network of small modules containing only about eight qubits is competitive in total qubit count versus a "course" network with modules containing many hundreds of qubits.
M. Garbey; C. Picard
2008-01-01
The goal of this paper is to present a versatile framework for solution verification of PDE's.We first generalize the Richardson Extrapolation technique to an optimized extrapolation solution procedure that constructs the best consistent solution from a set of two or three coarse grid solution in the discrete norm of choice. This technique generalizes the Least Square Extrapolation method introduced by one of the author and W. Shyy. We second establish the conditioning number of the problem in a reduced space that approximates the main feature of the numerical solution thanks to a sensitivity analysis. Overall our method produces an a posteriori error estimation in this reduced space of approximation. The key feature of our method is that our construction does not require an internal knowledge of the software neither the source code that produces the solution to be verified. It can be applied in principle as a postprocessing procedure to off the shelf commercial code. We demonstrate the robustness of our method with two steady problems that are separately an incompressible back step flow test case and a heat transfer problem for a battery. Our error estimate might be ultimately verified with a near by manufactured solution. While our procedure is systematic and requires numerous computation of residuals, one can take advantage of distributed computing to get quickly the error estimate.
JUN YEOB LEE
2014-10-01
Full Text Available During the development process of a thermal-hydraulic system code, a non-regression test (NRT must be performed repeatedly in order to prevent software regression. The NRT process, however, is time-consuming and labor-intensive. Thus, automation of this process is an ideal solution. In this study, we have developed a program to support an efficient NRT for the SPACE code and demonstrated its usability. This results in a high degree of efficiency for code development. The program was developed using the Visual Basic for Applications and designed so that it can be easily customized for the NRT of other computer codes.
Computer code to predict the heat of explosion of high energy materials
Muthurajan, H. [Armament Research and Development Establishment, Pashan, Pune 411021 (India)], E-mail: muthurajan_h@rediffmail.com; Sivabalan, R.; Pon Saravanan, N.; Talawar, M.B. [High Energy Materials Research Laboratory, Sutarwadi, Pune 411 021 (India)
2009-01-30
The computational approach to the thermochemical changes involved in the process of explosion of a high energy materials (HEMs) vis-a-vis its molecular structure aids a HEMs chemist/engineers to predict the important thermodynamic parameters such as heat of explosion of the HEMs. Such a computer-aided design will be useful in predicting the performance of a given HEM as well as in conceiving futuristic high energy molecules that have significant potential in the field of explosives and propellants. The software code viz., LOTUSES developed by authors predicts various characteristics of HEMs such as explosion products including balanced explosion reactions, density of HEMs, velocity of detonation, CJ pressure, etc. The new computational approach described in this paper allows the prediction of heat of explosion ({delta}H{sub e}) without any experimental data for different HEMs, which are comparable with experimental results reported in literature. The new algorithm which does not require any complex input parameter is incorporated in LOTUSES (version 1.5) and the results are presented in this paper. The linear regression analysis of all data point yields the correlation coefficient R{sup 2} = 0.9721 with a linear equation y = 0.9262x + 101.45. The correlation coefficient value 0.9721 reveals that the computed values are in good agreement with experimental values and useful for rapid hazard assessment of energetic materials.
Computer code to predict the heat of explosion of high energy materials.
Muthurajan, H; Sivabalan, R; Pon Saravanan, N; Talawar, M B
2009-01-30
The computational approach to the thermochemical changes involved in the process of explosion of a high energy materials (HEMs) vis-à-vis its molecular structure aids a HEMs chemist/engineers to predict the important thermodynamic parameters such as heat of explosion of the HEMs. Such a computer-aided design will be useful in predicting the performance of a given HEM as well as in conceiving futuristic high energy molecules that have significant potential in the field of explosives and propellants. The software code viz., LOTUSES developed by authors predicts various characteristics of HEMs such as explosion products including balanced explosion reactions, density of HEMs, velocity of detonation, CJ pressure, etc. The new computational approach described in this paper allows the prediction of heat of explosion (DeltaH(e)) without any experimental data for different HEMs, which are comparable with experimental results reported in literature. The new algorithm which does not require any complex input parameter is incorporated in LOTUSES (version 1.5) and the results are presented in this paper. The linear regression analysis of all data point yields the correlation coefficient R(2)=0.9721 with a linear equation y=0.9262x+101.45. The correlation coefficient value 0.9721 reveals that the computed values are in good agreement with experimental values and useful for rapid hazard assessment of energetic materials.
Implementation of discrete transfer radiation method into swift computational fluid dynamics code
Baburić Mario
2004-01-01
Full Text Available The Computational Fluid Dynamics (CFD has developed into a powerful tool widely used in science, technology and industrial design applications, when ever fluid flow, heat transfer, combustion, or other complicated physical processes, are involved. During decades of development of CFD codes scientists were writing their own codes, that had to include not only the model of processes that were of interest, but also a whole spectrum of necessary CFD procedures, numerical techniques, pre-processing and post-processing. That has arrested much of the scientist effort in work that has been copied many times over, and was not actually producing the added value. The arrival of commercial CFD codes brought relief to many engineers that could now use the user-function approach for mod el ling purposes, en trusting the application to do the rest of the work. This pa per shows the implementation of Discrete Transfer Radiation Method into AVL’s commercial CFD code SWIFT with the help of user defined functions. Few standard verification test cases were per formed first, and in order to check the implementation of the radiation method it self, where the comparisons with available analytic solution could be performed. After wards, the validation was done by simulating the combustion in the experimental furnace at IJmuiden (Netherlands, for which the experimental measurements were available. The importance of radiation prediction in such real-size furnaces is proved again to be substantial, where radiation itself takes the major fraction of over all heat transfer. The oil-combustion model used in simulations was the semi-empirical one that has been developed at the Power Engineering Department, and which is suit able for a wide range of typical oil flames.
NONE
1997-03-01
This Manual represents Revision 5 of the user documentation for the modular code system referred to as SCALE. The history of the SCALE code system dates back to 1969 when the current Computational Physics and Engineering Division at Oak Ridge National Laboratory (ORNL) began providing the transportation package certification staff at the U.S. Atomic Energy Commission with computational support in the use of the new KENO code for performing criticality safety assessments with the statistical Monte Carlo method. From 1969 to 1976 the certification staff relied on the ORNL staff to assist them in the correct use of codes and data for criticality, shielding, and heat transfer analyses of transportation packages. However, the certification staff learned that, with only occasional use of the codes, it was difficult to become proficient in performing the calculations often needed for an independent safety review. Thus, shortly after the move of the certification staff to the U.S. Nuclear Regulatory Commission (NRC), the NRC staff proposed the development of an easy-to-use analysis system that provided the technical capabilities of the individual modules with which they were familiar. With this proposal, the concept of the Standardized Computer Analyses for Licensing Evaluation (SCALE) code system was born. This volume consists of the section of the manual dealing with three of the functional modules in the code. Those are the Morse-SGC for the SCALE system, Heating 7.2, and KENO V.a. The manual describes the latest released versions of the codes.
The Statistical Mechanics of Ideal MHD Turbulence
Shebalin, John V.
2003-01-01
Turbulence is a universal, nonlinear phenomenon found in all energetic fluid and plasma motion. In particular. understanding magneto hydrodynamic (MHD) turbulence and incorporating its effects in the computation and prediction of the flow of ionized gases in space, for example, are great challenges that must be met if such computations and predictions are to be meaningful. Although a general solution to the "problem of turbulence" does not exist in closed form, numerical integrations allow us to explore the phase space of solutions for both ideal and dissipative flows. For homogeneous, incompressible turbulence, Fourier methods are appropriate, and phase space is defined by the Fourier coefficients of the physical fields. In the case of ideal MHD flows, a fairly robust statistical mechanics has been developed, in which the symmetry and ergodic properties of phase space is understood. A discussion of these properties will illuminate our principal discovery: Coherent structure and randomness co-exist in ideal MHD turbulence. For dissipative flows, as opposed to ideal flows, progress beyond the dimensional analysis of Kolmogorov has been difficult. Here, some possible future directions that draw on the ideal results will also be discussed. Our conclusion will be that while ideal turbulence is now well understood, real turbulence still presents great challenges.
Automated Development of Accurate Algorithms and Efficient Codes for Computational Aeroacoustics
Goodrich, John W.; Dyson, Rodger W.
1999-01-01
The simulation of sound generation and propagation in three space dimensions with realistic aircraft components is a very large time dependent computation with fine details. Simulations in open domains with embedded objects require accurate and robust algorithms for propagation, for artificial inflow and outflow boundaries, and for the definition of geometrically complex objects. The development, implementation, and validation of methods for solving these demanding problems is being done to support the NASA pillar goals for reducing aircraft noise levels. Our goal is to provide algorithms which are sufficiently accurate and efficient to produce usable results rapidly enough to allow design engineers to study the effects on sound levels of design changes in propulsion systems, and in the integration of propulsion systems with airframes. There is a lack of design tools for these purposes at this time. Our technical approach to this problem combines the development of new, algorithms with the use of Mathematica and Unix utilities to automate the algorithm development, code implementation, and validation. We use explicit methods to ensure effective implementation by domain decomposition for SPMD parallel computing. There are several orders of magnitude difference in the computational efficiencies of the algorithms which we have considered. We currently have new artificial inflow and outflow boundary conditions that are stable, accurate, and unobtrusive, with implementations that match the accuracy and efficiency of the propagation methods. The artificial numerical boundary treatments have been proven to have solutions which converge to the full open domain problems, so that the error from the boundary treatments can be driven as low as is required. The purpose of this paper is to briefly present a method for developing highly accurate algorithms for computational aeroacoustics, the use of computer automation in this process, and a brief survey of the algorithms that
Louis, J.F.
1977-12-01
Research and development in open-cycle coal-fired MHD power generation at Massachusetts Institute of Technology (MIT) is summarized. Progress is reported on the following tasks: (1) Establish chemical, thermal, and electrical data to guide materials selection, develop improved detail designs, and support performance analyses of MHD electrode modules and insulator materials; (2) parametrically investigate selected electrode properties of critical design importance in chemical, thermal, and electrical environments simulating a coal-fired MHD generator; (3) develop combustion data pertinent to the design of MHD combustors; this work is intended to determine the combustion characteristics of selected coal feedstock in terms of devolatilization kinetics, char characteristics, and combustion gas chemistry; (4) establish techniques for the analysis and understanding of critical MHD phenomena which have an important bearing on MHD generator performance; such phenomena include inter-electrode breakdown, time-dependent behavior, effective plasma properties and plasma inhomogeneities; (5) establish the operating characteristics of an MHD disk generator; (6) continue work on computer techniques for modeling and for design and cost analysis of MHD components and the overall system; (7) integrate the engineering data and design criteria, as applicable, which are developed in the above-listed tasks into a model of the MHD channel; (8) participate in technical support of the DOE MHD Project Office.
Proskuryakov, K.N.; Bogomazov, D.N.; Poliakov, N. [Moscow Power Engineering Institute (Technical University), Moscow (Russian Federation)
2007-07-01
The new special module to neutron-physic and thermal-hydraulic computer codes for coolant acoustical characteristics calculation is worked out. The Russian computer code Rainbow has been selected for joint use with a developed module. This code system provides the possibility of EFOCP (Eigen Frequencies of Oscillations of the Coolant Pressure) calculations in any coolant acoustical elements of primary circuits of NPP. EFOCP values have been calculated for transient and for stationary operating. The calculated results for nominal operating were compared with results of measured EFOCP. For example, this comparison was provided for the system: 'pressurizer + surge line' of a WWER-1000 reactor. The calculated result 0.58 Hz practically coincides with the result of measurement (0.6 Hz). The EFOCP variations in transients are also shown. The presented results are intended to be useful for NPP vibration-acoustical certification. There are no serious difficulties for using this module with other computer codes.
G. García Segura
2000-01-01
Full Text Available Se presenta un escenario auto consistente para explicar la morfolog a de las nebulosas planetarias. El escenario es consistente con la distribuci on Gal actica de los diferentes tipos morfol ogicos. Este trabajo resuelve, por medio de efectos MHD, algunas de las caracter sticas controversiales que aparecen en las nebulosas planetarias. Estas caracter sticas incluyen la presencia de ujos axisim etricos y colimados, con una cinem atica que aumenta linealmente con la distancia y la existencia de morfolog as asim etricas tales como las de las nebulosas con simetr a de punto.
Retallick, F.D.
1978-04-01
This document establishes criteria to be utilized for the design of a pilot-scale (150 to 300 MW thermal) open cycle, coal-fired MHD/steam plant. Criteria for this Engineering Test Facility (ETF) are presented relative to plant siting, plant engineering and operations, MHD-ETF testing, costing and scheduling.
MHD turbulence and distributed chaos
Bershadskii, A
2016-01-01
It is shown, using results of recent direct numerical simulations, that spectral properties of distributed chaos in MHD turbulence with zero mean magnetic field are similar to those of hydrodynamic turbulence. An exception is MHD spontaneous breaking of space translational symmetry, when the stretched exponential spectrum $\\exp(-k/k_{\\beta})^{\\beta}$ has $\\beta=4/7$.
Pon-Barry, Heather; Packard, Becky Wai-Ling; St. John, Audrey
2017-01-01
A dilemma within computer science departments is developing sustainable ways to expand capacity within introductory computer science courses while remaining committed to inclusive practices. Training near-peer mentors for peer code review is one solution. This paper describes the preparation of near-peer mentors for their role, with a focus on…
Evaluation of feedback in conductive MHD devices
Grinberg, G.K.
1977-01-01
A method is recommended for computing feedback and the self-energizing threshold of conducting MHD devices. Circuits of equivalent magnetizing currents are used for this purpose in addition to equivalent electrical circuits. This kind of an approach makes it possible to reflect the influence of R/sub m/ on the operation of the device. Dimensionless functions were found which determine the critical value of the Reynolds magnetic number. The computations demonstrated that the redistribution of the magnetic field in the machine's operating zone under the influence of an induced field must be considered.
Louis, J.F.
1977-08-01
Research progress in open-cycle coal-fired MHD power generation at Massachusetts Institute of Technology (MIT) is reported. The scope and objectives of the MIT program are to: (1) Establish chemical, thermal, and electrical data to guide materials selection, develop improved detail designs, and support performance analyses of MHD electrode modules and insulator materials; (2) establish basic mechanical properties to guide detail design and fabrication of high field strength superconducting magnets for MHD applications; (3) parametrically investigate selected electrode properties of critical design importance in chemical, thermal, and electrical environments simulating a coal-fired MHD generator; (4) develop combustion data pertinent to the design of MHD combustors; (5) establish techniques for the analysis and understanding of critical MHD phenomena which have an important bearing on MHD generator performance; such phenomena include inter-electrode breakdown, time-dependent behavior, effective plasma properties and plasma inhomogeneities; (6) establish the operating characteristics of an MHD disk generator; (7) continue work on computer techniques for modeling and for design and cost analysis of MHD components and the overall system; (8) integrate the engineering data and design criteria, as applicable, which are developed in the above-listed tasks into a model of the MHD channel; (9) Participate in technical support of the ERDA MHD Project Office; (10) participate in the US/USSR Cooperative Program in MHD Power Generation. (11) During the summer of 1976, a short-term task in U-25 electrode screening was temporarily added to the scope of the contract. This effort involved screening tests, in the MIT MHD simulation facility of electrode modules and configurations intended for tests in the Soviet U-25 generator.
Wood, Jerry R.; Schmidt, James F.; Steinke, Ronald J.; Chima, Rodrick V.; Kunik, William G.
1987-01-01
Increased emphasis on sustained supersonic or hypersonic cruise has revived interest in the supersonic throughflow fan as a possible component in advanced propulsion systems. Use of a fan that can operate with a supersonic inlet axial Mach number is attractive from the standpoint of reducing the inlet losses incurred in diffusing the flow from a supersonic flight Mach number to a subsonic one at the fan face. The design of the experiment using advanced computational codes to calculate the components required is described. The rotor was designed using existing turbomachinery design and analysis codes modified to handle fully supersonic axial flow through the rotor. A two-dimensional axisymmetric throughflow design code plus a blade element code were used to generate fan rotor velocity diagrams and blade shapes. A quasi-three-dimensional, thin shear layer Navier-Stokes code was used to assess the performance of the fan rotor blade shapes. The final design was stacked and checked for three-dimensional effects using a three-dimensional Euler code interactively coupled with a two-dimensional boundary layer code. The nozzle design in the expansion region was analyzed with a three-dimensional parabolized viscous code which corroborated the results from the Euler code. A translating supersonic diffuser was designed using these same codes.
Advances in Simulation of Wave Interactions with Extended MHD Phenomena
Batchelor, Donald B [ORNL; D' Azevedo, Eduardo [ORNL; Bateman, Glenn [ORNL; Bernholdt, David E [ORNL; Bonoli, P. [Massachusetts Institute of Technology (MIT); Bramley, Randall B [ORNL; Breslau, Joshua [ORNL; Elwasif, Wael R [ORNL; Foley, S. [Indiana University; Jaeger, Erwin Frederick [ORNL; Jardin, S. C. [Princeton Plasma Physics Laboratory (PPPL); Klasky, Scott A [ORNL; Kruger, Scott E [ORNL; Ku, Long-Poe [ORNL; McCune, Douglas [Princeton Plasma Physics Laboratory (PPPL); Ramos, J. [Massachusetts Institute of Technology (MIT); Schissel, David P [ORNL; Schnack, Dalton D [ORNL
2009-01-01
The Integrated Plasma Simulator (IPS) provides a framework within which some of the most advanced, massively-parallel fusion modeling codes can be interoperated to provide a detailed picture of the multi-physics processes involved in fusion experiments. The presentation will cover four topics: (1) recent improvements to the IPS, (2) application of the IPS for very high resolution simulations of ITER scenarios, (3) studies of resistive and ideal MHD stability in tokamak discharges using IPS facilities, and (4) the application of RF power in the electron cyclotron range of frequencies to control slowly growing MHD modes in tokamaks and initial evaluations of optimized location for RF power deposition.
Advances in Simulation of Wave Interaction with Extended MHD Phenomena
Batchelor, Donald B [ORNL; Abla, Gheni [ORNL; D' Azevedo, Ed F [ORNL; Bateman, Glenn [Lehigh University, Bethlehem, PA; Bernholdt, David E [ORNL; Berry, Lee A [ORNL; Bonoli, P. [Massachusetts Institute of Technology (MIT); Bramley, R [Indiana University; Breslau, Joshua [ORNL; Chance, M. [Princeton Plasma Physics Laboratory (PPPL); Chen, J. [Princeton Plasma Physics Laboratory (PPPL); Choi, M. [General Atomics; Elwasif, Wael R [ORNL; Foley, S. [Indiana University; Fu, GuoYong [Princeton Plasma Physics Laboratory (PPPL); Harvey, R. W. [CompX, Del Mar, CA; Jaeger, Erwin Frederick [ORNL; Jardin, S. C. [Princeton Plasma Physics Laboratory (PPPL); Jenkins, T [University of Wisconsin; Keyes, David E [Columbia University; Klasky, Scott A [ORNL; Kruger, Scott [Tech-X Corporation; Ku, Long-Poe [Princeton Plasma Physics Laboratory (PPPL); Lynch, Vickie E [ORNL; McCune, Douglas [Princeton Plasma Physics Laboratory (PPPL); Ramos, J. [Massachusetts Institute of Technology (MIT); Schissel, D. [General Atomics; Schnack, [University of Wisconsin; Wright, J. [Massachusetts Institute of Technology (MIT)
2009-01-01
The Integrated Plasma Simulator (IPS) provides a framework within which some of the most advanced, massively-parallel fusion modeling codes can be interoperated to provide a detailed picture of the multi-physics processes involved in fusion experiments. The presentation will cover four topics: 1) recent improvements to the IPS, 2) application of the IPS for very high resolution simulations of ITER scenarios, 3) studies of resistive and ideal MHD stability in tokamk discharges using IPS facilities, and 4) the application of RF power in the electron cyclotron range of frequencies to control slowly growing MHD modes in tokamaks and initial evaluations of optimized location for RF power deposition.
Advances in simulation of wave interactions with extended MHD phenomena
Batchelor, D; D' Azevedo, E; Bernholdt, D E; Berry, L; Elwasif, W; Jaeger, E [Oak Ridge National Laboratory (United States); Abla, G; Choi, M [General Atomics (United States); Bateman, G [Lehigh University (United States); Bonoli, P [Plasma Science and Fusion Center, Massachusetts Institute of Technology (United States); Bramley, R; Foley, S [Indiana University (United States); Breslau, J; Chance, M; Chen, J; Fu, G; Jardin, S [Princeton Plasma Physics Laboratory (United States); Harvey, R [CompX International (United States); Jenkins, T [University of Wisconsin (United States); Keyes, D, E-mail: batchelordb@ornl.go [Columbia University (United States)
2009-07-01
The Integrated Plasma Simulator (IPS) provides a framework within which some of the most advanced, massively-parallel fusion modeling codes can be interoperated to provide a detailed picture of the multi-physics processes involved in fusion experiments. The presentation will cover four topics: 1) recent improvements to the IPS, 2) application of the IPS for very high resolution simulations of ITER scenarios, 3) studies of resistive and ideal MHD stability in tokamk discharges using IPS facilities, and 4) the application of RF power in the electron cyclotron range of frequencies to control slowly growing MHD modes in tokamaks and initial evaluations of optimized location for RF power deposition.
M. Schüssler
Full Text Available Two aspects of solar MHD are discussed in relation to the work of the MHD simulation group at KIS. Photospheric magneto-convection, the nonlinear interaction of magnetic field and convection in a strongly stratified, radiating fluid, is a key process of general astrophysical relevance. Comprehensive numerical simulations including radiative transfer have significantly improved our understanding of the processes and have become an important tool for the interpretation of observational data. Examples of field intensification in the solar photosphere ('convective collapse' are shown. The second line of research is concerned with the dynamics of flux tubes in the convection zone, which has far-reaching implications for our understanding of the solar dynamo. Simulations indicate that the field strength in the region where the flux is stored before erupting to form sunspot groups is of the order of 10^{5} G, an order of magnitude larger than previous estimates based on equipartition with the kinetic energy of convective flows.
Key words. Solar physics · astrophysics and astronomy (photosphere and chromosphere; stellar interiors and dynamo theory; numerical simulation studies.
Smolentsev, S., E-mail: sergey@fusion.ucla.edu [University of California, Los Angeles (United States); Courtessole, C.; Abdou, M.; Sharafat, S. [University of California, Los Angeles (United States); Sahu, S. [Institute of Plasma Research (India); Sketchley, T. [University of California, Los Angeles (United States)
2016-10-15
Highlights: • Numerical studies were performed as a pre-experimental analysis to the experiment on MHD PbLi flows in a rectangular duct with a flow channel insert (FCI). • Dynamic testing of foam-based SiC foam-based CVD coated FCI has been performed using MaPLE facility at UCLA. • Two physical models were proposed to explain the experimental results and 3D and 2D computations performed using COMSOL, HIMAG and UCLA codes. • The obtained results suggest that more work on FCI development, fabrication and testing has to be done to assure good hermetic properties before the implementation in a fusion device. - Abstract: A flow channel insert (FCI) is the key element of the DCLL blanket concept. The FCI serves as electrical and thermal insulator to reduce the MHD pressure drop and to decouple the temperature-limited ferritic structure from the flowing hot lead-lithium (PbLi) alloy. The main focus of the paper is on numerical computations to simulate MHD flows in the first experiments on PbLi flows in a stainless steel rectangular duct with a foam-based silicon carbide (SiC) FCI. A single uninterrupted long-term (∼6500 h) test has recently been performed on a CVD coated FCI sample in the flowing PbLi in a magnetic field up to 1.5 T at the PbLi temperature of 300 °C using the MaPLE loop at UCLA. An unexpectedly high MHD pressure drop measured in this experiment suggests that a PbLi ingress into the FCI occurred in the course of the experiment, resulting in degradation of electroinsulating FCI properties. The ingress through the protective CVD layer was further confirmed by the post-experimental microscopic analysis of the FCI. The numerical modeling included 2D and 3D computations using HIMAG, COMSOL and a UCLA research code to address important flow features associated with the FCI finite length, fringing magnetic field, rounded FCI corners and also to predict changes in the MHD pressure drop in the unwanted event of a PbLi ingress. Two physical
Berna, G. A; Bohn, M. P.; Rausch, W. N.; Williford, R. E.; Lanning, D. D.
1981-01-01
FRAPCON-2 is a FORTRAN IV computer code that calculates the steady state response of light Mater reactor fuel rods during long-term burnup. The code calculates the temperature, pressure, deformation, and tai lure histories of a fuel rod as functions of time-dependent fuel rod power and coolant boundary conditions. The phenomena modeled by the code include (a) heat conduction through the fuel and cladding, (b) cladding elastic and plastic deformation, (c) fuel-cladding mechanical interaction, (d) fission gas release, (e} fuel rod internal gas pressure, (f) heat transfer between fuel and cladding, (g) cladding oxidation, and (h) heat transfer from cladding to coolant. The code contains necessary material properties, water properties, and heat transfer correlations. FRAPCON-2 is programmed for use on the CDC Cyber 175 and 176 computers. The FRAPCON-2 code Is designed to generate initial conditions for transient fuel rod analysis by either the FRAP-T6 computer code or the thermal-hydraulic code, RELAP4/MOD7 Version 2.
NONE
1997-03-01
This Manual represents Revision 5 of the user documentation for the modular code system referred to as SCALE. The history of the SCALE code system dates back to 1969 when the current Computational Physics and Engineering Division at Oak Ridge National Laboratory (ORNL) began providing the transportation package certification staff at the U. S. Atomic Energy Commission with computational support in the use of the new KENO code for performing criticality safety assessments with the statistical Monte Carlo method. From 1969 to 1976 the certification staff relied on the ORNL staff to assist them in the correct use of codes and data for criticality, shielding, and heat transfer analyses of transportation packages. However, the certification staff learned that, with only occasional use of the codes, it was difficult to become proficient in performing the calculations often needed for an independent safety review. Thus, shortly after the move of the certification staff to the U.S. Nuclear Regulatory Commission (NRC), the NRC staff proposed the development of an easy-to-use analysis system that provided the technical capabilities of the individual modules with which they were familiar. With this proposal, the concept of the Standardized Computer Analyses for Licensing Evaluation (SCALE) code system was born. This volume is part of the manual related to the control modules for the newest updated version of this computational package.
MULTI2D - a computer code for two-dimensional radiation hydrodynamics
Ramis, R.; Meyer-ter-Vehn, J.; Ramírez, J.
2009-06-01
Simulation of radiation hydrodynamics in two spatial dimensions is developed, having in mind, in particular, target design for indirectly driven inertial confinement energy (IFE) and the interpretation of related experiments. Intense radiation pulses by laser or particle beams heat high-Z target configurations of different geometries and lead to a regime which is optically thick in some regions and optically thin in others. A diffusion description is inadequate in this situation. A new numerical code has been developed which describes hydrodynamics in two spatial dimensions (cylindrical R-Z geometry) and radiation transport along rays in three dimensions with the 4 π solid angle discretized in direction. Matter moves on a non-structured mesh composed of trilateral and quadrilateral elements. Radiation flux of a given direction enters on two (one) sides of a triangle and leaves on the opposite side(s) in proportion to the viewing angles depending on the geometry. This scheme allows to propagate sharply edged beams without ray tracing, though at the price of some lateral diffusion. The algorithm treats correctly both the optically thin and optically thick regimes. A symmetric semi-implicit (SSI) method is used to guarantee numerical stability. Program summaryProgram title: MULTI2D Catalogue identifier: AECV_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AECV_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 151 098 No. of bytes in distributed program, including test data, etc.: 889 622 Distribution format: tar.gz Programming language: C Computer: PC (32 bits architecture) Operating system: Linux/Unix RAM: 2 Mbytes Word size: 32 bits Classification: 19.7 External routines: X-window standard library (libX11.so) and corresponding heading files (X11/*.h) are
Interface design of VSOP'94 computer code for safety analysis
Natsir, Khairina, E-mail: yenny@batan.go.id; Andiwijayakusuma, D.; Wahanani, Nursinta Adi [Center for Development of Nuclear Informatics - National Nuclear Energy Agency, PUSPIPTEK, Serpong, Tangerang, Banten (Indonesia); Yazid, Putranto Ilham [Center for Nuclear Technology, Material and Radiometry- National Nuclear Energy Agency, Jl. Tamansari No.71, Bandung 40132 (Indonesia)
2014-09-30
Today, most software applications, also in the nuclear field, come with a graphical user interface. VSOP'94 (Very Superior Old Program), was designed to simplify the process of performing reactor simulation. VSOP is a integrated code system to simulate the life history of a nuclear reactor that is devoted in education and research. One advantage of VSOP program is its ability to calculate the neutron spectrum estimation, fuel cycle, 2-D diffusion, resonance integral, estimation of reactors fuel costs, and integrated thermal hydraulics. VSOP also can be used to comparative studies and simulation of reactor safety. However, existing VSOP is a conventional program, which was developed using Fortran 65 and have several problems in using it, for example, it is only operated on Dec Alpha mainframe platforms and provide text-based output, difficult to use, especially in data preparation and interpretation of results. We develop a GUI-VSOP, which is an interface program to facilitate the preparation of data, run the VSOP code and read the results in a more user friendly way and useable on the Personal 'Computer (PC). Modifications include the development of interfaces on preprocessing, processing and postprocessing. GUI-based interface for preprocessing aims to provide a convenience way in preparing data. Processing interface is intended to provide convenience in configuring input files and libraries and do compiling VSOP code. Postprocessing interface designed to visualized the VSOP output in table and graphic forms. GUI-VSOP expected to be useful to simplify and speed up the process and analysis of safety aspects.
A computational model of cellular mechanisms of temporal coding in the medial geniculate body (MGB.
Cal F Rabang
Full Text Available Acoustic stimuli are often represented in the early auditory pathway as patterns of neural activity synchronized to time-varying features. This phase-locking predominates until the level of the medial geniculate body (MGB, where previous studies have identified two main, largely segregated response types: Stimulus-synchronized responses faithfully preserve the temporal coding from its afferent inputs, and Non-synchronized responses, which are not phase locked to the inputs, represent changes in temporal modulation by a rate code. The cellular mechanisms underlying this transformation from phase-locked to rate code are not well understood. We use a computational model of a MGB thalamocortical neuron to test the hypothesis that these response classes arise from inferior colliculus (IC excitatory afferents with divergent properties similar to those observed in brain slice studies. Large-conductance inputs exhibiting synaptic depression preserved input synchrony as short as 12.5 ms interclick intervals, while maintaining low firing rates and low-pass filtering responses. By contrast, small-conductance inputs with Mixed plasticity (depression of AMPA-receptor component and facilitation of NMDA-receptor component desynchronized afferent inputs, generated a click-rate dependent increase in firing rate, and high-pass filtered the inputs. Synaptic inputs with facilitation often permitted band-pass synchrony along with band-pass rate tuning. These responses could be tuned by changes in membrane potential, strength of the NMDA component, and characteristics of synaptic plasticity. These results demonstrate how the same synchronized input spike trains from the inferior colliculus can be transformed into different representations of temporal modulation by divergent synaptic properties.
Reduced gravity boiling and condensing experiments simulated with the COBRA/TRAC computer code
Cuta, Judith M.; Krotiuk, William
1988-01-01
A series of reduced-gravity two-phase flow experiments has been conducted with a boiler/condenser apparatus in the NASA KC-135 aircraft in order to obtain basic thermal-hydraulic data applicable to analytical design tools. Several test points from the KC-135 tests were selected for simulation by means of the COBRA/TRAC two-fluid, three-field thermal-hydraulic computer code; the points were chosen for a 25-90 percent void-fraction range. The possible causes for the lack of agreement noted between simulations and experiments are explored, with attention to the physical characteristics of two-phase flow in one-G and near-zero-G conditions.
Discrete logarithm computations over finite fields using Reed-Solomon codes
Augot, Daniel
2012-01-01
Cheng and Wan have related the decoding of Reed-Solomon codes to the computation of discrete logarithms over finite fields, with the aim of proving the hardness of their decoding. In this work, we experiment with solving the discrete logarithm over GF(q^h) using Reed-Solomon decoding. For fixed h and q going to infinity, we introduce an algorithm (RSDL) needing O (h! q^2) operations over GF(q), operating on a q x q matrix with (h+2) q non-zero coefficients. We give faster variants including an incremental version and another one that uses auxiliary finite fields that need not be subfields of GF(q^h); this variant is very practical for moderate values of q and h. We include some numerical results of our first implementations.
ACUTRI a computer code for assessing doses to the general public due to acute tritium releases
Yokoyama, S; Noguchi, H; Ryufuku, S; Sasaki, T
2002-01-01
Tritium, which is used as a fuel of a D-T burning fusion reactor, is the most important radionuclide for the safety assessment of a nuclear fusion experimental reactor such as ITER. Thus, a computer code, ACUTRI, which calculates the radiological impact of tritium released accidentally to the atmosphere, has been developed, aiming to be of use in a discussion of licensing of a fusion experimental reactor and an environmental safety evaluation method in Japan. ACUTRI calculates an individual tritium dose based on transfer models specific to tritium in the environment and ICRP dose models. In this calculation it is also possible to analyze statistically on meteorology in the same way as a conventional dose assessment method according to the meteorological guide of the Nuclear Safety Commission of Japan. A Gaussian plume model is used for calculating the atmospheric dispersion of tritium gas (HT) and/or tritiated water (HTO). The environmental pathway model in ACUTRI considers the following internal exposures: i...
Park, Chang Kyu; Jae, Moo Sung; Jo, Young Gyun; Park, Rae Jun; Kim, Jae Hwan; Ha, Jae Ju; Kang, Dae Il; Choi, Sun Young; Kim, Si Hwan [Korea Atomic Energy Res. Inst., Taejon (Korea, Republic of)
1994-07-01
We have surveyed new technologies and research results for the accident management of nuclear power plants. And, based on the concept of using the existing plant capabilities for accident management, both in-vessel and ex-vessel strategies were identified and analyzed. When assessing accident management strategies, their effectiveness, adverse effects, and their feasibility must be considered. We have developed a framework for assessing the strategies with these factors in mind. We have applied the developed framework to assessing the strategies, including the likelihood that the operator correctly diagnoses the situation and successfully implements the strategies. Finally, the cavity flooding strategy was assessed by applying it to the station blackout sequence, which have been identified as one of the major contributors to risk at the reference plant. The thermohydraulic analyses with sensitivity calculations have been performed using MAAP 4 computer code. (Author).
Bousquet, Nicolas
2010-01-01
This article deals with the estimation of a probability p of an undesirable event. Its occurence is formalized by the exceedance of a threshold reliability value by the unidimensional output of a time-consuming computer code G with multivariate probabilistic input X. When G is assumed monotonous with respect to X, the Monotonous Reliability Method was proposed by de Rocquigny (2009) in an engineering context to provide sequentially narrowing 100%-confidence bounds and a crude estimate of p, via deterministic or stochastic designs of experiments. The present article consists in a formalization and technical deepening of this idea, as a large basis for future theoretical and applied studies. Three kinds of results are especially emphasized. First, the bounds themselves remain too crude and conservative estimators of p for a dimension of X upper than 2. Second, a maximum-likelihood estimator of p can be easily built, presenting a high variance reduction with respect to a standard Monte Carlo case, but suffering ...
Finite Element Simulation Code for Computing Thermal Radiation from a Plasma
Nguyen, C. N.; Rappaport, H. L.
2004-11-01
A finite element code, ``THERMRAD,'' for computing thermal radiation from a plasma is under development. Radiation from plasma test particles is found in cylindrical geometry. Although the plasma equilibrium is assumed axisymmetric individual test particle excitation produces a non-axisymmetric electromagnetic response. Specially designed Whitney class basis functions are to be used to allow the solution to be solved on a two-dimensional grid. The basis functions enforce both a vanishing of the divergence of the electric field within grid elements where the complex index of refraction is assumed constant and continuity of tangential electric field across grid elements while allowing the normal component of the electric field to be discontinuous. An appropriate variational principle which incorporates the Sommerfeld radiation condition on the simulation boundary, as well as its discretization by the Rayleigh-Ritz technique is given. 1. ``Finte Element Method for Electromagnetics Problems,'' Volakis et al., Wiley, 1998.
Bade, W. L.; Yos, J. M.
1975-01-01
A computer program for calculating quasi-one-dimensional gas flow in axisymmetric and two-dimensional nozzles and rectangular channels is presented. Flow is assumed to start from a state of thermochemical equilibrium at a high temperature in an upstream reservoir. The program provides solutions based on frozen chemistry, chemical equilibrium, and nonequilibrium flow with finite reaction rates. Electronic nonequilibrium effects can be included using a two-temperature model. An approximate laminar boundary layer calculation is given for the shear and heat flux on the nozzle wall. Boundary layer displacement effects on the inviscid flow are considered also. Chemical equilibrium and transport property calculations are provided by subroutines. The code contains precoded thermochemical, chemical kinetic, and transport cross section data for high-temperature air, CO2-N2-Ar mixtures, helium, and argon. It provides calculations of the stagnation conditions on axisymmetric or two-dimensional models, and of the conditions on the flat surface of a blunt wedge. The primary purpose of the code is to describe the flow conditions and test conditions in electric arc heated wind tunnels.
ACUTRI: a computer code for assessing doses to the general public due to acute tritium releases
Yokoyama, Sumi; Noguchi, Hiroshi [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment; Ryufuku, Susumu; Sasaki, Toshihisa; Kurosawa, Naohiro [Visible Information Center, Inc., Tokai, Ibaraki (Japan)
2002-11-01
Tritium, which is used as a fuel of a D-T burning fusion reactor, is the most important radionuclide for the safety assessment of a nuclear fusion experimental reactor such as ITER. Thus, a computer code, ACUTRI, which calculates the radiological impact of tritium released accidentally to the atmosphere, has been developed, aiming to be of use in a discussion of licensing of a fusion experimental reactor and an environmental safety evaluation method in Japan. ACUTRI calculates an individual tritium dose based on transfer models specific to tritium in the environment and ICRP dose models. In this calculation it is also possible to analyze statistically on meteorology in the same way as a conventional dose assessment method according to the meteorological guide of the Nuclear Safety Commission of Japan. A Gaussian plume model is used for calculating the atmospheric dispersion of tritium gas (HT) and/or tritiated water (HTO). The environmental pathway model in ACUTRI considers the following internal exposures: inhalation from a primary plume (HT and/or HTO) released from the facilities and inhalation from a secondary plume (HTO) reemitted from the ground following deposition of HT and HTO. This report describes an outline of the ACUTRI code, a user guide and the results of test calculation. (author)
Coded aperture x-ray diffraction imaging with transmission computed tomography side-information
Odinaka, Ikenna; Greenberg, Joel A.; Kaganovsky, Yan; Holmgren, Andrew; Hassan, Mehadi; Politte, David G.; O'Sullivan, Joseph A.; Carin, Lawrence; Brady, David J.
2016-03-01
Coded aperture X-ray diffraction (coherent scatter spectral) imaging provides fast and dose-efficient measurements of the molecular structure of an object. The information provided is spatially-dependent and material-specific, and can be utilized in medical applications requiring material discrimination, such as tumor imaging. However, current coded aperture coherent scatter spectral imaging system assume a uniformly or weakly attenuating object, and are plagued by image degradation due to non-uniform self-attenuation. We propose accounting for such non-uniformities in the self-attenuation by utilizing an X-ray computed tomography (CT) image (reconstructed attenuation map). In particular, we present an iterative algorithm for coherent scatter spectral image reconstruction, which incorporates the attenuation map, at different stages, resulting in more accurate coherent scatter spectral images in comparison to their uncorrected counterpart. The algorithm is based on a spectrally grouped edge-preserving regularizer, where the neighborhood edge weights are determined by spatial distances and attenuation values.
Bertolotto, D.
2011-11-15
The current doctoral research is focused on the development and validation of a coupled computational tool, to combine the advantages of computational fluid dynamics (CFD) in analyzing complex flow fields and of state-of-the-art system codes employed for nuclear power plant (NPP) simulations. Such a tool can considerably enhance the analysis of NPP transient behavior, e.g. in the case of pressurized water reactor (PWR) accident scenarios such as Main Steam Line Break (MSLB) and boron dilution, in which strong coolant flow asymmetries and multi-dimensional mixing effects strongly influence the reactivity of the reactor core, as described in Chap. 1. To start with, a literature review on code coupling is presented in Chap. 2, together with the corresponding ongoing projects in the international community. Special reference is made to the framework in which this research has been carried out, i.e. the Paul Scherrer Institute's (PSI) project STARS (Steady-state and Transient Analysis Research for the Swiss reactors). In particular, the codes chosen for the coupling, i.e. the CFD code ANSYS CFX V11.0 and the system code US-NRC TRACE V5.0, are part of the STARS codes system. Their main features are also described in Chap. 2. The development of the coupled tool, named CFX/TRACE from the names of the two constitutive codes, has proven to be a complex and broad-based task, and therefore constraints had to be put on the target requirements, while keeping in mind a certain modularity to allow future extensions to be made with minimal efforts. After careful consideration, the coupling was defined to be on-line, parallel and with non-overlapping domains connected by an interface, which was developed through the Parallel Virtual Machines (PVM) software, as described in Chap. 3. Moreover, two numerical coupling schemes were implemented and tested: a sequential explicit scheme and a sequential semi-implicit scheme. Finally, it was decided that the coupling would be single
2017-04-13
AFRL-AFOSR-UK-TR-2017-0029 Automated and Assistive Tools for Accelerated Code migration of Scientific Computing on to Heterogeneous MultiCore Systems ...MultiCore Systems 5a. CONTRACT NUMBER FA8655-12-1-2021 5b. GRANT NUMBER Grant 12-2021 5c. PROGRAM ELEMENT NUMBER 61102F 6. AUTHOR(S...code for Heterogeneous multicore systems . The approach was based on the OmpSs programming model and the performance tools that constitute two strategic
Santoyo, E. [Universidad Nacional Autonoma de Mexico, Centro de Investigacion en Energia, Temixco (Mexico); Garcia, A.; Santoyo, S. [Unidad Geotermia, Inst. de Investigaciones Electricas, Temixco (Mexico); Espinosa, G. [Universidad Autonoma Metropolitana, Co. Vicentina (Mexico); Hernandez, I. [ITESM, Centro de Sistemas de Manufactura, Monterrey (Mexico)
2000-07-01
The development and application of the computer code STATIC{sub T}EMP, a useful tool for calculating static formation temperatures from actual bottomhole temperature data logged in geothermal wells is described. STATIC{sub T}EMP is based on five analytical methods which are the most frequently used in the geothermal industry. Conductive and convective heat flow models (radial, spherical/radial and cylindrical/radial) were selected. The computer code is a useful tool that can be reliably used in situ to determine static formation temperatures before or during the completion stages of geothermal wells (drilling and cementing). Shut-in time and bottomhole temperature measurements logged during well completion activities are required as input data. Output results can include up to seven computations of the static formation temperature by each wellbore temperature data set analysed. STATIC{sub T}EMP was written in Fortran-77 Microsoft language for MS-DOS environment using structured programming techniques. It runs on most IBM compatible personal computers. The source code and its computational architecture as well as the input and output files are described in detail. Validation and application examples on the use of this computer code with wellbore temperature data (obtained from specialised literature) and with actual bottomhole temperature data (taken from completion operations of some geothermal wells) are also presented. (Author)
Walitt, L.
1982-01-01
The VANS successive approximation numerical method was extended to the computation of three dimensional, viscous, transonic flows in turbomachines. A cross-sectional computer code, which conserves mass flux at each point of the cross-sectional surface of computation was developed. In the VANS numerical method, the cross-sectional computation follows a blade-to-blade calculation. Numerical calculations were made for an axial annular turbine cascade and a transonic, centrifugal impeller with splitter vanes. The subsonic turbine cascade computation was generated in blade-to-blade surface to evaluate the accuracy of the blade-to-blade mode of marching. Calculated blade pressures at the hub, mid, and tip radii of the cascade agreed with corresponding measurements. The transonic impeller computation was conducted to test the newly developed locally mass flux conservative cross-sectional computer code. Both blade-to-blade and cross sectional modes of calculation were implemented for this problem. A triplet point shock structure was computed in the inducer region of the impeller. In addition, time-averaged shroud static pressures generally agreed with measured shroud pressures. It is concluded that the blade-to-blade computation produces a useful engineering flow field in regions of subsonic relative flow; and cross-sectional computation, with a locally mass flux conservative continuity equation, is required to compute the shock waves in regions of supersonic relative flow.
A New MHD-assisted Stokes Inversion Technique
Riethmüller, T. L.; Solanki, S. K.; Barthol, P.; Gandorfer, A.; Gizon, L.; Hirzberger, J.; van Noort, M.; Blanco Rodríguez, J.; Del Toro Iniesta, J. C.; Orozco Suárez, D.; Schmidt, W.; Martínez Pillet, V.; Knölker, M.
2017-03-01
We present a new method of Stokes inversion of spectropolarimetric data and evaluate it by taking the example of a Sunrise/IMaX observation. An archive of synthetic Stokes profiles is obtained by the spectral synthesis of state-of-the-art magnetohydrodynamics (MHD) simulations and a realistic degradation to the level of the observed data. The definition of a merit function allows the archive to be searched for the synthetic Stokes profiles that best match the observed profiles. In contrast to traditional Stokes inversion codes, which solve the Unno-Rachkovsky equations for the polarized radiative transfer numerically and fit the Stokes profiles iteratively, the new technique provides the full set of atmospheric parameters. This gives us the ability to start an MHD simulation that takes the inversion result as an initial condition. After a relaxation process of half an hour solar time we obtain physically consistent MHD data sets with a target similar to the observation. The new MHD simulation is used to repeat the method in a second iteration, which further improves the match between observation and simulation, resulting in a factor of 2.2 lower mean {χ }2 value. One advantage of the new technique is that it provides the physical parameters on a geometrical height scale. It constitutes a first step toward inversions that give results consistent with the MHD equations.
GRHydro: A new open source general-relativistic magnetohydrodynamics code for the Einstein Toolkit
Moesta, Philipp; Faber, Joshua A; Haas, Roland; Noble, Scott C; Bode, Tanja; Loeffler, Frank; Ott, Christian D; Reisswig, Christian; Schnetter, Erik
2013-01-01
We present the new general-relativistic magnetohydrodynamics (GRMHD) capabilities of the Einstein Toolkit, an open-source community-driven numerical relativity and computational relativistic astrophysics code. The GRMHD extension of the Toolkit builds upon previous releases and implements the evolution of relativistic magnetised fluids in the ideal MHD limit in fully dynamical spacetimes using the same shock-capturing techniques previously applied to hydrodynamical evolution. In order to maintain the divergence-free character of the magnetic field, the code implements both hyperbolic divergence cleaning and constrained transport schemes. We present test results for a number of MHD tests in Minkowski and curved spacetimes. Minkowski tests include aligned and oblique planar shocks, cylindrical explosions, magnetic rotors, Alfv\\'en waves and advected loops, as well as a set of tests designed to study the response of the divergence cleaning scheme to numerically generated monopoles. We study the code's performanc...
MULTI-fs - A computer code for laser-plasma interaction in the femtosecond regime
Ramis, R.; Eidmann, K.; Meyer-ter-Vehn, J.; Hüller, S.
2012-03-01
The code MULTI-fs is a numerical tool devoted to the study of the interaction of ultrashort sub-picosecond laser pulses with matter in the intensity range from 10 11 to 10 17 W cm -2. Hydrodynamics is solved in one-dimensional geometry together with laser energy deposition and transport by thermal conduction and radiation. In contrast to long nanosecond pulses, short pulses generate steep gradient plasmas with typical scale lengths in the order of the laser wavelength and smaller. Under these conditions, Maxwell's equations are solved explicitly to obtain the light field. Concerning laser absorption, two different models for the electron-ion collision frequency are implemented to cover the regime of warm dense matter between high-temperature plasma and solid matter and also interaction with short-wave-length (VUV) light. MULTI-fs code is based on the MULTI radiation-hydrodynamic code [R. Ramis, R. Schmalz, J. Meyer-ter-Vehn, Comp. Phys. Comm. 49 (1988) 475] and most of the original features for the treatment of radiation are maintained. Program summaryProgram title: MULTI-fs Catalogue identifier: AEKT_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEKT_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 49 598 No. of bytes in distributed program, including test data, etc.: 443 771 Distribution format: tar.gz Programming language: FORTRAN Computer: PC (32 bits and 64 bits architecture) Operating system: Linux/Unix RAM: 1.6 MiB Classification: 19.13, 21.2 Subprograms used: Cat Id: AECV_v1_0; Title: MULTI2D; Reference: CPC 180 (2009) 977 Nature of problem: One-dimensional interaction of intense ultrashort (sub-picosecond) and ultraintense (up to 10 17 W cm -2) laser beams with matter. Solution method: The hydrodynamic motion coupled to laser propagation and
Kostin, Mikhail [FRIB, MSU; Mokhov, Nikolai [FNAL; Niita, Koji [RIST, Japan
2013-09-25
A parallel computing framework has been developed to use with general-purpose radiation transport codes. The framework was implemented as a C++ module that uses MPI for message passing. It is intended to be used with older radiation transport codes implemented in Fortran77, Fortran 90 or C. The module is significantly independent of radiation transport codes it can be used with, and is connected to the codes by means of a number of interface functions. The framework was developed and tested in conjunction with the MARS15 code. It is possible to use it with other codes such as PHITS, FLUKA and MCNP after certain adjustments. Besides the parallel computing functionality, the framework offers a checkpoint facility that allows restarting calculations with a saved checkpoint file. The checkpoint facility can be used in single process calculations as well as in the parallel regime. The framework corrects some of the known problems with the scheduling and load balancing found in the original implementations of the parallel computing functionality in MARS15 and PHITS. The framework can be used efficiently on homogeneous systems and networks of workstations, where the interference from the other users is possible.
McGrail, B.P.; Bacon, D.H.
1998-02-01
Planned performance assessments for the proposed disposal of low-activity waste (LAW) glass produced from remediation of wastes stored in underground tanks at Hanford, Washington will require calculations of radionuclide release rates from the subsurface disposal facility. These calculations will be done with the aid of computer codes. The available computer codes with suitable capabilities at the time Revision 0 of this document was prepared were ranked in terms of the feature sets implemented in the code that match a set of physical, chemical, numerical, and functional capabilities needed to assess release rates from the engineered system. The needed capabilities were identified from an analysis of the important physical and chemical processes expected to affect LAW glass corrosion and the mobility of radionuclides. This analysis was repeated in this report but updated to include additional processes that have been found to be important since Revision 0 was issued and to include additional codes that have been released. The highest ranked computer code was found to be the STORM code developed at PNNL for the US Department of Energy for evaluation of arid land disposal sites.
Keshavarz, Mohammad Hossein; Motamedoshariati, Hadi; Moghayadnia, Reza; Nazari, Hamid Reza; Azarniamehraban, Jamshid
2009-12-30
In this paper a new simple user-friendly computer code, in Visual Basic, has been introduced to evaluate detonation performance of high explosives and their thermochemical properties. The code is based on recently developed methods to obtain thermochemical and performance parameters of energetic materials, which can complement the computer outputs of the other thermodynamic chemical equilibrium codes. It can predict various important properties of high explosive including velocity of detonation, detonation pressure, heat of detonation, detonation temperature, Gurney velocity, adiabatic exponent and specific impulse of high explosives. It can also predict detonation performance of aluminized explosives that can have non-ideal behaviors. This code has been validated with well-known and standard explosives and compared the predicted results, where the predictions of desired properties were possible, with outputs of some computer codes. A large amount of data for detonation performance on different classes of explosives from C-NO(2), O-NO(2) and N-NO(2) energetic groups have also been generated and compared with well-known complex code BKW.
Keshavarz, Mohammad Hossein, E-mail: mhkeshavarz@mut-es.ac.ir [Department of Chemistry, Malek-ashtar University of Technology, Shahin-shahr P.O. Box 83145/115 (Iran, Islamic Republic of); Motamedoshariati, Hadi; Moghayadnia, Reza; Nazari, Hamid Reza; Azarniamehraban, Jamshid [Department of Chemistry, Malek-ashtar University of Technology, Shahin-shahr P.O. Box 83145/115 (Iran, Islamic Republic of)
2009-12-30
In this paper a new simple user-friendly computer code, in Visual Basic, has been introduced to evaluate detonation performance of high explosives and their thermochemical properties. The code is based on recently developed methods to obtain thermochemical and performance parameters of energetic materials, which can complement the computer outputs of the other thermodynamic chemical equilibrium codes. It can predict various important properties of high explosive including velocity of detonation, detonation pressure, heat of detonation, detonation temperature, Gurney velocity, adiabatic exponent and specific impulse of high explosives. It can also predict detonation performance of aluminized explosives that can have non-ideal behaviors. This code has been validated with well-known and standard explosives and compared the predicted results, where the predictions of desired properties were possible, with outputs of some computer codes. A large amount of data for detonation performance on different classes of explosives from C-NO{sub 2}, O-NO{sub 2} and N-NO{sub 2} energetic groups have also been generated and compared with well-known complex code BKW.
CATARACT: Computer code for improving power calculations at NREL's high-flux solar furnace
Scholl, K.; Bingham, C.; Lewandowski, A.
1994-01-01
The High-Flux Solar Furnace (HFSF), operated by the National Renewable Energy Laboratory, uses a camera-based, flux-mapping system to analyze the distribution and to determine total power at the focal point. The flux-mapping system consists of a diffusively reflecting plate with seven circular foil calorimeters, a charge-coupled device (CCD) camera, an IBM-compatible personal computer with a frame-grabber board, and commercial image analysis software. The calorimeters provide flux readings that are used to scale the image captured from the plate by the camera. The image analysis software can estimate total power incident on the plate by integrating under the 3-dimensional image. Because of the physical layout of the HFSF, the camera is positioned at a 20 angle to the flux mapping plate normal. The foreshortening of the captured images that results represents a systematic error in the power calculations because the software incorrectly assumes the image is parallel to the camera's array. We have written a FORTRAN computer program called CATARACT (camera/target angle correction) that we use to transform the original flux-mapper image to a plane that is normal to the camera's optical axis. A description of the code and the results of experiments performed to verify it are presented. Also presented are comparisons of the total power available from the HFSF as determined from the flux mapping system and theoretical considerations.
C.S. Ierotheou
2001-01-01
Full Text Available The shared-memory programming model can be an effective way to achieve parallelism on shared memory parallel computers. Historically however, the lack of a programming standard using directives and the limited scalability have affected its take-up. Recent advances in hardware and software technologies have resulted in improvements to both the performance of parallel programs with compiler directives and the issue of portability with the introduction of OpenMP. In this study, the Computer Aided Parallelisation Toolkit has been extended to automatically generate OpenMP-based parallel programs with nominal user assistance. We categorize the different loop types and show how efficient directives can be placed using the toolkit's in-depth interprocedural analysis. Examples are taken from the NAS parallel benchmarks and a number of real-world application codes. This demonstrates the great potential of using the toolkit to quickly parallelise serial programs as well as the good performance achievable on up to 300 processors for hybrid message passing-directive parallelisations.
A simulation of a pebble bed reactor core by the MCNP-4C computer code
Bakhshayesh Moshkbar Khalil
2009-01-01
Full Text Available Lack of energy is a major crisis of our century; the irregular increase of fossil fuel costs has forced us to search for novel, cheaper, and safer sources of energy. Pebble bed reactors - an advanced new generation of reactors with specific advantages in safety and cost - might turn out to be the desired candidate for the role. The calculation of the critical height of a pebble bed reactor at room temperature, while using the MCNP-4C computer code, is the main goal of this paper. In order to reduce the MCNP computing time compared to the previously proposed schemes, we have devised a new simulation scheme. Different arrangements of kernels in fuel pebble simulations were investigated and the best arrangement to decrease the MCNP execution time (while keeping the accuracy of the results, chosen. The neutron flux distribution and control rods worth, as well as their shadowing effects, have also been considered in this paper. All calculations done for the HTR-10 reactor core are in good agreement with experimental results.
Glassman, Arthur J.; Jones, Scott M.
1991-01-01
This analysis and this computer code apply to full, split, and dual expander cycles. Heat regeneration from the turbine exhaust to the pump exhaust is allowed. The combustion process is modeled as one of chemical equilibrium in an infinite-area or a finite-area combustor. Gas composition in the nozzle may be either equilibrium or frozen during expansion. This report, which serves as a users guide for the computer code, describes the system, the analysis methodology, and the program input and output. Sample calculations are included to show effects of key variables such as nozzle area ratio and oxidizer-to-fuel mass ratio.
Test results of a 40 kW Stirling engine and comparison with the NASA-Lewis computer code predictions
Allen, D.; Cairelli, J.
1985-01-01
A Stirling engine was tested without auxiliaries at NASA-Lewis. Three different regenerator configurations were tested with hydrogen. The test objectives were (1) to obtain steady-state and dynamic engine data, including indicated power, for validation of an existing computer model for this engine; and (2) to evaluate structurally the use of silicon carbide regenerators. This paper presents comparisons of the measured brake performance, indicated mean effective pressure, and cyclic pressure variations with those predicted by the code. The measured data tended to be lower than the computer code predictions. The silicon carbide foam regenerators appear to be structurally suitable, but the foam matrix tested severely reduced performance.
Hamada, Michael S [Los Alamos National Laboratory; Higdon, David M [Los Alamos National Laboratory
2009-01-01
In this paper, we present a generic example to illustrate various points about making future predictions of population performance using a biased performance computer code, physical performance data, and critical performance parameter data sampled from the population at various times. We show how the actual performance data help to correct the biased computer code and the impact of uncertainty especially when the prediction is made far from where the available data are taken. We also demonstrate how a Bayesian approach allows both inferences about the unknown parameters and predictions to be made in a consistent framework.
Grid cells generate an analog error-correcting code for singularly precise neural computation.
Sreenivasan, Sameet; Fiete, Ila
2011-09-11
Entorhinal grid cells in mammals fire as a function of animal location, with spatially periodic response patterns. This nonlocal periodic representation of location, a local variable, is unlike other neural codes. There is no theoretical explanation for why such a code should exist. We examined how accurately the grid code with noisy neurons allows an ideal observer to estimate location and found this code to be a previously unknown type of population code with unprecedented robustness to noise. In particular, the representational accuracy attained by grid cells over the coding range was in a qualitatively different class from what is possible with observed sensory and motor population codes. We found that a simple neural network can effectively correct the grid code. To the best of our knowledge, these results are the first demonstration that the brain contains, and may exploit, powerful error-correcting codes for analog variables.
Louis, J.F.
1977-10-01
Current research and development in open-cycle coal-fired MHD power generation at Massachusetts Institute of Technology is presented. Progress is reported on the following tasks: (1) Establish chemical, thermal, and electrical data to guide materials selection, develop improved detail designs, and support performance analyses of MHD electrode modules and insulator materials; (2) Parametrically investigate selected electrode properties of critical design importance in chemical, thermal, and electrical environments simulating a coal-fired MHD generator; (3) Develop combustion data pertinent to the design of MHD combustors; this work is intended to determine the combustion characteristics of selected coal feed stock in terms of devolatilization kinetics, char characteristics, and combustion gas chemistry; (4) Establish techniques for the analysis and understanding of critical MHD phenomena which have an important bearing on MHD generator performance; such phenomena include inter-electrode breakdown, time-dependent behavior, effective plasma properties and plasma inhomogeneities; (5) Establish the operating characteristics of an MHD disk generator; (6) Continue work on computer techniques for modeling and for design and cost analysis of MHD components and the overall system; (7) Integrate the engineering data and design criteria, as applicable, which are developed in the above-listed tasks into a model of the MHD channel.
Louis, J.F.
1976-10-01
Research progress on open-cycle coal-fired MHD power generation at Massachusetts Institute of Technology is detailed. Work is reported in the following areas: (1) Establish chemical, thermal, and electrical data to guide materials selection, develop improved detail designs, and support performance analyses of MHD electrode modules and insulator materials; (2) parametrically investigate selected electrode properties of critical design importance in chemical, thermal, and electrical environments simulating a coal-fired MHD generator; (3) develop combustion data pertinent to the design of MHD combustors; (4) establish techniques for the analysis and understanding of critical MHD phenomena which have an important bearing on MHD generator performance; (5) establish the operating characteristics of an MHD disk generator; (6) continue work on computer techniques for modeling and for design and cost analysis of MHD components and the overall system; (7) integrate the engineering data and design criteria, as applicable, which are developed in the above-listed tasks into a model of the MHD channel; (8) U-25 electrode screening tests.
A new class of codes for Boolean masking of cryptographic computations
Carlet, Claude; Kim, Jon-Lark; Solé, Patrick
2011-01-01
We introduce a new class of rate one half binary codes: complementary information set codes. A binary linear code of length 2n and dimension n is called a complementary information set code (CIS code for short) if it has two disjoint information sets. This class of codes contains self-dual codes as a subclass. It is connected to graph correlation immune Boolean functions of use in the security of hardware implementations of cryptographic primitives. Such codes permit to improve the cost of masking cryptographic algorithms against side channel attacks. In this paper we investigate this new class of codes: we give optimal or best known CIS codes of length < 132. We derive general constructions based on cyclic codes and on double circulant codes. We derive a Varshamov-Gilbert bound for long CIS codes, and show that they can all be classified in small lengths \\leq 12 by the building up construction. Some nonlinear S-boxes are constructed by using Z4-codes, based on the notion of dual distance of an unrestricte...
MHD Simulations of the Plasma Flow in the Magnetic Nozzle
Smith, T. E. R.; Keidar, M.; Sankaran, K.; olzin, K. A.
2013-01-01
The magnetohydrodynamic (MHD) flow of plasma through a magnetic nozzle is simulated by solving the governing equations for the plasma flow in the presence of an static magnetic field representing the applied nozzle. This work will numerically investigate the flow and behavior of the plasma as the inlet plasma conditions and magnetic nozzle field strength are varied. The MHD simulations are useful for addressing issues such as plasma detachment and to can be used to gain insight into the physical processes present in plasma flows found in thrusters that use magnetic nozzles. In the model, the MHD equations for a plasma, with separate temperatures calculated for the electrons and ions, are integrated over a finite cell volume with flux through each face computed for each of the conserved variables (mass, momentum, magnetic flux, energy) [1]. Stokes theorem is used to convert the area integrals over the faces of each cell into line integrals around the boundaries of each face. The state of the plasma is described using models of the ionization level, ratio of specific heats, thermal conductivity, and plasma resistivity. Anisotropies in current conduction due to Hall effect are included, and the system is closed using a real-gas equation of state to describe the relationship between the plasma density, temperature, and pressure.A separate magnetostatic solver is used to calculate the applied magnetic field, which is assumed constant for these calculations. The total magnetic field is obtained through superposition of the solution for the applied magnetic field and the self-consistently computed induced magnetic fields that arise as the flowing plasma reacts to the presence of the applied field. A solution for the applied magnetic field is represented in Fig. 1 (from Ref. [2]), exhibiting the classic converging-diverging field pattern. Previous research was able to demonstrate effects such as back-emf at a super-Alfvenic flow, which significantly alters the shape of the
Lin, J. W.; Erickson, T. A.
2011-12-01
Historically, the application of high-performance computing (HPC) to the atmospheric sciences has focused on using the increases in processor speed, storage, and parallelization to run longer simulations of larger and more complex models. Such a focus, however, has led to a user culture where code robustness and reusability is ignored or discouraged. Additionally, such a culture works against nurturing and growing connections between high-performance computational earth sciences and scientific users outside of that community. Given the explosion in computational power available to researchers unconnected with the traditional HPC centers, as well as in the number of quality tools available to conduct analysis and visualization, the programming insularity of the earth science modeling and analysis community acts as a formidible barrier to increasing the usefulness and robustness of computational earth science products. In this talk, we suggest adoption of best practices from the software engineering community, and in particular the open-source community, has the potential to improve the quality of code and increase the impact of earth sciences HPC. In particular, we will discuss the impact of practices such as unit testing and code review, the need and preconditions for code reusability, and the importance of APIs and open frameworks to enable scientific discovery across sub-disciplines. We will present examples of the cross-disciplinary fertilization possible with open APIs. Finally, we will discuss ways funding agencies and the computational earth sciences community can help encourage the adoption of such best practices.
Indo-Soviet experiment on an MHD generator test section at the Soviet U-O/sub 2/ facility
Ananthapadmanabhan, P.V.; Bapat, A.V.; Das, A.K. (Bhabha Atomic Research Centre, Bombay (India))
1982-09-01
This paper summarizes the major results of the joint Indo-Soviet experiment for testing the Indian MHD generator channel section, designed and fabricated at the Bhabha Atomic Research Centre, Bombay, which was carried out at the U-02 facility in Moscow, USSR, in May 1980. The total test duration was 65 hours and included electrophysical tests and life tests under applied electric fields. The main purpose of the tests was to substantiate the physical concepts, computer codes, design features and special processing techniques involved in the development of MHD generators for the Indian pilot plant at Tiruchirapalli. The experimental observations on the phenomena of heat transfer to the walls, gas dynamics in the channel, electrical characteristics of the generator and near-electrode processes including the analysis of arc spots correlate with the theoretical estimates based on present uderstanding of the physical processes occuring in similar MHD generators. The post-operational inspection of the channel section and extensive investigation of materials through microscopic analysis, chemical analysis and x-ray analysis are also reported in this paper. The joint test programme has clearly demonstrated the definite operating capability of the test section and has given sufficient information and encouragement for building better and improved channels for the future.
NONE
1997-03-01
This Manual represents Revision 5 of the user documentation for the modular code system referred to as SCALE. The history of the SCALE code system dates back to 1969 when the current Computational Physics and Engineering Division at Oak Ridge National Laboratory (ORNL) began providing the transportation package certification staff at the U.S. Atomic Energy Commission with computational support in the use of the new KENO code for performing criticality safety assessments with the statistical Monte Carlo method. From 1969 to 1976 the certification staff relied on the ORNL staff to assist them in the correct use of codes and data for criticality, shielding, and heat transfer analyses of transportation packages. However, the certification staff learned that, with only occasional use of the codes, it was difficult to become proficient in performing the calculations often needed for an independent safety review. Thus, shortly after the move of the certification staff to the U.S. Nuclear Regulatory Commission (NRC), the NRC staff proposed the development of an easy-to-use analysis system that provided the technical capabilities of the individual modules with which they were familiar. With this proposal, the concept of the Standardized Computer Analyses for Licensing Evaluation (SCALE) code system was born. This volume consists of the section of the manual dealing with eight of the functional modules in the code. Those are: BONAMI - resonance self-shielding by the Bondarenko method; NITAWL-II - SCALE system module for performing resonance shielding and working library production; XSDRNPM - a one-dimensional discrete-ordinates code for transport analysis; XSDOSE - a module for calculating fluxes and dose rates at points outside a shield; KENO IV/S - an improved monte carlo criticality program; COUPLE; ORIGEN-S - SCALE system module to calculate fuel depletion, actinide transmutation, fission product buildup and decay, and associated radiation source terms; ICE.
DIST: a computer code system for calculation of distribution ratios of solutes in the purex system
Tachimori, Shoichi [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment
1996-05-01
Purex is a solvent extraction process for reprocessing the spent nuclear fuel using tri n-butylphosphate (TBP). A computer code system DIST has been developed to calculate distribution ratios for the major solutes in the Purex process. The DIST system is composed of database storing experimental distribution data of U(IV), U(VI), Pu(III), Pu(IV), Pu(VI), Np(IV), Np(VI), HNO{sub 3} and HNO{sub 2}: DISTEX and of Zr(IV), Tc(VII): DISTEXFP and calculation programs to calculate distribution ratios of U(IV), U(VI), Pu(III), Pu(IV), Pu(VI), Np(IV), Np(VI), HNO{sub 3} and HNO{sub 2}(DIST1), and Zr(IV), Tc(VII)(DITS2). The DIST1 and DIST2 determine, by the best-fit procedures, the most appropriate values of many parameters put on empirical equations by using the DISTEX data which fulfill the assigned conditions and are applied to calculate distribution ratios of the respective solutes. Approximately 5,000 data were stored in the DISTEX and DISTEXFP. In the present report, the following items are described, 1) specific features of DIST1 and DIST2 codes and the examples of calculation 2) explanation of databases, DISTEX, DISTEXFP and a program DISTIN, which manages the data in the DISTEX and DISTEXFP by functions as input, search, correction and delete. and at the annex, 3) programs of DIST1, DIST2, and figure-drawing programs DIST1G and DIST2G 4) user manual for DISTIN. 5) source programs of DIST1 and DIST2. 6) the experimental data stored in the DISTEX and DISTEXFP. (author). 122 refs.
Kumar Parijat Tripathi
Full Text Available RNA-seq is a new tool to measure RNA transcript counts, using high-throughput sequencing at an extraordinary accuracy. It provides quantitative means to explore the transcriptome of an organism of interest. However, interpreting this extremely large data into biological knowledge is a problem, and biologist-friendly tools are lacking. In our lab, we developed Transcriptator, a web application based on a computational Python pipeline with a user-friendly Java interface. This pipeline uses the web services available for BLAST (Basis Local Search Alignment Tool, QuickGO and DAVID (Database for Annotation, Visualization and Integrated Discovery tools. It offers a report on statistical analysis of functional and Gene Ontology (GO annotation's enrichment. It helps users to identify enriched biological themes, particularly GO terms, pathways, domains, gene/proteins features and protein-protein interactions related informations. It clusters the transcripts based on functional annotations and generates a tabular report for functional and gene ontology annotations for each submitted transcript to the web server. The implementation of QuickGo web-services in our pipeline enable the users to carry out GO-Slim analysis, whereas the integration of PORTRAIT (Prediction of transcriptomic non coding RNA (ncRNA by ab initio methods helps to identify the non coding RNAs and their regulatory role in transcriptome. In summary, Transcriptator is a useful software for both NGS and array data. It helps the users to characterize the de-novo assembled reads, obtained from NGS experiments for non-referenced organisms, while it also performs the functional enrichment analysis of differentially expressed transcripts/genes for both RNA-seq and micro-array experiments. It generates easy to read tables and interactive charts for better understanding of the data. The pipeline is modular in nature, and provides an opportunity to add new plugins in the future. Web application is
Marcos Antonio Klunk
Full Text Available ABSTRACTDiagenetic reactions, characterized by the dissolution and precipitation of minerals at low temperatures, control the quality of sedimentary rocks as hydrocarbon reservoirs. Geochemical modeling, a tool used to understand diagenetic processes, is performed through computer codes based on thermodynamic and kinetic parameters. In a comparative study, we reproduced the diagenetic reactions observed in Snorre Field reservoir sandstones, Norwegian North Sea. These reactions had been previously modeled in the literature using DISSOL-THERMAL code. In this study, we modeled the diagenetic reactions in the reservoirs using Geochemist's Workbench (GWB and TOUGHREACT software, based on a convective-diffusive-reactive model and on the thermodynamic and kinetic parameters compiled for each reaction. TOUGHREACT and DISSOL-THERMAL modeling showed dissolution of quartz, K-feldspar and plagioclase in a similar temperature range from 25 to 80°C. In contrast, GWB modeling showed dissolution of albite, plagioclase and illite, as well as precipitation of quartz, K-feldspar and kaolinite in the same temperature range. The modeling generated by the different software for temperatures of 100, 120 and 140°C showed similarly the dissolution of quartz, K-feldspar, plagioclase and kaolinite, but differed in the precipitation of albite and illite. At temperatures of 150 and 160°C, GWB and TOUGHREACT produced different results from the DISSOL-THERMAL, except for the dissolution of quartz, plagioclase and kaolinite. The comparative study allows choosing the numerical modeling software whose results are closer to the diagenetic reactions observed in the petrographic analysis of the modeled reservoirs.
Spokoyny, Ilana; Chen, James Y; Raman, Rema; Ernstrom, Karin; Agrawal, Kunal; Modir, Royya F; Meyer, Dawn M; Meyer, Brett C
2016-12-01
Head computed tomography (CT) is critical for stroke code evaluations and often happens prior to completion of the neurological exam. Eye deviation on neuroimaging (DeyeCOM sign) has utility for predicting stroke diagnosis and correlates with National Institutes of Health Stroke Scale (NIHSS) gaze score. We further assessed the utility of the DeyeCOM sign, without complex caliper-based eye deviation calculations, but simply with a visual determination method. Patients with initial head CT and final diagnosis from an institutional review board-approved consecutive prospective registry of stroke codes at the University of California, San Diego, were included. Five stroke specialists and 1 neuroradiologist reviewed each CT. DeyeCOM+ patients were compared to DeyeCOM- patients (baseline characteristics, diagnosis, and NIHSS gaze score). Kappa statistics compared stroke specialists to neuroradiologist reads, and visual determination to caliper measurement of DeyeCOM sign. Of 181 patients, 46 were DeyeCOM+. Ischemic stroke was more commonly diagnosed in DeyeCOM+ patients compared to other diagnoses (P = .039). DeyeCOM+ patients were more likely to have an NIHSS gaze score of 1 or higher (P = .006). The NIHSS score of DeyeCOM+ stroke versus DeyeCOM- stroke patients was 8.3 ± 6.0 versus 6.7 ± 8.0 (P = .065). Functional outcomes were similar (P = .59). Stroke specialists had excellent agreement with the neuroradiologist (Κ = .89). Visual inspection had excellent agreement with the caliper method (Κ = .88). Using a time-sensitive visual determination of gaze deviation on imaging was predictive of ischemic stroke diagnosis and presence of NIHSS gaze score, and was consistent with the more complex caliper method. This study furthers the clinical utility of the DeyeCOM sign for predicting ischemic strokes. Copyright © 2016 National Stroke Association. Published by Elsevier Inc. All rights reserved.
Zuccaro, Antonio; Guarracino, Mario Rosario
2015-01-01
RNA-seq is a new tool to measure RNA transcript counts, using high-throughput sequencing at an extraordinary accuracy. It provides quantitative means to explore the transcriptome of an organism of interest. However, interpreting this extremely large data into biological knowledge is a problem, and biologist-friendly tools are lacking. In our lab, we developed Transcriptator, a web application based on a computational Python pipeline with a user-friendly Java interface. This pipeline uses the web services available for BLAST (Basis Local Search Alignment Tool), QuickGO and DAVID (Database for Annotation, Visualization and Integrated Discovery) tools. It offers a report on statistical analysis of functional and Gene Ontology (GO) annotation’s enrichment. It helps users to identify enriched biological themes, particularly GO terms, pathways, domains, gene/proteins features and protein—protein interactions related informations. It clusters the transcripts based on functional annotations and generates a tabular report for functional and gene ontology annotations for each submitted transcript to the web server. The implementation of QuickGo web-services in our pipeline enable the users to carry out GO-Slim analysis, whereas the integration of PORTRAIT (Prediction of transcriptomic non coding RNA (ncRNA) by ab initio methods) helps to identify the non coding RNAs and their regulatory role in transcriptome. In summary, Transcriptator is a useful software for both NGS and array data. It helps the users to characterize the de-novo assembled reads, obtained from NGS experiments for non-referenced organisms, while it also performs the functional enrichment analysis of differentially expressed transcripts/genes for both RNA-seq and micro-array experiments. It generates easy to read tables and interactive charts for better understanding of the data. The pipeline is modular in nature, and provides an opportunity to add new plugins in the future. Web application is freely
MHD squeezing flow between two infinite plates
Umar Khan
2014-03-01
Full Text Available Magneto hydrodynamic (MHD squeezing flow of a viscous fluid has been discussed. Conservation laws combined with similarity transformations have been used to formulate the flow mathematically that leads to a highly nonlinear ordinary differential equation. Analytical solution to the resulting differential equation is determined by employing Variation of Parameters Method (VPM. Runge–Kutta order-4 method is also used to solve the same problem for the sake of comparison. It is found that solution using VPM reduces the computational work yet maintains a very high level of accuracy. The influence of different parameters is also discussed and demonstrated graphically.
Blink, J.A.
1985-03-01
In this manual we describe the use of the FORIG computer code to solve isotope-generation and depletion problems in fusion and fission reactors. FORIG runs on a Cray-1 computer and accepts more extensive activation cross sections than ORIGEN2 from which it was adapted. This report is an updated and a combined version of the previous ORIGEN2 and FORIG manuals. 7 refs., 15 figs., 13 tabs.
Ruetten, H.J.; Haas, K.A.; Brockmann, H.; Ohlig, U.; Scherer, W.
2000-10-01
V.S.O.P. is a computer code system for the comprehensive numerical simulation of the physics of thermal reactors. It implies the setup of the reactor and of the fuel element, processing of cross sections, neutron spectrum evaluation, neutron diffusion calculation in two or three dimensions, fuel burnup, fuel shuffling, reactor control, thermal hydraulics and fuel cycle costs. The thermal hydraulics part (steady state and time-dependent) is restricted to HTRs and to two spatial dimensions. The code can simulate the reactor operation from the initial core towards the equilibrium core. V.S.O.P.(99) represents the further development of V.S.O.P. (97). Compared to its precursor, the code system has been improved in many details. Major extensions have been included concerning the thermal hydraulic sections. Beyond that, the many modules of the code-system have been condensed to only 2 executables in the ''99''-release of V.S.O.P., to be comfortably handled on a WINDOWS-PC or a UNIX-computer. The necessary data input as well as the handling and book-keeping of intermediate data sets has been condensed and simplified. A 64 MB memory should be available for the execution of the code. The hard disk requirement for the executables and the basic libraries associated with the code amounts to about 7 MB. (orig.)
Rathjen, K. A.
1977-01-01
A digital computer code CAVE (Conduction Analysis Via Eigenvalues), which finds application in the analysis of two dimensional transient heating of hypersonic vehicles is described. The CAVE is written in FORTRAN 4 and is operational on both IBM 360-67 and CDC 6600 computers. The method of solution is a hybrid analytical numerical technique that is inherently stable permitting large time steps even with the best of conductors having the finest of mesh size. The aerodynamic heating boundary conditions are calculated by the code based on the input flight trajectory or can optionally be calculated external to the code and then entered as input data. The code computes the network conduction and convection links, as well as capacitance values, given basic geometrical and mesh sizes, for four generations (leading edges, cooled panels, X-24C structure and slabs). Input and output formats are presented and explained. Sample problems are included. A brief summary of the hybrid analytical-numerical technique, which utilizes eigenvalues (thermal frequencies) and eigenvectors (thermal mode vectors) is given along with aerodynamic heating equations that have been incorporated in the code and flow charts.
High-Order Finite Difference GLM-MHD Schemes for Cell-Centered MHD
Mignone, A; Bodo, G
2010-01-01
We present and compare third- as well as fifth-order accurate finite difference schemes for the numerical solution of the compressible ideal MHD equations in multiple spatial dimensions. The selected methods lean on four different reconstruction techniques based on recently improved versions of the weighted essentially non-oscillatory (WENO) schemes, monotonicity preserving (MP) schemes as well as slope-limited polynomial reconstruction. The proposed numerical methods are highly accurate in smooth regions of the flow, avoid loss of accuracy in proximity of smooth extrema and provide sharp non-oscillatory transitions at discontinuities. We suggest a numerical formulation based on a cell-centered approach where all of the primary flow variables are discretized at the zone center. The divergence-free condition is enforced by augmenting the MHD equations with a generalized Lagrange multiplier yielding a mixed hyperbolic/parabolic correction, as in Dedner et al. (J. Comput. Phys. 175 (2002) 645-673). The resulting...
Source Term Model for Vortex Generator Vanes in a Navier-Stokes Computer Code
Waithe, Kenrick A.
2004-01-01
A source term model for an array of vortex generators was implemented into a non-proprietary Navier-Stokes computer code, OVERFLOW. The source term models the side force created by a vortex generator vane. The model is obtained by introducing a side force to the momentum and energy equations that can adjust its strength automatically based on the local flow. The model was tested and calibrated by comparing data from numerical simulations and experiments of a single low profile vortex generator vane on a flat plate. In addition, the model was compared to experimental data of an S-duct with 22 co-rotating, low profile vortex generators. The source term model allowed a grid reduction of about seventy percent when compared with the numerical simulations performed on a fully gridded vortex generator on a flat plate without adversely affecting the development and capture of the vortex created. The source term model was able to predict the shape and size of the stream-wise vorticity and velocity contours very well when compared with both numerical simulations and experimental data. The peak vorticity and its location were also predicted very well when compared to numerical simulations and experimental data. The circulation predicted by the source term model matches the prediction of the numerical simulation. The source term model predicted the engine fan face distortion and total pressure recovery of the S-duct with 22 co-rotating vortex generators very well. The source term model allows a researcher to quickly investigate different locations of individual or a row of vortex generators. The researcher is able to conduct a preliminary investigation with minimal grid generation and computational time.
Computational promoter analysis of mouse, rat and human antimicrobial peptide-coding genes
Kai Chikatoshi
2006-12-01
Full Text Available Abstract Background Mammalian antimicrobial peptides (AMPs are effectors of the innate immune response. A multitude of signals coming from pathways of mammalian pathogen/pattern recognition receptors and other proteins affect the expression of AMP-coding genes (AMPcgs. For many AMPcgs the promoter elements and transcription factors that control their tissue cell-specific expression have yet to be fully identified and characterized. Results Based upon the RIKEN full-length cDNA and public sequence data derived from human, mouse and rat, we identified 178 candidate AMP transcripts derived from 61 genes belonging to 29 AMP families. However, only for 31 mouse genes belonging to 22 AMP families we were able to determine true orthologous relationships with 30 human and 15 rat sequences. We screened the promoter regions of AMPcgs in the three species for motifs by an ab initio motif finding method and analyzed the derived promoter characteristics. Promoter models were developed for alpha-defensins, penk and zap AMP families. The results suggest a core set of transcription factors (TFs that regulate the transcription of AMPcg families in mouse, rat and human. The three most frequent core TFs groups include liver-, nervous system-specific and nuclear hormone receptors (NHRs. Out of 440 motifs analyzed, we found that three represent potentially novel TF-binding motifs enriched in promoters of AMPcgs, while the other four motifs appear to be species-specific. Conclusion Our large-scale computational analysis of promoters of 22 families of AMPcgs across three mammalian species suggests that their key transcriptional regulators are likely to be TFs of the liver-, nervous system-specific and NHR groups. The computationally inferred promoter elements and potential TF binding motifs provide a rich resource for targeted experimental validation of TF binding and signaling studies that aim at the regulation of mouse, rat or human AMPcgs.
Oster, C.A.
1976-02-01
DIGRTS is a computer program for calculating the concentration distribution of an inert gas diffusing through a composite solid which permits reversible trapping of the gas atoms. The program is coded entirely in FORTRAN IV. The composite solid can consist of up to ten regions. These regions are further subdivided in a manner so that the total number of nodes (subdivision points) is less than or equal to 200. The code can readily be modified should these limits be undesirable. The code permits only constant parameters for describing the diffusion process but the extension to allow time dependent parameters is clearly marked in the program. In addition to calculations of the concentration distributions, the time dependent release fraction is also computed. (auth)
Stimulus Specificity of Brain-Computer Interfaces Based on Code Modulation Visual Evoked Potentials.
Qingguo Wei
Full Text Available A brain-computer interface (BCI based on code modulated visual evoked potentials (c-VEP is among the fastest BCIs that have ever been reported, but it has not yet been given a thorough study. In this study, a pseudorandom binary M sequence and its time lag sequences are utilized for modulation of different stimuli and template matching is adopted as the method for target recognition. Five experiments were devised to investigate the effect of stimulus specificity on target recognition and we made an effort to find the optimal stimulus parameters for size, color and proximity of the stimuli, length of modulation sequence and its lag between two adjacent stimuli. By changing the values of these parameters and measuring classification accuracy of the c-VEP BCI, an optimal value of each parameter can be attained. Experimental results of ten subjects showed that stimulus size of visual angle 3.8°, white, spatial proximity of visual angle 4.8° center to center apart, modulation sequence of length 63 bits and the lag of 4 bits between adjacent stimuli yield individually superior performance. These findings provide a basis for determining stimulus presentation of a high-performance c-VEP based BCI system.
Assessment of Turbulent Shock-Boundary Layer Interaction Computations Using the OVERFLOW Code
Oliver, A. B.; Lillard, R. P.; Schwing, A. M.; Blaisdell, G> A.; Lyrintzis, A. S.
2007-01-01
The performance of two popular turbulence models, the Spalart-Allmaras model and Menter s SST model, and one relatively new model, Olsen & Coakley s Lag model, are evaluated using the OVERFLOWcode. Turbulent shock-boundary layer interaction predictions are evaluated with three different experimental datasets: a series of 2D compression ramps at Mach 2.87, a series of 2D compression ramps at Mach 2.94, and an axisymmetric coneflare at Mach 11. The experimental datasets include flows with no separation, moderate separation, and significant separation, and use several different experimental measurement techniques (including laser doppler velocimetry (LDV), pitot-probe measurement, inclined hot-wire probe measurement, preston tube skin friction measurement, and surface pressure measurement). Additionally, the OVERFLOW solutions are compared to the solutions of a second CFD code, DPLR. The predictions for weak shock-boundary layer interactions are in reasonable agreement with the experimental data. For strong shock-boundary layer interactions, all of the turbulence models overpredict the separation size and fail to predict the correct skin friction recovery distribution. In most cases, surface pressure predictions show too much upstream influence, however including the tunnel side-wall boundary layers in the computation improves the separation predictions.
Dual Coding Theory Explains Biphasic Collective Computation in Neural Decision-Making.
Daniels, Bryan C; Flack, Jessica C; Krakauer, David C
2017-01-01
A central question in cognitive neuroscience is how unitary, coherent decisions at the whole organism level can arise from the distributed behavior of a large population of neurons with only partially overlapping information. We address this issue by studying neural spiking behavior recorded from a multielectrode array with 169 channels during a visual motion direction discrimination task. It is well known that in this task there are two distinct phases in neural spiking behavior. Here we show Phase I is a distributed or incompressible phase in which uncertainty about the decision is substantially reduced by pooling information from many cells. Phase II is a redundant or compressible phase in which numerous single cells contain all the information present at the population level in Phase I, such that the firing behavior of a single cell is enough to predict the subject's decision. Using an empirically grounded dynamical modeling framework, we show that in Phase I large cell populations with low redundancy produce a slow timescale of information aggregation through critical slowing down near a symmetry-breaking transition. Our model indicates that increasing collective amplification in Phase II leads naturally to a faster timescale of information pooling and consensus formation. Based on our results and others in the literature, we propose that a general feature of collective computation is a "coding duality" in which there are accumulation and consensus formation processes distinguished by different timescales.
Investigation of station blackout scenario in VVER440/v230 with RELAP5 computer code
Gencheva, Rositsa Veselinova, E-mail: roseh@mail.bg; Stefanova, Antoaneta Emilova, E-mail: antoanet@inrne.bas.bg; Groudev, Pavlin Petkov, E-mail: pavlinpg@inrne.bas.bg
2015-12-15
Highlights: • We have modeled SBO in VVER440. • RELAP5/MOD3 computer code has been used. • Base case calculation has been done. • Fail case calculation has been done. • Operator and alternative operator actions have been investigated. - Abstract: During the development of symptom-based emergency operating procedures (SB-EOPs) for VVER440/v230 units at Kozloduy Nuclear Power Plant (NPP) a number of analyses have been performed using the RELAP5/MOD3 (Carlson et al., 1990). Some of them investigate the response of VVER440/v230 during the station blackout (SBO). The main purpose of the analyses presented in this paper is to identify the behavior of important VVER440 parameters in case of total station blackout. The RELAP5/MOD3 has been used to simulate the SBO in VVER440 NPP model (Fletcher and Schultz, 1995). This model was developed at the Institute for Nuclear Research and Nuclear Energy, Bulgarian Academy of Sciences (INRNE-BAS), Sofia, for analyses of operational occurrences, abnormal events and design based scenarios. The model provides a significant analytical capability for specialists working in the field of NPP safety.
Elfer, N.; Meibaum, R.; Olsen, G.
1995-01-01
A unique collection of computer codes, Space Debris Surfaces (SD_SURF), have been developed to assist in the design and analysis of space debris protection systems. SD_SURF calculates and summarizes a vehicle's vulnerability to space debris as a function of impact velocity and obliquity. An SD_SURF analysis will show which velocities and obliquities are the most probable to cause a penetration. This determination can help the analyst select a shield design that is best suited to the predominant penetration mechanism. The analysis also suggests the most suitable parameters for development or verification testing. The SD_SURF programs offer the option of either FORTRAN programs or Microsoft-EXCEL spreadsheets and macros. The FORTRAN programs work with BUMPERII. The EXCEL spreadsheets and macros can be used independently or with selected output from the SD_SURF FORTRAN programs. Examples will be presented of the interaction between space vehicle geometry, the space debris environment, and the penetration and critical damage ballistic limit surfaces of the shield under consideration.
A computational code for resolution of general compartment models applied to internal dosimetry
Claro, Thiago R.; Todo, Alberto S., E-mail: claro@usp.br, E-mail: astodo@ipen.br [Instituto de Pesquisas Energeticas e Nucleares (IPEN/CNEN-SP), Sao Paulo, SP (Brazil)
2011-07-01
The dose resulting from internal contamination can be estimated with the use of biokinetic models combined with experimental results obtained from bio analysis and the knowledge of the incorporation time. The biokinetics models can be represented by a set of compartments expressing the transportation, retention and elimination of radionuclides from the body. The ICRP publications, number 66, 78 and 100, present compartmental models for the respiratory tract, gastrointestinal tract and for systemic distribution for an array of radionuclides of interest for the radiological protection. The objective of this work is to develop a computational code for designing, visualization and resolution of compartmental models of any nature. There are available four different techniques for the resolution of system of differential equations, including semi-analytical and numerical methods. The software was developed in C{ne} programming, using a Microsoft Access database and XML standards for file exchange with other applications. Compartmental models for uranium, thorium and iodine radionuclides were generated for the validation of the CBT software. The models were subsequently solved by SSID software and the results compared with the values published in the issue 78 of ICRP. In all cases the system is in accordance with the values published by ICRP. (author)
V.S.O.P. (99/09) computer code system for reactor physics and fuel cycle simulation. Version 2009
Ruetten, H.J.; Haas, K.A.; Brockmann, H.; Ohlig, U.; Pohl, C.; Scherer, W.
2010-07-15
V.S.O.P. (99/ 09) represents the further development of V.S.O.P. (99/ 05). Compared to its precursor, the code system has been improved again in many details. The main motivation for this new code version was to update the basic nuclear libraries used by the code system. Thus, all cross section libraries involved in the code have now been based on ENDF/B-VII. V.S.O.P. is a computer code system for the comprehensive numerical simulation of the physics of thermal reactors. It implies the setup of the reactor and of the fuel element, processing of cross sections, neutron spectrum evaluation, neutron diffusion calculation in two or three dimensions, fuel burnup, fuel shuffling, reactor control, thermal hydraulics and fuel cycle costs. The thermal hydraulics part (steady state and time-dependent) is restricted to gas-cooled reactors and to two spatial dimensions. The code can simulate the reactor operation from the initial core towards the equilibrium core. This latest code version was developed and tested under the WINDOWS-XP - operating system. (orig.)
TRIM: A finite-volume MHD algorithm for an unstructured adaptive mesh
Schnack, D.D.; Lottati, I.; Mikic, Z. [Science Applications International Corp., San Diego, CA (United States)] [and others
1995-07-01
The authors describe TRIM, a MHD code which uses finite volume discretization of the MHD equations on an unstructured adaptive grid of triangles in the poloidal plane. They apply it to problems related to modeling tokamak toroidal plasmas. The toroidal direction is treated by a pseudospectral method. Care was taken to center variables appropriately on the mesh and to construct a self adjoint diffusion operator for cell centered variables.
Zizin, M. N.; Zimin, V. G.; Zizina, S. N.; Kryakvin, L. V.; Pitilimov, V. A.; Tereshonok, V. A.
2010-12-01
The ShIPR intellectual code system for mathematical simulation of nuclear reactors includes a set of computing modules implementing the preparation of macro cross sections on the basis of the two-group library of neutron-physics cross sections obtained for the SKETCH-N nodal code. This library is created by using the UNK code for 3D diffusion computation of first VVER-1000 fuel loadings. Computation of neutron fields in the ShIPR system is performed using the DP3 code in the two-group diffusion approximation in 3D triangular geometry. The efficiency of all groups of control rods for the first fuel loading of the third unit of the Kalinin Nuclear Power Plant is computed. The temperature, barometric, and density effects of reactivity as well as the reactivity coefficient due to the concentration of boric acid in the reactor were computed additionally. Results of computations are compared with the experiment.
李莉; 刘悦; 许欣洋; 夏新念
2012-01-01
A cylindrical model of linear MHD instabilities in tokamaks is presented. In the model, the cylindrical plasma is surrounded by a vacuum which is divided into inner and outer vacuum areas by a conducting wall. Linearized resistivity MHD equations with plasma viscosity are adopted to describe our model, and the equations are solved numerically as an initial value problem. Some of the results are used as benchmark tests for the code, and then a series of equilibrium current profiles are used to simulate the bootstrap current profiles in actual experiments with a bump on tail. Thus the effects of these kinds of profiles on MHD instabilities in tokamaks are revealed. From the analysis of the numerical results, it is found that more plasma can be confined when the center of the current bump is closer to the plasma surface, and a higher and narrower current bump has a better stabilizing effect on the MHD instabilities.
Ball, J.; Glowa, G.; Wren, J. [Atomic Energy of Canada Limited, Chalk River, Ontario (Canada); Ewig, F. [GRS Koln (Germany); Dickenson, S. [AEAT, (United Kingdom); Billarand, Y.; Cantrel, L. [IPSN (France); Rydl, A. [NRIR (Czech Republic); Royen, J. [OECD/NEA (France)
2001-11-01
This report describes the results of the second phase of International Standard Problem (ISP) 41, an iodine behaviour code comparison exercise. The first phase of the study, which was based on a simple Radioiodine Test Facility (RTF) experiment, demonstrated that all of the iodine behaviour codes had the capability to reproduce iodine behaviour for a narrow range of conditions (single temperature, no organic impurities, controlled pH steps). The current phase, a parametric study, was designed to evaluate the sensitivity of iodine behaviour codes to boundary conditions such as pH, dose rate, temperature and initial I{sup -} concentration. The codes used in this exercise were IODE(IPSN), IODE(NRIR), IMPAIR(GRS), INSPECT(AEAT), IMOD(AECL) and LIRIC(AECL). The parametric study described in this report identified several areas of discrepancy between the various codes. In general, the codes agree regarding qualitative trends, but their predictions regarding the actual amount of volatile iodine varied considerably. The largest source of the discrepancies between code predictions appears to be their different approaches to modelling the formation and destruction of organic iodides. A recommendation arising from this exercise is that an additional code comparison exercise be performed on organic iodide formation, against data obtained front intermediate-scale studies (two RTF (AECL, Canada) and two CAIMAN facility, (IPSN, France) experiments have been chosen). This comparison will allow each of the code users to realistically evaluate and improve the organic iodide behaviour sub-models within their codes. (author)
Ikushima, Takeshi [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment
1998-05-01
A computer code system CASKET (CASK thermal and structural analyses and Evaluation code system) for the thermal and structural analyses which are indispensable for radioactive material transport and/or storage cask designs has been developed. The CASKET is a simplified computer code system to perform parametric analyses on sensitivity evaluations in designing a cask and conducting its safety analysis. Main features of the CASKET are as follow: (1) it is capable to perform impact analysis of casks with shock absorbers, (2) it is capable to perform impact analysis of casks with fins. (3) puncture analysis of casks is capable, (4) rocking analysis of casks during seismic load is capable, (5) material property data library are provided for impact analysis of casks, (6) material property data library are provided for thermal analysis of casks, (7) fin energy absorption data library are provided for impact analysis of casks with fins are and (8) not only main frame computers (OS MSP) but also work stations (OS UNIX) and personal computers (OS Windows 3.1) are available. In the paper, brief illustrations of calculation methods are presented. Some calculation results are compared with experimental ones to confirm the computer programs are useful for thermal and structural analyses. (author)
Ikushima, Takeshi [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment
1998-05-01
A computer code system CASKET (CASK thermal and structural analyses and Evaluation code system) for the thermal and structural analyses which are indispensable for radioactive material transport and/or storage cask designs has been developed. The CASKET is a simplified computer code system to perform parametric analyses on sensitivity evaluations in designing a cask and conducting its safety analysis. Main features of the CASKET are as follow: (1) it is capable to perform impact analysis of casks with shock absorbers, (2) it is capable to perform impact analysis of casks with fins. (3) puncture analysis of casks is capable, (4) rocking analysis of casks during seismic load is capable, (5) material property data library are provided for impact analysis of casks, (6) material property data library are provided for thermal analysis of casks, (7) fin energy absorption data library are provided for impact analysis of casks with fins are and (8) not only main frame computers (OS MSP) but also work stations (OS UNIX) and personal computers (OS Windows 3.1) are available. In the paper, brief illustrations of calculation methods are presented. Some calculation results are compared with experimental ones to confirm the computer programs are useful for thermal and structural analyses. (author)
Weber, Sebastian; Austregesilo, Henrique; Bals, Christine; Band, Sebastian; Hollands, Thorsten; Koellein, Carsten; Lovasz, Liviusz; Pandazis, Peter; Schubert, Johann-Dietrich; Sonnenkalb, Martin
2016-10-15
In the framework of the reactor safety research program sponsored by the German Federal Ministry for Economic Affairs and Energy (BMWi), the computer code system ATHLET/ATHLET-CD has been further developed as an analysis tool for the simulation of accidents in nuclear power plants with pressurized and boiling water reactors as well as for the evaluation of accident management procedures. The main objective was to provide a mechanistic analysis tool for best estimate calculations of transients, accidents, and severe accidents with core degradation in light water reactors. With the continued development, the capability of the code system has been largely improved, allowing best estimate calculations of design and beyond design base accidents, and the simulation of advanced core degradation with enhanced model extent in a reasonable calculation time. ATHLET comprises inter alia a 6-equation model, models for the simulation of non-condensable gases and tracking of boron concentration, as well as additional component and process models for the complete system simulation. Among numerous model improvements, the code application has been extended to super critical pressures. The mechanistic description of the dynamic development of flow regimes on the basis of a transport equation for the interface area has been further developed. This ATHLET version is completely integrated in ATHLET-CD. ATHLET-CD further comprises dedicated models for the simulation of fuel and control assembly degradation for both pressurized and boiling water reactors, debris bed with melting in the core region, as well as fission product and aerosol release and transport in the cooling system, inclusive of decay of nuclide inventories and of chemical reactions in the gas phase. The continued development also concerned the modelling of absorber material release, of melting, melt relocation and freezing, and the interaction with the wall of the reactor pressure vessel. The following models were newly
Module description of TOKAMAK equilibrium code MEUDAS
Suzuki, Masaei; Hayashi, Nobuhiko; Matsumoto, Taro; Ozeki, Takahisa [Japan Atomic Energy Research Inst., Naka, Ibaraki (Japan). Naka Fusion Research Establishment
2002-01-01
The analysis of an axisymmetric MHD equilibrium serves as a foundation of TOKAMAK researches, such as a design of devices and theoretical research, the analysis of experiment result. For this reason, also in JAERI, an efficient MHD analysis code has been developed from start of TOKAMAK research. The free boundary equilibrium code ''MEUDAS'' which uses both the DCR method (Double-Cyclic-Reduction Method) and a Green's function can specify the pressure and the current distribution arbitrarily, and has been applied to the analysis of a broad physical subject as a code having rapidity and high precision. Also the MHD convergence calculation technique in ''MEUDAS'' has been built into various newly developed codes. This report explains in detail each module in ''MEUDAS'' for performing convergence calculation in solving the MHD equilibrium. (author)
MHD natural convection in open inclined square cavity with a heated circular cylinder
Hosain, Sheikh Anwar; Alim, M. A.; Saha, Satrajit Kumar
2017-06-01
MHD natural convection in open cavity becomes very important in many scientific and engineering problems, because of it's application in the design of electronic devices, solar thermal receivers, uncovered flat plate solar collectors having rows of vertical strips, geothermal reservoirs, etc. Several experiments and numerical investigations have been presented for describing the phenomenon of natural convection in open cavity for two decades. MHD natural convection and fluid flow in a two-dimensional open inclined square cavity with a heated circular cylinder was considered. The opposite wall to the opening side of the cavity was first kept to constant heat flux q, at the same time the surrounding fluid interacting with the aperture was maintained to an ambient temperature T∞. The top and bottom wall was kept to low and high temperature respectively. The fluid with different Prandtl numbers. The properties of the fluid are assumed to be constant. As a result a buoyancy force is created inside the cavity due to temperature difference and natural convection is formed inside the cavity. The Computational Fluid Dynamics (CFD) code are used to discretize the solution domain and represent the numerical result to graphical form.. Triangular meshes are used to obtain the solution of the problem. The streamlines and isotherms are produced, heat transfer parameter Nu are obtained. The results are presented in graphical as well as tabular form. The results show that heat flux decreases for increasing inclination of the cavity and the heat flux is a increasing function of Prandtl number Pr and decreasing function of Hartmann number Ha. It is observed that fluid moves counterclockwise around the cylinder in the cavity. Various recirculations are formed around the cylinder. The almost all isotherm lines are concentrated at the right lower corner of the cavity. The object of this work is to develop a Mathematical model regarding the effect of MHD natural convection flow around
MHD Integrated Topping Cycle Project
1992-03-01
The Magnetohydrodynamics (MHD) Integrated Topping Cycle (ITC) Project represents the culmination of the proof-of-concept (POC) development stage in the US Department of Energy (DOE) program to advance MHD technology to early commercial development stage utility power applications. The project is a joint effort, combining the skills of three topping cycle component developers: TRW, Avco/TDS, and Westinghouse. TRW, the prime contractor and system integrator, is responsible for the 50 thermal megawatt (50 MW{sub t}) slagging coal combustion subsystem. Avco/TDS is responsible for the MHD channel subsystem (nozzle, channel, diffuser, and power conditioning circuits), and Westinghouse is responsible for the current consolidation subsystem. The ITC Project will advance the state-of-the-art in MHD power systems with the design, construction, and integrated testing of 50 MW{sub t} power train components which are prototypical of the equipment that will be used in an early commercial scale MHD utility retrofit. Long duration testing of the integrated power train at the Component Development and Integration Facility (CDIF) in Butte, Montana will be performed, so that by the early 1990's, an engineering data base on the reliability, availability, maintainability and performance of the system will be available to allow scaleup of the prototypical designs to the next development level. This Sixteenth Quarterly Technical Progress Report covers the period May 1, 1991 to July 31, 1991.
MHD Integrated Topping Cycle Project
1992-03-01
The Magnetohydrodynamics (MHD) Integrated Topping Cycle (ITC) Project represents the culmination of the proof-of-concept (POC) development stage in the US Department of Energy (DOE) program to advance MHD technology to early commercial development stage utility power applications. The project is a joint effort, combining the skills of three topping cycle component developers: TRW, Avco/TDS, and Westinghouse. TRW, the prime contractor and system integrator, is responsible for the 50 thermal megawatt (50 MW{sub t}) slagging coal combustion subsystem. Avco/TDS is responsible for the MHD channel subsystem (nozzle, channel, diffuser, and power conditioning circuits), and Westinghouse is responsible for the current consolidation subsystem. The ITC Project will advance the state-of-the-art in MHD power systems with the design, construction, and integrated testing of 50 MW{sub t} power train components which are prototypical of the equipment that will be used in an early commercial scale MHD utility retrofit. Long duration testing of the integrated power train at the Component Development and Integration Facility (CDIF) in Butte, Montana will be performed, so that by the early 1990's, an engineering data base on the reliability, availability, maintainability and performance of the system will be available to allow scaleup of the prototypical designs to the next development level. This Sixteenth Quarterly Technical Progress Report covers the period May 1, 1991 to July 31, 1991.
Computing ITG turbulence with a full-f semi-Lagrangian code
Grandgirard, V.; Sarazin, Y.; Garbet, X.; Dif-Pradalier, G.; Ghendrih, Ph. [CEA Cadarache, CEA/DSM/DRFC, Assoc Euratom CEA, Cadarache (France); Crouseilles, N.; Latu, G.; Sonnendruecker, E. [Univ Strasbourg 1, IRMA, F-67084 Strasbourg (France); Besse, N.; Bertrand, P. [Univ Nancy 1, LPMIA, BP 239, F-54506 Vandoeuvre Les Nancy (France)
2008-07-01
This paper addresses non-linear global gyrokinetic simulations of ion temperature gradient (ITG) driven turbulence with the GYSELA code. The particularity of GYSELA code is to use a semi-Lagrangian (SL) scheme for the full distribution function. The 4D non-linear drift-kinetic version of the code already shows the interest in such a SL method which exhibits good properties of energy conservation. The code has been upgraded to run 5D toroidal simulations. Linear benchmarks and non-linear results are presented. (authors)
Computing ITG turbulence with a full- f semi-Lagrangian code
Grandgirard, V.; Sarazin, Y.; Garbet, X.; Dif-Pradalier, G.; Ghendrih, Ph.; Crouseilles, N.; Latu, G.; Sonnendrücker, E.; Besse, N.; Bertrand, P.
2008-02-01
This paper addresses non-linear global gyrokinetic simulations of ion temperature gradient (ITG) driven turbulence with the G YSELA code. The particularity of G YSELA code is to use a semi-Lagrangian (SL) scheme for the full distribution function. The 4D non-linear drift-kinetic version of the code already shows the interest in such a SL method which exhibits good properties of energy conservation. The code has been upgrated to run 5D toroidal simulations. Linear benchmarks and non-linear results are presented.
Shershnev, Anton A.; Kudryavtsev, Alexey N.; Kashkovsky, Alexander V.; Khotyanovsky, Dmitry V.
2016-10-01
The present paper describes HyCFS code, developed for numerical simulation of compressible high-speed flows on hybrid CPU/GPU (Central Processing Unit / Graphical Processing Unit) computational clusters on the basis of full unsteady Navier-Stokes equations, using modern shock capturing high-order TVD (Total Variation Diminishing) and WENO (Weighted Essentially Non-Oscillatory) schemes on general curvilinear structured grids. We discuss the specific features of hybrid architecture and details of program implementation and present the results of code verification.
潘卫; 陈燎原; 董家齐; 沈勇; 张锦华
2009-01-01
将理想全二维磁流体(MHD)稳定性数值代码GATO成功移植到HL-2A高性能计算系统,并用其模拟研究了HL-2A装置一次典型弹丸注入实验的等离子体的理想MHD稳定性.利用HL-2A装置第4050次放电的实验数据,应用EFIT代码重建了几个时间片的平衡位形,然后应用GATO代码对每个平衡位形的MHD稳定性进行了计算.通过对结果进行分析比较得出初步结论,由于弹丸注入而造成的反剪切位形的致稳作用,可以提高等离子体的约束性能.
Population coding and decoding in a neural field: a computational study.
Wu, Si; Amari, Shun-Ichi; Nakahara, Hiroyuki
2002-05-01
This study uses a neural field model to investigate computational aspects of population coding and decoding when the stimulus is a single variable. A general prototype model for the encoding process is proposed, in which neural responses are correlated, with strength specified by a gaussian function of their difference in preferred stimuli. Based on the model, we study the effect of correlation on the Fisher information, compare the performances of three decoding methods that differ in the amount of encoding information being used, and investigate the implementation of the three methods by using a recurrent network. This study not only rediscovers main results in existing literatures in a unified way, but also reveals important new features, especially when the neural correlation is strong. As the neural correlation of firing becomes larger, the Fisher information decreases drastically. We confirm that as the width of correlation increases, the Fisher information saturates and no longer increases in proportion to the number of neurons. However, we prove that as the width increases further--wider than (sqrt)2 times the effective width of the turning function--the Fisher information increases again, and it increases without limit in proportion to the number of neurons. Furthermore, we clarify the asymptotic efficiency of the maximum likelihood inference (MLI) type of decoding methods for correlated neural signals. It shows that when the correlation covers a nonlocal range of population (excepting the uniform correlation and when the noise is extremely small), the MLI type of method, whose decoding error satisfies the Cauchy-type distribution, is not asymptotically efficient. This implies that the variance is no longer adequate to measure decoding accuracy.
Hu, Chia-Chang
2005-12-01
A novel space-time adaptive near-far robust code-synchronization array detector for asynchronous DS-CDMA systems is developed in this paper. There are the same basic requirements that are needed by the conventional matched filter of an asynchronous DS-CDMA system. For the real-time applicability, a computationally efficient architecture of the proposed detector is developed that is based on the concept of the multistage Wiener filter (MWF) of Goldstein and Reed. This multistage technique results in a self-synchronizing detection criterion that requires no inversion or eigendecomposition of a covariance matrix. As a consequence, this detector achieves a complexity that is only a linear function of the size of antenna array ([InlineEquation not available: see fulltext.]), the rank of the MWF ([InlineEquation not available: see fulltext.]), the system processing gain ([InlineEquation not available: see fulltext.]), and the number of samples in a chip interval ([InlineEquation not available: see fulltext.]), that is,[InlineEquation not available: see fulltext.]. The complexity of the equivalent detector based on the minimum mean-squared error (MMSE) or the subspace-based eigenstructure analysis is a function of[InlineEquation not available: see fulltext.]. Moreover, this multistage scheme provides a rapid adaptive convergence under limited observation-data support. Simulations are conducted to evaluate the performance and convergence behavior of the proposed detector with the size of the[InlineEquation not available: see fulltext.]-element antenna array, the amount of the[InlineEquation not available: see fulltext.]-sample support, and the rank of the[InlineEquation not available: see fulltext.]-stage MWF. The performance advantage of the proposed detector over other DS-CDMA detectors is investigated as well.
Chia-Chang Hu
2005-04-01
Full Text Available A novel space-time adaptive near-far robust code-synchronization array detector for asynchronous DS-CDMA systems is developed in this paper. There are the same basic requirements that are needed by the conventional matched filter of an asynchronous DS-CDMA system. For the real-time applicability, a computationally efficient architecture of the proposed detector is developed that is based on the concept of the multistage Wiener filter (MWF of Goldstein and Reed. This multistage technique results in a self-synchronizing detection criterion that requires no inversion or eigendecomposition of a covariance matrix. As a consequence, this detector achieves a complexity that is only a linear function of the size of antenna array (J, the rank of the MWF (M, the system processing gain (N, and the number of samples in a chip interval (S, that is, Ã°ÂÂ’Âª(JMNS. The complexity of the equivalent detector based on the minimum mean-squared error (MMSE or the subspace-based eigenstructure analysis is a function of Ã°ÂÂ’Âª((JNS3. Moreover, this multistage scheme provides a rapid adaptive convergence under limited observation-data support. Simulations are conducted to evaluate the performance and convergence behavior of the proposed detector with the size of the J-element antenna array, the amount of the L-sample support, and the rank of the M-stage MWF. The performance advantage of the proposed detector over other DS-CDMA detectors is investigated as well.
Comparison of AMOS computer code wakefield real part impedances with analytic results
Mayhall, D J; Nelson, S D
2000-11-30
We have performed eleven AMOS (Azimuthal Mode Simulator)[1] code runs with a simple, right circular cylindrical accelerating cavity inserted into a circular, cylindrical, lossless beam pipe to calculate the real part of the n = 1(dipole) transverse wakefield impedance of this structure. We have compared this wakefield impedance in units of ohms/m(Wm) over the frequency range of 0-1 GHz to analytic predictions from Equation (2.3.8) of Briggs et al[2]. The results from Equation (2.3.8) were converted from the CGS units of statohms to the MKS units of ohms({Omega}) and then multiplied by (2{pi}f)/c = {Omega}/c = 2{pi}/{lambda}, where f is the frequency in Hz, c is the speed of light in vacuum in m/sec, {omega} is the angular frequency in radians/sec, and {lambda} is the wavelength in m. The dipole transverse wakefield impedance written to file from AMOS must be multiplied by c/o to convert it from units of {Omega}/m to units of {Omega}. The agreement between the AMOS runs and the analytic predictions are excellent for computational grids with square cells (dz = dr) and good for grids with rectangular cells (dz < dr). The quantity dz is the fixed-size axial grid spacing, and dr is the fixed-size radial grid spacing. We have also performed one AMOS run for the same geometry to calculate the real part of the n = 0(monopole) longitudinal wakefield impedance of this structure. We have compared this wakefield impedance in units of {Omega} with analytic predictions from Equation (1.4.8) of Briggs et al[1] converted to the MKS units of {Omega}. The agreement between the two results is excellent in this case. For the monopole longitudinal wakefield impedance written to file from AMOS, nothing must be done to convert the results to units of {Omega}. In each case, the computer calculations were carried out to 50 nsec of simulation time.
Bandini, B.R. [Los Alamos National Lab., NM (United States)
1990-05-01
No present light water reactor accident analysis code employs both high state of the art neutronics and thermal-hydraulics computational algorithms. Adding a modern three-dimensional neutron kinetics model to the present TRAC-PFI/MOD2 code would create a fully up to date pressurized water reactor accident evaluation code. After reviewing several options, it was decided that the Nodal Expansion Method would best provide the basis for this multidimensional transient neutronic analysis capability. Steady-state and transient versions of the Nodal Expansion Method were coded in both three-dimensional Cartesian and cylindrical geometries. In stand-alone form this method of solving the few group neutron diffusion equations was shown to yield efficient and accurate results for a variety of steady-state and transient benchmark problems. The Nodal Expansion Method was then incorporated into TRAC-PFl/MOD2. The combined NEM/TRAC code results agreed well with the EPRI-ARROTTA core-only transient analysis code when modelling a severe PWR control rod ejection accident.
[Is the Furness-Moore Code applicable for computer and telex?].
Rötzscher, K
1979-12-01
The extent to which the identification of disaster victims could be improved with the code for dental findings developed by Furness and Moore (1969) was studied. This code records the most important data for a set of teeth in 12 digits.
On the performance of a 2D unstructured computational rheology code on a GPU
Pereira, S.P.; Vuik, K.; Pinho, F.T.; Nobrega, J.M.
2013-01-01
The present work explores the massively parallel capabilities of the most advanced architecture of graphics processing units (GPUs) code named “Fermi”, on a two-dimensional unstructured cell-centred finite volume code. We use the SIMPLE algorithm to solve the continuity and momentum equations that