WorldWideScience

Sample records for particle mesh simulations

  1. A software framework for the portable parallelization of particle-mesh simulations

    DEFF Research Database (Denmark)

    Sbalzarini, I.F.; Walther, Jens Honore; Polasek, B.

    2006-01-01

    Abstract: We present a software framework for the transparent and portable parallelization of simulations using particle-mesh methods. Particles are used to transport physical properties and a mesh is required in order to reinitialize the distorted particle locations, ensuring the convergence...

  2. Value for money in particle-mesh plasma simulations

    International Nuclear Information System (INIS)

    Eastwood, J.W.

    1976-01-01

    The established particle-mesh method of simulating a collisionless plasma is discussed. Problems are outlined, and it is stated that given constraints on mesh size and particle number, the only way to adjust the compromise between dispersive forces, collision time and heating time is by altering the force calculating cycle. In 'value for money', schemes, matching of parts of the force calculation cycle is optimized. Interparticle forces are considered. Optimized combinations of elements of the force calculation cycle are compared. Following sections cover the dispersion relation, and comparisons with other schemes. (U.K.)

  3. Simulation study for high resolution alpha particle spectrometry with mesh type collimator

    International Nuclear Information System (INIS)

    Park, Seunghoon; Kwak, Sungwoo; Kang, Hanbyeol; Shin, Jungki; Park, Iljin

    2014-01-01

    An alpha particle spectrometry with a mesh type collimator plays a crucial role in identifying specific radionuclide in a radioactive source collected from the atmosphere or environment. The energy resolution is degraded without collimation because particles with a high angle have a longer path to travel in the air. Therefore, collision with the background increases. The collimator can cut out particles which traveling at a high angle. As a result, an energy distribution with high resolution can be obtained. Therefore, the mesh type collimator is simulated for high resolution alpha particle spectrometry. In conclusion, the collimator can improve resolution. With collimator, the collimator is a role of cutting out particles with a high angle, so, low energy tail and broadened energy distribution can be reduced. The mesh diameter is found out as an important factor to control resolution and counting efficiency. Therefore, a target particle, for example, 235 U, can be distinguished by a detector with a collimator under a mixture of various nuclides, for example: 232 U, 238 U, and 232 Th

  4. A regularized vortex-particle mesh method for large eddy simulation

    DEFF Research Database (Denmark)

    Spietz, Henrik Juul; Walther, Jens Honore; Hejlesen, Mads Mølholm

    We present recent developments of the remeshed vortex particle-mesh method for simulating incompressible fluid flow. The presented method relies on a parallel higher-order FFT based solver for the Poisson equation. Arbitrary high order is achieved through regularization of singular Green’s function...... solutions to the Poisson equation and recently we have derived novel high order solutions for a mixture of open and periodic domains. With this approach the simulated variables may formally be viewed as the approximate solution to the filtered Navier Stokes equations, hence we use the method for Large Eddy...

  5. An expert system for automatic mesh generation for Sn particle transport simulation in parallel environment

    International Nuclear Information System (INIS)

    Apisit, Patchimpattapong; Alireza, Haghighat; Shedlock, D.

    2003-01-01

    An expert system for generating an effective mesh distribution for the SN particle transport simulation has been developed. This expert system consists of two main parts: 1) an algorithm for generating an effective mesh distribution in a serial environment, and 2) an algorithm for inference of an effective domain decomposition strategy for parallel computing. For the first part, the algorithm prepares an effective mesh distribution considering problem physics and the spatial differencing scheme. For the second part, the algorithm determines a parallel-performance-index (PPI), which is defined as the ratio of the granularity to the degree-of-coupling. The parallel-performance-index provides expected performance of an algorithm depending on computing environment and resources. A large index indicates a high granularity algorithm with relatively low coupling among processors. This expert system has been successfully tested within the PENTRAN (Parallel Environment Neutral-Particle Transport) code system for simulating real-life shielding problems. (authors)

  6. An expert system for automatic mesh generation for Sn particle transport simulation in parallel environment

    Energy Technology Data Exchange (ETDEWEB)

    Apisit, Patchimpattapong [Electricity Generating Authority of Thailand, Office of Corporate Planning, Bangkruai, Nonthaburi (Thailand); Alireza, Haghighat; Shedlock, D. [Florida Univ., Department of Nuclear and Radiological Engineering, Gainesville, FL (United States)

    2003-07-01

    An expert system for generating an effective mesh distribution for the SN particle transport simulation has been developed. This expert system consists of two main parts: 1) an algorithm for generating an effective mesh distribution in a serial environment, and 2) an algorithm for inference of an effective domain decomposition strategy for parallel computing. For the first part, the algorithm prepares an effective mesh distribution considering problem physics and the spatial differencing scheme. For the second part, the algorithm determines a parallel-performance-index (PPI), which is defined as the ratio of the granularity to the degree-of-coupling. The parallel-performance-index provides expected performance of an algorithm depending on computing environment and resources. A large index indicates a high granularity algorithm with relatively low coupling among processors. This expert system has been successfully tested within the PENTRAN (Parallel Environment Neutral-Particle Transport) code system for simulating real-life shielding problems. (authors)

  7. A regularized vortex-particle mesh method for large eddy simulation

    Science.gov (United States)

    Spietz, H. J.; Walther, J. H.; Hejlesen, M. M.

    2017-11-01

    We present recent developments of the remeshed vortex particle-mesh method for simulating incompressible fluid flow. The presented method relies on a parallel higher-order FFT based solver for the Poisson equation. Arbitrary high order is achieved through regularization of singular Green's function solutions to the Poisson equation and recently we have derived novel high order solutions for a mixture of open and periodic domains. With this approach the simulated variables may formally be viewed as the approximate solution to the filtered Navier Stokes equations, hence we use the method for Large Eddy Simulation by including a dynamic subfilter-scale model based on test-filters compatible with the aforementioned regularization functions. Further the subfilter-scale model uses Lagrangian averaging, which is a natural candidate in light of the Lagrangian nature of vortex particle methods. A multiresolution variation of the method is applied to simulate the benchmark problem of the flow past a square cylinder at Re = 22000 and the obtained results are compared to results from the literature.

  8. An immersed interface vortex particle-mesh solver

    Science.gov (United States)

    Marichal, Yves; Chatelain, Philippe; Winckelmans, Gregoire

    2014-11-01

    An immersed interface-enabled vortex particle-mesh (VPM) solver is presented for the simulation of 2-D incompressible viscous flows, in the framework of external aerodynamics. Considering the simulation of free vortical flows, such as wakes and jets, vortex particle-mesh methods already provide a valuable alternative to standard CFD methods, thanks to the interesting numerical properties arising from its Lagrangian nature. Yet, accounting for solid bodies remains challenging, despite the extensive research efforts that have been made for several decades. The present immersed interface approach aims at improving the consistency and the accuracy of one very common technique (based on Lighthill's model) for the enforcement of the no-slip condition at the wall in vortex methods. Targeting a sharp treatment of the wall calls for substantial modifications at all computational levels of the VPM solver. More specifically, the solution of the underlying Poisson equation, the computation of the diffusion term and the particle-mesh interpolation are adapted accordingly and the spatial accuracy is assessed. The immersed interface VPM solver is subsequently validated on the simulation of some challenging impulsively started flows, such as the flow past a cylinder and that past an airfoil. Research Fellow (PhD student) of the F.R.S.-FNRS of Belgium.

  9. Cell-centered particle weighting algorithm for PIC simulations in a non-uniform 2D axisymmetric mesh

    Science.gov (United States)

    Araki, Samuel J.; Wirz, Richard E.

    2014-09-01

    Standard area weighting methods for particle-in-cell simulations result in systematic errors on particle densities for a non-uniform mesh in cylindrical coordinates. These errors can be significantly reduced by using weighted cell volumes for density calculations. A detailed description on the corrected volume calculations and cell-centered weighting algorithm in a non-uniform mesh is provided. The simple formulas for the corrected volume can be used for any type of quadrilateral and/or triangular mesh in cylindrical coordinates. Density errors arising from the cell-centered weighting algorithm are computed for radial density profiles of uniform, linearly decreasing, and Bessel function in an adaptive Cartesian mesh and an unstructured mesh. For all the density profiles, it is shown that the weighting algorithm provides a significant improvement for density calculations. However, relatively large density errors may persist at outermost cells for monotonically decreasing density profiles. A further analysis has been performed to investigate the effect of the density errors in potential calculations, and it is shown that the error at the outermost cell does not propagate into the potential solution for the density profiles investigated.

  10. Monte Carlo charged-particle tracking and energy deposition on a Lagrangian mesh.

    Science.gov (United States)

    Yuan, J; Moses, G A; McKenty, P W

    2005-10-01

    A Monte Carlo algorithm for alpha particle tracking and energy deposition on a cylindrical computational mesh in a Lagrangian hydrodynamics code used for inertial confinement fusion (ICF) simulations is presented. The straight line approximation is used to follow propagation of "Monte Carlo particles" which represent collections of alpha particles generated from thermonuclear deuterium-tritium (DT) reactions. Energy deposition in the plasma is modeled by the continuous slowing down approximation. The scheme addresses various aspects arising in the coupling of Monte Carlo tracking with Lagrangian hydrodynamics; such as non-orthogonal severely distorted mesh cells, particle relocation on the moving mesh and particle relocation after rezoning. A comparison with the flux-limited multi-group diffusion transport method is presented for a polar direct drive target design for the National Ignition Facility. Simulations show the Monte Carlo transport method predicts about earlier ignition than predicted by the diffusion method, and generates higher hot spot temperature. Nearly linear speed-up is achieved for multi-processor parallel simulations.

  11. Mesh Refinement for Particle-In-Cell Plasma Simulations: Applications to - and benefits for - Heavy-Ion-Fusion

    International Nuclear Information System (INIS)

    Vay, J.-L.; Colella, P.; McCorquodale, P.; Van Straalen, B.; Friedman, A.; Grote, D.P.

    2002-01-01

    The numerical simulation of the driving beams in a heavy ion fusion power plant is a challenging task, and simulation of the power plant as a whole, or even of the driver, is not yet possible. Despite the rapid progress in computer power, past and anticipated, one must consider the use of the most advanced numerical techniques, if we are to reach our goal expeditiously. One of the difficulties of these simulations resides in the disparity of scales, in time and in space, which must be resolved. When these disparities are in distinctive zones of the simulation region, a method which has proven to be effective in other areas (e.g., fluid dynamics simulations) is the mesh refinement technique. They discuss the challenges posed by the implementation of this technique into plasma simulations (due to the presence of particles and electromagnetic waves). They will present the prospects for and projected benefits of its application to heavy ion fusion. In particular to the simulation of the ion source and the final beam propagation in the chamber. A collaboration project is under way at LBNL between the Applied Numerical Algorithms Group (ANAG) and the HIF group to couple the Adaptive Mesh Refinement (AMR) library (CHOMBO) developed by the ANAG group to the Particle-In-Cell accelerator code WARP developed by the HIF-VNL. They describe their progress and present their initial findings

  12. Model of Random Polygon Particles for Concrete and Mesh Automatic Subdivision

    Institute of Scientific and Technical Information of China (English)

    2001-01-01

    In order to study the constitutive behavior of concrete in mesoscopic level, a new method is proposed in this paper. This method uses random polygon particles to simulate full grading broken aggregates of concrete. Based on computational geometry, we carry out the automatic generation of the triangle finite element mesh for the model of random polygon particles of concrete. The finite element mesh generated in this paper is also applicable to many other numerical methods.

  13. Introducing a distributed unstructured mesh into gyrokinetic particle-in-cell code, XGC

    Science.gov (United States)

    Yoon, Eisung; Shephard, Mark; Seol, E. Seegyoung; Kalyanaraman, Kaushik

    2017-10-01

    XGC has shown good scalability for large leadership supercomputers. The current production version uses a copy of the entire unstructured finite element mesh on every MPI rank. Although an obvious scalability issue if the mesh sizes are to be dramatically increased, the current approach is also not optimal with respect to data locality of particles and mesh information. To address these issues we have initiated the development of a distributed mesh PIC method. This approach directly addresses the base scalability issue with respect to mesh size and, through the use of a mesh entity centric view of the particle mesh relationship, provides opportunities to address data locality needs of many core and GPU supported heterogeneous systems. The parallel mesh PIC capabilities are being built on the Parallel Unstructured Mesh Infrastructure (PUMI). The presentation will first overview the form of mesh distribution used and indicate the structures and functions used to support the mesh, the particles and their interaction. Attention will then focus on the node-level optimizations being carried out to ensure performant operation of all PIC operations on the distributed mesh. Partnership for Edge Physics Simulation (EPSI) Grant No. DE-SC0008449 and Center for Extended Magnetohydrodynamic Modeling (CEMM) Grant No. DE-SC0006618.

  14. Simulations of a single vortex ring using an unbounded, regularized particle-mesh based vortex method

    DEFF Research Database (Denmark)

    Hejlesen, Mads Mølholm; Spietz, Henrik J.; Walther, Jens Honore

    2014-01-01

    , unbounded particle-mesh based vortex method is used to simulate the instability, transition to turbulence and eventual destruction of a single vortex ring. From the simulation data a novel method on analyzing the dynamics of the enstrophy is presented based on the alignment of the vorticity vector...... with the principal axis of the strain rate tensor. We find that the dynamics of the enstrophy density is dominated by the local flow deformation and axis of rotation, which is used to infer some concrete tendencies related to the topology of the vorticity field....

  15. N-body simulations for f(R) gravity using a self-adaptive particle-mesh code

    International Nuclear Information System (INIS)

    Zhao Gongbo; Koyama, Kazuya; Li Baojiu

    2011-01-01

    We perform high-resolution N-body simulations for f(R) gravity based on a self-adaptive particle-mesh code MLAPM. The chameleon mechanism that recovers general relativity on small scales is fully taken into account by self-consistently solving the nonlinear equation for the scalar field. We independently confirm the previous simulation results, including the matter power spectrum, halo mass function, and density profiles, obtained by Oyaizu et al.[Phys. Rev. D 78, 123524 (2008)] and Schmidt et al.[Phys. Rev. D 79, 083518 (2009)], and extend the resolution up to k∼20 h/Mpc for the measurement of the matter power spectrum. Based on our simulation results, we discuss how the chameleon mechanism affects the clustering of dark matter and halos on full nonlinear scales.

  16. Multiphase flow modelling of volcanic ash particle settling in water using adaptive unstructured meshes

    Science.gov (United States)

    Jacobs, C. T.; Collins, G. S.; Piggott, M. D.; Kramer, S. C.; Wilson, C. R. G.

    2013-02-01

    Small-scale experiments of volcanic ash particle settling in water have demonstrated that ash particles can either settle slowly and individually, or rapidly and collectively as a gravitationally unstable ash-laden plume. This has important implications for the emplacement of tephra deposits on the seabed. Numerical modelling has the potential to extend the results of laboratory experiments to larger scales and explore the conditions under which plumes may form and persist, but many existing models are computationally restricted by the fixed mesh approaches that they employ. In contrast, this paper presents a new multiphase flow model that uses an adaptive unstructured mesh approach. As a simulation progresses, the mesh is optimized to focus numerical resolution in areas important to the dynamics and decrease it where it is not needed, thereby potentially reducing computational requirements. Model verification is performed using the method of manufactured solutions, which shows the correct solution convergence rates. Model validation and application considers 2-D simulations of plume formation in a water tank which replicate published laboratory experiments. The numerically predicted settling velocities for both individual particles and plumes, as well as instability behaviour, agree well with experimental data and observations. Plume settling is clearly hindered by the presence of a salinity gradient, and its influence must therefore be taken into account when considering particles in bodies of saline water. Furthermore, individual particles settle in the laminar flow regime while plume settling is shown (by plume Reynolds numbers greater than unity) to be in the turbulent flow regime, which has a significant impact on entrainment and settling rates. Mesh adaptivity maintains solution accuracy while providing a substantial reduction in computational requirements when compared to the same simulation performed using a fixed mesh, highlighting the benefits of an

  17. N-body simulations for f(R) gravity using a self-adaptive particle-mesh code

    Science.gov (United States)

    Zhao, Gong-Bo; Li, Baojiu; Koyama, Kazuya

    2011-02-01

    We perform high-resolution N-body simulations for f(R) gravity based on a self-adaptive particle-mesh code MLAPM. The chameleon mechanism that recovers general relativity on small scales is fully taken into account by self-consistently solving the nonlinear equation for the scalar field. We independently confirm the previous simulation results, including the matter power spectrum, halo mass function, and density profiles, obtained by Oyaizu [Phys. Rev. DPRVDAQ1550-7998 78, 123524 (2008)10.1103/PhysRevD.78.123524] and Schmidt [Phys. Rev. DPRVDAQ1550-7998 79, 083518 (2009)10.1103/PhysRevD.79.083518], and extend the resolution up to k˜20h/Mpc for the measurement of the matter power spectrum. Based on our simulation results, we discuss how the chameleon mechanism affects the clustering of dark matter and halos on full nonlinear scales.

  18. Application of particle-mesh Ewald summation to ONIOM theory

    International Nuclear Information System (INIS)

    Kobayashi, Osamu; Nanbu, Shinkoh

    2015-01-01

    Highlights: • Particle-mesh Ewald sum is extended to ONIOM scheme. • Non-adiabatic MD simulation in solution is performed. • The behavior of excited (Z)-penta-2,4-dieniminium cation in methanol is simulated. • The difference between gas phase and solution is predicted. - Abstract: We extended a particle mesh Ewald (PME) summation method to the ONIOM (our Own N-layered Integrated molecular Orbitals and molecular Mechanics) scheme (PME-ONIOM) to validate the simulation in solution. This took the form of a nonadiabatic ab initio molecular dynamics (MD) simulation in which the Zhu-Nakamura trajectory surface hopping (ZN-TSH) method was performed for the photoisomerization of a (Z)-penta-2,4-dieniminium cation (protonated Schiff base, PSB3) electronically excited to the S 1 state in a methanol solution. We also calculated a nonadiabatic ab initio MD simulation with only minimum image convention (MI-ONIOM). The lifetime determined by PME-ONIOM-MD was 3.483 ps. The MI-ONIOM-MD lifetime of 0.4642 ps was much shorter than those of PME-ONIOM-MD and the experimentally determined excited state lifetime. The difference eminently illustrated the accurate treatment of the long-range solvation effect, which destines the electronically excited PSB3 for staying in S 1 at the pico-second or the femto-second time scale.

  19. Charged particle tracking through electrostatic wire meshes using the finite element method

    Energy Technology Data Exchange (ETDEWEB)

    Devlin, L. J.; Karamyshev, O.; Welsch, C. P., E-mail: carsten.welsch@cockcroft.ac.uk [The Cockcroft Institute, Daresbury Laboratory, Warrington (United Kingdom); Department of Physics, University of Liverpool, Liverpool (United Kingdom)

    2016-06-15

    Wire meshes are used across many disciplines to accelerate and focus charged particles, however, analytical solutions are non-exact and few codes exist which simulate the exact fields around a mesh with physical sizes. A tracking code based in Matlab-Simulink using field maps generated using finite element software has been developed which tracks electrons or ions through electrostatic wire meshes. The fields around such a geometry are presented as an analytical expression using several basic assumptions, however, it is apparent that computational calculations are required to obtain realistic values of electric potential and fields, particularly when multiple wire meshes are deployed. The tracking code is flexible in that any quantitatively describable particle distribution can be used for both electrons and ions as well as other benefits such as ease of export to other programs for analysis. The code is made freely available and physical examples are highlighted where this code could be beneficial for different applications.

  20. Vortex particle-mesh simulations of vertical axis wind turbine flows: from the airfoil performance to the very far wake

    Directory of Open Access Journals (Sweden)

    P. Chatelain

    2017-06-01

    Full Text Available A vortex particle-mesh (VPM method with immersed lifting lines has been developed and validated. Based on the vorticity–velocity formulation of the Navier–Stokes equations, it combines the advantages of a particle method and of a mesh-based approach. The immersed lifting lines handle the creation of vorticity from the blade elements and its early development. Large-eddy simulation (LES of vertical axis wind turbine (VAWT flows is performed. The complex wake development is captured in detail and over up to 15 diameters downstream: from the blades to the near-wake coherent vortices and then through the transitional ones to the fully developed turbulent far wake (beyond 10 rotor diameters. The statistics and topology of the mean flow are studied. The computational sizes also allow insights into the detailed unsteady vortex dynamics and topological flow features, such as a recirculation region influenced by the tip speed ratio and the rotor geometry.

  1. Merging for Particle-Mesh Complex Particle Kinetic Modeling of the Multiple Plasma Beams

    Science.gov (United States)

    Lipatov, Alexander S.

    2011-01-01

    We suggest a merging procedure for the Particle-Mesh Complex Particle Kinetic (PMCPK) method in case of inter-penetrating flow (multiple plasma beams). We examine the standard particle-in-cell (PIC) and the PMCPK methods in the case of particle acceleration by shock surfing for a wide range of the control numerical parameters. The plasma dynamics is described by a hybrid (particle-ion-fluid-electron) model. Note that one may need a mesh if modeling with the computation of an electromagnetic field. Our calculations use specified, time-independent electromagnetic fields for the shock, rather than self-consistently generated fields. While a particle-mesh method is a well-verified approach, the CPK method seems to be a good approach for multiscale modeling that includes multiple regions with various particle/fluid plasma behavior. However, the CPK method is still in need of a verification for studying the basic plasma phenomena: particle heating and acceleration by collisionless shocks, magnetic field reconnection, beam dynamics, etc.

  2. Mesh refinement for particle-in-cell plasma simulations: Applications to - and benefits for HIF (heavy ion fusion)

    International Nuclear Information System (INIS)

    Vay, J.L.; Colella, P.; McCorquodale, P.; Van Straalen, B.; Friedman, A.; Grote, D.P.

    2002-01-01

    The numerical simulation of the driving beams in a heavy ion fusion power plant is a challenging task, and simulation of the power plant as a whole, or even of the drive,r is not yet possible. Despite the rapid progress in computer power, past and anticipated, one must consider the use of the most advanced numerical techniques, if they are to reach the goal expeditiously. One of the difficulties of these simulations resides in the disparity of scales, in time and in space, which must be resolved. When these disparities are in distinctive zones of the simulation region, a method which has proven to be effective in other areas (e.g., fluid dynamics simulations) is the mesh refinement technique. They discuss the challenges posed by the implementation of this technique into plasma simulations (due to the presence of particles and electromagnetic waves). They will present the prospects for and projected benefits of its application to heavy ion fusion, in particular to the simulation of the ion source and the final beam propagation in the chamber

  3. PowderSim: Lagrangian Discrete and Mesh-Free Continuum Simulation Code for Cohesive Soils

    Science.gov (United States)

    Johnson, Scott; Walton, Otis; Settgast, Randolph

    2013-01-01

    PowderSim is a calculation tool that combines a discrete-element method (DEM) module, including calibrated interparticle-interaction relationships, with a mesh-free, continuum, SPH (smoothed-particle hydrodynamics) based module that utilizes enhanced, calibrated, constitutive models capable of mimicking both large deformations and the flow behavior of regolith simulants and lunar regolith under conditions anticipated during in situ resource utilization (ISRU) operations. The major innovation introduced in PowderSim is to use a mesh-free method (SPH-based) with a calibrated and slightly modified critical-state soil mechanics constitutive model to extend the ability of the simulation tool to also address full-scale engineering systems in the continuum sense. The PowderSim software maintains the ability to address particle-scale problems, like size segregation, in selected regions with a traditional DEM module, which has improved contact physics and electrostatic interaction models.

  4. Progress in the study of mesh refinement for particle-in-cell plasma simulations and its application to heavy ion fusion

    International Nuclear Information System (INIS)

    Vay, J.-L.; Friedman, A.; Grote, D.P.

    2002-01-01

    The numerical simulation of the driving beams in a heavy ion fusion power plant is a challenging task, and, despite rapid progress in computer power, one must consider the use of the most advanced numerical techniques. One of the difficulties of these simulations resides in the disparity of scales in time and in space which must be resolved. When these disparities are in distinctive zones of the simulation region, a method which has proven to be effective in other areas (e.g. fluid dynamics simulations) is the Adaptive-Mesh-Refinement (AMR) technique. We follow in this article the progress accomplished in the last few months in the merging of the AMR technique with Particle-In-Cell (PIC) method. This includes a detailed modeling of the Lampel-Tiefenback solution for the one-dimensional diode using novel techniques to suppress undesirable numerical oscillations and an AMR patch to follow the head of the particle distribution. We also report new results concerning the modeling of ion sources using the axisymmetric WARPRZ-AMR prototype showing the utility of an AMR patch resolving the emitter vicinity and the beam edge

  5. Enriching Triangle Mesh Animations with Physically Based Simulation.

    Science.gov (United States)

    Li, Yijing; Xu, Hongyi; Barbic, Jernej

    2017-10-01

    We present a system to combine arbitrary triangle mesh animations with physically based Finite Element Method (FEM) simulation, enabling control over the combination both in space and time. The input is a triangle mesh animation obtained using any method, such as keyframed animation, character rigging, 3D scanning, or geometric shape modeling. The input may be non-physical, crude or even incomplete. The user provides weights, specified using a minimal user interface, for how much physically based simulation should be allowed to modify the animation in any region of the model, and in time. Our system then computes a physically-based animation that is constrained to the input animation to the amount prescribed by these weights. This permits smoothly turning physics on and off over space and time, making it possible for the output to strictly follow the input, to evolve purely based on physically based simulation, and anything in between. Achieving such results requires a careful combination of several system components. We propose and analyze these components, including proper automatic creation of simulation meshes (even for non-manifold and self-colliding undeformed triangle meshes), converting triangle mesh animations into animations of the simulation mesh, and resolving collisions and self-collisions while following the input.

  6. Mesh refinement of simulation with the AID riser transmission gamma

    International Nuclear Information System (INIS)

    Lima Filho, Hilario J.B. de; Benachour, Mohand; Dantas, Carlos C.; Brito, Marcio F.P.; Santos, Valdemir A. dos

    2013-01-01

    Type reactors Circulating Fluidized Bed (CFBR) vertical, in which the particulate and gaseous phases have flows upward (riser) have been widely used in gasification processes, combustion and fluid catalytic cracking (FCC). These biphasic reactors (gas-solid) efficiency depends largely on their hydrodynamic characteristics, and shows different behaviors in the axial and radial directions. The solids axial distribution is observed by the higher concentration in the base, getting more diluted toward the top. Radially, the solids concentration is characterized as core-annular, in which the central region is highly diluted, consisting of dispersed particles and fluid. In the present work developed a two-dimensional geometry (2D) techniques through simulations in computational fluid dynamics (CFD) to predict the gas-solid flow in the riser type CFBR through transient modeling, based on the kinetic theory of granular flow . The refinement of computational meshes provide larger amounts of information on the parameters studied, but may increase the processing time of the simulations. A minimum number of cells applied to the mesh construction was obtained by testing five meshes. The validation of the hydrodynamic parameters was performed using a range of 241Am source and detector NaI (Tl). The numerical results were provided consistent with the experimental data, indicating that the refined computational mesh in a controlled manner, improve the approximation of the expected results. (author)

  7. Parallel adaptive simulations on unstructured meshes

    International Nuclear Information System (INIS)

    Shephard, M S; Jansen, K E; Sahni, O; Diachin, L A

    2007-01-01

    This paper discusses methods being developed by the ITAPS center to support the execution of parallel adaptive simulations on unstructured meshes. The paper first outlines the ITAPS approach to the development of interoperable mesh, geometry and field services to support the needs of SciDAC application in these areas. The paper then demonstrates the ability of unstructured adaptive meshing methods built on such interoperable services to effectively solve important physics problems. Attention is then focused on ITAPs' developing ability to solve adaptive unstructured mesh problems on massively parallel computers

  8. Water Flow Simulation using Smoothed Particle Hydrodynamics (SPH)

    Science.gov (United States)

    Vu, Bruce; Berg, Jared; Harris, Michael F.

    2014-01-01

    Simulation of water flow from the rainbird nozzles has been accomplished using the Smoothed Particle Hydrodynamics (SPH). The advantage of using SPH is that no meshing is required, thus the grid quality is no longer an issue and accuracy can be improved.

  9. High-fidelity meshes from tissue samples for diffusion MRI simulations.

    Science.gov (United States)

    Panagiotaki, Eleftheria; Hall, Matt G; Zhang, Hui; Siow, Bernard; Lythgoe, Mark F; Alexander, Daniel C

    2010-01-01

    This paper presents a method for constructing detailed geometric models of tissue microstructure for synthesizing realistic diffusion MRI data. We construct three-dimensional mesh models from confocal microscopy image stacks using the marching cubes algorithm. Random-walk simulations within the resulting meshes provide synthetic diffusion MRI measurements. Experiments optimise simulation parameters and complexity of the meshes to achieve accuracy and reproducibility while minimizing computation time. Finally we assess the quality of the synthesized data from the mesh models by comparison with scanner data as well as synthetic data from simple geometric models and simplified meshes that vary only in two dimensions. The results support the extra complexity of the three-dimensional mesh compared to simpler models although sensitivity to the mesh resolution is quite robust.

  10. Adaptive and dynamic meshing methods for numerical simulations

    Science.gov (United States)

    Acikgoz, Nazmiye

    For the numerical simulation of many problems of engineering interest, it is desirable to have an automated mesh adaption tool capable of producing high quality meshes with an affordably low number of mesh points. This is important especially for problems, which are characterized by anisotropic features of the solution and require mesh clustering in the direction of high gradients. Another significant issue in meshing emerges in the area of unsteady simulations with moving boundaries or interfaces, where the motion of the boundary has to be accommodated by deforming the computational grid. Similarly, there exist problems where current mesh needs to be adapted to get more accurate solutions because either the high gradient regions are initially predicted inaccurately or they change location throughout the simulation. To solve these problems, we propose three novel procedures. For this purpose, in the first part of this work, we present an optimization procedure for three-dimensional anisotropic tetrahedral grids based on metric-driven h-adaptation. The desired anisotropy in the grid is dictated by a metric that defines the size, shape, and orientation of the grid elements throughout the computational domain. Through the use of topological and geometrical operators, the mesh is iteratively adapted until the final mesh minimizes a given objective function. In this work, the objective function measures the distance between the metric of each simplex and a target metric, which can be either user-defined (a-priori) or the result of a-posteriori error analysis. During the adaptation process, one tries to decrease the metric-based objective function until the final mesh is compliant with the target within a given tolerance. However, in regions such as corners and complex face intersections, the compliance condition was found to be very difficult or sometimes impossible to satisfy. In order to address this issue, we propose an optimization process based on an ad

  11. A multilevel particle method for gas dynamics: application to multi-fluids simulation

    International Nuclear Information System (INIS)

    Weynans, Lisl

    2006-12-01

    In inertial confinement fusion, laser implosions require to know hydrodynamic flow in presence of shocks. This work is devoted to the evaluation of the ability of a particle-mesh method, inspired from Vortex-In-Cell methods, to simulate gas dynamics, especially multi-fluids. First, we develop a particle method, associated with a conservative re-meshing step, which is performed with high order interpolating kernels. We study theoretically and numerically this method. This analysis gives evidence of a strong relationship between the particle method and high order Lax-Wendroff-like finite difference schemes. We introduce a new scheme for the advection of particles. Then we implement a multilevel technique, inspired from AMR, which allows us to increase locally the accuracy of the computations. Finally we develop a level set-like technique, discretized on the particles, to simulate the interface between compressible flows. We use the multilevel technique to improve the interface resolution and the conservation of partial masses. (author)

  12. Crack growth simulation for plural crack using hexahedral mesh generation technique

    International Nuclear Information System (INIS)

    Orita, Y; Wada, Y; Kikuchi, M

    2010-01-01

    This paper describes a surface crack growth simulation using a new mesh generation technique. The generated mesh is constituted of all hexahedral elements. Hexahedral elements are suitable for an analysis of fracture mechanics parameters, i.e. stress intensity factor. The advantage of a hexahedral mesh is good accuracy of an analysis and less number of degrees of freedoms than a tetrahedral mesh. In this study, a plural crack growth simulation is computed using the hexahedral mesh and its distribution of stress intensity factor is investigated.

  13. Vortex Particle-Mesh simulations of Vertical Axis Wind Turbine flows: from the blade aerodynamics to the very far wake

    Science.gov (United States)

    Chatelain, P.; Duponcheel, M.; Caprace, D.-G.; Marichal, Y.; Winckelmans, G.

    2016-09-01

    A Vortex Particle-Mesh (VPM) method with immersed lifting lines has been developed and validated. Based on the vorticity-velocity formulation of the Navier-Stokes equations, it combines the advantages of a particle method and of a mesh-based approach. The immersed lifting lines handle the creation of vorticity from the blade elements and its early development. LES of Vertical Axis Wind Turbine (VAWT) flows are performed. The complex wake development is captured in details and over very long distances: from the blades to the near wake coherent vortices, then through the transitional ones to the fully developed turbulent far wake (beyond 10 rotor diameters). The statistics and topology of the mean flow are studied. The computational sizes also allow insights into the detailed unsteady vortex dynamics, including some unexpected topological flow features.

  14. Vortex Particle-Mesh simulations of Vertical Axis Wind Turbine flows: from the blade aerodynamics to the very far wake

    International Nuclear Information System (INIS)

    Chatelain, P; Duponcheel, M; Caprace, D-G; Winckelmans, G; Marichal, Y

    2016-01-01

    A Vortex Particle-Mesh (VPM) method with immersed lifting lines has been developed and validated. Based on the vorticity-velocity formulation of the Navier-Stokes equations, it combines the advantages of a particle method and of a mesh-based approach. The immersed lifting lines handle the creation of vorticity from the blade elements and its early development. LES of Vertical Axis Wind Turbine (VAWT) flows are performed. The complex wake development is captured in details and over very long distances: from the blades to the near wake coherent vortices, then through the transitional ones to the fully developed turbulent far wake (beyond 10 rotor diameters). The statistics and topology of the mean flow are studied. The computational sizes also allow insights into the detailed unsteady vortex dynamics, including some unexpected topological flow features. (paper)

  15. Finite Element in Angle Unit Sphere Meshing for Charged Particle Transport.

    Energy Technology Data Exchange (ETDEWEB)

    Ortega, Mario Ivan [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Drumm, Clifton R. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2017-10-01

    Finite element in angle formulations of the charged particle transport equation require the discretization of the unit sphere. In Sceptre, a three-dimensional surface mesh of a sphere is transformed into a two-dimensional mesh. Projection of a sphere onto a two-dimensional surface is well studied with map makers spending the last few centuries attempting to create maps that preserve proportion and area. Using these techniques, various meshing schemes for the unit sphere were investigated.

  16. Finite element simulation of impact response of wire mesh screens

    Directory of Open Access Journals (Sweden)

    Wang Caizheng

    2015-01-01

    Full Text Available In this paper, the response of wire mesh screens to low velocity impact with blunt objects is investigated using finite element (FE simulation. The woven wire mesh is modelled with homogeneous shell elements with equivalent smeared mechanical properties. The mechanical behaviour of the woven wire mesh was determined experimentally with tensile tests on steel wire mesh coupons to generate the data for the smeared shell material used in the FE. The effects of impacts with a low mass (4 kg and a large mass (40 kg providing the same impact energy are studied. The joint between the wire mesh screen and the aluminium frame surrounding it is modelled using contact elements with friction between the corresponding elements. Damage to the screen of different types compromising its structural integrity, such as mesh separation and pulling out from the surrounding frame is modelled. The FE simulation is validated with results of impact tests conducted on woven steel wire screen meshes.

  17. Current achievements and future perspectives on particle simulation technologies for fluid dynamics and heat transfer

    International Nuclear Information System (INIS)

    Koshizuka, Seiichi

    2011-01-01

    The Moving Particle Semi-implicit (MPS) method is one of the particle methods in which continuum mechanics is analyzed using the concept of particles. Since meshes are not used, large deformation of free surfaces and material interfaces can be simulated without the problems of mesh distortion. Thus, the MPS method has been applied to multiphase flow analysis in nuclear engineering. The advantages of the particle methods are also useful for applications in other engineering fields: ship engineering, civil engineering, microflow, biomechanics, visualization, etc. In this review, calculation examples are described and classified. Commercial codes have been released and applied in industries. The particle methods are also used in TV programs, movies, and computer games. Combinations of numerical techniques for multiphysics problems, fast calculations, and high-quality visualizations are expected to lead to real-time particle simulations for various new applications in the near future. (author)

  18. Particle simulation of 3D galactic hydrodynamics on the ICL DAP

    International Nuclear Information System (INIS)

    Johns, T.C.; Nelson, A.H.

    1985-01-01

    A non-self-gravitating galactic hydrodynamics code based on a quasi-particle technique and making use of a mesh for force evaluation and sorting purposes is described. The short-range nature of the interparticle pressure forces, coupled with the use of a mesh allows a particularly fast algorithm. The 3D representation of the galaxy is mapped onto the ''3D'' main store of ICL DAP in a natural way, the 2 spatial dimensions in the plane of the galaxy becoming the 2 dimensions of the processor plane on the DAP and the third dimension varying within individual processor storage elements. This leads to a fairly straightforward implementation and a high degree of parallelism in the crucial parts of the code. The particle shuffling which is necessary after each timestep is facilitated by the use of a parallel variant of the bitonic sorting algorithm. Some results of simulations using a 63x63x16 mesh and about 50,000 particles to follow the evolution of a model disk galaxy are presented

  19. Impact of Variable-Resolution Meshes on Regional Climate Simulations

    Science.gov (United States)

    Fowler, L. D.; Skamarock, W. C.; Bruyere, C. L.

    2014-12-01

    The Model for Prediction Across Scales (MPAS) is currently being used for seasonal-scale simulations on globally-uniform and regionally-refined meshes. Our ongoing research aims at analyzing simulations of tropical convective activity and tropical cyclone development during one hurricane season over the North Atlantic Ocean, contrasting statistics obtained with a variable-resolution mesh against those obtained with a quasi-uniform mesh. Analyses focus on the spatial distribution, frequency, and intensity of convective and grid-scale precipitations, and their relative contributions to the total precipitation as a function of the horizontal scale. Multi-month simulations initialized on May 1st 2005 using ERA-Interim re-analyses indicate that MPAS performs satisfactorily as a regional climate model for different combinations of horizontal resolutions and transitions between the coarse and refined meshes. Results highlight seamless transitions for convection, cloud microphysics, radiation, and land-surface processes between the quasi-uniform and locally- refined meshes, despite the fact that the physics parameterizations were not developed for variable resolution meshes. Our goal of analyzing the performance of MPAS is twofold. First, we want to establish that MPAS can be successfully used as a regional climate model, bypassing the need for nesting and nudging techniques at the edges of the computational domain as done in traditional regional climate modeling. Second, we want to assess the performance of our convective and cloud microphysics parameterizations as the horizontal resolution varies between the lower-resolution quasi-uniform and higher-resolution locally-refined areas of the global domain.

  20. Numerical analysis of splashing fluid using hybrid method of mesh-based and particle-based modelings

    International Nuclear Information System (INIS)

    Tanaka, Nobuatsu; Ogawara, Takuya; Kaneda, Takeshi; Maseguchi, Ryo

    2009-01-01

    In order to simulate splashing and scattering fluid behaviors, we developed a hybrid method of mesh-based model for large-scale continuum fluid and particle-based model for small-scale discrete fluid particles. As for the solver of the continuum fluid, we adopt the CIVA RefIned Multiphase SimulatiON (CRIMSON) code to evaluate two phase flow behaviors based on the recent computational fluid dynamics (CFD) techniques. The phase field model has been introduced to the CRIMSON in order to solve the problem of loosing phase interface sharpness in long-term calculation. As for the solver of the discrete fluid droplets, we applied the idea of Smoothed Particle Hydrodynamics (SPH) method. Both continuum fluid and discrete fluid interact each other through drag interaction force. We verified our method by applying it to a popular benchmark problem of collapse of water column problems, especially focusing on the splashing and scattering fluid behaviors after the column collided against the wall. We confirmed that the gross splashing and scattering behaviors were well reproduced by the introduction of particle model while the detailed behaviors of the particles were slightly different from the experimental results. (author)

  1. Influence of mesh non-orthogonality on numerical simulation of buoyant jet flows

    International Nuclear Information System (INIS)

    Ishigaki, Masahiro; Abe, Satoshi; Sibamoto, Yasuteru; Yonomoto, Taisuke

    2017-01-01

    Highlights: • Influence of mesh non-orthogonality on numerical solution of buoyant jet flows. • Buoyant jet flows are simulated with hexahedral and prismatic meshes. • Jet instability with prismatic meshes may be overestimated compared to that with hexahedral meshes. • Modified solvers that can reduce the influence of mesh non-orthogonality and reduce computation time are proposed. - Abstract: In the present research, we discuss the influence of mesh non-orthogonality on numerical solution of a type of buoyant flow. Buoyant jet flows are simulated numerically with hexahedral and prismatic mesh elements in an open source Computational Fluid Dynamics (CFD) code called “OpenFOAM”. Buoyant jet instability obtained with the prismatic meshes may be overestimated compared to that obtained with the hexahedral meshes when non-orthogonal correction is not applied in the code. Although the non-orthogonal correction method can improve the instability generated by mesh non-orthogonality, it may increase computation time required to reach a convergent solution. Thus, we propose modified solvers that can reduce the influence of mesh non-orthogonality and reduce the computation time compared to the existing solvers in OpenFOAM. It is demonstrated that calculations for a buoyant jet with a large temperature difference are performed faster by the modified solver.

  2. Influence of mesh non-orthogonality on numerical simulation of buoyant jet flows

    Energy Technology Data Exchange (ETDEWEB)

    Ishigaki, Masahiro, E-mail: ishigaki.masahiro@jaea.go.jp; Abe, Satoshi; Sibamoto, Yasuteru; Yonomoto, Taisuke

    2017-04-01

    Highlights: • Influence of mesh non-orthogonality on numerical solution of buoyant jet flows. • Buoyant jet flows are simulated with hexahedral and prismatic meshes. • Jet instability with prismatic meshes may be overestimated compared to that with hexahedral meshes. • Modified solvers that can reduce the influence of mesh non-orthogonality and reduce computation time are proposed. - Abstract: In the present research, we discuss the influence of mesh non-orthogonality on numerical solution of a type of buoyant flow. Buoyant jet flows are simulated numerically with hexahedral and prismatic mesh elements in an open source Computational Fluid Dynamics (CFD) code called “OpenFOAM”. Buoyant jet instability obtained with the prismatic meshes may be overestimated compared to that obtained with the hexahedral meshes when non-orthogonal correction is not applied in the code. Although the non-orthogonal correction method can improve the instability generated by mesh non-orthogonality, it may increase computation time required to reach a convergent solution. Thus, we propose modified solvers that can reduce the influence of mesh non-orthogonality and reduce the computation time compared to the existing solvers in OpenFOAM. It is demonstrated that calculations for a buoyant jet with a large temperature difference are performed faster by the modified solver.

  3. Gear Mesh Loss-of-Lubrication Experiments and Analytical Simulation

    Science.gov (United States)

    Handschuh, Robert F.; Polly, Joseph; Morales, Wilfredo

    2011-01-01

    An experimental program to determine the loss-of-lubrication (LOL) characteristics of spur gears in an aerospace simulation test facility has been completed. Tests were conducted using two different emergency lubricant types: (1) an oil mist system (two different misted lubricants) and (2) a grease injection system (two different grease types). Tests were conducted using a NASA Glenn test facility normally used for conducting contact fatigue. Tests were run at rotational speeds up to 10000 rpm using two different gear designs and two different gear materials. For the tests conducted using an air-oil misting system, a minimum lubricant injection rate was determined to permit the gear mesh to operate without failure for at least 1 hr. The tests allowed an elevated steady state temperature to be established. A basic 2-D heat transfer simulation has been developed to investigate temperatures of a simulated gear as a function of frictional behavior. The friction (heat generation source) between the meshing surfaces is related to the position in the meshing cycle, the load applied, and the amount of lubricant in the contact. Experimental conditions will be compared to those from the 2-D simulation.

  4. Simulating galactic dust grain evolution on a moving mesh

    Science.gov (United States)

    McKinnon, Ryan; Vogelsberger, Mark; Torrey, Paul; Marinacci, Federico; Kannan, Rahul

    2018-05-01

    Interstellar dust is an important component of the galactic ecosystem, playing a key role in multiple galaxy formation processes. We present a novel numerical framework for the dynamics and size evolution of dust grains implemented in the moving-mesh hydrodynamics code AREPO suited for cosmological galaxy formation simulations. We employ a particle-based method for dust subject to dynamical forces including drag and gravity. The drag force is implemented using a second-order semi-implicit integrator and validated using several dust-hydrodynamical test problems. Each dust particle has a grain size distribution, describing the local abundance of grains of different sizes. The grain size distribution is discretised with a second-order piecewise linear method and evolves in time according to various dust physical processes, including accretion, sputtering, shattering, and coagulation. We present a novel scheme for stochastically forming dust during stellar evolution and new methods for sub-cycling of dust physics time-steps. Using this model, we simulate an isolated disc galaxy to study the impact of dust physical processes that shape the interstellar grain size distribution. We demonstrate, for example, how dust shattering shifts the grain size distribution to smaller sizes resulting in a significant rise of radiation extinction from optical to near-ultraviolet wavelengths. Our framework for simulating dust and gas mixtures can readily be extended to account for other dynamical processes relevant in galaxy formation, like magnetohydrodynamics, radiation pressure, and thermo-chemical processes.

  5. Combining spray nozzle simulators with meshes: characterization of rainfall intensity and drop properties

    Science.gov (United States)

    Carvalho, Sílvia C. P.; de Lima, João L. M. P.; de Lima, M. Isabel P.

    2013-04-01

    Rainfall simulators can be a powerful tool to increase our understanding of hydrological and geomorphological processes. Nevertheless, rainfall simulators' design and operation might be rather demanding, for achieving specific rainfall intensity distributions and drop characteristics. The pressurized simulators have some advantages over the non-pressurized simulators: drops do not rely on gravity to reach terminal velocity, but are sprayed out under pressure; pressurized simulators also yield a broad range of drop sizes in comparison with drop-formers simulators. The main purpose of this study was to explore in the laboratory the potential of combining spray nozzle simulators with meshes in order to change rainfall characteristics (rainfall intensity and diameters and fall speed of drops). Different types of spray nozzles were tested, such as single full-cone and multiple full-cone nozzles. The impact of the meshes on the simulated rain was studied by testing different materials (i.e. plastic and steel meshes), square apertures and wire thicknesses, and different vertical distances between the nozzle and the meshes underneath. The diameter and fall speed of the rain drops were measured using a Laser Precipitation Monitor (Thies Clima). The rainfall intensity range and coefficients of uniformity of the sprays and the drop size distribution, fall speed and kinetic energy were analysed. Results show that when meshes intercept drop trajectories the spatial distribution of rainfall intensity and the drop size distribution are affected. As the spray nozzles generate typically small drop sizes and narrow drop size distributions, meshes can be used to promote the formation of bigger drops and random their landing positions.

  6. Worlds largest particle physics laboratory selects Proxim Wireless Mesh

    CERN Multimedia

    2007-01-01

    "Proxim Wireless has announced that the European Organization for Nuclear Research (CERN), the world's largest particle physics laboratory and the birthplace of the World Wide Web, is using it's ORiNOCO AP-4000 mesh access points to extend the range of the laboratory's Wi-Fi network and to provide continuous monitoring of the lab's calorimeters" (1/2 page)

  7. Fluid flow and heat transfer investigation of pebble bed reactors using mesh adaptive large-eddy simulation

    International Nuclear Information System (INIS)

    Pavlidis, D.; Lathouwers, D.

    2011-01-01

    A computational fluid dynamics model with anisotropic mesh adaptivity is used to investigate coolant flow and heat transfer in pebble bed reactors. A novel method for implicitly incorporating solid boundaries based on multi-fluid flow modelling is adopted. The resulting model is able to resolve and simulate flow and heat transfer in randomly packed beds, regardless of the actual geometry, starting off with arbitrarily coarse meshes. The model is initially evaluated using an orderly stacked square channel of channel-height-to-particle diameter ratio of unity for a range of Reynolds numbers. The model is then applied to the face-centred cubical geometry. Coolant flow and heat transfer patterns are investigated. (author)

  8. Tetrahedral-Mesh Simulation of Turbulent Flows with the Space-Time Conservative Schemes

    Science.gov (United States)

    Chang, Chau-Lyan; Venkatachari, Balaji; Cheng, Gary C.

    2015-01-01

    Direct numerical simulations of turbulent flows are predominantly carried out using structured, hexahedral meshes despite decades of development in unstructured mesh methods. Tetrahedral meshes offer ease of mesh generation around complex geometries and the potential of an orientation free grid that would provide un-biased small-scale dissipation and more accurate intermediate scale solutions. However, due to the lack of consistent multi-dimensional numerical formulations in conventional schemes for triangular and tetrahedral meshes at the cell interfaces, numerical issues exist when flow discontinuities or stagnation regions are present. The space-time conservative conservation element solution element (CESE) method - due to its Riemann-solver-free shock capturing capabilities, non-dissipative baseline schemes, and flux conservation in time as well as space - has the potential to more accurately simulate turbulent flows using unstructured tetrahedral meshes. To pave the way towards accurate simulation of shock/turbulent boundary-layer interaction, a series of wave and shock interaction benchmark problems that increase in complexity, are computed in this paper with triangular/tetrahedral meshes. Preliminary computations for the normal shock/turbulence interactions are carried out with a relatively coarse mesh, by direct numerical simulations standards, in order to assess other effects such as boundary conditions and the necessity of a buffer domain. The results indicate that qualitative agreement with previous studies can be obtained for flows where, strong shocks co-exist along with unsteady waves that display a broad range of scales, with a relatively compact computational domain and less stringent requirements for grid clustering near the shock. With the space-time conservation properties, stable solutions without any spurious wave reflections can be obtained without a need for buffer domains near the outflow/farfield boundaries. Computational results for the

  9. Coupling of a 3-D vortex particle-mesh method with a finite volume near-wall solver

    Science.gov (United States)

    Marichal, Y.; Lonfils, T.; Duponcheel, M.; Chatelain, P.; Winckelmans, G.

    2011-11-01

    This coupling aims at improving the computational efficiency of high Reynolds number bluff body flow simulations by using two complementary methods and exploiting their respective advantages in distinct parts of the domain. Vortex particle methods are particularly well suited for free vortical flows such as wakes or jets (the computational domain -with non zero vorticity- is then compact and dispersion errors are negligible). Finite volume methods, however, can handle boundary layers much more easily due to anisotropic mesh refinement. In the present approach, the vortex method is used in the whole domain (overlapping domain technique) but its solution is highly underresolved in the vicinity of the wall. It thus has to be corrected by the near-wall finite volume solution at each time step. Conversely, the vortex method provides the outer boundary conditions for the near-wall solver. A parallel multi-resolution vortex particle-mesh approach is used here along with an Immersed Boundary method in order to take the walls into account. The near-wall flow is solved by OpenFOAM® using the PISO algorithm. We validate the methodology on the flow past a sphere at a moderate Reynolds number. F.R.S. - FNRS Research Fellow.

  10. Integration of Heat Transfer, Stress, and Particle Trajectory Simulation

    Energy Technology Data Exchange (ETDEWEB)

    Thuc Bui; Michael Read; Lawrence ives

    2012-05-17

    Calabazas Creek Research, Inc. developed and currently markets Beam Optics Analyzer (BOA) in the United States and abroad. BOA is a 3D, charged particle optics code that solves the electric and magnetic fields with and without the presence of particles. It includes automatic and adaptive meshing to resolve spatial scales ranging from a few millimeters to meters. It is fully integrated with CAD packages, such as SolidWorks, allowing seamless geometry updates. The code includes iterative procedures for optimization, including a fully functional, graphical user interface. Recently, time dependent, particle in cell capability was added, pushing particles synchronically under quasistatic electromagnetic fields to obtain particle bunching under RF conditions. A heat transfer solver was added during this Phase I program. Completed tasks include: (1) Added a 3D finite element heat transfer solver with adaptivity; (2) Determined the accuracy of the linear heat transfer field solver to provide the basis for development of higher order solvers in Phase II; (3) Provided more accurate and smoother power density fields; and (4) Defined the geometry using the same CAD model, while maintaining different meshes, and interfacing the power density field between the particle simulator and heat transfer solvers. These objectives were achieved using modern programming techniques and algorithms. All programming was in C++ and parallelization in OpenMP, utilizing state-of-the-art multi-core technology. Both x86 and x64 versions are supported. The GUI design and implementation used Microsoft Foundation Class.

  11. Interoperable mesh and geometry tools for advanced petascale simulations

    International Nuclear Information System (INIS)

    Diachin, L; Bauer, A; Fix, B; Kraftcheck, J; Jansen, K; Luo, X; Miller, M; Ollivier-Gooch, C; Shephard, M S; Tautges, T; Trease, H

    2007-01-01

    SciDAC applications have a demonstrated need for advanced software tools to manage the complexities associated with sophisticated geometry, mesh, and field manipulation tasks, particularly as computer architectures move toward the petascale. The Center for Interoperable Technologies for Advanced Petascale Simulations (ITAPS) will deliver interoperable and interchangeable mesh, geometry, and field manipulation services that are of direct use to SciDAC applications. The premise of our technology development goal is to provide such services as libraries that can be used with minimal intrusion into application codes. To develop these technologies, we focus on defining a common data model and data-structure neutral interfaces that unify a number of different services such as mesh generation and improvement, front tracking, adaptive mesh refinement, shape optimization, and solution transfer operations. We highlight the use of several ITAPS services in SciDAC applications

  12. Nyx: Adaptive mesh, massively-parallel, cosmological simulation code

    Science.gov (United States)

    Almgren, Ann; Beckner, Vince; Friesen, Brian; Lukic, Zarija; Zhang, Weiqun

    2017-12-01

    Nyx code solves equations of compressible hydrodynamics on an adaptive grid hierarchy coupled with an N-body treatment of dark matter. The gas dynamics in Nyx use a finite volume methodology on an adaptive set of 3-D Eulerian grids; dark matter is represented as discrete particles moving under the influence of gravity. Particles are evolved via a particle-mesh method, using Cloud-in-Cell deposition/interpolation scheme. Both baryonic and dark matter contribute to the gravitational field. In addition, Nyx includes physics for accurately modeling the intergalactic medium; in optically thin limits and assuming ionization equilibrium, the code calculates heating and cooling processes of the primordial-composition gas in an ionizing ultraviolet background radiation field.

  13. A novel partitioning method for block-structured adaptive meshes

    Science.gov (United States)

    Fu, Lin; Litvinov, Sergej; Hu, Xiangyu Y.; Adams, Nikolaus A.

    2017-07-01

    We propose a novel partitioning method for block-structured adaptive meshes utilizing the meshless Lagrangian particle concept. With the observation that an optimum partitioning has high analogy to the relaxation of a multi-phase fluid to steady state, physically motivated model equations are developed to characterize the background mesh topology and are solved by multi-phase smoothed-particle hydrodynamics. In contrast to well established partitioning approaches, all optimization objectives are implicitly incorporated and achieved during the particle relaxation to stationary state. Distinct partitioning sub-domains are represented by colored particles and separated by a sharp interface with a surface tension model. In order to obtain the particle relaxation, special viscous and skin friction models, coupled with a tailored time integration algorithm are proposed. Numerical experiments show that the present method has several important properties: generation of approximately equal-sized partitions without dependence on the mesh-element type, optimized interface communication between distinct partitioning sub-domains, continuous domain decomposition which is physically localized and implicitly incremental. Therefore it is particularly suitable for load-balancing of high-performance CFD simulations.

  14. A novel partitioning method for block-structured adaptive meshes

    Energy Technology Data Exchange (ETDEWEB)

    Fu, Lin, E-mail: lin.fu@tum.de; Litvinov, Sergej, E-mail: sergej.litvinov@aer.mw.tum.de; Hu, Xiangyu Y., E-mail: xiangyu.hu@tum.de; Adams, Nikolaus A., E-mail: nikolaus.adams@tum.de

    2017-07-15

    We propose a novel partitioning method for block-structured adaptive meshes utilizing the meshless Lagrangian particle concept. With the observation that an optimum partitioning has high analogy to the relaxation of a multi-phase fluid to steady state, physically motivated model equations are developed to characterize the background mesh topology and are solved by multi-phase smoothed-particle hydrodynamics. In contrast to well established partitioning approaches, all optimization objectives are implicitly incorporated and achieved during the particle relaxation to stationary state. Distinct partitioning sub-domains are represented by colored particles and separated by a sharp interface with a surface tension model. In order to obtain the particle relaxation, special viscous and skin friction models, coupled with a tailored time integration algorithm are proposed. Numerical experiments show that the present method has several important properties: generation of approximately equal-sized partitions without dependence on the mesh-element type, optimized interface communication between distinct partitioning sub-domains, continuous domain decomposition which is physically localized and implicitly incremental. Therefore it is particularly suitable for load-balancing of high-performance CFD simulations.

  15. Flow simulation of a Pelton bucket using finite volume particle method

    International Nuclear Information System (INIS)

    Vessaz, C; Jahanbakhsh, E; Avellan, F

    2014-01-01

    The objective of the present paper is to perform an accurate numerical simulation of the high-speed water jet impinging on a Pelton bucket. To reach this goal, the Finite Volume Particle Method (FVPM) is used to discretize the governing equations. FVPM is an arbitrary Lagrangian-Eulerian method, which combines attractive features of Smoothed Particle Hydrodynamics and conventional mesh-based Finite Volume Method. This method is able to satisfy free surface and no-slip wall boundary conditions precisely. The fluid flow is assumed weakly compressible and the wall boundary is represented by one layer of particles located on the bucket surface. In the present study, the simulations of the flow in a stationary bucket are investigated for three different impinging angles: 72°, 90° and 108°. The particles resolution is first validated by a convergence study. Then, the FVPM results are validated with available experimental data and conventional grid-based Volume Of Fluid simulations. It is shown that the wall pressure field is in good agreement with the experimental and numerical data. Finally, the torque evolution and water sheet location are presented for a simulation of five rotating Pelton buckets

  16. The numerical simulation study of hemodynamics of the new dense-mesh stent

    Science.gov (United States)

    Ma, Jiali; Yuan, Zhishan; Yu, Xuebao; Feng, Zhaowei; Miao, Weidong; Xu, Xueli; Li, Juntao

    2017-09-01

    The treatment of aortic aneurysm in new dense mesh stent is based on the principle of hemodynamic changes. But the mechanism is not yet very clear. This paper analyzed and calculated the hemodynamic situation before and after the new dense mesh stent implanting by the method of numerical simulation. The results show the dense mesh stent changed and impacted the blood flow in the aortic aneurysm. The changes include significant decrement of blood velocity, pressure and shear forces, while ensuring blood can supply branches, which means the new dense mesh stent's hemodynamic mechanism in the treatment of aortic aneurysm is clearer. It has very important significance in developing new dense mesh stent in order to cure aortic aneurysm.

  17. Vectorization of a particle code used in the simulation of rarefied hypersonic flow

    Science.gov (United States)

    Baganoff, D.

    1990-01-01

    A limitation of the direct simulation Monte Carlo (DSMC) method is that it does not allow efficient use of vector architectures that predominate in current supercomputers. Consequently, the problems that can be handled are limited to those of one- and two-dimensional flows. This work focuses on a reformulation of the DSMC method with the objective of designing a procedure that is optimized to the vector architectures found on machines such as the Cray-2. In addition, it focuses on finding a better balance between algorithmic complexity and the total number of particles employed in a simulation so that the overall performance of a particle simulation scheme can be greatly improved. Simulations of the flow about a 3D blunt body are performed with 10 to the 7th particles and 4 x 10 to the 5th mesh cells. Good statistics are obtained with time averaging over 800 time steps using 4.5 h of Cray-2 single-processor CPU time.

  18. CFD simulation of an internal spin-filter: evidence of lateral migration and exchange flow through the mesh.

    Science.gov (United States)

    Figueredo-Cardero, Alvio; Chico, Ernesto; Castilho, Leda R; Medronho, Ricardo A

    2009-11-01

    In the present work Computational Fluid Dynamics (CFD) was used to study the flow field and particle dynamics in an internal spin-filter (SF) bioreactor system. Evidence of a radial exchange flow through the filter mesh was detected, with a magnitude up to 130-fold higher than the perfusion flow, thus significantly contributing to radial drag. The exchange flow magnitude was significantly influenced by the filter rotation rate, but not by the perfusion flow, within the ranges evaluated. Previous reports had only given indirect evidences of this exchange flow phenomenon in spin-filters, but the current simulations were able to quantify and explain it. Flow pattern inside the spin-filter bioreactor resembled a typical Taylor-Couette flow, with vortices being formed in the annular gap and eventually penetrating the internal volume of the filter, thus being the probable reason for the significant exchange flow observed. The simulations also showed that cells become depleted in the vicinity of the mesh due to lateral particle migration. Cell concentration near the filter was approximately 50% of the bulk concentration, explaining why cell separation achieved in SFs is not solely due to size exclusion. The results presented indicate the power of CFD techniques to study and better understand spin-filter systems, aiming at the establishment of effective design, operation and scale-up criteria.

  19. Numerical simulation of 3D unsteady flow in a rotating pump by dynamic mesh technique

    International Nuclear Information System (INIS)

    Huang, S; Guo, J; Yang, F X

    2013-01-01

    In this paper, the numerical simulation of unsteady flow for three kinds of typical rotating pumps, roots blower, roto-jet pump and centrifugal pump, were performed using the three-dimensional Dynamic Mesh technique. In the unsteady simulation, all the computational domains, as stationary, were set in one inertial reference frame. The motions of the solid boundaries were defined by the Profile file in FLUENT commercial code, in which the rotational orientation and speed of the rotors were specified. Three methods (Spring-based Smoothing, Dynamic Layering and Local Re-meshing) were used to achieve mesh deformation and re-meshing. The unsteady solutions of flow field and pressure distribution were solved. After a start-up stage, the flow parameters exhibit time-periodic behaviour corresponding to blade passing frequency of rotor. This work shows that Dynamic Mesh technique could achieve numerical simulation of three-dimensional unsteady flow field in various kinds of rotating pumps and have a strong versatility and broad application prospects

  20. Finite element formulation of fluctuating hydrodynamics for fluids filled with rigid particles using boundary fitted meshes

    Energy Technology Data Exchange (ETDEWEB)

    De Corato, M., E-mail: marco.decorato@unina.it [Dipartimento di Ingegneria Chimica, dei Materiali e della Produzione Industriale, Università di Napoli Federico II, Piazzale Tecchio 80, 80125 Napoli (Italy); Slot, J.J.M., E-mail: j.j.m.slot@tue.nl [Department of Mathematics and Computer Science, Eindhoven University of Technology, PO Box 513, 5600 MB Eindhoven (Netherlands); Hütter, M., E-mail: m.huetter@tue.nl [Department of Mechanical Engineering, Eindhoven University of Technology, PO Box 513, 5600 MB Eindhoven (Netherlands); D' Avino, G., E-mail: gadavino@unina.it [Dipartimento di Ingegneria Chimica, dei Materiali e della Produzione Industriale, Università di Napoli Federico II, Piazzale Tecchio 80, 80125 Napoli (Italy); Maffettone, P.L., E-mail: pierluca.maffettone@unina.it [Dipartimento di Ingegneria Chimica, dei Materiali e della Produzione Industriale, Università di Napoli Federico II, Piazzale Tecchio 80, 80125 Napoli (Italy); Hulsen, M.A., E-mail: m.a.hulsen@tue.nl [Department of Mechanical Engineering, Eindhoven University of Technology, PO Box 513, 5600 MB Eindhoven (Netherlands)

    2016-07-01

    In this paper, we present a finite element implementation of fluctuating hydrodynamics with a moving boundary fitted mesh for treating the suspended particles. The thermal fluctuations are incorporated into the continuum equations using the Landau and Lifshitz approach [1]. The proposed implementation fulfills the fluctuation–dissipation theorem exactly at the discrete level. Since we restrict the equations to the creeping flow case, this takes the form of a relation between the diffusion coefficient matrix and friction matrix both at the particle and nodal level of the finite elements. Brownian motion of arbitrarily shaped particles in complex confinements can be considered within the present formulation. A multi-step time integration scheme is developed to correctly capture the drift term required in the stochastic differential equation (SDE) describing the evolution of the positions of the particles. The proposed approach is validated by simulating the Brownian motion of a sphere between two parallel plates and the motion of a spherical particle in a cylindrical cavity. The time integration algorithm and the fluctuating hydrodynamics implementation are then applied to study the diffusion and the equilibrium probability distribution of a confined circle under an external harmonic potential.

  1. Mechanical behaviour of synthetic surgical meshes: finite element simulation of the herniated abdominal wall.

    Science.gov (United States)

    Hernández-Gascón, B; Peña, E; Melero, H; Pascual, G; Doblaré, M; Ginebra, M P; Bellón, J M; Calvo, B

    2011-11-01

    The material properties of meshes used in hernia surgery contribute to the overall mechanical behaviour of the repaired abdominal wall. The mechanical response of a surgical mesh has to be defined since the haphazard orientation of an anisotropic mesh can lead to inconsistent surgical outcomes. This study was designed to characterize the mechanical behaviour of three surgical meshes (Surgipro®, Optilene® and Infinit®) and to describe a mechanical constitutive law that accurately reproduces the experimental results. Finally, through finite element simulation, the behaviour of the abdominal wall was modelled before and after surgical mesh implant. Uniaxial loading of mesh samples in two perpendicular directions revealed the isotropic response of Surgipro® and the anisotropic behaviour of Optilene® and Infinit®. A phenomenological constitutive law was used to reproduce the measured experimental curves. To analyze the mechanical effect of the meshes once implanted in the abdomen, finite element simulation of the healthy and partially herniated repaired rabbit abdominal wall served to reproduce wall behaviour before and after mesh implant. In all cases, maximal displacements were lower and maximal principal stresses higher in the implanted abdomen than the intact wall model. Despite the fact that no mesh showed a behaviour that perfectly matched that of abdominal muscle, the Infinit® mesh was able to best comply with the biomechanics of the abdominal wall. Copyright © 2011 Acta Materialia Inc. Published by Elsevier Ltd. All rights reserved.

  2. Integration of Heat Transfer, Stress, and Particle Trajectory Simulation. Final report

    International Nuclear Information System (INIS)

    Bui, Thuc; Read, Michael; Ives, Lawrence

    2012-01-01

    Calabazas Creek Research, Inc. developed and currently markets Beam Optics Analyzer (BOA) in the United States and abroad. BOA is a 3D, charged particle optics code that solves the electric and magnetic fields with and without the presence of particles. It includes automatic and adaptive meshing to resolve spatial scales ranging from a few millimeters to meters. It is fully integrated with CAD packages, such as SolidWorks, allowing seamless geometry updates. The code includes iterative procedures for optimization, including a fully functional, graphical user interface. Recently, time dependent, particle in cell capability was added, pushing particles synchronically under quasistatic electromagnetic fields to obtain particle bunching under RF conditions. A heat transfer solver was added during this Phase I program. Completed tasks include: (1) Added a 3D finite element heat transfer solver with adaptivity; (2) Determined the accuracy of the linear heat transfer field solver to provide the basis for development of higher order solvers in Phase II; (3) Provided more accurate and smoother power density fields; and (4) Defined the geometry using the same CAD model, while maintaining different meshes, and interfacing the power density field between the particle simulator and heat transfer solvers. These objectives were achieved using modern programming techniques and algorithms. All programming was in C++ and parallelization in OpenMP, utilizing state-of-the-art multi-core technology. Both x86 and x64 versions are supported. The GUI design and implementation used Microsoft Foundation Class.

  3. Combining the Vortex Particle-Mesh method with a Multi-Body System solver for the simulation of self-propelled articulated swimmers

    Science.gov (United States)

    Bernier, Caroline; Gazzola, Mattia; Ronsse, Renaud; Chatelain, Philippe

    2017-11-01

    We present a 2D fluid-structure interaction simulation method with a specific focus on articulated and actuated structures. The proposed algorithm combines a viscous Vortex Particle-Mesh (VPM) method based on a penalization technique and a Multi-Body System (MBS) solver. The hydrodynamic forces and moments acting on the structure parts are not computed explicitly from the surface stresses; they are rather recovered from the projection and penalization steps within the VPM method. The MBS solver accounts for the body dynamics via the Euler-Lagrange formalism. The deformations of the structure are dictated by the hydrodynamic efforts and actuation torques. Here, we focus on simplified swimming structures composed of neutrally buoyant ellipses connected by virtual joints. The joints are actuated through a simple controller in order to reproduce the swimming patterns of an eel-like swimmer. The method enables to recover the histories of torques applied on each hinge along the body. The method is verified on several benchmarks: an impulsively started elastically mounted cylinder and free swimming articulated fish-like structures. Validation will be performed by means of an experimental swimming robot that reproduces the 2D articulated ellipses.

  4. Algebraic mesh generation for large scale viscous-compressible aerodynamic simulation

    International Nuclear Information System (INIS)

    Smith, R.E.

    1984-01-01

    Viscous-compressible aerodynamic simulation is the numerical solution of the compressible Navier-Stokes equations and associated boundary conditions. Boundary-fitted coordinate systems are well suited for the application of finite difference techniques to the Navier-Stokes equations. An algebraic approach to boundary-fitted coordinate systems is one where an explicit functional relation describes a mesh on which a solution is obtained. This approach has the advantage of rapid-precise mesh control. The basic mathematical structure of three algebraic mesh generation techniques is described. They are transfinite interpolation, the multi-surface method, and the two-boundary technique. The Navier-Stokes equations are transformed to a computational coordinate system where boundary-fitted coordinates can be applied. Large-scale computation implies that there is a large number of mesh points in the coordinate system. Computation of viscous compressible flow using boundary-fitted coordinate systems and the application of this computational philosophy on a vector computer are presented

  5. Full Core Multiphysics Simulation with Offline Mesh Deformation

    Energy Technology Data Exchange (ETDEWEB)

    Merzari, E. [Argonne National Lab. (ANL), Argonne, IL (United States); Shemon, E. R. [Argonne National Lab. (ANL), Argonne, IL (United States); Yu, Y. [Argonne National Lab. (ANL), Argonne, IL (United States); Thomas, J. W. [Argonne National Lab. (ANL), Argonne, IL (United States); Obabko, A. [Argonne National Lab. (ANL), Argonne, IL (United States); Jain, Rajeev [Argonne National Lab. (ANL), Argonne, IL (United States); Mahadevan, Vijay [Argonne National Lab. (ANL), Argonne, IL (United States); Solberg, Jerome [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Ferencz, R. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Whitesides, R. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2015-12-21

    In this report, building on previous reports issued in FY13 we describe our continued efforts to integrate thermal/hydraulics, neutronics, and structural mechanics modeling codes to perform coupled analysis of a representative fast sodium-cooled reactor core. The focus of the present report is a full core simulation with off-line mesh deformation.

  6. Fuel-steel mixing and radial mesh effects in power excursion simulations

    International Nuclear Information System (INIS)

    Chen, X.-N.; Rineiski, A.; Gabrielli, F.; Andriolo, L.; Vezzoni, B.; Li, R.; Maschek, W.; Kiefhaber, E.

    2016-01-01

    Highlights: • Fuel-steel mixing and radial mesh effects are significant on power excursion. • The earliest power peak is reduced and retarded by these two effects. • Unprotected loss of coolant transients in ESFR core are calculated. - Abstract: This paper deals with SIMMER-III once-through simulations of the earliest power excursion initiated by an unprotected loss of flow (ULOF) in the Working Horse design of the European Sodium Cooled Fast Reactor (ESFR). Since the sodium void effect is strictly positive in this core and dominant in the transient, a power excursion is initiated by sodium boiling in the ULOF case. Two major effects, namely (1) reactivity effects due to fuel-steel mixing after melting and (2) the radial mesh size, which were not considered originally in SIMMER simulations for ESFR, are studied. The first effect concerns the reactivity difference between the heterogeneous fuel/clad/wrapper configuration and the homogeneous mixture of steel and fuel. The full core homogenization (due to melting) effect is −2 $, though a smaller effect takes place in case of partial core melting. The second effect is due to the SIMMER sub-assembly (SA) coarse mesh treatment, where a simultaneous sodium boiling onset in all SAs belonging to one ring leads to an overestimated reactivity ramp. For investigating the influence of fuel/steel mixing effects, a lumped “homogenization” reactivity feedback has been introduced, being proportional to the molten steel mass. For improving the coarse mesh treatment, we employ finer radial meshes to take the subchannel effects into account, where the side and interior channels have different coolant velocities and temperatures. The simulation results show that these two effects have significant impacts on the earliest power excursion after the sodium boiling.

  7. A particle finite element method for machining simulations

    Science.gov (United States)

    Sabel, Matthias; Sator, Christian; Müller, Ralf

    2014-07-01

    The particle finite element method (PFEM) appears to be a convenient technique for machining simulations, since the geometry and topology of the problem can undergo severe changes. In this work, a short outline of the PFEM-algorithm is given, which is followed by a detailed description of the involved operations. The -shape method, which is used to track the topology, is explained and tested by a simple example. Also the kinematics and a suitable finite element formulation are introduced. To validate the method simple settings without topological changes are considered and compared to the standard finite element method for large deformations. To examine the performance of the method, when dealing with separating material, a tensile loading is applied to a notched plate. This investigation includes a numerical analysis of the different meshing parameters, and the numerical convergence is studied. With regard to the cutting simulation it is found that only a sufficiently large number of particles (and thus a rather fine finite element discretisation) leads to converged results of process parameters, such as the cutting force.

  8. In vitro bioactivity of 3D Ti-mesh with bioceramic coatings in simulated body fluid

    Directory of Open Access Journals (Sweden)

    Wei Yi

    2014-09-01

    Full Text Available 3D Ti-mesh has been coated with bioceramics under different coating conditions, such as material compositions and micro-porosity, using a dip casting method. Hydroxyapatite (HA, micro-HA particles (HAp, a bioglass (BG and their different mixtures together with polymer additives were used to control HA-coating microstructures. Layered composites with the following coating-to-substrate designs, such as BG/Ti, HA + BG/BG/Ti and HAp + BG/BG/Ti, were fabricated. The bioactivity of these coated composites and the uncoated Ti-mesh substrate was then investigated in a simulated body fluid (SBF. The Ti-mesh substrate and BG/Ti composite did not induce biomimetic apatite deposition when they were immersed in SBF for the selected BG, a pressable dental ceramic, used in this study. After seven days in SBF, an apatite layer was formed on both HA + BG/BG/Ti and HAp + BG/BG/Ti composites. The difference is the apatite layer on the HAp + BG/BG/Ti composite was rougher and contained more micro-pores, while the apatite layer on the HA + BG/BG/Ti composite was dense and smooth. The formation of biomimetic apatite, being more bioresorbable, is favored for bone regeneration.

  9. Reference Computational Meshing Strategy for Computational Fluid Dynamics Simulation of Departure from Nucleate BoilingReference Computational Meshing Strategy for Computational Fluid Dynamics Simulation of Departure from Nucleate Boiling

    Energy Technology Data Exchange (ETDEWEB)

    Pointer, William David [ORNL

    2017-08-01

    The objective of this effort is to establish a strategy and process for generation of suitable computational mesh for computational fluid dynamics simulations of departure from nucleate boiling in a 5 by 5 fuel rod assembly held in place by PWR mixing vane spacer grids. This mesh generation process will support ongoing efforts to develop, demonstrate and validate advanced multi-phase computational fluid dynamics methods that enable more robust identification of dryout conditions and DNB occurrence.Building upon prior efforts and experience, multiple computational meshes were developed using the native mesh generation capabilities of the commercial CFD code STAR-CCM+. These meshes were used to simulate two test cases from the Westinghouse 5 by 5 rod bundle facility. The sensitivity of predicted quantities of interest to the mesh resolution was then established using two evaluation methods, the Grid Convergence Index method and the Least Squares method. This evaluation suggests that the Least Squares method can reliably establish the uncertainty associated with local parameters such as vector velocity components at a point in the domain or surface averaged quantities such as outlet velocity magnitude. However, neither method is suitable for characterization of uncertainty in global extrema such as peak fuel surface temperature, primarily because such parameters are not necessarily associated with a fixed point in space. This shortcoming is significant because the current generation algorithm for identification of DNB event conditions relies on identification of such global extrema. Ongoing efforts to identify DNB based on local surface conditions will address this challenge

  10. A hybridized discontinuous Galerkin framework for high-order particle-mesh operator splitting of the incompressible Navier-Stokes equations

    Science.gov (United States)

    Maljaars, Jakob M.; Labeur, Robert Jan; Möller, Matthias

    2018-04-01

    A generic particle-mesh method using a hybridized discontinuous Galerkin (HDG) framework is presented and validated for the solution of the incompressible Navier-Stokes equations. Building upon particle-in-cell concepts, the method is formulated in terms of an operator splitting technique in which Lagrangian particles are used to discretize an advection operator, and an Eulerian mesh-based HDG method is employed for the constitutive modeling to account for the inter-particle interactions. Key to the method is the variational framework provided by the HDG method. This allows to formulate the projections between the Lagrangian particle space and the Eulerian finite element space in terms of local (i.e. cellwise) ℓ2-projections efficiently. Furthermore, exploiting the HDG framework for solving the constitutive equations results in velocity fields which excellently approach the incompressibility constraint in a local sense. By advecting the particles through these velocity fields, the particle distribution remains uniform over time, obviating the need for additional quality control. The presented methodology allows for a straightforward extension to arbitrary-order spatial accuracy on general meshes. A range of numerical examples shows that optimal convergence rates are obtained in space and, given the particular time stepping strategy, second-order accuracy is obtained in time. The model capabilities are further demonstrated by presenting results for the flow over a backward facing step and for the flow around a cylinder.

  11. Go with the Flow. Moving meshes and solution monitoring for compressible flow simulation

    NARCIS (Netherlands)

    van Dam, A.

    2009-01-01

    The simulation of time-dependent physical problems, such as flows of some kind, places high demands on the domain discretization in order to obtain high accuracy of the numerical solution. We present a moving mesh method in which the mesh points automatically move towards regions where high spatial

  12. Numerical simulation for quenching meshes with TONUS platform

    International Nuclear Information System (INIS)

    Bin, Chen; Hongxing, Yu

    2009-01-01

    For mitigation of hydrogen risks during severe accidents to protect the integrity of containment, PAR and ignitors are used in current advanced nuclear power plants. But multiple combustions induced by ignitors and consequent DDT phenomena are not practically eliminated. An innovative design call 'quenching meshes' is considered to confine hydrogen flame within one compartment by metallic meshes, so that hazardous flame propagation can be prevented. The numerical simulation results based on discretization of the full Navier-Stokes equations with global one-step reaction represented by Arrhenius laminar combustion model have shown the possibility of flame quenching 'numerically'. This is achieved via multiplication of the combustion rate expression by a Heaviside function having an ignition temperature as a parameter. Qualitative behavior of the computed flow shows that the flame velocity diminishes while passing through a quenching mesh, while qualitative analysis based on the energy balance reveals the mechanism of flame quenching. All the above analysis has been performed for a stoichiometric mixture and normal initial pressure and temperature for initial conditions. For further research we would like to suggest the investigation of the influence of the mixture composition, initial pressure and/or temperature on the quenching criteria

  13. The Space-Time Conservative Schemes for Large-Scale, Time-Accurate Flow Simulations with Tetrahedral Meshes

    Science.gov (United States)

    Venkatachari, Balaji Shankar; Streett, Craig L.; Chang, Chau-Lyan; Friedlander, David J.; Wang, Xiao-Yen; Chang, Sin-Chung

    2016-01-01

    Despite decades of development of unstructured mesh methods, high-fidelity time-accurate simulations are still predominantly carried out on structured, or unstructured hexahedral meshes by using high-order finite-difference, weighted essentially non-oscillatory (WENO), or hybrid schemes formed by their combinations. In this work, the space-time conservation element solution element (CESE) method is used to simulate several flow problems including supersonic jet/shock interaction and its impact on launch vehicle acoustics, and direct numerical simulations of turbulent flows using tetrahedral meshes. This paper provides a status report for the continuing development of the space-time conservation element solution element (CESE) numerical and software framework under the Revolutionary Computational Aerosciences (RCA) project. Solution accuracy and large-scale parallel performance of the numerical framework is assessed with the goal of providing a viable paradigm for future high-fidelity flow physics simulations.

  14. A particle-based model to simulate the micromechanics of single-plant parenchyma cells and aggregates

    International Nuclear Information System (INIS)

    Van Liedekerke, P; Tijskens, E; Smeedts, B; Ramon, H; Ghysels, P; Samaey, G; Roose, D

    2010-01-01

    This paper is concerned with addressing how plant tissue mechanics is related to the micromechanics of cells. To this end, we propose a mesh-free particle method to simulate the mechanics of both individual plant cells (parenchyma) and cell aggregates in response to external stresses. The model considers two important features in the plant cell: (1) the cell protoplasm, the interior liquid phase inducing hydrodynamic phenomena, and (2) the cell wall material, a viscoelastic solid material that contains the protoplasm. In this particle framework, the cell fluid is modeled by smoothed particle hydrodynamics (SPH), a mesh-free method typically used to address problems with gas and fluid dynamics. In the solid phase (cell wall) on the other hand, the particles are connected by pairwise interactions holding them together and preventing the fluid to penetrate the cell wall. The cell wall hydraulic conductivity (permeability) is built in as well through the SPH formulation. Although this model is also meant to be able to deal with dynamic and even violent situations (leading to cell wall rupture or cell–cell debonding), we have concentrated on quasi-static conditions. The results of single-cell compression simulations show that the conclusions found by analytical models and experiments can be reproduced at least qualitatively. Relaxation tests revealed that plant cells have short relaxation times (1 µs–10 µs) compared to mammalian cells. Simulations performed on cell aggregates indicated an influence of the cellular organization to the tissue response, as was also observed in experiments done on tissues with a similar structure

  15. Mesh-free Hamiltonian implementation of two dimensional Darwin model

    Science.gov (United States)

    Siddi, Lorenzo; Lapenta, Giovanni; Gibbon, Paul

    2017-08-01

    A new approach to Darwin or magnetoinductive plasma simulation is presented, which combines a mesh-free field solver with a robust time-integration scheme avoiding numerical divergence errors in the solenoidal field components. The mesh-free formulation employs an efficient parallel Barnes-Hut tree algorithm to speed up the computation of fields summed directly from the particles, avoiding the necessity of divergence cleaning procedures typically required by particle-in-cell methods. The time-integration scheme employs a Hamiltonian formulation of the Lorentz force, circumventing the development of violent numerical instabilities associated with time differentiation of the vector potential. It is shown that a semi-implicit scheme converges rapidly and is robust to further numerical instabilities which can develop from a dominant contribution of the vector potential to the canonical momenta. The model is validated by various static and dynamic benchmark tests, including a simulation of the Weibel-like filamentation instability in beam-plasma interactions.

  16. Simulation of geothermal water extraction in heterogeneous reservoirs using dynamic unstructured mesh optimisation

    Science.gov (United States)

    Salinas, P.; Pavlidis, D.; Jacquemyn, C.; Lei, Q.; Xie, Z.; Pain, C.; Jackson, M.

    2017-12-01

    It is well known that the pressure gradient into a production well increases with decreasing distance to the well. To properly capture the local pressure drawdown into the well a high grid or mesh resolution is required; moreover, the location of the well must be captured accurately. In conventional simulation models, the user must interact with the model to modify grid resolution around wells of interest, and the well location is approximated on a grid defined early in the modelling process.We report a new approach for improved simulation of near wellbore flow in reservoir scale models through the use of dynamic mesh optimisation and the recently presented double control volume finite element method. Time is discretized using an adaptive, implicit approach. Heterogeneous geologic features are represented as volumes bounded by surfaces. Within these volumes, termed geologic domains, the material properties are constant. Up-, cross- or down-scaling of material properties during dynamic mesh optimization is not required, as the properties are uniform within each geologic domain. A given model typically contains numerous such geologic domains. Wells are implicitly coupled with the domain, and the fluid flows is modelled inside the wells. The method is novel for two reasons. First, a fully unstructured tetrahedral mesh is used to discretize space, and the spatial location of the well is specified via a line vector, ensuring its location even if the mesh is modified during the simulation. The well location is therefore accurately captured, the approach allows complex well trajectories and wells with many laterals to be modelled. Second, computational efficiency is increased by use of dynamic mesh optimization, in which an unstructured mesh adapts in space and time to key solution fields (preserving the geometry of the geologic domains), such as pressure, velocity or temperature, this also increases the quality of the solutions by placing higher resolution where required

  17. Kinetic solvers with adaptive mesh in phase space

    Science.gov (United States)

    Arslanbekov, Robert R.; Kolobov, Vladimir I.; Frolova, Anna A.

    2013-12-01

    An adaptive mesh in phase space (AMPS) methodology has been developed for solving multidimensional kinetic equations by the discrete velocity method. A Cartesian mesh for both configuration (r) and velocity (v) spaces is produced using a “tree of trees” (ToT) data structure. The r mesh is automatically generated around embedded boundaries, and is dynamically adapted to local solution properties. The v mesh is created on-the-fly in each r cell. Mappings between neighboring v-space trees is implemented for the advection operator in r space. We have developed algorithms for solving the full Boltzmann and linear Boltzmann equations with AMPS. Several recent innovations were used to calculate the discrete Boltzmann collision integral with dynamically adaptive v mesh: the importance sampling, multipoint projection, and variance reduction methods. We have developed an efficient algorithm for calculating the linear Boltzmann collision integral for elastic and inelastic collisions of hot light particles in a Lorentz gas. Our AMPS technique has been demonstrated for simulations of hypersonic rarefied gas flows, ion and electron kinetics in weakly ionized plasma, radiation and light-particle transport through thin films, and electron streaming in semiconductors. We have shown that AMPS allows minimizing the number of cells in phase space to reduce the computational cost and memory usage for solving challenging kinetic problems.

  18. Numerical simulation of deformation of dynamic mesh in the human vocal tract model

    Directory of Open Access Journals (Sweden)

    Řidký Václav

    2015-01-01

    Full Text Available Numerical simulation of the acoustic signal generation in the human vocal tract is a very complex problem. The computational mesh is not static; it is deformed due to vibration of vocal folds. Movement of vocal folds is in this case prescribed as function of translation and rotation. A new boundary condition for the 2DOF motion of the vocal folds was implemented in OpenFOAM, an open-source software package based on finite volume method Work is focused on the dynamic mesh and deformation of structured meshes in the computation a package OpenFOAM. These methods are compared with focus onquality of the mesh (non-orthogonality, aspect ratio and skewness.

  19. Regional Community Climate Simulations with variable resolution meshes in the Community Earth System Model

    Science.gov (United States)

    Zarzycki, C. M.; Gettelman, A.; Callaghan, P.

    2017-12-01

    Accurately predicting weather extremes such as precipitation (floods and droughts) and temperature (heat waves) requires high resolution to resolve mesoscale dynamics and topography at horizontal scales of 10-30km. Simulating such resolutions globally for climate scales (years to decades) remains computationally impractical. Simulating only a small region of the planet is more tractable at these scales for climate applications. This work describes global simulations using variable-resolution static meshes with multiple dynamical cores that target the continental United States using developmental versions of the Community Earth System Model version 2 (CESM2). CESM2 is tested in idealized, aquaplanet and full physics configurations to evaluate variable mesh simulations against uniform high and uniform low resolution simulations at resolutions down to 15km. Different physical parameterization suites are also evaluated to gauge their sensitivity to resolution. Idealized variable-resolution mesh cases compare well to high resolution tests. More recent versions of the atmospheric physics, including cloud schemes for CESM2, are more stable with respect to changes in horizontal resolution. Most of the sensitivity is due to sensitivity to timestep and interactions between deep convection and large scale condensation, expected from the closure methods. The resulting full physics model produces a comparable climate to the global low resolution mesh and similar high frequency statistics in the high resolution region. Some biases are reduced (orographic precipitation in the western United States), but biases do not necessarily go away at high resolution (e.g. summertime JJA surface Temp). The simulations are able to reproduce uniform high resolution results, making them an effective tool for regional climate studies and are available in CESM2.

  20. ADAPTIVE MESH REFINEMENT SIMULATIONS OF GALAXY FORMATION: EXPLORING NUMERICAL AND PHYSICAL PARAMETERS

    International Nuclear Information System (INIS)

    Hummels, Cameron B.; Bryan, Greg L.

    2012-01-01

    We carry out adaptive mesh refinement cosmological simulations of Milky Way mass halos in order to investigate the formation of disk-like galaxies in a Λ-dominated cold dark matter model. We evolve a suite of five halos to z = 0 and find a gas disk formation in each; however, in agreement with previous smoothed particle hydrodynamics simulations (that did not include a subgrid feedback model), the rotation curves of all halos are centrally peaked due to a massive spheroidal component. Our standard model includes radiative cooling and star formation, but no feedback. We further investigate this angular momentum problem by systematically modifying various simulation parameters including: (1) spatial resolution, ranging from 1700 to 212 pc; (2) an additional pressure component to ensure that the Jeans length is always resolved; (3) low star formation efficiency, going down to 0.1%; (4) fixed physical resolution as opposed to comoving resolution; (5) a supernova feedback model that injects thermal energy to the local cell; and (6) a subgrid feedback model which suppresses cooling in the immediate vicinity of a star formation event. Of all of these, we find that only the last (cooling suppression) has any impact on the massive spheroidal component. In particular, a simulation with cooling suppression and feedback results in a rotation curve that, while still peaked, is considerably reduced from our standard runs.

  1. A DAFT DL_POLY distributed memory adaptation of the Smoothed Particle Mesh Ewald method

    Science.gov (United States)

    Bush, I. J.; Todorov, I. T.; Smith, W.

    2006-09-01

    The Smoothed Particle Mesh Ewald method [U. Essmann, L. Perera, M.L. Berkowtz, T. Darden, H. Lee, L.G. Pedersen, J. Chem. Phys. 103 (1995) 8577] for calculating long ranged forces in molecular simulation has been adapted for the parallel molecular dynamics code DL_POLY_3 [I.T. Todorov, W. Smith, Philos. Trans. Roy. Soc. London 362 (2004) 1835], making use of a novel 3D Fast Fourier Transform (DAFT) [I.J. Bush, The Daresbury Advanced Fourier transform, Daresbury Laboratory, 1999] that perfectly matches the Domain Decomposition (DD) parallelisation strategy [W. Smith, Comput. Phys. Comm. 62 (1991) 229; M.R.S. Pinches, D. Tildesley, W. Smith, Mol. Sim. 6 (1991) 51; D. Rapaport, Comput. Phys. Comm. 62 (1991) 217] of the DL_POLY_3 code. In this article we describe software adaptations undertaken to import this functionality and provide a review of its performance.

  2. MHD simulations on an unstructured mesh

    International Nuclear Information System (INIS)

    Strauss, H.R.; Park, W.

    1996-01-01

    We describe work on a full MHD code using an unstructured mesh. MH3D++ is an extension of the PPPL MH3D resistive full MHD code. MH3D++ replaces the structured mesh and finite difference / fourier discretization of MH3D with an unstructured mesh and finite element / fourier discretization. Low level routines which perform differential operations, solution of PDEs such as Poisson's equation, and graphics, are encapsulated in C++ objects to isolate the finite element operations from the higher level code. The high level code is the same, whether it is run in structured or unstructured mesh versions. This allows the unstructured mesh version to be benchmarked against the structured mesh version. As a preliminary example, disruptions in DIIID reverse shear equilibria are studied numerically with the MH3D++ code. Numerical equilibria were first produced starting with an EQDSK file containing equilibrium data of a DIII-D L-mode negative central shear discharge. Using these equilibria, the linearized equations are time advanced to get the toroidal mode number n = 1 linear growth rate and eigenmode, which is resistively unstable. The equilibrium and linear mode are used to initialize 3D nonlinear runs. An example shows poloidal slices of 3D pressure surfaces: initially, on the left, and at an intermediate time, on the right

  3. Documentation for MeshKit - Reactor Geometry (&mesh) Generator

    Energy Technology Data Exchange (ETDEWEB)

    Jain, Rajeev [Argonne National Lab. (ANL), Argonne, IL (United States); Mahadevan, Vijay [Argonne National Lab. (ANL), Argonne, IL (United States)

    2015-09-30

    This report gives documentation for using MeshKit’s Reactor Geometry (and mesh) Generator (RGG) GUI and also briefly documents other algorithms and tools available in MeshKit. RGG is a program designed to aid in modeling and meshing of complex/large hexagonal and rectilinear reactor cores. RGG uses Argonne’s SIGMA interfaces, Qt and VTK to produce an intuitive user interface. By integrating a 3D view of the reactor with the meshing tools and combining them into one user interface, RGG streamlines the task of preparing a simulation mesh and enables real-time feedback that reduces accidental scripting mistakes that could waste hours of meshing. RGG interfaces with MeshKit tools to consolidate the meshing process, meaning that going from model to mesh is as easy as a button click. This report is designed to explain RGG v 2.0 interface and provide users with the knowledge and skills to pilot RGG successfully. Brief documentation of MeshKit source code, tools and other algorithms available are also presented for developers to extend and add new algorithms to MeshKit. RGG tools work in serial and parallel and have been used to model complex reactor core models consisting of conical pins, load pads, several thousands of axially varying material properties of instrumentation pins and other interstices meshes.

  4. Generation of hybrid meshes for the simulation of petroleum reservoirs; Generation de maillages hybrides pour la simulation de reservoirs petroliers

    Energy Technology Data Exchange (ETDEWEB)

    Balaven-Clermidy, S.

    2001-12-01

    Oil reservoir simulations study multiphase flows in porous media. These flows are described and evaluated through numerical schemes on a discretization of the reservoir domain. In this thesis, we were interested in this spatial discretization and a new kind of hybrid mesh has been proposed where the radial nature of flows in the vicinity of wells is directly taken into account in the geometry. Our modular approach described wells and their drainage area through radial circular meshes. These well meshes are inserted in a structured reservoir mesh (a Corner Point Geometry mesh) made up with hexahedral cells. Finally, in order to generate a global conforming mesh, proper connections are realized between the different kinds of meshes through unstructured transition ones. To compute these transition meshes that we want acceptable in terms of finite volume methods, an automatic method based on power diagrams has been developed. Our approach can deal with a homogeneous anisotropic medium and allows the user to insert vertical or horizontal wells as well as secondary faults in the reservoir mesh. Our work has been implemented, tested and validated in 2D and 2D1/2. It can also be extended in 3D when the geometrical constraints are simplicial ones: points, segments and triangles. (author)

  5. Dynamic Mesh CFD Simulations of Orion Parachute Pendulum Motion During Atmospheric Entry

    Science.gov (United States)

    Halstrom, Logan D.; Schwing, Alan M.; Robinson, Stephen K.

    2016-01-01

    This paper demonstrates the usage of computational fluid dynamics to study the effects of pendulum motion dynamics of the NASAs Orion Multi-Purpose Crew Vehicle parachute system on the stability of the vehicles atmospheric entry and decent. Significant computational fluid dynamics testing has already been performed at NASAs Johnson Space Center, but this study sought to investigate the effect of bulk motion of the parachute, such as pitching, on the induced aerodynamic forces. Simulations were performed with a moving grid geometry oscillating according to the parameters observed in flight tests. As with the previous simulations, OVERFLOW computational fluid dynamics tool is used with the assumption of rigid, non-permeable geometry. Comparison to parachute wind tunnel tests is included for a preliminary validation of the dynamic mesh model. Results show qualitative differences in the flow fields of the static and dynamic simulations and quantitative differences in the induced aerodynamic forces, suggesting that dynamic mesh modeling of the parachute pendulum motion may uncover additional dynamic effects.

  6. Adaptive mesh refinement and adjoint methods in geophysics simulations

    Science.gov (United States)

    Burstedde, Carsten

    2013-04-01

    It is an ongoing challenge to increase the resolution that can be achieved by numerical geophysics simulations. This applies to considering sub-kilometer mesh spacings in global-scale mantle convection simulations as well as to using frequencies up to 1 Hz in seismic wave propagation simulations. One central issue is the numerical cost, since for three-dimensional space discretizations, possibly combined with time stepping schemes, a doubling of resolution can lead to an increase in storage requirements and run time by factors between 8 and 16. A related challenge lies in the fact that an increase in resolution also increases the dimensionality of the model space that is needed to fully parametrize the physical properties of the simulated object (a.k.a. earth). Systems that exhibit a multiscale structure in space are candidates for employing adaptive mesh refinement, which varies the resolution locally. An example that we found well suited is the mantle, where plate boundaries and fault zones require a resolution on the km scale, while deeper area can be treated with 50 or 100 km mesh spacings. This approach effectively reduces the number of computational variables by several orders of magnitude. While in this case it is possible to derive the local adaptation pattern from known physical parameters, it is often unclear what are the most suitable criteria for adaptation. We will present the goal-oriented error estimation procedure, where such criteria are derived from an objective functional that represents the observables to be computed most accurately. Even though this approach is well studied, it is rarely used in the geophysics community. A related strategy to make finer resolution manageable is to design methods that automate the inference of model parameters. Tweaking more than a handful of numbers and judging the quality of the simulation by adhoc comparisons to known facts and observations is a tedious task and fundamentally limited by the turnaround times

  7. DEM GPU studies of industrial scale particle simulations for granular flow civil engineering applications

    Science.gov (United States)

    Pizette, Patrick; Govender, Nicolin; Wilke, Daniel N.; Abriak, Nor-Edine

    2017-06-01

    The use of the Discrete Element Method (DEM) for industrial civil engineering industrial applications is currently limited due to the computational demands when large numbers of particles are considered. The graphics processing unit (GPU) with its highly parallelized hardware architecture shows potential to enable solution of civil engineering problems using discrete granular approaches. We demonstrate in this study the pratical utility of a validated GPU-enabled DEM modeling environment to simulate industrial scale granular problems. As illustration, the flow discharge of storage silos using 8 and 17 million particles is considered. DEM simulations have been performed to investigate the influence of particle size (equivalent size for the 20/40-mesh gravel) and induced shear stress for two hopper shapes. The preliminary results indicate that the shape of the hopper significantly influences the discharge rates for the same material. Specifically, this work shows that GPU-enabled DEM modeling environments can model industrial scale problems on a single portable computer within a day for 30 seconds of process time.

  8. A Tissue Relevance and Meshing Method for Computing Patient-Specific Anatomical Models in Endoscopic Sinus Surgery Simulation

    Science.gov (United States)

    Audette, M. A.; Hertel, I.; Burgert, O.; Strauss, G.

    This paper presents on-going work on a method for determining which subvolumes of a patient-specific tissue map, extracted from CT data of the head, are relevant to simulating endoscopic sinus surgery of that individual, and for decomposing these relevant tissues into triangles and tetrahedra whose mesh size is well controlled. The overall goal is to limit the complexity of the real-time biomechanical interaction while ensuring the clinical relevance of the simulation. Relevant tissues are determined as the union of the pathology present in the patient, of critical tissues deemed to be near the intended surgical path or pathology, and of bone and soft tissue near the intended path, pathology or critical tissues. The processing of tissues, prior to meshing, is based on the Fast Marching method applied under various guises, in a conditional manner that is related to tissue classes. The meshing is based on an adaptation of a meshing method of ours, which combines the Marching Tetrahedra method and the discrete Simplex mesh surface model to produce a topologically faithful surface mesh with well controlled edge and face size as a first stage, and Almost-regular Tetrahedralization of the same prescribed mesh size as a last stage.

  9. A Novel Haptic Interactive Approach to Simulation of Surgery Cutting Based on Mesh and Meshless Models

    Science.gov (United States)

    Liu, Peter X.; Lai, Pinhua; Xu, Shaoping; Zou, Yanni

    2018-01-01

    In the present work, the majority of implemented virtual surgery simulation systems have been based on either a mesh or meshless strategy with regard to soft tissue modelling. To take full advantage of the mesh and meshless models, a novel coupled soft tissue cutting model is proposed. Specifically, the reconstructed virtual soft tissue consists of two essential components. One is associated with surface mesh that is convenient for surface rendering and the other with internal meshless point elements that is used to calculate the force feedback during cutting. To combine two components in a seamless way, virtual points are introduced. During the simulation of cutting, the Bezier curve is used to characterize smooth and vivid incision on the surface mesh. At the same time, the deformation of internal soft tissue caused by cutting operation can be treated as displacements of the internal point elements. Furthermore, we discussed and proved the stability and convergence of the proposed approach theoretically. The real biomechanical tests verified the validity of the introduced model. And the simulation experiments show that the proposed approach offers high computational efficiency and good visual effect, enabling cutting of soft tissue with high stability. PMID:29850006

  10. Conservation Properties of the Hamiltonian Particle-Mesh method for the Quasi-Geostrophic Equations on a sphere

    NARCIS (Netherlands)

    H. Thorsdottir (Halldora)

    2011-01-01

    htmlabstractThe Hamiltonian particle-mesh (HPM) method is used to solve the Quasi-Geostrophic model generalized to a sphere, using the Spherepack modeling package to solve the Helmholtz equation on a colatitude-longitude grid with spherical harmonics. The predicted energy conservation of a

  11. Energy mesh optimization for multi-level calculation schemes

    International Nuclear Information System (INIS)

    Mosca, P.; Taofiki, A.; Bellier, P.; Prevost, A.

    2011-01-01

    The industrial calculations of third generation nuclear reactors are based on sophisticated strategies of homogenization and collapsing at different spatial and energetic levels. An important issue to ensure the quality of these calculation models is the choice of the collapsing energy mesh. In this work, we show a new approach to generate optimized energy meshes starting from the SHEM 281-group library. The optimization model is applied on 1D cylindrical cells and consists of finding an energy mesh which minimizes the errors between two successive collision probability calculations. The former is realized over the fine SHEM mesh with Livolant-Jeanpierre self-shielded cross sections and the latter is performed with collapsed cross sections over the energy mesh being optimized. The optimization is done by the particle swarm algorithm implemented in the code AEMC and multigroup flux solutions are obtained from standard APOLLO2 solvers. By this new approach, a set of new optimized meshes which encompass from 10 to 50 groups has been defined for PWR and BWR calculations. This set will allow users to adapt the energy detail of the solution to the complexity of the calculation (assembly, multi-assembly, two-dimensional whole core). Some preliminary verifications, in which the accuracy of the new meshes is measured compared to a direct 281-group calculation, show that the 30-group optimized mesh offers a good compromise between simulation time and accuracy for a standard 17 x 17 UO 2 assembly with and without control rods. (author)

  12. A Momentum-Exchange/Fictitious Domain-Lattice Boltzmann Method for Solving Particle Suspensions

    Energy Technology Data Exchange (ETDEWEB)

    Jeon, Seok Yun; Yoon, Joon Yong [Hanyang Univ., Seoul (Korea, Republic of); Kim, Chul Kyu [Korea Institute of Civil Engineering and Building Technology, Goyang (Korea, Republic of); Shin, Myung Seob [Korea Intellectual Property Office(KIPO), Daejeon (Korea, Republic of)

    2016-06-15

    This study presents a Lattice Boltzmann Method (LBM) coupled with a momentum-exchange approach/fictitious domain (MEA/FD) method for the simulation of particle suspensions. The method combines the advantages of the LB and the FD methods by using two unrelated meshes, namely, a Eulerian mesh for the flow domain and a Lagrangian mesh for the solid domain. The rigid body conditions are enforced by the momentum-exchange scheme in which the desired value of velocity is imposed directly in the particle inner domain by introducing a pseudo body force to satisfy the constraint of rigid body motion, which is the key idea of a fictitious domain (FD) method. The LB-MEA/FD method has been validated by simulating two different cases, and the results have been compared with those through other methods. The numerical evidence illustrated the capability and robustness of the present method for simulating particle suspensions.

  13. Solving kinetic equations with adaptive mesh in phase space for rarefied gas dynamics and plasma physics (Invited)

    International Nuclear Information System (INIS)

    Kolobov, Vladimir; Arslanbekov, Robert; Frolova, Anna

    2014-01-01

    The paper describes an Adaptive Mesh in Phase Space (AMPS) technique for solving kinetic equations with deterministic mesh-based methods. The AMPS technique allows automatic generation of adaptive Cartesian mesh in both physical and velocity spaces using a Tree-of-Trees data structure. We illustrate advantages of AMPS for simulations of rarefied gas dynamics and electron kinetics on low temperature plasmas. In particular, we consider formation of the velocity distribution functions in hypersonic flows, particle kinetics near oscillating boundaries, and electron kinetics in a radio-frequency sheath. AMPS provide substantial savings in computational cost and increased efficiency of the mesh-based kinetic solvers

  14. Solving kinetic equations with adaptive mesh in phase space for rarefied gas dynamics and plasma physics (Invited)

    Energy Technology Data Exchange (ETDEWEB)

    Kolobov, Vladimir [CFD Research Corporation, Huntsville, AL 35805, USA and The University of Alabama in Huntsville, Huntsville, AL 35805 (United States); Arslanbekov, Robert [CFD Research Corporation, Huntsville, AL 35805 (United States); Frolova, Anna [Computing Center of the Russian Academy of Sciences, Moscow, 119333 (Russian Federation)

    2014-12-09

    The paper describes an Adaptive Mesh in Phase Space (AMPS) technique for solving kinetic equations with deterministic mesh-based methods. The AMPS technique allows automatic generation of adaptive Cartesian mesh in both physical and velocity spaces using a Tree-of-Trees data structure. We illustrate advantages of AMPS for simulations of rarefied gas dynamics and electron kinetics on low temperature plasmas. In particular, we consider formation of the velocity distribution functions in hypersonic flows, particle kinetics near oscillating boundaries, and electron kinetics in a radio-frequency sheath. AMPS provide substantial savings in computational cost and increased efficiency of the mesh-based kinetic solvers.

  15. Simulating control rod and fuel assembly motion using moving meshes

    Energy Technology Data Exchange (ETDEWEB)

    Gilbert, D. [Department of Electrical and Computer Engineering, McMaster University, 1280 Main Street West, Hamilton Ontario, L8S 4K1 (Canada)], E-mail: gilbertdw1@gmail.com; Roman, J.E. [Departamento de Sistemas Informaticos y Computacion, Universidad Politecnica de Valencia, Camino de Vera s/n, 46022 Valencia (Spain); Garland, Wm. J. [Department of Engineering Physics, McMaster University, 1280 Main Street West, Hamilton Ontario, L8S 4K1 (Canada); Poehlman, W.F.S. [Department of Computing and Software, McMaster University, 1280 Main Street West, Hamilton Ontario, L8S 4K1 (Canada)

    2008-02-15

    A prerequisite for designing a transient simulation experiment which includes the motion of control and fuel assemblies is the careful verification of a steady state model which computes k{sub eff} versus assembly insertion distance. Previous studies in nuclear engineering have usually approached the problem of the motion of control rods with the use of nonlinear nodal models. Nodal methods employ special approximations for the leading and trailing cells of the moving assemblies to avoid the rod cusping problem which results from the naive volume weighted cell cross-section approximation. A prototype framework called the MOOSE has been developed for modeling moving components in the presence of diffusion phenomena. A linear finite difference model is constructed, solutions for which are computed by SLEPc, a high performance parallel eigenvalue solver. Design techniques for the implementation of a patched non-conformal mesh which links groups of sub-meshes that can move relative to one another are presented. The generation of matrices which represent moving meshes which conserve neutron current at their boundaries, and the performance of the framework when applied to model reactivity insertion experiments is also discussed.

  16. Modeling and analysis of large-eddy simulations of particle-laden turbulent boundary layer flows

    KAUST Repository

    Rahman, Mustafa M.

    2017-01-05

    We describe a framework for the large-eddy simulation of solid particles suspended and transported within an incompressible turbulent boundary layer (TBL). For the fluid phase, the large-eddy simulation (LES) of incompressible turbulent boundary layer employs stretched spiral vortex subgrid-scale model and a virtual wall model similar to the work of Cheng, Pullin & Samtaney (J. Fluid Mech., 2015). This LES model is virtually parameter free and involves no active filtering of the computed velocity field. Furthermore, a recycling method to generate turbulent inflow is implemented. For the particle phase, the direct quadrature method of moments (DQMOM) is chosen in which the weights and abscissas of the quadrature approximation are tracked directly rather than the moments themselves. The numerical method in this framework is based on a fractional-step method with an energy-conservative fourth-order finite difference scheme on a staggered mesh. This code is parallelized based on standard message passing interface (MPI) protocol and is designed for distributed-memory machines. It is proposed to utilize this framework to examine transport of particles in very large-scale simulations. The solver is validated using the well know result of Taylor-Green vortex case. A large-scale sandstorm case is simulated and the altitude variations of number density along with its fluctuations are quantified.

  17. Accurate halo-galaxy mocks from automatic bias estimation and particle mesh gravity solvers

    Science.gov (United States)

    Vakili, Mohammadjavad; Kitaura, Francisco-Shu; Feng, Yu; Yepes, Gustavo; Zhao, Cheng; Chuang, Chia-Hsun; Hahn, ChangHoon

    2017-12-01

    Reliable extraction of cosmological information from clustering measurements of galaxy surveys requires estimation of the error covariance matrices of observables. The accuracy of covariance matrices is limited by our ability to generate sufficiently large number of independent mock catalogues that can describe the physics of galaxy clustering across a wide range of scales. Furthermore, galaxy mock catalogues are required to study systematics in galaxy surveys and to test analysis tools. In this investigation, we present a fast and accurate approach for generation of mock catalogues for the upcoming galaxy surveys. Our method relies on low-resolution approximate gravity solvers to simulate the large-scale dark matter field, which we then populate with haloes according to a flexible non-linear and stochastic bias model. In particular, we extend the PATCHY code with an efficient particle mesh algorithm to simulate the dark matter field (the FASTPM code), and with a robust MCMC method relying on the EMCEE code for constraining the parameters of the bias model. Using the haloes in the BigMultiDark high-resolution N-body simulation as a reference catalogue, we demonstrate that our technique can model the bivariate probability distribution function (counts-in-cells), power spectrum and bispectrum of haloes in the reference catalogue. Specifically, we show that the new ingredients permit us to reach percentage accuracy in the power spectrum up to k ∼ 0.4 h Mpc-1 (within 5 per cent up to k ∼ 0.6 h Mpc-1) with accurate bispectra improving previous results based on Lagrangian perturbation theory.

  18. Improvement of the homogeneity of high mobility In{sub 2}O{sub 3}:H films by sputtering through a mesh electrode studied by Monte Carlo simulation and thin film analysis

    Energy Technology Data Exchange (ETDEWEB)

    Scherg-Kurmes, Harald; Hafez, Ahmad; Szyszka, Bernd [Technische Universitaet Berlin, Einsteinufer 25, 10587, Berlin (Germany); Siemers, Michael; Pflug, Andreas [Fraunhofer IST, Bienroder Weg 54E, 38108, Braunschweig (Germany); Schlatmann, Rutger [Helmholtz Zentrum Berlin, PVcomB, Schwarzschildstr. 3, 12489, Berlin (Germany); Rech, Bernd [Helmholtz Zentrum Berlin, Institute for Silicon Photovoltaics, Kekulestrasse 5, 12489, Berlin (Germany)

    2016-09-15

    Hydrogen-doped indium oxide (IOH) is a transparent conductive oxide offering great potential to optoelectronic applications because of its high mobility of over 100 cm{sup 2} V{sup -1}s{sup -1}. In films deposited statically by RF magnetron sputtering, a small area directly opposing the target center with a higher resistivity and lower crystallinity than the rest of the film has been found via hall- and XRD-measurements, which we attribute to plasma damage. In order to investigate the distribution of particle energies during the sputtering process we have simulated the RF-sputtering deposition process of IOH by particle-in-cell Monte Carlo (PICMC) simulation. At the surface of ceramic sputtering targets, negatively charged oxygen ions are created. These ions are accelerated toward the substrate in the plasma sheath with energies up to 150 eV. They damage the growing film and reduce its crystallinity. The influence of a negatively biased mesh inside the sputtering chamber on particle energies and distributions has been simulated and investigated. We found that the mesh decreased the high-energetic oxygen ion density at the substrate, thus enabling a more homogeneous IOH film growth. The theoretical results have been verified by XRD X-ray diffractometry (XRD), 4-point probe, and hall measurements of statically deposited IOH films on glass. (copyright 2016 WILEY-VCH Verlag GmbH and Co. KGaA, Weinheim)

  19. BOA, Beam Optics Analyzer A Particle-In-Cell Code

    International Nuclear Information System (INIS)

    Bui, Thuc

    2007-01-01

    The program was tasked with implementing time dependent analysis of charges particles into an existing finite element code with adaptive meshing, called Beam Optics Analyzer (BOA). BOA was initially funded by a DOE Phase II program to use the finite element method with adaptive meshing to track particles in unstructured meshes. It uses modern programming techniques, state-of-the-art data structures, so that new methods, features and capabilities are easily added and maintained. This Phase II program was funded to implement plasma simulations in BOA and extend its capabilities to model thermal electrons, secondary emissions, self magnetic field and implement a more comprehensive post-processing and feature-rich GUI. The program was successful in implementing thermal electrons, secondary emissions, and self magnetic field calculations. The BOA GUI was also upgraded significantly, and CCR is receiving interest from the microwave tube and semiconductor equipment industry for the code. Implementation of PIC analysis was partially successful. Computational resource requirements for modeling more than 2000 particles begin to exceed the capability of most readily available computers. Modern plasma analysis typically requires modeling of approximately 2 million particles or more. The problem is that tracking many particles in an unstructured mesh that is adapting becomes inefficient. In particular memory requirements become excessive. This probably makes particle tracking in unstructured meshes currently unfeasible with commonly available computer resources. Consequently, Calabazas Creek Research, Inc. is exploring hybrid codes where the electromagnetic fields are solved on the unstructured, adaptive mesh while particles are tracked on a fixed mesh. Efficient interpolation routines should be able to transfer information between nodes of the two meshes. If successfully developed, this could provide high accuracy and reasonable computational efficiency.

  20. Finite element meshing approached as a global minimization process

    Energy Technology Data Exchange (ETDEWEB)

    WITKOWSKI,WALTER R.; JUNG,JOSEPH; DOHRMANN,CLARK R.; LEUNG,VITUS J.

    2000-03-01

    The ability to generate a suitable finite element mesh in an automatic fashion is becoming the key to being able to automate the entire engineering analysis process. However, placing an all-hexahedron mesh in a general three-dimensional body continues to be an elusive goal. The approach investigated in this research is fundamentally different from any other that is known of by the authors. A physical analogy viewpoint is used to formulate the actual meshing problem which constructs a global mathematical description of the problem. The analogy used was that of minimizing the electrical potential of a system charged particles within a charged domain. The particles in the presented analogy represent duals to mesh elements (i.e., quads or hexes). Particle movement is governed by a mathematical functional which accounts for inter-particles repulsive, attractive and alignment forces. This functional is minimized to find the optimal location and orientation of each particle. After the particles are connected a mesh can be easily resolved. The mathematical description for this problem is as easy to formulate in three-dimensions as it is in two- or one-dimensions. The meshing algorithm was developed within CoMeT. It can solve the two-dimensional meshing problem for convex and concave geometries in a purely automated fashion. Investigation of the robustness of the technique has shown a success rate of approximately 99% for the two-dimensional geometries tested. Run times to mesh a 100 element complex geometry were typically in the 10 minute range. Efficiency of the technique is still an issue that needs to be addressed. Performance is an issue that is critical for most engineers generating meshes. It was not for this project. The primary focus of this work was to investigate and evaluate a meshing algorithm/philosophy with efficiency issues being secondary. The algorithm was also extended to mesh three-dimensional geometries. Unfortunately, only simple geometries were tested

  1. A 3d particle simulation code for heavy ion fusion accelerator studies

    International Nuclear Information System (INIS)

    Friedman, A.; Bangerter, R.O.; Callahan, D.A.; Grote, D.P.; Langdon, A.B.; Haber, I.

    1990-01-01

    We describe WARP, a new particle-in-cell code being developed and optimized for ion beam studies in true geometry. We seek to model transport around bends, axial compression with strong focusing, multiple beamlet interaction, and other inherently 3d processes that affect emittance growth. Constraints imposed by memory and running time are severe. Thus, we employ only two 3d field arrays (ρ and φ), and difference φ directly on each particle to get E, rather than interpolating E from three meshes; use of a single 3d array is feasible. A new method for PIC simulation of bent beams follows the beam particles in a family of rotated laboratory frames, thus ''straightening'' the bends. We are also incorporating an envelope calculation, an (r, z) model, and 1d (axial) model within WARP. The BASIS development and run-time system is used, providing a powerful interactive environment in which the user has access to all variables in the code database. 10 refs., 3 figs

  2. Enhanced stopping of macro-particles in particle-in-cell simulations

    International Nuclear Information System (INIS)

    May, J.; Tonge, J.; Ellis, I.; Mori, W. B.; Fiuza, F.; Fonseca, R. A.; Silva, L. O.; Ren, C.

    2014-01-01

    We derive an equation for energy transfer from relativistic charged particles to a cold background plasma appropriate for finite-size particles that are used in particle-in-cell simulation codes. Expressions for one-, two-, and three-dimensional particles are presented, with special attention given to the two-dimensional case. This energy transfer is due to the electric field of the wake set up in the background plasma by the relativistic particle. The enhanced stopping is dependent on the q 2 /m, where q is the charge and m is the mass of the relativistic particle, and therefore simulation macro-particles with large charge but identical q/m will stop more rapidly. The stopping power also depends on the effective particle shape of the macro-particle. These conclusions are verified in particle-in-cell simulations. We present 2D simulations of test particles, relaxation of high-energy tails, and integrated fast ignition simulations showing that the enhanced drag on macro-particles may adversely affect the results of these simulations in a wide range of high-energy density plasma scenarios. We also describe a particle splitting algorithm which can potentially overcome this problem and show its effect in controlling the stopping of macro-particles

  3. Numerical simulation of flow induced by a pitched blade turbine. Comparison of the sliding mesh technique and an averaged source term method

    Energy Technology Data Exchange (ETDEWEB)

    Majander, E.O.J.; Manninen, M.T. [VTT Energy, Espoo (Finland)

    1996-12-31

    The flow induced by a pitched blade turbine was simulated using the sliding mesh technique. The detailed geometry of the turbine was modelled in a computational mesh rotating with the turbine and the geometry of the reactor including baffles was modelled in a stationary co-ordinate system. Effects of grid density were investigated. Turbulence was modelled by using the standard k-{epsilon} model. Results were compared to experimental observations. Velocity components were found to be in good agreement with the measured values throughout the tank. Averaged source terms were calculated from the sliding mesh simulations in order to investigate the reliability of the source term approach. The flow field in the tank was then simulated in a simple grid using these source terms. Agreement with the results of the sliding mesh simulations was good. Commercial CFD-code FLUENT was used in all simulations. (author)

  4. Numerical simulation of flow induced by a pitched blade turbine. Comparison of the sliding mesh technique and an averaged source term method

    Energy Technology Data Exchange (ETDEWEB)

    Majander, E O.J.; Manninen, M T [VTT Energy, Espoo (Finland)

    1997-12-31

    The flow induced by a pitched blade turbine was simulated using the sliding mesh technique. The detailed geometry of the turbine was modelled in a computational mesh rotating with the turbine and the geometry of the reactor including baffles was modelled in a stationary co-ordinate system. Effects of grid density were investigated. Turbulence was modelled by using the standard k-{epsilon} model. Results were compared to experimental observations. Velocity components were found to be in good agreement with the measured values throughout the tank. Averaged source terms were calculated from the sliding mesh simulations in order to investigate the reliability of the source term approach. The flow field in the tank was then simulated in a simple grid using these source terms. Agreement with the results of the sliding mesh simulations was good. Commercial CFD-code FLUENT was used in all simulations. (author)

  5. An adaptive mesh refinement-multiphase lattice Boltzmann flux solver for simulation of complex binary fluid flows

    Science.gov (United States)

    Yuan, H. Z.; Wang, Y.; Shu, C.

    2017-12-01

    This paper presents an adaptive mesh refinement-multiphase lattice Boltzmann flux solver (AMR-MLBFS) for effective simulation of complex binary fluid flows at large density ratios. In this method, an AMR algorithm is proposed by introducing a simple indicator on the root block for grid refinement and two possible statuses for each block. Unlike available block-structured AMR methods, which refine their mesh by spawning or removing four child blocks simultaneously, the present method is able to refine its mesh locally by spawning or removing one to four child blocks independently when the refinement indicator is triggered. As a result, the AMR mesh used in this work can be more focused on the flow region near the phase interface and its size is further reduced. In each block of mesh, the recently proposed MLBFS is applied for the solution of the flow field and the level-set method is used for capturing the fluid interface. As compared with existing AMR-lattice Boltzmann models, the present method avoids both spatial and temporal interpolations of density distribution functions so that converged solutions on different AMR meshes and uniform grids can be obtained. The proposed method has been successfully validated by simulating a static bubble immersed in another fluid, a falling droplet, instabilities of two-layered fluids, a bubble rising in a box, and a droplet splashing on a thin film with large density ratios and high Reynolds numbers. Good agreement with the theoretical solution, the uniform-grid result, and/or the published data has been achieved. Numerical results also show its effectiveness in saving computational time and virtual memory as compared with computations on uniform meshes.

  6. MHD simulations on an unstructured mesh

    International Nuclear Information System (INIS)

    Strauss, H.R.; Park, W.; Belova, E.; Fu, G.Y.; Sugiyama, L.E.

    1998-01-01

    Two reasons for using an unstructured computational mesh are adaptivity, and alignment with arbitrarily shaped boundaries. Two codes which use finite element discretization on an unstructured mesh are described. FEM3D solves 2D and 3D RMHD using an adaptive grid. MH3D++, which incorporates methods of FEM3D into the MH3D generalized MHD code, can be used with shaped boundaries, which might be 3D

  7. Automatic, unstructured mesh optimization for simulation and assessment of tide- and surge-driven hydrodynamics in a longitudinal estuary: St. Johns River

    Science.gov (United States)

    Bacopoulos, Peter

    2018-05-01

    A localized truncation error analysis with complex derivatives (LTEA+CD) is applied recursively with advanced circulation (ADCIRC) simulations of tides and storm surge for finite element mesh optimization. Mesh optimization is demonstrated with two iterations of LTEA+CD for tidal simulation in the lower 200 km of the St. Johns River, located in northeast Florida, and achieves more than an over 50% decrease in the number of mesh nodes, relating to a twofold increase in efficiency, at a zero cost to model accuracy. The recursively generated meshes using LTEA+CD lead to successive reductions in the global cumulative truncation error associated with the model mesh. Tides are simulated with root mean square error (RMSE) of 0.09-0.21 m and index of agreement (IA) values generally in the 80s and 90s percentage ranges. Tidal currents are simulated with RMSE of 0.09-0.23 m s-1 and IA values of 97% and greater. Storm tide due to Hurricane Matthew 2016 is simulated with RMSE of 0.09-0.33 m and IA values of 75-96%. Analysis of the LTEA+CD results shows the M2 constituent to dominate the node spacing requirement in the St. Johns River, with the M4 and M6 overtides and the STEADY constituent contributing some. Friction is the predominant physical factor influencing the target element size distribution, especially along the main river stem, while frequency (inertia) and Coriolis (rotation) are supplementary contributing factors. The combination of interior- and boundary-type computational molecules, providing near-full coverage of the model domain, renders LTEA+CD an attractive mesh generation/optimization tool for complex coastal and estuarine domains. The mesh optimization procedure using LTEA+CD is automatic and extensible to other finite element-based numerical models. Discussion is provided on the scope of LTEA+CD, the starting point (mesh) of the procedure, the user-specified scaling of the LTEA+CD results, and the iteration (termination) of LTEA+CD for mesh optimization.

  8. Implicit Geometry Meshing for the simulation of Rotary Friction Welding

    Science.gov (United States)

    Schmicker, D.; Persson, P.-O.; Strackeljan, J.

    2014-08-01

    The simulation of Rotary Friction Welding (RFW) is a challenging task, since it states a coupled problem of phenomena like large plastic deformations, heat flux, contact and friction. In particular the mesh generation and its restoration when using a Lagrangian description of motion is of significant severity. In this regard Implicit Geometry Meshing (IGM) algorithms are promising alternatives to the more conventional explicit methods. Because of the implicit description of the geometry during remeshing, the IGM procedure turns out to be highly robust and generates spatial discretizations of high quality regardless of the complexity of the flash shape and its inclusions. A model for efficient RFW simulation is presented, which is based on a Carreau fluid law, an Augmented Lagrange approach in mapping the incompressible deformations, a penalty contact approach, a fully regularized Coulomb-/fluid friction law and a hybrid time integration strategy. The implementation of the IGM algorithm using 6-node triangular finite elements is described in detail. The techniques are demonstrated on a fairly complex friction welding problem, demonstrating the performance and the potentials of the proposed method. The techniques are general and straight-forward to implement, and offer the potential of successful adoption to a wide range of other engineering problems.

  9. Reinforcing effect of glass-fiber mesh on complete dentures in a test model with a simulated oral mucosa.

    Science.gov (United States)

    Yu, Sang-Hui; Oh, Seunghan; Cho, Hye-Won; Bae, Ji-Myung

    2017-11-01

    Studies that evaluated the strength of complete dentures reinforced with glass-fiber mesh or metal mesh on a cast with a simulated oral mucosa are lacking. The purpose of this in vitro study was to compare the mechanical properties of maxillary complete dentures reinforced with glass-fiber mesh with those of metal mesh in a new test model, using a simulated oral mucosa. Complete dentures reinforced with 2 types of glass-fiber mesh, SES mesh (SES) and glass cloth (GC) and metal mesh (metal) were fabricated. Complete dentures without any reinforcement were prepared as a control (n=10). The complete dentures were located on a cast with a simulated oral mucosa, and a load was applied on the posterior artificial teeth bilaterally. The fracture load, elastic modulus, and toughness of a complete denture were measured using a universal testing machine at a crosshead speed of 5 mm/min. The fracture load and elastic modulus were analyzed using 1-way analysis of variance, and the toughness was analyzed with the Kruskal-Wallis test (α=.05). The Tukey multiple range test was used as a post hoc test. The fracture load and toughness of the SES group was significantly higher than that of the metal and control groups (P<.05) but not significantly different from that of the GC group. The elastic modulus of the metal group was significantly higher than that of the control group (P<.05), and no significant differences were observed in the SES and GC groups. Compared with the control group, the fracture load and toughness of the SES and GC groups were higher, while those of the metal group were not significantly different. Copyright © 2017 Editorial Council for the Journal of Prosthetic Dentistry. Published by Elsevier Inc. All rights reserved.

  10. Simulation of bluff-body flows using iterative penalization in a multiresolution particle-mesh vortex method

    DEFF Research Database (Denmark)

    Spietz, Henrik Juul; Hejlesen, Mads Mølholm; Walther, Jens Honore

    in the oncoming flow. This may lead to structural instability e.g. when the shedding frequency aligns with the natural frequency of the structure. Fluid structure interaction must especially be considered when designing long span bridges. A three dimensional vortex-in-cell method is applied for the direct......The ability to predict aerodynamic forces, due to the interaction of a fluid flow with a solid body, is central in many fields of engineering and is necessary to identify error-prone structural designs. In bluff-body flows the aerodynamic forces oscillate due to vortex shedding and variations...... numerical simulation of the flow past a bodies of arbitrary shape. Vortex methods use a simple formulation where only the trajectories of discrete vortex particles are simulated. The Lagrangian formulation eliminates the CFL type condition that Eulerian methods have to satisfy. This allows vortex methods...

  11. Lattice-Boltzmann Method with Dynamic Grid Refinement for Simulating Particle Deposition on a Single Fibre

    Directory of Open Access Journals (Sweden)

    Helmut Schomburg

    2013-03-01

    Full Text Available In this work a numerical approach to predict the deposition behaviour of nano-scale particles on the surface of a single fibre by resolving the resulting dendrite-like particle structures in detail is presented. The gas flow simulation is carried out by a two-dimensional Lattice-Boltzmann method, which is coupled with a Lagrangian approach for the particle motion. To decrease calculation time and system requirements the Lattice-Boltzmann model is extended to allow for local grid refinement. Because of the a priori unknown location of deposition, the simulation procedure starts on a coarse mesh which is then locally refined in a fully adaptive way in regions of accumulated particles. After each deposition the fluid flow is recalculated in order to resolve the coupling of the flow with the growing particle structures correctly. For the purpose of avoiding unphysical blocking of flow by growing particle dendrites the Lattice-Boltzmann method is extended to permeable cells in these regions using the Brinkmann equation. This extended deposition model is compared to simpler approaches, where the deposit has no retroaction on the flow or is treated as a solid structure. It is clear that the permeable model is most realistic and allows considering the particle deposition on a fibre as two-dimensional problem. Comprehensive simulations were conducted for analysing the importance of different parameters, i.e. free-stream velocity and particle diameter on the deposit structure. The results of this sensitivity analysis agree qualitatively well with former published numerical and experimental results. Finally the structure of the particle deposit was quantitatively characterised by using a modified fractal dimension.

  12. MeshVoro: A Three-Dimensional Voronoi Mesh Building Tool for the TOUGH Family of Codes

    Energy Technology Data Exchange (ETDEWEB)

    Freeman, C. M.; Boyle, K. L.; Reagan, M.; Johnson, J.; Rycroft, C.; Moridis, G. J.

    2013-09-30

    Few tools exist for creating and visualizing complex three-dimensional simulation meshes, and these have limitations that restrict their application to particular geometries and circumstances. Mesh generation needs to trend toward ever more general applications. To that end, we have developed MeshVoro, a tool that is based on the Voro (Rycroft 2009) library and is capable of generating complex threedimensional Voronoi tessellation-based (unstructured) meshes for the solution of problems of flow and transport in subsurface geologic media that are addressed by the TOUGH (Pruess et al. 1999) family of codes. MeshVoro, which includes built-in data visualization routines, is a particularly useful tool because it extends the applicability of the TOUGH family of codes by enabling the scientifically robust and relatively easy discretization of systems with challenging 3D geometries. We describe several applications of MeshVoro. We illustrate the ability of the tool to straightforwardly transform a complex geological grid into a simulation mesh that conforms to the specifications of the TOUGH family of codes. We demonstrate how MeshVoro can describe complex system geometries with a relatively small number of grid blocks, and we construct meshes for geometries that would have been practically intractable with a standard Cartesian grid approach. We also discuss the limitations and appropriate applications of this new technology.

  13. Visualization of Octree Adaptive Mesh Refinement (AMR) in Astrophysical Simulations

    Science.gov (United States)

    Labadens, M.; Chapon, D.; Pomaréde, D.; Teyssier, R.

    2012-09-01

    Computer simulations are important in current cosmological research. Those simulations run in parallel on thousands of processors, and produce huge amount of data. Adaptive mesh refinement is used to reduce the computing cost while keeping good numerical accuracy in regions of interest. RAMSES is a cosmological code developed by the Commissariat à l'énergie atomique et aux énergies alternatives (English: Atomic Energy and Alternative Energies Commission) which uses Octree adaptive mesh refinement. Compared to grid based AMR, the Octree AMR has the advantage to fit very precisely the adaptive resolution of the grid to the local problem complexity. However, this specific octree data type need some specific software to be visualized, as generic visualization tools works on Cartesian grid data type. This is why the PYMSES software has been also developed by our team. It relies on the python scripting language to ensure a modular and easy access to explore those specific data. In order to take advantage of the High Performance Computer which runs the RAMSES simulation, it also uses MPI and multiprocessing to run some parallel code. We would like to present with more details our PYMSES software with some performance benchmarks. PYMSES has currently two visualization techniques which work directly on the AMR. The first one is a splatting technique, and the second one is a custom ray tracing technique. Both have their own advantages and drawbacks. We have also compared two parallel programming techniques with the python multiprocessing library versus the use of MPI run. The load balancing strategy has to be smartly defined in order to achieve a good speed up in our computation. Results obtained with this software are illustrated in the context of a massive, 9000-processor parallel simulation of a Milky Way-like galaxy.

  14. A multiresolution remeshed particle vortex method using patches

    DEFF Research Database (Denmark)

    Rasmussen, Johannes Tophøj; Cottet, George-Henri; Walther, Jens Honore

    vortex particle-mesh VIC algorithm interpolates particle vorticity to a mesh, solves a Poisson equation for the stream function using FFTs and calculates velocities as the curl of the stream function. With both vorticity and velocity available on the mesh, values of the substantial derivative of particle...... implementation with patches of varying resolution, is applied to the two-dimensional flow past a cylinder. The vorticity field can be divided into two regions, an arbitrary patch of vorticity and the remaining exterior vorticity field. Due to the linearity of the Poisson equation the velocity field corresponding...... to the total vorticity field is the sum of the free space solutions to the Poisson equation to each region. Hereby the flow on the patch can be simulated at a higher resolution, while including the influence from the coarser exterior region. Particles are remeshed and interpolated only to the region from which...

  15. Influence of mesh structure on 2D full shallow water equations and SCS Curve Number simulation of rainfall/runoff events

    Science.gov (United States)

    Caviedes-Voullième, Daniel; García-Navarro, Pilar; Murillo, Javier

    2012-07-01

    SummaryHydrological simulation of rain-runoff processes is often performed with lumped models which rely on calibration to generate storm hydrographs and study catchment response to rain. In this paper, a distributed, physically-based numerical model is used for runoff simulation in a mountain catchment. This approach offers two advantages. The first is that by using shallow-water equations for runoff flow, there is less freedom to calibrate routing parameters (as compared to, for example, synthetic hydrograph methods). The second, is that spatial distributions of water depth and velocity can be obtained. Furthermore, interactions among the various hydrological processes can be modeled in a physically-based approach which may depend on transient and spatially distributed factors. On the other hand, the undertaken numerical approach relies on accurate terrain representation and mesh selection, which also affects significantly the computational cost of the simulations. Hence, we investigate the response of a gauged catchment with this distributed approach. The methodology consists of analyzing the effects that the mesh has on the simulations by using a range of meshes. Next, friction is applied to the model and the response to variations and interaction with the mesh is studied. Finally, a first approach with the well-known SCS Curve Number method is studied to evaluate its behavior when coupled with a shallow-water model for runoff flow. The results show that mesh selection is of great importance, since it may affect the results in a magnitude as large as physical factors, such as friction. Furthermore, results proved to be less sensitive to roughness spatial distribution than to mesh properties. Finally, the results indicate that SCS-CN may not be suitable for simulating hydrological processes together with a shallow-water model.

  16. Systemic characterization and evaluation of particle packings as initial sets for discrete element simulations

    Science.gov (United States)

    Morfa, Carlos Recarey; Cortés, Lucía Argüelles; Farias, Márcio Muniz de; Morales, Irvin Pablo Pérez; Valera, Roberto Roselló; Oñate, Eugenio

    2018-07-01

    A methodology that comprises several characterization properties for particle packings is proposed in this paper. The methodology takes into account factors such as dimension and shape of particles, space occupation, homogeneity, connectivity and isotropy, among others. This classification and integration of several properties allows to carry out a characterization process to systemically evaluate the particle packings in order to guarantee the quality of the initial meshes in discrete element simulations, in both the micro- and the macroscales. Several new properties were created, and improvements in existing ones are presented. Properties from other disciplines were adapted to be used in the evaluation of particle systems. The methodology allows to easily characterize media at the level of the microscale (continuous geometries—steels, rocks microstructures, etc., and discrete geometries) and the macroscale. A global, systemic and integral system for characterizing and evaluating particle sets, based on fuzzy logic, is presented. Such system allows researchers to have a unique evaluation criterion based on the aim of their research. Examples of applications are shown.

  17. Systemic characterization and evaluation of particle packings as initial sets for discrete element simulations

    Science.gov (United States)

    Morfa, Carlos Recarey; Cortés, Lucía Argüelles; Farias, Márcio Muniz de; Morales, Irvin Pablo Pérez; Valera, Roberto Roselló; Oñate, Eugenio

    2017-10-01

    A methodology that comprises several characterization properties for particle packings is proposed in this paper. The methodology takes into account factors such as dimension and shape of particles, space occupation, homogeneity, connectivity and isotropy, among others. This classification and integration of several properties allows to carry out a characterization process to systemically evaluate the particle packings in order to guarantee the quality of the initial meshes in discrete element simulations, in both the micro- and the macroscales. Several new properties were created, and improvements in existing ones are presented. Properties from other disciplines were adapted to be used in the evaluation of particle systems. The methodology allows to easily characterize media at the level of the microscale (continuous geometries—steels, rocks microstructures, etc., and discrete geometries) and the macroscale. A global, systemic and integral system for characterizing and evaluating particle sets, based on fuzzy logic, is presented. Such system allows researchers to have a unique evaluation criterion based on the aim of their research. Examples of applications are shown.

  18. MUSIC: a mesh-unrestricted simulation code

    International Nuclear Information System (INIS)

    Bonalumi, R.A.; Rouben, B.; Dastur, A.R.; Dondale, C.S.; Li, H.Y.H.

    1978-01-01

    A general formalism to solve the G-group neutron diffusion equation is described. The G-group flux is represented by complementing an ''asymptotic'' mode with (G-1) ''transient'' modes. A particular reduction-to-one-group technique gives a high computational efficiency. MUSIC, a 2-group code using the above formalism, is presented. MUSIC is demonstrated on a fine-mesh calculation and on 2 coarse-mesh core calculations: a heavy-water reactor (HWR) problem and the 2-D lightwater reactor (LWR) IAEA benchmark. Comparison is made to finite-difference results

  19. Adaptive hybrid mesh refinement for multiphysics applications

    International Nuclear Information System (INIS)

    Khamayseh, Ahmed; Almeida, Valmor de

    2007-01-01

    The accuracy and convergence of computational solutions of mesh-based methods is strongly dependent on the quality of the mesh used. We have developed methods for optimizing meshes that are comprised of elements of arbitrary polygonal and polyhedral type. We present in this research the development of r-h hybrid adaptive meshing technology tailored to application areas relevant to multi-physics modeling and simulation. Solution-based adaptation methods are used to reposition mesh nodes (r-adaptation) or to refine the mesh cells (h-adaptation) to minimize solution error. The numerical methods perform either the r-adaptive mesh optimization or the h-adaptive mesh refinement method on the initial isotropic or anisotropic meshes to equidistribute weighted geometric and/or solution error function. We have successfully introduced r-h adaptivity to a least-squares method with spherical harmonics basis functions for the solution of the spherical shallow atmosphere model used in climate modeling. In addition, application of this technology also covers a wide range of disciplines in computational sciences, most notably, time-dependent multi-physics, multi-scale modeling and simulation

  20. New methods to interpolate large volume of data from points or particles (Mesh-Free) methods application for its scientific visualization

    International Nuclear Information System (INIS)

    Reyes Lopez, Y.; Yervilla Herrera, H.; Viamontes Esquivel, A.; Recarey Morfa, C. A.

    2009-01-01

    In the following paper we developed a new method to interpolate large volumes of scattered data, focused mainly on the results of the Mesh free Methods, Points Methods and the Particles Methods application. Through this one, we use local radial basis function as interpolating functions. We also use over-tree as the data structure that allows to accelerate the localization of the data that influences to interpolate the values at a new point, speeding up the application of scientific visualization techniques to generate images from large data volumes from the application of Mesh-free Methods, Points and Particle Methods, in the resolution of diverse models of physics-mathematics. As an example, the results obtained after applying this method using the local interpolation functions of Shepard are shown. (Author) 22 refs

  1. Revisiting the use of the immersed-boundary lattice-Boltzmann method for simulations of suspended particles

    Science.gov (United States)

    Mountrakis, L.; Lorenz, E.; Hoekstra, A. G.

    2017-07-01

    The immersed-boundary lattice-Boltzmann method (IB-LBM) is increasingly being used in simulations of dense suspensions. These systems are computationally very expensive and can strongly benefit from lower resolutions that still maintain the desired accuracy for the quantities of interest. IB-LBM has a number of free parameters that have to be defined, often without exact knowledge of the tradeoffs, since their behavior in low resolutions is not well understood. Such parameters are the lattice constant Δ x , the number of vertices Nv, the interpolation kernel ϕ , and the LBM relaxation time τ . We investigate the effect of these IB-LBM parameters on a number of straightforward but challenging benchmarks. The systems considered are (a) the flow of a single sphere in shear flow, (b) the collision of two spheres in shear flow, and (c) the lubrication interaction of two spheres. All benchmarks are performed in three dimensions. The first two systems are used for determining two effective radii: the hydrodynamic radius rhyd and the particle interaction radius rinter. The last system is used to establish the numerical robustness of the lubrication forces, used to probe the hydrodynamic interactions in the limit of small gaps. Our results show that lower spatial resolutions result in larger hydrodynamic and interaction radii, while surface densities should be chosen above two vertices per LU2 result to prevent fluid penetration in underresolved meshes. Underresolved meshes also failed to produce the migration of particles toward the center of the domain due to lift forces in Couette flow, mostly noticeable for IBM-kernel ϕ2. Kernel ϕ4, despite being more robust toward mesh resolution, produces a notable membrane thickness, leading to the breakdown of the lubrication forces in larger gaps, and its use in dense suspensions where the mean particle distances are small can result in undesired behavior. rhyd is measured to be different from rinter, suggesting that there is

  2. Unstructured mesh adaptivity for urban flooding modelling

    Science.gov (United States)

    Hu, R.; Fang, F.; Salinas, P.; Pain, C. C.

    2018-05-01

    Over the past few decades, urban floods have been gaining more attention due to their increase in frequency. To provide reliable flooding predictions in urban areas, various numerical models have been developed to perform high-resolution flood simulations. However, the use of high-resolution meshes across the whole computational domain causes a high computational burden. In this paper, a 2D control-volume and finite-element flood model using adaptive unstructured mesh technology has been developed. This adaptive unstructured mesh technique enables meshes to be adapted optimally in time and space in response to the evolving flow features, thus providing sufficient mesh resolution where and when it is required. It has the advantage of capturing the details of local flows and wetting and drying front while reducing the computational cost. Complex topographic features are represented accurately during the flooding process. For example, the high-resolution meshes around the buildings and steep regions are placed when the flooding water reaches these regions. In this work a flooding event that happened in 2002 in Glasgow, Scotland, United Kingdom has been simulated to demonstrate the capability of the adaptive unstructured mesh flooding model. The simulations have been performed using both fixed and adaptive unstructured meshes, and then results have been compared with those published 2D and 3D results. The presented method shows that the 2D adaptive mesh model provides accurate results while having a low computational cost.

  3. Enhanced Computer Aided Simulation of Meshing and Contact With Application for Spiral Bevel Gear Drives

    National Research Council Canada - National Science Library

    Litvin, F

    1999-01-01

    An integrated tooth contact analysis (TCA) computer program for the simulation of meshing and contact of gear drives that calculates transmission errors and shift of hearing contact for misaligned gear drives has been developed...

  4. Impact of mesh and DEM resolutions in SEM simulation of 3D seismic response

    NARCIS (Netherlands)

    Khan, Saad; van der Meijde, M.; van der Werff, H.M.A.; Shafique, Muhammad

    2017-01-01

    This study shows that the resolution of a digital elevation model (DEM) and model mesh strongly influences 3D simulations of seismic response. Topographic heterogeneity scatters seismic waves and causes variation in seismic response (am-plification and deamplification of seismic amplitudes) at the

  5. Direct numerical simulations of agglomeration of circular colloidal particles in two-dimensional shear flow

    International Nuclear Information System (INIS)

    Choi, Young Joon; Djilali, Ned

    2016-01-01

    Colloidal agglomeration of nanoparticles in shear flow is investigated by solving the fluid-particle and particle-particle interactions in a 2D system. We use an extended finite element method in which the dynamics of the particles is solved in a fully coupled manner with the flow, allowing an accurate description of the fluid-particle interfaces without the need of boundary-fitted meshes or of empirical correlations to account for the hydrodynamic interactions between the particles. Adaptive local mesh refinement using a grid deformation method is incorporated with the fluid-structure interaction algorithm, and the particle-particle interaction at the microscopic level is modeled using the Lennard-Jones potential. Motivated by the process used in fabricating fuel cell catalysts from a colloidal ink, the model is applied to investigate agglomeration of colloidal particles under external shear flow in a sliding bi-periodic Lees-Edwards frame with varying shear rates and particle fraction ratios. Both external shear and particle fraction are found to have a crucial impact on the structure formation of colloidal particles in a suspension. Segregation intensity and graph theory are used to analyze the underlying agglomeration patterns and structures, and three agglomeration regimes are identified

  6. Parallel Performance Optimizations on Unstructured Mesh-based Simulations

    Energy Technology Data Exchange (ETDEWEB)

    Sarje, Abhinav; Song, Sukhyun; Jacobsen, Douglas; Huck, Kevin; Hollingsworth, Jeffrey; Malony, Allen; Williams, Samuel; Oliker, Leonid

    2015-01-01

    © The Authors. Published by Elsevier B.V. This paper addresses two key parallelization challenges the unstructured mesh-based ocean modeling code, MPAS-Ocean, which uses a mesh based on Voronoi tessellations: (1) load imbalance across processes, and (2) unstructured data access patterns, that inhibit intra- and inter-node performance. Our work analyzes the load imbalance due to naive partitioning of the mesh, and develops methods to generate mesh partitioning with better load balance and reduced communication. Furthermore, we present methods that minimize both inter- and intranode data movement and maximize data reuse. Our techniques include predictive ordering of data elements for higher cache efficiency, as well as communication reduction approaches. We present detailed performance data when running on thousands of cores using the Cray XC30 supercomputer and show that our optimization strategies can exceed the original performance by over 2×. Additionally, many of these solutions can be broadly applied to a wide variety of unstructured grid-based computations.

  7. An eFTD-VP framework for efficiently generating patient-specific anatomically detailed facial soft tissue FE mesh for craniomaxillofacial surgery simulation.

    Science.gov (United States)

    Zhang, Xiaoyan; Kim, Daeseung; Shen, Shunyao; Yuan, Peng; Liu, Siting; Tang, Zhen; Zhang, Guangming; Zhou, Xiaobo; Gateno, Jaime; Liebschner, Michael A K; Xia, James J

    2018-04-01

    Accurate surgical planning and prediction of craniomaxillofacial surgery outcome requires simulation of soft tissue changes following osteotomy. This can only be achieved by using an anatomically detailed facial soft tissue model. The current state-of-the-art of model generation is not appropriate to clinical applications due to the time-intensive nature of manual segmentation and volumetric mesh generation. The conventional patient-specific finite element (FE) mesh generation methods are to deform a template FE mesh to match the shape of a patient based on registration. However, these methods commonly produce element distortion. Additionally, the mesh density for patients depends on that of the template model. It could not be adjusted to conduct mesh density sensitivity analysis. In this study, we propose a new framework of patient-specific facial soft tissue FE mesh generation. The goal of the developed method is to efficiently generate a high-quality patient-specific hexahedral FE mesh with adjustable mesh density while preserving the accuracy in anatomical structure correspondence. Our FE mesh is generated by eFace template deformation followed by volumetric parametrization. First, the patient-specific anatomically detailed facial soft tissue model (including skin, mucosa, and muscles) is generated by deforming an eFace template model. The adaptation of the eFace template model is achieved by using a hybrid landmark-based morphing and dense surface fitting approach followed by a thin-plate spline interpolation. Then, high-quality hexahedral mesh is constructed by using volumetric parameterization. The user can control the resolution of hexahedron mesh to best reflect clinicians' need. Our approach was validated using 30 patient models and 4 visible human datasets. The generated patient-specific FE mesh showed high surface matching accuracy, element quality, and internal structure matching accuracy. They can be directly and effectively used for clinical

  8. Simulating non-Newtonian flows with the moving particle semi-implicit method with an SPH kernel

    International Nuclear Information System (INIS)

    Xiang, Hao; Chen, Bin

    2015-01-01

    The moving particle semi-implicit (MPS) method and smoothed particle hydrodynamics (SPH) are commonly used mesh-free particle methods for free surface flows. The MPS method has superiority in incompressible flow simulation and simple programing. However, the crude kernel function is not accurate enough for the discretization of the divergence of the shear stress tensor by the particle inconsistency when the MPS method is extended to non-Newtonian flows. This paper presents an improved MPS method with an SPH kernel to simulate non-Newtonian flows. To improve the consistency of the partial derivative, the SPH cubic spline kernel and the Taylor series expansion are combined with the MPS method. This approach is suitable for all non-Newtonian fluids that can be described with τ  = μ(|γ|) Δ (where τ is the shear stress tensor, μ is the viscosity, |γ| is the shear rate, and Δ is the strain tensor), e.g., the Casson and Cross fluids. Two examples are simulated including the Newtonian Poiseuille flow and container filling process of the Cross fluid. The results of Poiseuille flow are more accurate than the traditional MPS method, and different filling processes are obtained with good agreement with previous results, which verified the validation of the new algorithm. For the Cross fluid, the jet fracture length can be correlated with We 0.28 Fr 0.78 (We is the Weber number, Fr is the Froude number). (paper)

  9. Interoperable mesh components for large-scale, distributed-memory simulations

    International Nuclear Information System (INIS)

    Devine, K; Leung, V; Diachin, L; Miller, M

    2009-01-01

    SciDAC applications have a demonstrated need for advanced software tools to manage the complexities associated with sophisticated geometry, mesh, and field manipulation tasks, particularly as computer architectures move toward the petascale. In this paper, we describe a software component - an abstract data model and programming interface - designed to provide support for parallel unstructured mesh operations. We describe key issues that must be addressed to successfully provide high-performance, distributed-memory unstructured mesh services and highlight some recent research accomplishments in developing new load balancing and MPI-based communication libraries appropriate for leadership class computing. Finally, we give examples of the use of parallel adaptive mesh modification in two SciDAC applications.

  10. 3D Adaptive Mesh Refinement Simulations of Pellet Injection in Tokamaks

    International Nuclear Information System (INIS)

    Samtaney, S.; Jardin, S.C.; Colella, P.; Martin, D.F.

    2003-01-01

    We present results of Adaptive Mesh Refinement (AMR) simulations of the pellet injection process, a proven method of refueling tokamaks. AMR is a computationally efficient way to provide the resolution required to simulate realistic pellet sizes relative to device dimensions. The mathematical model comprises of single-fluid MHD equations with source terms in the continuity equation along with a pellet ablation rate model. The numerical method developed is an explicit unsplit upwinding treatment of the 8-wave formulation, coupled with a MAC projection method to enforce the solenoidal property of the magnetic field. The Chombo framework is used for AMR. The role of the E x B drift in mass redistribution during inside and outside pellet injections is emphasized

  11. Studies of Particle Wake Potentials in Plasmas

    Science.gov (United States)

    Ellis, Ian; Graziani, Frank; Glosli, James; Strozzi, David; Surh, Michael; Richards, David; Decyk, Viktor; Mori, Warren

    2011-10-01

    Fast Ignition studies require a detailed understanding of electron scattering, stopping, and energy deposition in plasmas with variable values for the number of particles within a Debye sphere. Presently there is disagreement in the literature concerning the proper description of these processes. Developing and validating proper descriptions requires studying the processes using first-principle electrostatic simulations and possibly including magnetic fields. We are using the particle-particle particle-mesh (PPPM) code ddcMD and the particle-in-cell (PIC) code BEPS to perform these simulations. As a starting point in our study, we examine the wake of a particle passing through a plasma in 3D electrostatic simulations performed with ddcMD and with BEPS using various cell sizes. In this poster, we compare the wakes we observe in these simulations with each other and predictions from Vlasov theory. Prepared by LLNL under Contract DE-AC52-07NA27344 and by UCLA under Grant DE-FG52-09NA29552.

  12. Simulation of the Beam-Beam Effects in e+e- Storage Rings with a Method of Reducing the Region of Mesh

    Energy Technology Data Exchange (ETDEWEB)

    Cai, Yunhai

    2000-08-31

    A highly accurate self-consistent particle code to simulate the beam-beam collision in e{sup +}e{sup -} storage rings has been developed. It adopts a method of solving the Poisson equation with an open boundary. The method consists of two steps: assigning the potential on a finite boundary using the Green's function, and then solving the potential inside the boundary with a fast Poisson solver. Since the solution of the Poisson's equation is unique, the authors solution is exactly the same as the one obtained by simply using the Green's function. The method allows us to select much smaller region of mesh and therefore increase the resolution of the solver. The better resolution makes more accurate the calculation of the dynamics in the core of the beams. The luminosity simulated with this method agrees quantitatively with the measurement for the PEP-II B-factory ring in the linear and nonlinear beam current regimes, demonstrating its predictive capability in detail.

  13. Numerical simulation of a flow-like landslide using the particle finite element method

    Science.gov (United States)

    Zhang, Xue; Krabbenhoft, Kristian; Sheng, Daichao; Li, Weichao

    2015-01-01

    In this paper, an actual landslide process that occurred in Southern China is simulated by a continuum approach, the particle finite element method (PFEM). The PFEM attempts to solve the boundary-value problems in the framework of solid mechanics, satisfying the governing equations including momentum conservation, displacement-strain relation, constitutive relation as well as the frictional contact between the sliding mass and the slip surface. To warrant the convergence behaviour of solutions, the problem is formulated as a mathematical programming problem, while the particle finite element procedure is employed to tackle the issues of mesh distortion and free-surface evolution. The whole procedure of the landslide, from initiation, sliding to deposition, is successfully reproduced by the continuum approach. It is shown that the density of the mass has little influence on the sliding process in the current landslide, whereas both the geometry and the roughness of the slip surface play important roles. Comparative studies are also conducted where a satisfactory agreement is obtained.

  14. Micromagnetic simulations of spinel ferrite particles

    International Nuclear Information System (INIS)

    Dantas, Christine C.; Gama, Adriana M.

    2010-01-01

    This paper presents the results of simulations of the magnetization field ac response (at 2-12 GHz) of various submicron ferrite particles (cylindrical dots). The ferrites in the present simulations have the spinel structure, expressed here by M 1 - n Zn n Fe 2 O 4 (where M stands for a divalent metal), and the parameters chosen were the following: (a) for n=0: M={Fe, Mn, Co, Ni, Mg, Cu }; (b) for n=0.1: M = {Fe, Mg} (mixed ferrites). These runs represent full 3D micromagnetic (one-particle) ferrite simulations. We find evidences of confined spin waves in all simulations, as well as a complex behavior nearby the main resonance peak in the case of the M = {Mg, Cu} ferrites. A comparison of the n=0 and n=0.1 cases for fixed M reveals a significant change in the spectra in M = Mg ferrites, but only a minor change in the M=Fe case. An additional larger scale simulation of a 3 by 3 particle array was performed using similar conditions of the Fe 3 O 4 (magnetite; n=0, M = Fe) one-particle simulation. We find that the main resonance peak of the Fe 3 O 4 one-particle simulation is disfigured in the corresponding 3 by 3 particle simulation, indicating the extent to which dipolar interactions are able to affect the main resonance peak in that magnetic compound.

  15. Extended particle-based simulation for magnetic-aligned compaction of hard magnetic particles

    Energy Technology Data Exchange (ETDEWEB)

    Soda, Rikio; Takagi, Kenta; Ozaki, Kimihiro, E-mail: r-soda@aist.go.jp

    2015-12-15

    In order to understand the magnetic-aligned compaction process, we develop a three-dimensional (3D) discrete element method for simulating the motion of hard magnetic particles subjected to strong compression and magnetic fields. The proposed simulation model also considers the exact magnetic force involved via the calculation of the magnetic moment. First, to validate the simulation model, single-action compaction in the absence of a magnetic field was calculated. The calculated compaction curves are in good quantitative agreement with experimental ones. Based on this simulation model, the alignment behavior of Nd–Fe–B particles during compression under the application of a static magnetic field. The developed simulation model enables the visualization of particle behavior including the misorientation of the magnetization easy axis, which provided the quantitative relationships between applied pressure and particle misorientation. - Highlights: • A practical 3D DEM simulation technique for magnetic-aligned compaction was developed. • An extended simulation model was introduced for hard magnetic particles. • Magnetic-aligned compaction was simulated using the developed simulation model.

  16. High-Fidelity RF Gun Simulations with the Parallel 3D Finite Element Particle-In-Cell Code Pic3P

    Energy Technology Data Exchange (ETDEWEB)

    Candel, A; Kabel, A.; Lee, L.; Li, Z.; Limborg, C.; Ng, C.; Schussman, G.; Ko, K.; /SLAC

    2009-06-19

    SLAC's Advanced Computations Department (ACD) has developed the first parallel Finite Element 3D Particle-In-Cell (PIC) code, Pic3P, for simulations of RF guns and other space-charge dominated beam-cavity interactions. Pic3P solves the complete set of Maxwell-Lorentz equations and thus includes space charge, retardation and wakefield effects from first principles. Pic3P uses higher-order Finite Elementmethods on unstructured conformal meshes. A novel scheme for causal adaptive refinement and dynamic load balancing enable unprecedented simulation accuracy, aiding the design and operation of the next generation of accelerator facilities. Application to the Linac Coherent Light Source (LCLS) RF gun is presented.

  17. From 4D Medical Images (CT, MRI, and Ultrasound to 4D Structured Mesh Models of the Left Ventricular Endocardium for Patient-Specific Simulations

    Directory of Open Access Journals (Sweden)

    Federico Canè

    2018-01-01

    Full Text Available With cardiovascular disease (CVD remaining the primary cause of death worldwide, early detection of CVDs becomes essential. The intracardiac flow is an important component of ventricular function, motion kinetics, wash-out of ventricular chambers, and ventricular energetics. Coupling between Computational Fluid Dynamics (CFD simulations and medical images can play a fundamental role in terms of patient-specific diagnostic tools. From a technical perspective, CFD simulations with moving boundaries could easily lead to negative volumes errors and the sudden failure of the simulation. The generation of high-quality 4D meshes (3D in space + time with 1-to-1 vertex becomes essential to perform a CFD simulation with moving boundaries. In this context, we developed a semiautomatic morphing tool able to create 4D high-quality structured meshes starting from a segmented 4D dataset. To prove the versatility and efficiency, the method was tested on three different 4D datasets (Ultrasound, MRI, and CT by evaluating the quality and accuracy of the resulting 4D meshes. Furthermore, an estimation of some physiological quantities is accomplished for the 4D CT reconstruction. Future research will aim at extending the region of interest, further automation of the meshing algorithm, and generating structured hexahedral mesh models both for the blood and myocardial volume.

  18. Monte Carlo Simulation for Particle Detectors

    CERN Document Server

    Pia, Maria Grazia

    2012-01-01

    Monte Carlo simulation is an essential component of experimental particle physics in all the phases of its life-cycle: the investigation of the physics reach of detector concepts, the design of facilities and detectors, the development and optimization of data reconstruction software, the data analysis for the production of physics results. This note briefly outlines some research topics related to Monte Carlo simulation, that are relevant to future experimental perspectives in particle physics. The focus is on physics aspects: conceptual progress beyond current particle transport schemes, the incorporation of materials science knowledge relevant to novel detection technologies, functionality to model radiation damage, the capability for multi-scale simulation, quantitative validation and uncertainty quantification to determine the predictive power of simulation. The R&D on simulation for future detectors would profit from cooperation within various components of the particle physics community, and synerg...

  19. Advanced visualization technology for terascale particle accelerator simulations

    International Nuclear Information System (INIS)

    Ma, K-L; Schussman, G.; Wilson, B.; Ko, K.; Qiang, J.; Ryne, R.

    2002-01-01

    This paper presents two new hardware-assisted rendering techniques developed for interactive visualization of the terascale data generated from numerical modeling of next generation accelerator designs. The first technique, based on a hybrid rendering approach, makes possible interactive exploration of large-scale particle data from particle beam dynamics modeling. The second technique, based on a compact texture-enhanced representation, exploits the advanced features of commodity graphics cards to achieve perceptually effective visualization of the very dense and complex electromagnetic fields produced from the modeling of reflection and transmission properties of open structures in an accelerator design. Because of the collaborative nature of the overall accelerator modeling project, the visualization technology developed is for both desktop and remote visualization settings. We have tested the techniques using both time varying particle data sets containing up to one billion particle s per time step and electromagnetic field data sets with millions of mesh elements

  20. Enhancements to the Combinatorial Geometry Particle Tracker in the Mercury Monte Carlo Transport Code: Embedded Meshes and Domain Decomposition

    International Nuclear Information System (INIS)

    Greenman, G.M.; O'Brien, M.J.; Procassini, R.J.; Joy, K.I.

    2009-01-01

    Two enhancements to the combinatorial geometry (CG) particle tracker in the Mercury Monte Carlo transport code are presented. The first enhancement is a hybrid particle tracker wherein a mesh region is embedded within a CG region. This method permits efficient calculations of problems with contain both large-scale heterogeneous and homogeneous regions. The second enhancement relates to the addition of parallelism within the CG tracker via spatial domain decomposition. This permits calculations of problems with a large degree of geometric complexity, which are not possible through particle parallelism alone. In this method, the cells are decomposed across processors and a particles is communicated to an adjacent processor when it tracks to an interprocessor boundary. Applications that demonstrate the efficacy of these new methods are presented

  1. Simulation of transients with space-dependent feedback by coarse mesh flux expansion method

    International Nuclear Information System (INIS)

    Langenbuch, S.; Maurer, W.; Werner, W.

    1975-01-01

    For the simulation of the time-dependent behaviour of large LWR-cores, even the most efficient Finite-Difference (FD) methods require a prohibitive amount of computing time in order to achieve results of acceptable accuracy. Static CM-solutions computed with a mesh-size corresponding to the fuel element structure (about 20 cm) are at least as accurate as FD-solutions computed with about 5 cm mesh-size. For 3d-calculations this results in a reduction of storage requirements by a factor 60 and of computing costs by a factor 40, relative to FD-methods. These results have been obtained for pure neutronic calculations, where feedback is not taken into account. In this paper it is demonstrated that the method retains its accuracy also in kinetic calculations, even in the presence of strong space dependent feedback. (orig./RW) [de

  2. Visual interrogation of gyrokinetic particle simulations

    International Nuclear Information System (INIS)

    Jones, Chad; Ma, K-L; Sanderson, Allen; Myers, Lee Roy Jr

    2007-01-01

    Gyrokinetic particle simulations are critical to the study of anomalous energy transport associated with plasma microturbulence in magnetic confinement fusion experiments. The simulations are conducted on massively parallel computers and produce large quantities of particles, variables, and time steps, thus presenting a formidable challenge to data analysis tasks. We present two new visualization techniques for scientists to improve their understanding of the time-varying, multivariate particle data. One technique allows scientists to examine correlations in multivariate particle data with tightly coupled views of the data in both physical space and variable space, and to visually identify and track features of interest. The second technique, built into SCIRun, allows scientists to perform range-based queries over a series of time slices and visualize the resulting particles using glyphs. The ability to navigate the multiple dimensions of the particle data, as well as query individual or a collection of particles, enables scientists to not only validate their simulations but also discover new phenomena in their data

  3. Data parallel sorting for particle simulation

    Science.gov (United States)

    Dagum, Leonardo

    1992-01-01

    Sorting on a parallel architecture is a communications intensive event which can incur a high penalty in applications where it is required. In the case of particle simulation, only integer sorting is necessary, and sequential implementations easily attain the minimum performance bound of O (N) for N particles. Parallel implementations, however, have to cope with the parallel sorting problem which, in addition to incurring a heavy communications cost, can make the minimun performance bound difficult to attain. This paper demonstrates how the sorting problem in a particle simulation can be reduced to a merging problem, and describes an efficient data parallel algorithm to solve this merging problem in a particle simulation. The new algorithm is shown to be optimal under conditions usual for particle simulation, and its fieldwise implementation on the Connection Machine is analyzed in detail. The new algorithm is about four times faster than a fieldwise implementation of radix sort on the Connection Machine.

  4. A unified monolithic approach for multi-fluid flows and fluid-structure interaction using the Particle Finite Element Method with fixed mesh

    Science.gov (United States)

    Becker, P.; Idelsohn, S. R.; Oñate, E.

    2015-06-01

    This paper describes a strategy to solve multi-fluid and fluid-structure interaction (FSI) problems using Lagrangian particles combined with a fixed finite element (FE) mesh. Our approach is an extension of the fluid-only PFEM-2 (Idelsohn et al., Eng Comput 30(2):2-2, 2013; Idelsohn et al., J Numer Methods Fluids, 2014) which uses explicit integration over the streamlines to improve accuracy. As a result, the convective term does not appear in the set of equations solved on the fixed mesh. Enrichments in the pressure field are used to improve the description of the interface between phases.

  5. Simulation of 2-D Compressible Flows on a Moving Curvilinear Mesh with an Implicit-Explicit Runge-Kutta Method

    KAUST Repository

    AbuAlSaud, Moataz

    2012-07-01

    The purpose of this thesis is to solve unsteady two-dimensional compressible Navier-Stokes equations for a moving mesh using implicit explicit (IMEX) Runge- Kutta scheme. The moving mesh is implemented in the equations using Arbitrary Lagrangian Eulerian (ALE) formulation. The inviscid part of the equation is explicitly solved using second-order Godunov method, whereas the viscous part is calculated implicitly. We simulate subsonic compressible flow over static NACA-0012 airfoil at different angle of attacks. Finally, the moving mesh is examined via oscillating the airfoil between angle of attack = 0 and = 20 harmonically. It is observed that the numerical solution matches the experimental and numerical results in the literature to within 20%.

  6. Finite-element 3D simulation tools for high-current relativistic electron beams

    Science.gov (United States)

    Humphries, Stanley; Ekdahl, Carl

    2002-08-01

    The DARHT second-axis injector is a challenge for computer simulations. Electrons are subject to strong beam-generated forces. The fields are fully three-dimensional and accurate calculations at surfaces are critical. We describe methods applied in OmniTrak, a 3D finite-element code suite that can address DARHT and the full range of charged-particle devices. The system handles mesh generation, electrostatics, magnetostatics and self-consistent particle orbits. The MetaMesh program generates meshes of conformal hexahedrons to fit any user geometry. The code has the unique ability to create structured conformal meshes with cubic logic. Organized meshes offer advantages in speed and memory utilization in the orbit and field solutions. OmniTrak is a versatile charged-particle code that handles 3D electric and magnetic field solutions on independent meshes. The program can update both 3D field solutions from the calculated beam space-charge and current-density. We shall describe numerical methods for orbit tracking on a hexahedron mesh. Topics include: 1) identification of elements along the particle trajectory, 2) fast searches and adaptive field calculations, 3) interpolation methods to terminate orbits on material surfaces, 4) automatic particle generation on multiple emission surfaces to model space-charge-limited emission and field emission, 5) flexible Child law algorithms, 6) implementation of the dual potential model for 3D magnetostatics, and 7) assignment of charge and current from model particle orbits for self-consistent fields.

  7. Modeling of magnetic particle suspensions for simulations

    CERN Document Server

    Satoh, Akira

    2017-01-01

    The main objective of the book is to highlight the modeling of magnetic particles with different shapes and magnetic properties, to provide graduate students and young researchers information on the theoretical aspects and actual techniques for the treatment of magnetic particles in particle-based simulations. In simulation, we focus on the Monte Carlo, molecular dynamics, Brownian dynamics, lattice Boltzmann and stochastic rotation dynamics (multi-particle collision dynamics) methods. The latter two simulation methods can simulate both the particle motion and the ambient flow field simultaneously. In general, specialized knowledge can only be obtained in an effective manner under the supervision of an expert. The present book is written to play such a role for readers who wish to develop the skill of modeling magnetic particles and develop a computer simulation program using their own ability. This book is therefore a self-learning book for graduate students and young researchers. Armed with this knowledge,...

  8. Resolution convergence in cosmological hydrodynamical simulations using adaptive mesh refinement

    Science.gov (United States)

    Snaith, Owain N.; Park, Changbom; Kim, Juhan; Rosdahl, Joakim

    2018-06-01

    We have explored the evolution of gas distributions from cosmological simulations carried out using the RAMSES adaptive mesh refinement (AMR) code, to explore the effects of resolution on cosmological hydrodynamical simulations. It is vital to understand the effect of both the resolution of initial conditions (ICs) and the final resolution of the simulation. Lower initial resolution simulations tend to produce smaller numbers of low-mass structures. This will strongly affect the assembly history of objects, and has the same effect of simulating different cosmologies. The resolution of ICs is an important factor in simulations, even with a fixed maximum spatial resolution. The power spectrum of gas in simulations using AMR diverges strongly from the fixed grid approach - with more power on small scales in the AMR simulations - even at fixed physical resolution and also produces offsets in the star formation at specific epochs. This is because before certain times the upper grid levels are held back to maintain approximately fixed physical resolution, and to mimic the natural evolution of dark matter only simulations. Although the impact of hold-back falls with increasing spatial and IC resolutions, the offsets in the star formation remain down to a spatial resolution of 1 kpc. These offsets are of the order of 10-20 per cent, which is below the uncertainty in the implemented physics but are expected to affect the detailed properties of galaxies. We have implemented a new grid-hold-back approach to minimize the impact of hold-back on the star formation rate.

  9. Performance of the hybrid wireless mesh protocol for wireless mesh networks

    DEFF Research Database (Denmark)

    Boye, Magnus; Staalhagen, Lars

    2010-01-01

    Wireless mesh networks offer a new way of providing end-user access and deploying network infrastructure. Though mesh networks offer a price competitive solution to wired networks, they also come with a set of new challenges such as optimal path selection, channel utilization, and load balancing....... and proactive. Two scenarios of different node density are considered for both path selection modes. The results presented in this paper are based on a simulation model of the HWMP specification in the IEEE 802.11s draft 4.0 implemented in OPNET Modeler....

  10. Electrical Resistance Imaging of Two-Phase Flow With a Mesh Grouping Technique Based On Particle Swarm Optimization

    International Nuclear Information System (INIS)

    Lee, Bo An; Kim, Bong Seok; Ko, Min Seok; Kim, Kyung Young; Kim, Sin

    2014-01-01

    An electrical resistance tomography (ERT) technique combining the particle swarm optimization (PSO) algorithm with the Gauss-Newton method is applied to the visualization of two-phase flows. In the ERT, the electrical conductivity distribution, namely the conductivity values of pixels (numerical meshes) comprising the domain in the context of a numerical image reconstruction algorithm, is estimated with the known injected currents through the electrodes attached on the domain boundary and the measured potentials on those electrodes. In spite of many favorable characteristics of ERT such as no radiation, low cost, and high temporal resolution compared to other tomography techniques, one of the major drawbacks of ERT is low spatial resolution due to the inherent ill-posedness of conventional image reconstruction algorithms. In fact, the number of known data is much less than that of the unknowns (meshes). Recalling that binary mixtures like two-phase flows consist of only two substances with distinct electrical conductivities, this work adopts the PSO algorithm for mesh grouping to reduce the number of unknowns. In order to verify the enhanced performance of the proposed method, several numerical tests are performed. The comparison between the proposed algorithm and conventional Gauss-Newton method shows significant improvements in the quality of reconstructed images

  11. ELECTRICAL RESISTANCE IMAGING OF TWO-PHASE FLOW WITH A MESH GROUPING TECHNIQUE BASED ON PARTICLE SWARM OPTIMIZATION

    Directory of Open Access Journals (Sweden)

    BO AN LEE

    2014-02-01

    Full Text Available An electrical resistance tomography (ERT technique combining the particle swarm optimization (PSO algorithm with the Gauss-Newton method is applied to the visualization of two-phase flows. In the ERT, the electrical conductivity distribution, namely the conductivity values of pixels (numerical meshes comprising the domain in the context of a numerical image reconstruction algorithm, is estimated with the known injected currents through the electrodes attached on the domain boundary and the measured potentials on those electrodes. In spite of many favorable characteristics of ERT such as no radiation, low cost, and high temporal resolution compared to other tomography techniques, one of the major drawbacks of ERT is low spatial resolution due to the inherent ill-posedness of conventional image reconstruction algorithms. In fact, the number of known data is much less than that of the unknowns (meshes. Recalling that binary mixtures like two-phase flows consist of only two substances with distinct electrical conductivities, this work adopts the PSO algorithm for mesh grouping to reduce the number of unknowns. In order to verify the enhanced performance of the proposed method, several numerical tests are performed. The comparison between the proposed algorithm and conventional Gauss-Newton method shows significant improvements in the quality of reconstructed images.

  12. Electrical Resistance Imaging of Two-Phase Flow With a Mesh Grouping Technique Based On Particle Swarm Optimization

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Bo An; Kim, Bong Seok; Ko, Min Seok; Kim, Kyung Young; Kim, Sin [Jeju National Univ., Jeju (Korea, Republic of)

    2014-02-15

    An electrical resistance tomography (ERT) technique combining the particle swarm optimization (PSO) algorithm with the Gauss-Newton method is applied to the visualization of two-phase flows. In the ERT, the electrical conductivity distribution, namely the conductivity values of pixels (numerical meshes) comprising the domain in the context of a numerical image reconstruction algorithm, is estimated with the known injected currents through the electrodes attached on the domain boundary and the measured potentials on those electrodes. In spite of many favorable characteristics of ERT such as no radiation, low cost, and high temporal resolution compared to other tomography techniques, one of the major drawbacks of ERT is low spatial resolution due to the inherent ill-posedness of conventional image reconstruction algorithms. In fact, the number of known data is much less than that of the unknowns (meshes). Recalling that binary mixtures like two-phase flows consist of only two substances with distinct electrical conductivities, this work adopts the PSO algorithm for mesh grouping to reduce the number of unknowns. In order to verify the enhanced performance of the proposed method, several numerical tests are performed. The comparison between the proposed algorithm and conventional Gauss-Newton method shows significant improvements in the quality of reconstructed images.

  13. Simple rANDom wALk simulation

    Energy Technology Data Exchange (ETDEWEB)

    2016-05-20

    SANDAL is a particle-mesh simulation mini-app. The simple simulated physics propagates a set of particles in a constant, 2D wind field with Gaussian turbulence. SANDAL demonstrates the feasibility of formulating a computational physics problem using an alternative, modern design. Specifically, it is implemented using relational tables and queries, rather than array-based data model. It also is implemented with functional language design and cloud deployment, Scala and Apache Spark, rather than Fortran and MPI+X.

  14. The application of the mesh-free method in the numerical simulations of the higher-order continuum structures

    Energy Technology Data Exchange (ETDEWEB)

    Sun, Yuzhou, E-mail: yuzhousun@126.com; Chen, Gensheng; Li, Dongxia [School of Civil Engineering and Architecture, Zhongyuan University of Technology, Zhengzhou (China)

    2016-06-08

    This paper attempts to study the application of mesh-free method in the numerical simulations of the higher-order continuum structures. A high-order bending beam considers the effect of the third-order derivative of deflections, and can be viewed as a one-dimensional higher-order continuum structure. The moving least-squares method is used to construct the shape function with the high-order continuum property, the curvature and the third-order derivative of deflections are directly interpolated with nodal variables and the second- and third-order derivative of the shape function, and the mesh-free computational scheme is establish for beams. The coupled stress theory is introduced to describe the special constitutive response of the layered rock mass in which the bending effect of thin layer is considered. The strain and the curvature are directly interpolated with the nodal variables, and the mesh-free method is established for the layered rock mass. The good computational efficiency is achieved based on the developed mesh-free method, and some key issues are discussed.

  15. Three-dimensional dynamic rupture simulation with a high-order discontinuous Galerkin method on unstructured tetrahedral meshes

    KAUST Repository

    Pelties, Christian

    2012-02-18

    Accurate and efficient numerical methods to simulate dynamic earthquake rupture and wave propagation in complex media and complex fault geometries are needed to address fundamental questions in earthquake dynamics, to integrate seismic and geodetic data into emerging approaches for dynamic source inversion, and to generate realistic physics-based earthquake scenarios for hazard assessment. Modeling of spontaneous earthquake rupture and seismic wave propagation by a high-order discontinuous Galerkin (DG) method combined with an arbitrarily high-order derivatives (ADER) time integration method was introduced in two dimensions by de la Puente et al. (2009). The ADER-DG method enables high accuracy in space and time and discretization by unstructured meshes. Here we extend this method to three-dimensional dynamic rupture problems. The high geometrical flexibility provided by the usage of tetrahedral elements and the lack of spurious mesh reflections in the ADER-DG method allows the refinement of the mesh close to the fault to model the rupture dynamics adequately while concentrating computational resources only where needed. Moreover, ADER-DG does not generate spurious high-frequency perturbations on the fault and hence does not require artificial Kelvin-Voigt damping. We verify our three-dimensional implementation by comparing results of the SCEC TPV3 test problem with two well-established numerical methods, finite differences, and spectral boundary integral. Furthermore, a convergence study is presented to demonstrate the systematic consistency of the method. To illustrate the capabilities of the high-order accurate ADER-DG scheme on unstructured meshes, we simulate an earthquake scenario, inspired by the 1992 Landers earthquake, that includes curved faults, fault branches, and surface topography. Copyright 2012 by the American Geophysical Union.

  16. Fully non-linear multi-species Fokker-Planck-Landau collisions for gyrokinetic particle-in-cell simulations of fusion plasma

    Science.gov (United States)

    Hager, Robert; Yoon, E. S.; Ku, S.; D'Azevedo, E. F.; Worley, P. H.; Chang, C. S.

    2015-11-01

    We describe the implementation, and application of a time-dependent, fully nonlinear multi-species Fokker-Planck-Landau collision operator based on the single-species work of Yoon and Chang [Phys. Plasmas 21, 032503 (2014)] in the full-function gyrokinetic particle-in-cell codes XGC1 [Ku et al., Nucl. Fusion 49, 115021 (2009)] and XGCa. XGC simulations include the pedestal and scrape-off layer, where significant deviations of the particle distribution function from a Maxwellian can occur. Thus, in order to describe collisional effects on neoclassical and turbulence physics accurately, the use of a non-linear collision operator is a necessity. Our collision operator is based on a finite volume method using the velocity-space distribution functions sampled from the marker particles. Since the same fine configuration space mesh is used for collisions and the Poisson solver, the workload due to collisions can be comparable to or larger than the workload due to particle motion. We demonstrate that computing time spent on collisions can be kept affordable by applying advanced parallelization strategies while conserving mass, momentum, and energy to reasonable accuracy. We also show results of production scale XGCa simulations in the H-mode pedestal and compare to conventional theory. Work supported by US DOE OFES and OASCR.

  17. Discrete particle noise in particle-in-cell simulations of plasma microturbulence

    International Nuclear Information System (INIS)

    Nevins, W.M.; Hammett, G.W.; Dimits, A.M.; Dorland, W.; Shumaker, D.E.

    2005-01-01

    Recent gyrokinetic simulations of electron temperature gradient (ETG) turbulence with the global particle-in-cell (PIC) code GTC [Z. Lin et al., Proceedings of the 20th Fusion Energy Conference, Vilamoura, Portugal, 2004 (IAEA, Vienna, 2005)] yielded different results from earlier flux-tube continuum code simulations [F. Jenko and W. Dorland, Phys. Rev. Lett. 89, 225001 (2002)] despite similar plasma parameters. Differences between the simulation results were attributed to insufficient phase-space resolution and novel physics associated with global simulation models. The results of the global PIC code are reproduced here using the flux-tube PIC code PG3EQ [A. M. Dimits et al., Phys. Rev. Lett. 77, 71 (1996)], thereby eliminating global effects as the cause of the discrepancy. The late-time decay of the ETG turbulence and the steady-state heat transport observed in these PIC simulations are shown to result from discrete particle noise. Discrete particle noise is a numerical artifact, so both these PG3EQ simulations and, by inference, the GTC simulations that they reproduced have little to say about steady-state ETG turbulence and the associated anomalous heat transport. In the course of this work several diagnostics are developed to retrospectively test whether a particular PIC simulation is dominated by discrete particle noise

  18. Effect of particle size on mixing degree in dispensation.

    Science.gov (United States)

    Nakamura, Hitoshi; Yanagihara, Yoshitsugu; Sekiguchi, Hiroko; Ohtani, Michiteru; Kariya, Satoru; Uchino, Katsuyoshi; Suzuki, Hiroshi; Iga, Tatsuji

    2004-03-01

    By using lactose colored with erythrocin, we examined the effect of particle size on mixing degree during the preparation of triturations with a mortar and pestle. We used powders with different distributions of particle sizes, i.e., powder that passed through 32-mesh but was trapped on a 42-mesh sieve (32/42-mesh powder), powder that passed through a 42-mesh sieve but was trapped on a 60-mesh sieve (42/60-mesh powder), powder that passed through a 60-mesh sieve but was trapped on a 100-mesh sieve (60/100-mesh powder), and powder that passes through a 100-mesh sieve (> 100-mesh powder). The mixing degree of colored powder and non-colored powder whose distribution of particle sizes was the same as that of the colored powder was excellent. The coefficient of variation (CV) value of the mixing degree was 6.08% after 40 rotations when colored powder was mixed with non-colored powder that both passed through a 100-mesh sieve. The CV value of the mixing degree was low in the case of mixing of colored and non-colored powders with different particle size distributions. After mixing, about 50% of 42/60-mesh powder had become smaller particles, whereas the distribution of particle sizes was not influenced by the mixing of 60/100-mesh powder. It was suggested that the mixing degree is affected by distribution of particle sizes. It may be important to determine the mixing degrees for drugs with narrow therapeutic ranges.

  19. Adaptive mesh refinement for shocks and material interfaces

    Energy Technology Data Exchange (ETDEWEB)

    Dai, William Wenlong [Los Alamos National Laboratory

    2010-01-01

    There are three kinds of adaptive mesh refinement (AMR) in structured meshes. Block-based AMR sometimes over refines meshes. Cell-based AMR treats cells cell by cell and thus loses the advantage of the nature of structured meshes. Patch-based AMR is intended to combine advantages of block- and cell-based AMR, i.e., the nature of structured meshes and sharp regions of refinement. But, patch-based AMR has its own difficulties. For example, patch-based AMR typically cannot preserve symmetries of physics problems. In this paper, we will present an approach for a patch-based AMR for hydrodynamics simulations. The approach consists of clustering, symmetry preserving, mesh continuity, flux correction, communications, management of patches, and load balance. The special features of this patch-based AMR include symmetry preserving, efficiency of refinement across shock fronts and material interfaces, special implementation of flux correction, and patch management in parallel computing environments. To demonstrate the capability of the AMR framework, we will show both two- and three-dimensional hydrodynamics simulations with many levels of refinement.

  20. Numerical Study of Charged Inertial Particles in Turbulence using a Coupled Fluid-P3M Approach

    Science.gov (United States)

    Yao, Yuan; Capecelatro, Jesse

    2017-11-01

    Non-trivial interactions between charged particles and turbulence play an important role in many engineering and environmental flows, including clouds, fluidized bed reactors, charged hydrocarbon sprays and dusty plasmas. Due to the long-range nature of electrostatic forces, Coulomb interactions in systems with many particles must be handled carefully to avoid O(N2) computations. The particle-mesh (PM) method is typically employed in Eulerian-Lagrangian (EL) simulations as it avoids computing direct pairwise sums, but it fails to capture short-range interactions that are anticipated to be important when particles cluster. In this presentation, the particle-particle-particle-mesh (P3M) method that scales with O(NlogN) is implemented within a EL framework to simulate charged particles accurately in a tractable manner. The EL-P3M method is used to assess the competition between drag and Coulomb forces for a range of Stokes numbers and charges. Simulations of like- and oppositely-charged particles suspended in a two-dimensional Taylor-Green vortex and three-dimensional homogeneous isotropic turbulence are reported. One-point and two-point statistics obtained using PM and P3M are compared to assess the effect of added accuracy on collision rate and clustering.

  1. Electromagnetic ''particle-in-cell'' plasma simulation

    International Nuclear Information System (INIS)

    Langdon, A.B.

    1985-01-01

    ''PIC'' simulation tracks particles through electromagnetic fields calculated self-consistently from the charge and current densities of the particles themselves, external sources, and boundaries. Already used extensively in plasma physics, such simulations have become useful in the design of accelerators and their r.f. sources. 5 refs

  2. M3D project for simulation studies of plasmas

    International Nuclear Information System (INIS)

    Park, W.; Belova, E.V.; Fu, G.Y.; Sugiyama, L.E.

    1998-01-01

    The M3D (Multi-level 3D) project carries out simulation studies of plasmas of various regimes using multi-levels of physics, geometry, and mesh schemes in one code package. This paper and papers by Strauss, Sugiyama, and Belova in this workshop describe the project, and present examples of current applications. The currently available physics models of the M3D project are MHD, two-fluids, gyrokinetic hot particle/MHD hybrid, and gyrokinetic particle ion/two-fluid hybrid models. The code can be run with both structured and unstructured meshes

  3. Molecular Dynamic Studies of Particle Wake Potentials in Plasmas

    Science.gov (United States)

    Ellis, Ian; Graziani, Frank; Glosli, James; Strozzi, David; Surh, Michael; Richards, David; Decyk, Viktor; Mori, Warren

    2010-11-01

    Fast Ignition studies require a detailed understanding of electron scattering, stopping, and energy deposition in plasmas with variable values for the number of particles within a Debye sphere. Presently there is disagreement in the literature concerning the proper description of these processes. Developing and validating proper descriptions requires studying the processes using first-principle electrostatic simulations and possibly including magnetic fields. We are using the particle-particle particle-mesh (P^3M) code ddcMD to perform these simulations. As a starting point in our study, we examined the wake of a particle passing through a plasma. In this poster, we compare the wake observed in 3D ddcMD simulations with that predicted by Vlasov theory and those observed in the electrostatic PIC code BEPS where the cell size was reduced to .03λD.

  4. On Reducing Delay in Mesh-Based P2P Streaming: A Mesh-Push Approach

    Science.gov (United States)

    Liu, Zheng; Xue, Kaiping; Hong, Peilin

    The peer-assisted streaming paradigm has been widely employed to distribute live video data on the internet recently. In general, the mesh-based pull approach is more robust and efficient than the tree-based push approach. However, pull protocol brings about longer streaming delay, which is caused by the handshaking process of advertising buffer map message, sending request message and scheduling of the data block. In this paper, we propose a new approach, mesh-push, to address this issue. Different from the traditional pull approach, mesh-push implements block scheduling algorithm at sender side, where the block transmission is initiated by the sender rather than by the receiver. We first formulate the optimal upload bandwidth utilization problem, then present the mesh-push approach, in which a token protocol is designed to avoid block redundancy; a min-cost flow model is employed to derive the optimal scheduling for the push peer; and a push peer selection algorithm is introduced to reduce control overhead. Finally, we evaluate mesh-push through simulation, the results of which show mesh-push outperforms the pull scheduling in streaming delay, and achieves comparable delivery ratio at the same time.

  5. Mesh Processing in Medical Image Analysis

    DEFF Research Database (Denmark)

    The following topics are dealt with: mesh processing; medical image analysis; interactive freeform modeling; statistical shape analysis; clinical CT images; statistical surface recovery; automated segmentation; cerebral aneurysms; and real-time particle-based representation....

  6. Simulation of halo particles with Simpsons

    International Nuclear Information System (INIS)

    Machida, Shinji

    2003-01-01

    Recent code improvements and some simulation results of halo particles with Simpsons will be presented. We tried to identify resonance behavior of halo particles by looking at tune evolution of individual macro particle

  7. Simulation of halo particles with Simpsons

    Science.gov (United States)

    Machida, Shinji

    2003-12-01

    Recent code improvements and some simulation results of halo particles with Simpsons will be presented. We tried to identify resonance behavior of halo particles by looking at tune evolution of individual macro particle.

  8. Gyrokinetic particle simulation of neoclassical transport

    International Nuclear Information System (INIS)

    Lin, Z.; Tang, W.M.; Lee, W.W.

    1995-01-01

    A time varying weighting (δf ) scheme for gyrokinetic particle simulation is applied to a steady-state, multispecies simulation of neoclassical transport. Accurate collision operators conserving momentum and energy are developed and implemented. Simulation results using these operators are found to agree very well with neoclassical theory. For example, it is dynamically demonstrated that like-particle collisions produce no particle flux and that the neoclassical fluxes are ambipolar for an ion--electron plasma. An important physics feature of the present scheme is the introduction of toroidal flow to the simulations. Simulation results are in agreement with the existing analytical neoclassical theory. The poloidal electric field associated with toroidal mass flow is found to enhance density gradient-driven electron particle flux and the bootstrap current while reducing temperature gradient-driven flux and current. Finally, neoclassical theory in steep gradient profile relevant to the edge regime is examined by taking into account finite banana width effects. In general, in the present work a valuable new capability for studying important aspects of neoclassical transport inaccessible by conventional analytical calculation processes is demonstrated. copyright 1995 American Institute of Physics

  9. VARIABLE MESH STIFFNESS OF SPUR GEAR TEETH USING ...

    African Journals Online (AJOL)

    gear engagement. A gear mesh kinematic simulation ... model is appropnate for VMS of a spur gear tooth. The assumptions for ... This process has been continued until one complete tooth meshing cycle is ..... Element Method. Using MATLAB,.

  10. On Using Particle Finite Element for Hydrodynamics Problems Solving

    Directory of Open Access Journals (Sweden)

    E. V. Davidova

    2015-01-01

    Full Text Available The aim of the present research is to develop software for the Particle Finite Element Method (PFEM and its verification on the model problem of viscous incompressible flow simulation in a square cavity. The Lagrangian description of the medium motion is used: the nodes of the finite element mesh move together with the fluid that allows to consider them as particles of the medium. Mesh cells deform when in time-stepping procedure, so it is necessary to reconstruct the mesh to provide stability of the finite element numerical procedure.Meshing algorithm allows us to obtain the mesh, which satisfies the Delaunay criteria: it is called \\the possible triangles method". This algorithm is based on the well-known Fortune method of Voronoi diagram constructing for a certain set of points in the plane. The graphical representation of the possible triangles method is shown. It is suitable to use generalization of Delaunay triangulation in order to construct meshes with polygonal cells in case of multiple nodes close to be lying on the same circle.The viscous incompressible fluid flow is described by the Navier | Stokes equations and the mass conservation equation with certain initial and boundary conditions. A fractional steps method, which allows us to avoid non-physical oscillations of the pressure, provides the timestepping procedure. Using the finite element discretization and the Bubnov | Galerkin method allows us to carry out spatial discretization.For form functions calculation of finite element mesh with polygonal cells, \

  11. Many Drops Interactions I: Simulation of Coalescence, Flocculation and Fragmentation of Multiple Colliding Drops with Smoothed Particle Hydrodynamics

    Directory of Open Access Journals (Sweden)

    Alejandro Acevedo-Malavé

    2012-06-01

    Full Text Available Smoothed Particle Hydrodynamics (SPH is a Lagrangian mesh-free formalism and has been useful to model continuous fluid. This formalism is employed to solve the Navier-Stokes equations by replacing the fluid with a set of particles. These particles are interpolation points from which properties of the fluid can be determined. In this study, the SPH method is applied to simulate the hydrodynamics interaction of many drops, showing some settings for the coalescence, fragmentation and flocculation problem of equally sized liquid drops in three-dimensional spaces. For small velocities the drops interact only through their deformed surfaces and the flocculation of the droplets arises. This result is very different if the collision velocity is large enough for the fragmentation of droplets takes place. We observe that for velocities around 15 mm/ms the coalescence of droplets occurs. The velocity vector fields formed inside the drops during the collision process are shown.

  12. Numerical simulation of the laminar hydrogen flame in the presence of a quenching mesh

    International Nuclear Information System (INIS)

    Kudriakov, S.; Studer, E.; Bin, C.

    2011-01-01

    Recent studies of J.H. Song et al., and S.Y. Yang et al. have been concentrated on mitigation measures against hydrogen risk. The authors have proposed installation of quenching meshes between compartments or around the essential equipment in order to contain hydrogen flames. Preliminary tests were conducted which demonstrated the possibility of flame extinction using metallic meshes of specific size. Considerable amount of numerical and theoretical work on flame quenching phenomenon has been performed in the second half of the last century and several techniques and models have been proposed to predict the quenching phenomenon of the laminar flame system. Most of these models appreciated the importance of heat loss to the surroundings as a primary cause of extinguishment, in particular, the heat transfer by conduction to the containing wall. The supporting simulations predict flame-quenching structure either between parallel plates (quenching distance) or inside a tube of a certain diameter (quenching diameter). In the present study the flame quenching is investigated assuming the laminar hydrogen flame propagating towards a quenching mesh using two-dimensional configuration and the earlier developed models. It is shown that due to a heat loss to a metallic grid the flame can be quenched numerically. (authors)

  13. Radiation in Particle Simulations

    International Nuclear Information System (INIS)

    More, R.; Graziani, F.; Glosli, J.; Surh, M.

    2010-01-01

    Hot dense radiative (HDR) plasmas common to Inertial Confinement Fusion (ICF) and stellar interiors have high temperature (a few hundred eV to tens of keV), high density (tens to hundreds of g/cc) and high pressure (hundreds of megabars to thousands of gigabars). Typically, such plasmas undergo collisional, radiative, atomic and possibly thermonuclear processes. In order to describe HDR plasmas, computational physicists in ICF and astrophysics use atomic-scale microphysical models implemented in various simulation codes. Experimental validation of the models used to describe HDR plasmas are difficult to perform. Direct Numerical Simulation (DNS) of the many-body interactions of plasmas is a promising approach to model validation but, previous work either relies on the collisionless approximation or ignores radiation. We present four methods that attempt a new numerical simulation technique to address a currently unsolved problem: the extension of molecular dynamics to collisional plasmas including emission and absorption of radiation. The first method applies the Lienard-Weichert solution of Maxwell's equations for a classical particle whose motion is assumed to be known. The second method expands the electromagnetic field in normal modes (planewaves in a box with periodic boundary-conditions) and solves the equation for wave amplitudes coupled to the particle motion. The third method is a hybrid molecular dynamics/Monte Carlo (MD/MC) method which calculates radiation emitted or absorbed by electron-ion pairs during close collisions. The fourth method is a generalization of the third method to include small clusters of particles emitting radiation during close encounters: one electron simultaneously hitting two ions, two electrons simultaneously hitting one ion, etc. This approach is inspired by the virial expansion method of equilibrium statistical mechanics. Using a combination of these methods we believe it is possible to do atomic-scale particle simulations of

  14. Toward An Unstructured Mesh Database

    Science.gov (United States)

    Rezaei Mahdiraji, Alireza; Baumann, Peter Peter

    2014-05-01

    Unstructured meshes are used in several application domains such as earth sciences (e.g., seismology), medicine, oceanography, cli- mate modeling, GIS as approximate representations of physical objects. Meshes subdivide a domain into smaller geometric elements (called cells) which are glued together by incidence relationships. The subdivision of a domain allows computational manipulation of complicated physical structures. For instance, seismologists model earthquakes using elastic wave propagation solvers on hexahedral meshes. The hexahedral con- tains several hundred millions of grid points and millions of hexahedral cells. Each vertex node in the hexahedrals stores a multitude of data fields. To run simulation on such meshes, one needs to iterate over all the cells, iterate over incident cells to a given cell, retrieve coordinates of cells, assign data values to cells, etc. Although meshes are used in many application domains, to the best of our knowledge there is no database vendor that support unstructured mesh features. Currently, the main tool for querying and manipulating unstructured meshes are mesh libraries, e.g., CGAL and GRAL. Mesh li- braries are dedicated libraries which includes mesh algorithms and can be run on mesh representations. The libraries do not scale with dataset size, do not have declarative query language, and need deep C++ knowledge for query implementations. Furthermore, due to high coupling between the implementations and input file structure, the implementations are less reusable and costly to maintain. A dedicated mesh database offers the following advantages: 1) declarative querying, 2) ease of maintenance, 3) hiding mesh storage structure from applications, and 4) transparent query optimization. To design a mesh database, the first challenge is to define a suitable generic data model for unstructured meshes. We proposed ImG-Complexes data model as a generic topological mesh data model which extends incidence graph model to multi

  15. All-Particle Multiscale Computation of Hypersonic Rarefied Flow

    Science.gov (United States)

    Jun, E.; Burt, J. M.; Boyd, I. D.

    2011-05-01

    This study examines a new hybrid particle scheme used as an alternative means of multiscale flow simulation. The hybrid particle scheme employs the direct simulation Monte Carlo (DSMC) method in rarefied flow regions and the low diffusion (LD) particle method in continuum flow regions. The numerical procedures of the low diffusion particle method are implemented within an existing DSMC algorithm. The performance of the LD-DSMC approach is assessed by studying Mach 10 nitrogen flow over a sphere with a global Knudsen number of 0.002. The hybrid scheme results show good overall agreement with results from standard DSMC and CFD computation. Subcell procedures are utilized to improve computational efficiency and reduce sensitivity to DSMC cell size in the hybrid scheme. This makes it possible to perform the LD-DSMC simulation on a much coarser mesh that leads to a significant reduction in computation time.

  16. Development and verification of unstructured adaptive mesh technique with edge compatibility

    International Nuclear Information System (INIS)

    Ito, Kei; Ohshima, Hiroyuki; Kunugi, Tomoaki

    2010-01-01

    In the design study of the large-sized sodium-cooled fast reactor (JSFR), one key issue is suppression of gas entrainment (GE) phenomena at a gas-liquid interface. Therefore, the authors have been developed a high-precision CFD algorithm to evaluate the GE phenomena accurately. The CFD algorithm has been developed on unstructured meshes to establish an accurate modeling of JSFR system. For two-phase interfacial flow simulations, a high-precision volume-of-fluid algorithm is employed. It was confirmed that the developed CFD algorithm could reproduce the GE phenomena in a simple GE experiment. Recently, the authors have been developed an important technique for the simulation of the GE phenomena in JSFR. That is an unstructured adaptive mesh technique which can apply fine cells dynamically to the region where the GE occurs in JSFR. In this paper, as a part of the development, a two-dimensional unstructured adaptive mesh technique is discussed. In the two-dimensional adaptive mesh technique, each cell is refined isotropically to reduce distortions of the mesh. In addition, connection cells are formed to eliminate the edge incompatibility between refined and non-refined cells. The two-dimensional unstructured adaptive mesh technique is verified by solving well-known lid-driven cavity flow problem. As a result, the two-dimensional unstructured adaptive mesh technique succeeds in providing a high-precision solution, even though poor-quality distorted initial mesh is employed. In addition, the simulation error on the two-dimensional unstructured adaptive mesh is much less than the error on the structured mesh with a larger number of cells. (author)

  17. Computational mesh generation for vascular structures with deformable surfaces

    International Nuclear Information System (INIS)

    Putter, S. de; Laffargue, F.; Breeuwer, M.; Vosse, F.N. van de; Gerritsen, F.A.; Philips Medical Systems, Best

    2006-01-01

    Computational blood flow and vessel wall mechanics simulations for vascular structures are becoming an important research tool for patient-specific surgical planning and intervention. An important step in the modelling process for patient-specific simulations is the creation of the computational mesh based on the segmented geometry. Most known solutions either require a large amount of manual processing or lead to a substantial difference between the segmented object and the actual computational domain. We have developed a chain of algorithms that lead to a closely related implementation of image segmentation with deformable models and 3D mesh generation. The resulting processing chain is very robust and leads both to an accurate geometrical representation of the vascular structure as well as high quality computational meshes. The chain of algorithms has been tested on a wide variety of shapes. A benchmark comparison of our mesh generation application with five other available meshing applications clearly indicates that the new approach outperforms the existing methods in the majority of cases. (orig.)

  18. Multi-phase Volume Segmentation with Tetrahedral Mesh

    DEFF Research Database (Denmark)

    Nguyen Trung, Tuan; Dahl, Vedrana Andersen; Bærentzen, Jakob Andreas

    Volume segmentation is efficient for reconstructing material structure, which is important for several analyses, e.g. simulation with finite element method, measurement of quantitative information like surface area, surface curvature, volume, etc. We are concerned about the representations of the 3......D volumes, which can be categorized into two groups: fixed voxel grids [1] and unstructured meshes [2]. Among these two representations, the voxel grids are more popular since manipulating a fixed grid is easier than an unstructured mesh, but they are less efficient for quantitative measurements....... In many cases, the voxel grids are converted to explicit meshes, however the conversion may reduce the accuracy of the segmentations, and the effort for meshing is also not trivial. On the other side, methods using unstructured meshes have difficulty in handling topology changes. To reduce the complexity...

  19. Particle-in-cell simulations of Hall plasma thrusters

    Science.gov (United States)

    Miranda, Rodrigo; Ferreira, Jose Leonardo; Martins, Alexandre

    2016-07-01

    Hall plasma thrusters can be modelled using particle-in-cell (PIC) simulations. In these simulations, the plasma is described by a set of equations which represent a coupled system of charged particles and electromagnetic fields. The fields are computed using a spatial grid (i.e., a discretization in space), whereas the particles can move continuously in space. Briefly, the particle and fields dynamics are computed as follows. First, forces due to electric and magnetic fields are employed to calculate the velocities and positions of particles. Next, the velocities and positions of particles are used to compute the charge and current densities at discrete positions in space. Finally, these densities are used to solve the electromagnetic field equations in the grid, which are interpolated at the position of the particles to obtain the acting forces, and restart this cycle. We will present numerical simulations using software for PIC simulations to study turbulence, wave and instabilities that arise in Hall plasma thrusters. We have sucessfully reproduced a numerical simulation of a SPT-100 Hall thruster using a two-dimensional (2D) model. In addition, we are developing a 2D model of a cylindrical Hall thruster. The results of these simulations will contribute to improve the performance of plasma thrusters to be used in Cubesats satellites currenty in development at the Plasma Laboratory at University of Brasília.

  20. IMPOSING A LAGRANGIAN PARTICLE FRAMEWORK ON AN EULERIAN HYDRODYNAMICS INFRASTRUCTURE IN FLASH

    International Nuclear Information System (INIS)

    Dubey, A.; Daley, C.; Weide, K.; Graziani, C.; ZuHone, J.; Ricker, P. M.

    2012-01-01

    In many astrophysical simulations, both Eulerian and Lagrangian quantities are of interest. For example, in a galaxy cluster merger simulation, the intracluster gas can have Eulerian discretization, while dark matter can be modeled using particles. FLASH, a component-based scientific simulation code, superimposes a Lagrangian framework atop an adaptive mesh refinement Eulerian framework to enable such simulations. The discretization of the field variables is Eulerian, while the Lagrangian entities occur in many different forms including tracer particles, massive particles, charged particles in particle-in-cell mode, and Lagrangian markers to model fluid-structure interactions. These widely varying roles for Lagrangian entities are possible because of the highly modular, flexible, and extensible architecture of the Lagrangian framework. In this paper, we describe the Lagrangian framework in FLASH in the context of two very different applications, Type Ia supernovae and galaxy cluster mergers, which use the Lagrangian entities in fundamentally different ways.

  1. Imposing a Lagrangian Particle Framework on an Eulerian Hydrodynamics Infrastructure in Flash

    Science.gov (United States)

    Dubey, A.; Daley, C.; ZuHone, J.; Ricker, P. M.; Weide, K.; Graziani, C.

    2012-01-01

    In many astrophysical simulations, both Eulerian and Lagrangian quantities are of interest. For example, in a galaxy cluster merger simulation, the intracluster gas can have Eulerian discretization, while dark matter can be modeled using particles. FLASH, a component-based scientific simulation code, superimposes a Lagrangian framework atop an adaptive mesh refinement Eulerian framework to enable such simulations. The discretization of the field variables is Eulerian, while the Lagrangian entities occur in many different forms including tracer particles, massive particles, charged particles in particle-in-cell mode, and Lagrangian markers to model fluid structure interactions. These widely varying roles for Lagrangian entities are possible because of the highly modular, flexible, and extensible architecture of the Lagrangian framework. In this paper, we describe the Lagrangian framework in FLASH in the context of two very different applications, Type Ia supernovae and galaxy cluster mergers, which use the Lagrangian entities in fundamentally different ways.

  2. Global Particle-in-Cell Simulations of Mercury's Magnetosphere

    Science.gov (United States)

    Schriver, D.; Travnicek, P. M.; Lapenta, G.; Amaya, J.; Gonzalez, D.; Richard, R. L.; Berchem, J.; Hellinger, P.

    2017-12-01

    Spacecraft observations of Mercury's magnetosphere have shown that kinetic ion and electron particle effects play a major role in the transport, acceleration, and loss of plasma within the magnetospheric system. Kinetic processes include reconnection, the breakdown of particle adiabaticity and wave-particle interactions. Because of the vast range in spatial scales involved in magnetospheric dynamics, from local electron Debye length scales ( meters) to solar wind/planetary magnetic scale lengths (tens to hundreds of planetary radii), fully self-consistent kinetic simulations of a global planetary magnetosphere remain challenging. Most global simulations of Earth's and other planet's magnetosphere are carried out using MHD, enhanced MHD (e.g., Hall MHD), hybrid, or a combination of MHD and particle in cell (PIC) simulations. Here, 3D kinetic self-consistent hybrid (ion particle, electron fluid) and full PIC (ion and electron particle) simulations of the solar wind interaction with Mercury's magnetosphere are carried out. Using the implicit PIC and hybrid simulations, Mercury's relatively small, but highly kinetic magnetosphere will be examined to determine how the self-consistent inclusion of electrons affects magnetic reconnection, particle transport and acceleration of plasma at Mercury. Also the spatial and energy profiles of precipitating magnetospheric ions and electrons onto Mercury's surface, which can strongly affect the regolith in terms of space weathering and particle outflow, will be examined with the PIC and hybrid codes. MESSENGER spacecraft observations are used both to initiate and validate the global kinetic simulations to achieve a deeper understanding of the role kinetic physics play in magnetospheric dynamics.

  3. Optimization-based Fluid Simulation on Unstructured Meshes

    DEFF Research Database (Denmark)

    Misztal, Marek Krzysztof; Bridson, Robert; Erleben, Kenny

    2010-01-01

    for solving the fluid dynamics equations as well as direct access to the interface geometry data, making in- clusion of a new surface energy term feasible. Furthermore, using an unstructured mesh makes it straightforward to handle curved solid boundaries and gives us a possibility to explore several fluid...

  4. Visualization of particle in cell simulation

    International Nuclear Information System (INIS)

    Chen Ming; Cheng Yinhui

    2003-01-01

    This paper is trying to provide a new technique of the visualization for the Particle In Cell simulation, which takes effect by using the MATLAB external interface, so the real-time obsevation of particles came easier and more efficient. With this method, state of the particles, considered as 'particle cloud' can be found in the image produced automatically and their movement can be predicted. (authors)

  5. Particle-transport simulation with the Monte Carlo method

    International Nuclear Information System (INIS)

    Carter, L.L.; Cashwell, E.D.

    1975-01-01

    Attention is focused on the application of the Monte Carlo method to particle transport problems, with emphasis on neutron and photon transport. Topics covered include sampling methods, mathematical prescriptions for simulating particle transport, mechanics of simulating particle transport, neutron transport, and photon transport. A literature survey of 204 references is included. (GMT)

  6. Opfront: mesh

    DEFF Research Database (Denmark)

    2015-01-01

    Mesh generation and visualization software based on the CGAL library. Folder content: drawmesh Visualize slices of the mesh (surface/volumetric) as wireframe on top of an image (3D). drawsurf Visualize surfaces of the mesh (surface/volumetric). img2mesh Convert isosurface in image to volumetric m...... mesh (medit format). img2off Convert isosurface in image to surface mesh (off format). off2mesh Convert surface mesh (off format) to volumetric mesh (medit format). reduce Crop and resize 3D and stacks of images. data Example data to test the library on...

  7. Adaptive moving mesh methods for simulating one-dimensional groundwater problems with sharp moving fronts

    Science.gov (United States)

    Huang, W.; Zheng, Lingyun; Zhan, X.

    2002-01-01

    Accurate modelling of groundwater flow and transport with sharp moving fronts often involves high computational cost, when a fixed/uniform mesh is used. In this paper, we investigate the modelling of groundwater problems using a particular adaptive mesh method called the moving mesh partial differential equation approach. With this approach, the mesh is dynamically relocated through a partial differential equation to capture the evolving sharp fronts with a relatively small number of grid points. The mesh movement and physical system modelling are realized by solving the mesh movement and physical partial differential equations alternately. The method is applied to the modelling of a range of groundwater problems, including advection dominated chemical transport and reaction, non-linear infiltration in soil, and the coupling of density dependent flow and transport. Numerical results demonstrate that sharp moving fronts can be accurately and efficiently captured by the moving mesh approach. Also addressed are important implementation strategies, e.g. the construction of the monitor function based on the interpolation error, control of mesh concentration, and two-layer mesh movement. Copyright ?? 2002 John Wiley and Sons, Ltd.

  8. Deploy production sliding mesh capability with linear solver benchmarking.

    Energy Technology Data Exchange (ETDEWEB)

    Domino, Stefan P. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Thomas, Stephen [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Barone, Matthew F. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Williams, Alan B. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Ananthan, Shreyas [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Knaus, Robert C. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Overfelt, James [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Sprague, Mike [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Rood, Jon [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2018-02-01

    Wind applications require the ability to simulate rotating blades. To support this use-case, a novel design-order sliding mesh algorithm has been developed and deployed. The hybrid method combines the control volume finite element methodology (CVFEM) with concepts found within a discontinuous Galerkin (DG) finite element method (FEM) to manage a sliding mesh. The method has been demonstrated to be design-order for the tested polynomial basis (P=1 and P=2) and has been deployed to provide production simulation capability for a Vestas V27 (225 kW) wind turbine. Other stationary and canonical rotating ow simulations are also presented. As the majority of wind-energy applications are driving extensive usage of hybrid meshes, a foundational study that outlines near-wall numerical behavior for a variety of element topologies is presented. Results indicate that the proposed nonlinear stabilization operator (NSO) is an effective stabilization methodology to control Gibbs phenomena at large cell Peclet numbers. The study also provides practical mesh resolution guidelines for future analysis efforts. Application-driven performance and algorithmic improvements have been carried out to increase robustness of the scheme on hybrid production wind energy meshes. Specifically, the Kokkos-based Nalu Kernel construct outlined in the FY17/Q4 ExaWind milestone has been transitioned to the hybrid mesh regime. This code base is exercised within a full V27 production run. Simulation timings for parallel search and custom ghosting are presented. As the low-Mach application space requires implicit matrix solves, the cost of matrix reinitialization has been evaluated on a variety of production meshes. Results indicate that at low element counts, i.e., fewer than 100 million elements, matrix graph initialization and preconditioner setup times are small. However, as mesh sizes increase, e.g., 500 million elements, simulation time associated with \\setup-up" costs can increase to nearly 50% of

  9. Parallel Finite Element Particle-In-Cell Code for Simulations of Space-charge Dominated Beam-Cavity Interactions

    International Nuclear Information System (INIS)

    Candel, A.; Kabel, A.; Ko, K.; Lee, L.; Li, Z.; Limborg, C.; Ng, C.; Prudencio, E.; Schussman, G.; Uplenchwar, R.

    2007-01-01

    Over the past years, SLAC's Advanced Computations Department (ACD) has developed the parallel finite element (FE) particle-in-cell code Pic3P (Pic2P) for simulations of beam-cavity interactions dominated by space-charge effects. As opposed to standard space-charge dominated beam transport codes, which are based on the electrostatic approximation, Pic3P (Pic2P) includes space-charge, retardation and boundary effects as it self-consistently solves the complete set of Maxwell-Lorentz equations using higher-order FE methods on conformal meshes. Use of efficient, large-scale parallel processing allows for the modeling of photoinjectors with unprecedented accuracy, aiding the design and operation of the next-generation of accelerator facilities. Applications to the Linac Coherent Light Source (LCLS) RF gun are presented

  10. High-resolution Hydrodynamic Simulation of Tidal Detonation of a Helium White Dwarf by an Intermediate Mass Black Hole

    Science.gov (United States)

    Tanikawa, Ataru

    2018-05-01

    We demonstrate tidal detonation during a tidal disruption event (TDE) of a helium (He) white dwarf (WD) with 0.45 M ⊙ by an intermediate mass black hole using extremely high-resolution simulations. Tanikawa et al. have shown tidal detonation in results of previous studies from unphysical heating due to low-resolution simulations, and such unphysical heating occurs in three-dimensional (3D) smoothed particle hydrodynamics (SPH) simulations even with 10 million SPH particles. In order to avoid such unphysical heating, we perform 3D SPH simulations up to 300 million SPH particles, and 1D mesh simulations using flow structure in the 3D SPH simulations for 1D initial conditions. The 1D mesh simulations have higher resolutions than the 3D SPH simulations. We show that tidal detonation occurs and confirm that this result is perfectly converged with different space resolution in both 3D SPH and 1D mesh simulations. We find that detonation waves independently arise in leading parts of the WD, and yield large amounts of 56Ni. Although detonation waves are not generated in trailing parts of the WD, the trailing parts would receive detonation waves generated in the leading parts and would leave large amounts of Si group elements. Eventually, this He WD TDE would synthesize 56Ni of 0.30 M ⊙ and Si group elements of 0.08 M ⊙, and could be observed as a luminous thermonuclear transient comparable to SNe Ia.

  11. Optimization of Particle Search Algorithm for CFD-DEM Simulations

    Directory of Open Access Journals (Sweden)

    G. Baryshev

    2013-09-01

    Full Text Available Discrete element method has numerous applications in particle physics. However, simulating particles as discrete entities can become costly for large systems. In time-driven DEM simulation most computation time is taken by contact search stage. We propose an efficient collision detection method which is based on sorting particles by their coordinates. Using multiple sorting criteria allows minimizing number of potential neighbours and defines fitness of this approach for simulation of massive systems in 3D. This method is compared to a common approach that consists of placing particles onto a grid of cells. Advantage of the new approach is independence of simulation parameters upon particle radius and domain size.

  12. Numerical simulation of evolutionary erodible bedforms using the particle finite element method

    Science.gov (United States)

    Bravo, Rafael; Becker, Pablo; Ortiz, Pablo

    2017-07-01

    This paper presents a numerical strategy for the simulation of flows with evolutionary erodible boundaries. The fluid equations are fully resolved in 3D, while the sediment transport is modelled using the Exner equation and solved with an explicit Lagrangian procedure based on a fixed 2D mesh. Flow and sediment are coupled in geometry by deforming the fluid mesh in the vertical direction and in velocities with the experimental sediment flux computed using the Meyer Peter Müller model. A comparison with real experiments on channels is performed, giving good agreement.

  13. Plasma physics via particle simulation

    International Nuclear Information System (INIS)

    Birdsall, C.K.

    1981-01-01

    Plasmas are studied by following the motion of many particles in applied and self fields, analytically, experimentally and computationally. Plasmas for magnetic fusion energy devices are very hot, nearly collisionless and magnetized, with scale lengths of many ion gyroradii and Debye lengths. The analytic studies of such plasmas are very difficult as the plasma is nonuniform, anisotropic and nonlinear. The experimental studies have become very expensive in time and money, as the size, density and temperature approach fusion reactor values. Computational studies using many particles and/or fluids have complemented both theories and experiments for many years and have progressed to fully three dimensional electromagnetic models, albeit with hours of running times on the fastest largest computers. Particle simulation methods are presented in some detail, showing particle advance from acceleration to velocity to position, followed by calculation of the fields from charge and current densities and then further particle advance, and so on. Limitations due to the time stepping and use of a spatial grid are given, to avoid inaccuracies and instabilities. Examples are given for an electrostatic program in one dimension of an orbit averaging program, and for a three dimensional electromagnetic program. Applications of particle simulations of plasmas in magnetic and inertial fusion devices continue to grow, as well as to plasmas and beams in peripheral devices, such as sources, accelerators, and converters. (orig.)

  14. A parallel direct-forcing fictitious domain method for simulating microswimmers

    Science.gov (United States)

    Gao, Tong; Lin, Zhaowu

    2017-11-01

    We present a 3D parallel direct-forcing fictitious domain method for simulating swimming micro-organisms at small Reynolds numbers. We treat the motile micro-swimmers as spherical rigid particles using the ``Squirmer'' model. The particle dynamics are solved on the moving Larangian meshes that overlay upon a fixed Eulerian mesh for solving the fluid motion, and the momentum exchange between the two phases is resolved by distributing pseudo body-forces over the particle interior regions which constrain the background fictitious fluids to follow the particle movement. While the solid and fluid subproblems are solved separately, no inner-iterations are required to enforce numerical convergence. We demonstrate the accuracy and robustness of the method by comparing our results with the existing analytical and numerical studies for various cases of single particle dynamics and particle-particle interactions. We also perform a series of numerical explorations to obtain statistical and rheological measurements to characterize the dynamics and structures of Squirmer suspensions. NSF DMS 1619960.

  15. Euler-Lagrange Simulations of Shock Wave-Particle Cloud Interaction

    Science.gov (United States)

    Koneru, Rahul; Rollin, Bertrand; Ouellet, Frederick; Park, Chanyoung; Balachandar, S.

    2017-11-01

    Numerical experiments of shock interacting with an evolving and fixed cloud of particles are performed. In these simulations we use Eulerian-Lagrangian approach along with state-of-the-art point-particle force and heat transfer models. As validation, we use Sandia Multiphase Shock Tube experiments and particle-resolved simulations. The particle curtain upon interaction with the shock wave is expected to experience Kelvin-Helmholtz (KH) and Richtmyer-Meshkov (RM) instabilities. In the simulations evolving the particle cloud, the initial volume fraction profile matches with that of Sandia Multiphase Shock Tube experiments, and the shock Mach number is limited to M =1.66. Measurements of particle dispersion are made at different initial volume fractions. A detailed analysis of the influence of initial conditions on the evolution of the particle cloudis presented. The early time behavior of the models is studied in the fixed bed simulations at varying volume fractions and shock Mach numbers.The mean gas quantities are measured in the context of 1-way and 2-way coupled simulations. This work was supported by the U.S. Department of Energy, National Nuclear Security Administration, Advanced Simulation and Computing Program, as a Cooperative Agreement under the Predictive Science Academic Alliance Program, Contract No. DE-NA0002378.

  16. Numerical simulation of particle settling and cohesion in liquid

    Energy Technology Data Exchange (ETDEWEB)

    Johno, Y; Nakashima, K; Shigematsu, T; Ono, B [SASEBO National College of Technology, 1-1 Okishin, Sasebo, Nagasaki, 857-1193 (Japan); Satomi, M, E-mail: yjohno@post.cc.sasebo.ac.j [Sony Semiconductor Kyushu Corporation, Kikuchigun, Kumamoto (Japan)

    2009-02-01

    In this study, the motions of particles and particle clusters in liquid were numerically simulated. The particles of two sizes (Dp=40mum and 20mum) settle while repeating cohesion and dispersion, and finally the sediment of particles are formed at the bottom of a hexahedron container which is filled up with pure water. The flow field was solved with the Navier-Stokes equations and the particle motions were solved with the Lagrangian-type motion equations, where the interaction between fluid and particles due to drag forces were taken into account. The collision among particles was calculated using Distinct Element Method (DEM), and the effects of cohesive forces by van der Waals force acting on particle contact points were taken into account. Numerical simulations were performed under conditions in still flow and in shear flow. It was found that the simulation results enable us to know the state of the particle settling and the particle condensation.

  17. Second-order particle-in-cell (PIC) computational method in the one-dimensional variable Eulerian mesh system

    International Nuclear Information System (INIS)

    Pyun, J.J.

    1981-01-01

    As part of an effort to incorporate the variable Eulerian mesh into the second-order PIC computational method, a truncation error analysis was performed to calculate the second-order error terms for the variable Eulerian mesh system. The results that the maximum mesh size increment/decrement is limited to be α(Δr/sub i/) 2 where Δr/sub i/ is a non-dimensional mesh size of the ith cell, and α is a constant of order one. The numerical solutions of Burgers' equation by the second-order PIC method in the variable Eulerian mesh system wer compared with its exact solution. It was found that the second-order accuracy in the PIC method was maintained under the above condition. Additional problems were analyzed using the second-order PIC methods in both variable and uniform Eulerian mesh systems. The results indicate that the second-order PIC method in the variable Eulerian mesh system can provide substantial computational time saving with no loss in accuracy

  18. Direct numerical simulation of bubbles with adaptive mesh refinement with distributed algorithms

    International Nuclear Information System (INIS)

    Talpaert, Arthur

    2017-01-01

    This PhD work presents the implementation of the simulation of two-phase flows in conditions of water-cooled nuclear reactors, at the scale of individual bubbles. To achieve that, we study several models for Thermal-Hydraulic flows and we focus on a technique for the capture of the thin interface between liquid and vapour phases. We thus review some possible techniques for adaptive Mesh Refinement (AMR) and provide algorithmic and computational tools adapted to patch-based AMR, which aim is to locally improve the precision in regions of interest. More precisely, we introduce a patch-covering algorithm designed with balanced parallel computing in mind. This approach lets us finely capture changes located at the interface, as we show for advection test cases as well as for models with hyperbolic-elliptic coupling. The computations we present also include the simulation of the incompressible Navier-Stokes system, which models the shape changes of the interface between two non-miscible fluids. (author) [fr

  19. THM-GTRF: New Spider meshes, New Hydra-TH runs

    Energy Technology Data Exchange (ETDEWEB)

    Bakosi, Jozsef [Los Alamos National Laboratory; Christon, Mark A. [Los Alamos National Laboratory; Francois, Marianne M. [Los Alamos National Laboratory; Lowrie, Robert B. [Los Alamos National Laboratory; Nourgaliev, Robert [Los Alamos National Laboratory

    2012-06-20

    Progress is reported on computational capabilities for the grid-to-rod-fretting (GTRF) problem of pressurized water reactors. Numeca's Hexpress/Hybrid mesh generator is demonstrated as an excellent alternative to generating computational meshes for complex flow geometries, such as in GTRF. Mesh assessment is carried out using standard industrial computational fluid dynamics practices. Hydra-TH, a simulation code developed at LANL for reactor thermal-hydraulics, is demonstrated on hybrid meshes, containing different element types. A series of new Hydra-TH calculations has been carried out collecting turbulence statistics. Preliminary results on the newly generated meshes are discussed; full analysis will be documented in the L3 milestone, THM.CFD.P5.05, Sept. 2012.

  20. Multi-Algorithm Particle Simulations with Spatiocyte.

    Science.gov (United States)

    Arjunan, Satya N V; Takahashi, Koichi

    2017-01-01

    As quantitative biologists get more measurements of spatially regulated systems such as cell division and polarization, simulation of reaction and diffusion of proteins using the data is becoming increasingly relevant to uncover the mechanisms underlying the systems. Spatiocyte is a lattice-based stochastic particle simulator for biochemical reaction and diffusion processes. Simulations can be performed at single molecule and compartment spatial scales simultaneously. Molecules can diffuse and react in 1D (filament), 2D (membrane), and 3D (cytosol) compartments. The implications of crowded regions in the cell can be investigated because each diffusing molecule has spatial dimensions. Spatiocyte adopts multi-algorithm and multi-timescale frameworks to simulate models that simultaneously employ deterministic, stochastic, and particle reaction-diffusion algorithms. Comparison of light microscopy images to simulation snapshots is supported by Spatiocyte microscopy visualization and molecule tagging features. Spatiocyte is open-source software and is freely available at http://spatiocyte.org .

  1. High performance stream computing for particle beam transport simulations

    International Nuclear Information System (INIS)

    Appleby, R; Bailey, D; Higham, J; Salt, M

    2008-01-01

    Understanding modern particle accelerators requires simulating charged particle transport through the machine elements. These simulations can be very time consuming due to the large number of particles and the need to consider many turns of a circular machine. Stream computing offers an attractive way to dramatically improve the performance of such simulations by calculating the simultaneous transport of many particles using dedicated hardware. Modern Graphics Processing Units (GPUs) are powerful and affordable stream computing devices. The results of simulations of particle transport through the booster-to-storage-ring transfer line of the DIAMOND synchrotron light source using an NVidia GeForce 7900 GPU are compared to the standard transport code MAD. It is found that particle transport calculations are suitable for stream processing and large performance increases are possible. The accuracy and potential speed gains are compared and the prospects for future work in the area are discussed

  2. Accurate reaction-diffusion operator splitting on tetrahedral meshes for parallel stochastic molecular simulations

    Energy Technology Data Exchange (ETDEWEB)

    Hepburn, I.; De Schutter, E., E-mail: erik@oist.jp [Computational Neuroscience Unit, Okinawa Institute of Science and Technology Graduate University, Onna, Okinawa 904 0495 (Japan); Theoretical Neurobiology & Neuroengineering, University of Antwerp, Antwerp 2610 (Belgium); Chen, W. [Computational Neuroscience Unit, Okinawa Institute of Science and Technology Graduate University, Onna, Okinawa 904 0495 (Japan)

    2016-08-07

    Spatial stochastic molecular simulations in biology are limited by the intense computation required to track molecules in space either in a discrete time or discrete space framework, which has led to the development of parallel methods that can take advantage of the power of modern supercomputers in recent years. We systematically test suggested components of stochastic reaction-diffusion operator splitting in the literature and discuss their effects on accuracy. We introduce an operator splitting implementation for irregular meshes that enhances accuracy with minimal performance cost. We test a range of models in small-scale MPI simulations from simple diffusion models to realistic biological models and find that multi-dimensional geometry partitioning is an important consideration for optimum performance. We demonstrate performance gains of 1-3 orders of magnitude in the parallel implementation, with peak performance strongly dependent on model specification.

  3. Simulation of gas mixing and transport in a multi-compartment geometry using the GOTHIC containment code and relatively coarse meshes

    International Nuclear Information System (INIS)

    Andreani, Michele; Paladino, Domenico

    2010-01-01

    The recently concluded OECD SETH project included twenty-four experiments on basic flows and gas transport and mixing driven by jets and plumes in two, large, connected vessels of the PANDA facility. The experiments featured injection of saturated or superheated steam, or a mixture of steam and helium in one vessel and venting from the same vessel or from the connected one. These tests have been especially designed for providing an extensive data base for the assessment of three-dimensional codes, including CFD codes. In particular, one of the goals of the analytical activities associated with the experiments was to evaluate the detail of the model (mesh) necessary for capturing the various phenomena. This work reports an overview of the results obtained for these experimental data using the advanced containment code GOTHIC and relatively coarse meshes, which are coarser than the ones typically used for the simulation with commercial CFD codes, but are still representative of the models which are currently affordable for a full containment analysis. In general, the phenomena were correctly represented in the simulations with GOTHIC, and the agreement of the results with the data was in most cases pretty good, in some cases excellent. Only for a few tests (or particular phenomena occurring in some tests) the simulations showed noticeable discrepancies with the experimental data, which could be referred to either an insufficiently detailed mesh or to lack of specialized models for local effects.

  4. A collision model in plasma particle simulations

    International Nuclear Information System (INIS)

    Ma Yanyun; Chang Wenwei; Yin Yan; Yue Zongwu; Cao Lihua; Liu Daqing

    2000-01-01

    In order to offset the collisional effects reduced by using finite-size particles, β particle clouds are used in particle simulation codes (β is the ratio of charge or mass of modeling particles to real ones). The method of impulse approximation (strait line orbit approximation) is used to analyze the scattering cross section of β particle clouds plasmas. The authors can obtain the relation of the value of a and β and scattering cross section (a is the radius of β particle cloud). By using this relation the authors can determine the value of a and β so that the collisional effects of the modeling system is correspondent with the real one. The authors can also adjust the values of a and β so that the authors can enhance or reduce the collisional effects fictitiously. The results of simulation are in good agreement with the theoretical ones

  5. Fragmentation of Millimeter-Size Hypervelocity Projectiles on Combined Mesh-Plate Bumpers

    Directory of Open Access Journals (Sweden)

    Aleksandr Cherniaev

    2017-01-01

    Full Text Available This numerical study evaluates the concept of a combined mesh-plate bumper as a shielding system protecting unmanned spacecraft from small (1 mm orbital debris impacts. Two-component bumpers consisting of an external layer of woven mesh (aluminum or steel directly applied to a surface of the aluminum plate are considered. Results of numerical modeling with a projectile velocity of 7 km/s indicate that, in comparison to the steel mesh-combined bumper, the combination of aluminum mesh and aluminum plate provides better fragmentation of small hypervelocity projectiles. At the same time, none of the combined mesh/plate bumpers provide a significant increase of ballistic properties as compared to an aluminum plate bumper. This indicates that the positive results reported in the literature for bumpers with metallic meshes and large projectiles are not scalable down to millimeter-sized particles. Based on this investigation’s results, a possible modification of the combined mesh/plate bumper is proposed for the future study.

  6. A Generalized Weight-Based Particle-In-Cell Simulation Scheme

    International Nuclear Information System (INIS)

    Lee, W.W.; Jenkins, T.G.; Ethier, S.

    2010-01-01

    A generalized weight-based particle simulation scheme suitable for simulating magnetized plasmas, where the zeroth-order inhomogeneity is important, is presented. The scheme is an extension of the perturbative simulation schemes developed earlier for particle-in-cell (PIC) simulations. The new scheme is designed to simulate both the perturbed distribution ((delta)f) and the full distribution (full-F) within the same code. The development is based on the concept of multiscale expansion, which separates the scale lengths of the background inhomogeneity from those associated with the perturbed distributions. The potential advantage for such an arrangement is to minimize the particle noise by using (delta)f in the linear stage stage of the simulation, while retaining the flexibility of a full-F capability in the fully nonlinear stage of the development when signals associated with plasma turbulence are at a much higher level than those from the intrinsic particle noise.

  7. Partially linearized algorithms in gyrokinetic particle simulation

    Energy Technology Data Exchange (ETDEWEB)

    Dimits, A.M.; Lee, W.W.

    1990-10-01

    In this paper, particle simulation algorithms with time-varying weights for the gyrokinetic Vlasov-Poisson system have been developed. The primary purpose is to use them for the removal of the selected nonlinearities in the simulation of gradient-driven microturbulence so that the relative importance of the various nonlinear effects can be assessed. It is hoped that the use of these procedures will result in a better understanding of the transport mechanisms and scaling in tokamaks. Another application of these algorithms is for the improvement of the numerical properties of the simulation plasma. For instance, implementations of such algorithms (1) enable us to suppress the intrinsic numerical noise in the simulation, and (2) also make it possible to regulate the weights of the fast-moving particles and, in turn, to eliminate the associated high frequency oscillations. Examples of their application to drift-type instabilities in slab geometry are given. We note that the work reported here represents the first successful use of the weighted algorithms in particle codes for the nonlinear simulation of plasmas.

  8. Partially linearized algorithms in gyrokinetic particle simulation

    International Nuclear Information System (INIS)

    Dimits, A.M.; Lee, W.W.

    1990-10-01

    In this paper, particle simulation algorithms with time-varying weights for the gyrokinetic Vlasov-Poisson system have been developed. The primary purpose is to use them for the removal of the selected nonlinearities in the simulation of gradient-driven microturbulence so that the relative importance of the various nonlinear effects can be assessed. It is hoped that the use of these procedures will result in a better understanding of the transport mechanisms and scaling in tokamaks. Another application of these algorithms is for the improvement of the numerical properties of the simulation plasma. For instance, implementations of such algorithms (1) enable us to suppress the intrinsic numerical noise in the simulation, and (2) also make it possible to regulate the weights of the fast-moving particles and, in turn, to eliminate the associated high frequency oscillations. Examples of their application to drift-type instabilities in slab geometry are given. We note that the work reported here represents the first successful use of the weighted algorithms in particle codes for the nonlinear simulation of plasmas

  9. Simulation of concentration distribution of urban particles under wind

    Science.gov (United States)

    Chen, Yanghou; Yang, Hangsheng

    2018-02-01

    The concentration of particulate matter in the air is too high, which seriously affects people’s health. The concentration of particles in densely populated towns is also high. Understanding the distribution of particles in the air helps to remove them passively. The concentration distribution of particles in urban streets is simulated by using the FLUENT software. The simulation analysis based on Discrete Phase Modelling (DPM) of FLUENT. Simulation results show that the distribution of the particles is caused by different layout of buildings. And it is pointed out that in the windward area of the building and the leeward sides of the high-rise building are the areas with high concentration of particles. Understanding the concentration of particles in different areas is also helpful for people to avoid and reduce the concentration of particles in high concentration areas.

  10. Macroscale particle simulation of externally driven magnetic reconnection

    International Nuclear Information System (INIS)

    Murakami, Sadayoshi; Sato, Tetsuya.

    1991-09-01

    Externally driven reconnection, assuming an anomalous particle collision model, is numerically studied by means of a 2.5D macroscale particle simulation code in which the field and particle motions are solved self-consistently. Explosive magnetic reconnection and energy conversion are observed as a result of slow shock formation. Electron and ion distribution functions exhibit large bulk acceleration and heating of the plasma. Simulation runs with different collision parameters suggest that the development of reconnection, particle acceleration and heating do not significantly depend on the parameters of the collision model. (author)

  11. Full particle simulations of quasi-perpendicular shocks

    Science.gov (United States)

    Lembège, B.

    This tutorial-style review is dedicated to the different strategies and constraints used for analysing the dynamics of a collisionless shocks with full particle simulations. Main results obtained with such simulations can be found in published materials (recent references are provided in this text); these will be only quoted herein in order to illustrate a few aspects of these simulations. Thanks to the large improvement of super computers, full particle simulations reveal to be quite helpful for analyzing in details the dynamics of collisionless shocks. The main characteristics of such codes can be shortly reminded as follows: one resolves the full set of Poisson and Maxwell's equations without any approximation. Two approaches are commonly used for resolving this equation's set, more precisely the space derivatives: (i) the finite difference approach and (ii) the use of FFT's (Fast Fourier Transform). Two advantages of approach (ii) are that FFT's are highly optimized in supercomputers libraries, and these allow to separate all fields components into two groups: the longitudinal electrostatic component El (solution of Poisson equation) and the transverse electromagnetic components Et and Bt solutions of the Maxwell's equations (so called "fields pusher"). Such a separation is quite helpful in the post processing stage necessary for the data analysis, as will be explained in the presentation. both ions and electrons populations are treated as individual finite-size particles and suffer the effects of all fields via the Lorentz force, so called "particle pusher", which is applied to each particle. Because of the large number of particles commonly used, the particle pusher represents the most expensive part of the calculations on which most efforts of optimisation needs to be performed (in terms of "vectorisation" or of "parallelism"). Relativistic effects may be included in this force via the use of particle momemtum. Each particle has three velocity components (vx

  12. A Coulomb collision algorithm for weighted particle simulations

    Science.gov (United States)

    Miller, Ronald H.; Combi, Michael R.

    1994-01-01

    A binary Coulomb collision algorithm is developed for weighted particle simulations employing Monte Carlo techniques. Charged particles within a given spatial grid cell are pair-wise scattered, explicitly conserving momentum and implicitly conserving energy. A similar algorithm developed by Takizuka and Abe (1977) conserves momentum and energy provided the particles are unweighted (each particle representing equal fractions of the total particle density). If applied as is to simulations incorporating weighted particles, the plasma temperatures equilibrate to an incorrect temperature, as compared to theory. Using the appropriate pairing statistics, a Coulomb collision algorithm is developed for weighted particles. The algorithm conserves energy and momentum and produces the appropriate relaxation time scales as compared to theoretical predictions. Such an algorithm is necessary for future work studying self-consistent multi-species kinetic transport.

  13. Mesh Excision: Is Total Mesh Excision Necessary?

    Science.gov (United States)

    Wolff, Gillian F; Winters, J Christian; Krlin, Ryan M

    2016-04-01

    Nearly 29% of women will undergo a secondary, repeat operation for pelvic organ prolapse (POP) symptom recurrence following a primary repair, as reported by Abbott et al. (Am J Obstet Gynecol 210:163.e1-163.e1, 2014). In efforts to decrease the rates of failure, graft materials have been utilized to augment transvaginal repairs. Following the success of using polypropylene mesh (PPM) for stress urinary incontinence (SUI), the use of PPM in the transvaginal repair of POP increased. However, in recent years, significant concerns have been raised about the safety of PPM mesh. Complications, some specific to mesh, such as exposures, erosion, dyspareunia, and pelvic pain, have been reported with increased frequency. In the current literature, there is not substantive evidence to suggest that PPM has intrinsic properties that warrant total mesh removal in the absence of complications. There are a number of complications that can occur after transvaginal mesh placement that do warrant surgical intervention after failure of conservative therapy. In aggregate, there are no high-quality controlled studies that clearly demonstrate that total mesh removal is consistently more likely to achieve pain reduction. In the cases of obstruction and erosion, it seems clear that definitive removal of the offending mesh is associated with resolution of symptoms in the majority of cases and reasonable practice. There are a number of complications that can occur with removal of mesh, and patients should be informed of this as they formulate a choice of treatment. We will review these considerations as we examine the clinical question of whether total versus partial removal of mesh is necessary for the resolution of complications following transvaginal mesh placement.

  14. Fluid flow and heat transfer investigation of pebble bed reactors using mesh-adaptive LES

    International Nuclear Information System (INIS)

    Pavlidis, Dimitrios; Lathouwers, Danny

    2013-01-01

    The very high temperature reactor is one of the designs currently being considered for nuclear power generation. One its variants is the pebble bed reactor in which the coolant passes through complex geometries (pores) at high Reynolds numbers. A computational fluid dynamics model with anisotropic mesh adaptivity is used to investigate coolant flow and heat transfer in such reactors. A novel method for implicitly incorporating solid boundaries based on multi-fluid flow modelling is adopted. The resulting model is able to resolve and simulate flow and heat transfer in randomly packed beds, regardless of the actual geometry, starting off with arbitrarily coarse meshes. The model is initially evaluated using an orderly stacked square channel of channel-height-to-particle diameter ratio of unity for a range of Reynolds numbers. The model is then applied to the face-centred cubical geometry. coolant flow and heat transfer patterns are investigated

  15. Track-structure simulations for charged particles.

    Science.gov (United States)

    Dingfelder, Michael

    2012-11-01

    Monte Carlo track-structure simulations provide a detailed and accurate picture of radiation transport of charged particles through condensed matter of biological interest. Liquid water serves as a surrogate for soft tissue and is used in most Monte Carlo track-structure codes. Basic theories of radiation transport and track-structure simulations are discussed and differences compared to condensed history codes highlighted. Interaction cross sections for electrons, protons, alpha particles, and light and heavy ions are required input data for track-structure simulations. Different calculation methods, including the plane-wave Born approximation, the dielectric theory, and semi-empirical approaches are presented using liquid water as a target. Low-energy electron transport and light ion transport are discussed as areas of special interest.

  16. Direct numerical simulations of particle-laden density currents with adaptive, discontinuous finite elements

    Directory of Open Access Journals (Sweden)

    S. D. Parkinson

    2014-09-01

    Full Text Available High-resolution direct numerical simulations (DNSs are an important tool for the detailed analysis of turbidity current dynamics. Models that resolve the vertical structure and turbulence of the flow are typically based upon the Navier–Stokes equations. Two-dimensional simulations are known to produce unrealistic cohesive vortices that are not representative of the real three-dimensional physics. The effect of this phenomena is particularly apparent in the later stages of flow propagation. The ideal solution to this problem is to run the simulation in three dimensions but this is computationally expensive. This paper presents a novel finite-element (FE DNS turbidity current model that has been built within Fluidity, an open source, general purpose, computational fluid dynamics code. The model is validated through re-creation of a lock release density current at a Grashof number of 5 × 106 in two and three dimensions. Validation of the model considers the flow energy budget, sedimentation rate, head speed, wall normal velocity profiles and the final deposit. Conservation of energy in particular is found to be a good metric for measuring model performance in capturing the range of dynamics on a range of meshes. FE models scale well over many thousands of processors and do not impose restrictions on domain shape, but they are computationally expensive. The use of adaptive mesh optimisation is shown to reduce the required element count by approximately two orders of magnitude in comparison with fixed, uniform mesh simulations. This leads to a substantial reduction in computational cost. The computational savings and flexibility afforded by adaptivity along with the flexibility of FE methods make this model well suited to simulating turbidity currents in complex domains.

  17. Predicting mesh density for adaptive modelling of the global atmosphere.

    Science.gov (United States)

    Weller, Hilary

    2009-11-28

    The shallow water equations are solved using a mesh of polygons on the sphere, which adapts infrequently to the predicted future solution. Infrequent mesh adaptation reduces the cost of adaptation and load-balancing and will thus allow for more accurate mapping on adaptation. We simulate the growth of a barotropically unstable jet adapting the mesh every 12 h. Using an adaptation criterion based largely on the gradient of the vorticity leads to a mesh with around 20 per cent of the cells of a uniform mesh that gives equivalent results. This is a similar proportion to previous studies of the same test case with mesh adaptation every 1-20 min. The prediction of the mesh density involves solving the shallow water equations on a coarse mesh in advance of the locally refined mesh in order to estimate where features requiring higher resolution will grow, decay or move to. The adaptation criterion consists of two parts: that resolved on the coarse mesh, and that which is not resolved and so is passively advected on the coarse mesh. This combination leads to a balance between resolving features controlled by the large-scale dynamics and maintaining fine-scale features.

  18. Bmad: A relativistic charged particle simulation library

    International Nuclear Information System (INIS)

    Sagan, D.

    2006-01-01

    Bmad is a subroutine library for simulating relativistic charged particle beams in high-energy accelerators and storage rings. Bmad can be used to study both single and multi-particle beam dynamics using routines to track both particles and macroparticles. Bmad has various tracking algorithms including Runge-Kutta and symplectic (Lie algebraic) integration. Various effects such as wakefields, and radiation excitation and damping can be simulated. Bmad has been developed in a modular, object-oriented fashion to maximize flexibility. Interface routines allow Bmad to be called from C/C++ as well as Fortran programs. Bmad is well documented. Every routine is individually annotated, and there is an extensive manual

  19. Investigating the settling dynamics of cohesive silt particles with particle-resolving simulations

    Science.gov (United States)

    Sun, Rui; Xiao, Heng; Sun, Honglei

    2018-01-01

    The settling of cohesive sediment is ubiquitous in aquatic environments, and the study of the settling process is important for both engineering and environmental reasons. In the settling process, the silt particles show behaviors that are different from non-cohesive particles due to the influence of inter-particle cohesive force. For instance, the flocs formed in the settling process of cohesive silt can loosen the packing, and thus the structural densities of cohesive silt beds are much smaller than that of non-cohesive sand beds. While there is a consensus that cohesive behaviors depend on the characteristics of sediment particles (e.g., Bond number, particle size distribution), little is known about the exact influence of these characteristics on the cohesive behaviors. In addition, since the cohesive behaviors of the silt are caused by the inter-particle cohesive forces, the motions of and the contacts among silt particles should be resolved to study these cohesive behaviors in the settling process. However, studies of the cohesive behaviors of silt particles in the settling process based on particle-resolving approach are still lacking. In the present work, three-dimensional settling process is investigated numerically by using CFD-DEM (Computational Fluid Dynamics-Discrete Element Method). The inter-particle collision force, the van der Waals force, and the fluid-particle interaction forces are considered. The numerical model is used to simulate the hindered settling process of silt based on the experimental setup in the literature. The results obtained in the simulations, including the structural densities of the beds, the characteristic lines, and the particle terminal velocity, are in good agreement with the experimental observations in the literature. To the authors' knowledge, this is the first time that the influences of non-dimensional Bond number and particle polydispersity on the structural densities of silt beds have been investigated separately

  20. THREE-DIMENSIONAL ADAPTIVE MESH REFINEMENT SIMULATIONS OF LONG-DURATION GAMMA-RAY BURST JETS INSIDE MASSIVE PROGENITOR STARS

    Energy Technology Data Exchange (ETDEWEB)

    Lopez-Camara, D.; Lazzati, Davide [Department of Physics, NC State University, 2401 Stinson Drive, Raleigh, NC 27695-8202 (United States); Morsony, Brian J. [Department of Astronomy, University of Wisconsin-Madison, 2535 Sterling Hall, 475 N. Charter Street, Madison, WI 53706-1582 (United States); Begelman, Mitchell C., E-mail: dlopezc@ncsu.edu [JILA, University of Colorado, 440 UCB, Boulder, CO 80309-0440 (United States)

    2013-04-10

    We present the results of special relativistic, adaptive mesh refinement, 3D simulations of gamma-ray burst jets expanding inside a realistic stellar progenitor. Our simulations confirm that relativistic jets can propagate and break out of the progenitor star while remaining relativistic. This result is independent of the resolution, even though the amount of turbulence and variability observed in the simulations is greater at higher resolutions. We find that the propagation of the jet head inside the progenitor star is slightly faster in 3D simulations compared to 2D ones at the same resolution. This behavior seems to be due to the fact that the jet head in 3D simulations can wobble around the jet axis, finding the spot of least resistance to proceed. Most of the average jet properties, such as density, pressure, and Lorentz factor, are only marginally affected by the dimensionality of the simulations and therefore results from 2D simulations can be considered reliable.

  1. Lattice Boltzmann method used to simulate particle motion in a conduit

    Directory of Open Access Journals (Sweden)

    Dolanský Jindřich

    2017-06-01

    Full Text Available A three-dimensional numerical simulation of particle motion in a pipe with a rough bed is presented. The simulation based on the Lattice Boltzmann Method (LBM employs the hybrid diffuse bounce-back approach to model moving boundaries. The bed of the pipe is formed by stationary spherical particles of the same size as the moving particles. Particle movements are induced by gravitational and hydrodynamic forces. To evaluate the hydrodynamic forces, the Momentum Exchange Algorithm is used. The LBM unified computational frame makes it possible to simulate both the particle motion and the fluid flow and to study mutual interactions of the carrier liquid flow and particles and the particle–bed and particle–particle collisions. The trajectories of simulated and experimental particles are compared. The Particle Tracking method is used to track particle motion. The correctness of the applied approach is assessed.

  2. Adaptation of an unstructured-mesh, finite-element ocean model to the simulation of ocean circulation beneath ice shelves

    Science.gov (United States)

    Kimura, Satoshi; Candy, Adam S.; Holland, Paul R.; Piggott, Matthew D.; Jenkins, Adrian

    2013-07-01

    Several different classes of ocean model are capable of representing floating glacial ice shelves. We describe the incorporation of ice shelves into Fluidity-ICOM, a nonhydrostatic finite-element ocean model with the capacity to utilize meshes that are unstructured and adaptive in three dimensions. This geometric flexibility offers several advantages over previous approaches. The model represents melting and freezing on all ice-shelf surfaces including vertical faces, treats the ice shelf topography as continuous rather than stepped, and does not require any smoothing of the ice topography or any of the additional parameterisations of the ocean mixed layer used in isopycnal or z-coordinate models. The model can also represent a water column that decreases to zero thickness at the 'grounding line', where the floating ice shelf is joined to its tributary ice streams. The model is applied to idealised ice-shelf geometries in order to demonstrate these capabilities. In these simple experiments, arbitrarily coarsening the mesh outside the ice-shelf cavity has little effect on the ice-shelf melt rate, while the mesh resolution within the cavity is found to be highly influential. Smoothing the vertical ice front results in faster flow along the smoothed ice front, allowing greater exchange with the ocean than in simulations with a realistic ice front. A vanishing water-column thickness at the grounding line has little effect in the simulations studied. We also investigate the response of ice shelf basal melting to variations in deep water temperature in the presence of salt stratification.

  3. PART 2: LARGE PARTICLE MODELLING Simulation of particle filtration processes in deformable media

    Directory of Open Access Journals (Sweden)

    Gernot Boiger

    2008-06-01

    Full Text Available In filtration processes it is necessary to consider both, the interaction of thefluid with the solid parts as well as the effect of particles carried in the fluidand accumulated on the solid. While part 1 of this paper deals with themodelling of fluid structure interaction effects, the accumulation of dirtparticles will be addressed in this paper. A closer look is taken on theimplementation of a spherical, LAGRANGIAN particle model suitable forsmall and large particles. As dirt accumulates in the fluid stream, it interactswith the surrounding filter fibre structure and over time causes modificationsof the filter characteristics. The calculation of particle force interactioneffects is necessary for an adequate simulation of this situation. A detailedDiscrete Phase Lagrange Model was developed to take into account thetwo-way coupling of the fluid and accumulated particles. The simulation oflarge particles and the fluid-structure interaction is realised in a single finitevolume flow solver on the basis of the OpenSource software OpenFoam.

  4. Understanding bulk behavior of particulate materials from particle scale simulations

    Science.gov (United States)

    Deng, Xiaoliang

    Particulate materials play an increasingly significant role in various industries, such as pharmaceutical manufacturing, food, mining, and civil engineering. The objective of this research is to better understand bulk behaviors of particulate materials from particle scale simulations. Packing properties of assembly of particles are investigated first, focusing on the effects of particle size, surface energy, and aspect ratio on the coordination number, porosity, and packing structures. The simulation results show that particle sizes, surface energy, and aspect ratio all influence the porosity of packing to various degrees. The heterogeneous force networks within particle assembly under external compressive loading are investigated as well. The results show that coarse-coarse contacts dominate the strong network and coarse-fine contacts dominate the total network. Next, DEM models are developed to simulate the particle dynamics inside a conical screen mill (comil) and magnetically assisted impaction mixer (MAIM), both are important particle processing devices. For comil, the mean residence time (MRT), spatial distribution of particles, along with the collision dynamics between particles as well as particle and vessel geometries are examined as a function of the various operating parameters such as impeller speed, screen hole size, open area, and feed rate. The simulation results can help better understand dry coating experimental results using comil. For MAIM system, the magnetic force is incorporated into the contact model, allowing to describe the interactions between magnets. The simulation results reveal the connections between homogeneity of mixture and particle scale variables such as size of magnets and surface energy of non-magnets. In particular, at the fixed mass ratio of magnets to non-magnets and surface energy the smaller magnets lead to better homogeneity of mixing, which is in good agreement with previously published experimental results. Last but not

  5. Interactive methods for exploring particle simulation data

    Energy Technology Data Exchange (ETDEWEB)

    Co, Christopher S.; Friedman, Alex; Grote, David P.; Vay, Jean-Luc; Bethel, E. Wes; Joy, Kenneth I.

    2004-05-01

    In this work, we visualize high-dimensional particle simulation data using a suite of scatter plot-based visualizations coupled with interactive selection tools. We use traditional 2D and 3D projection scatter plots as well as a novel oriented disk rendering style to convey various information about the data. Interactive selection tools allow physicists to manually classify ''interesting'' sets of particles that are highlighted across multiple, linked views of the data. The power of our application is the ability to correspond new visual representations of the simulation data with traditional, well understood visualizations. This approach supports the interactive exploration of the high-dimensional space while promoting discovery of new particle behavior.

  6. Viscosity of dilute suspensions of rodlike particles: A numerical simulation method

    Science.gov (United States)

    Yamamoto, Satoru; Matsuoka, Takaaki

    1994-02-01

    The recently developed simulation method, named as the particle simulation method (PSM), is extended to predict the viscosity of dilute suspensions of rodlike particles. In this method a rodlike particle is modeled by bonded spheres. Each bond has three types of springs for stretching, bending, and twisting deformation. The rod model can therefore deform by changing the bond distance, bond angle, and torsion angle between paired spheres. The rod model can represent a variety of rigidity by modifying the bond parameters related to Young's modulus and the shear modulus of the real particle. The time evolution of each constituent sphere of the rod model is followed by molecular-dynamics-type approach. The intrinsic viscosity of a suspension of rodlike particles is derived from calculating an increased energy dissipation for each sphere of the rod model in a viscous fluid. With and without deformation of the particle, the motion of the rodlike particle was numerically simulated in a three-dimensional simple shear flow at a low particle Reynolds number and without Brownian motion of particles. The intrinsic viscosity of the suspension of rodlike particles was investigated on orientation angle, rotation orbit, deformation, and aspect ratio of the particle. For the rigid rodlike particle, the simulated rotation orbit compared extremely well with theoretical one which was obtained for a rigid ellipsoidal particle by use of Jeffery's equation. The simulated dependence of the intrinsic viscosity on various factors was also identical with that of theories for suspensions of rigid rodlike particles. For the flexible rodlike particle, the rotation orbit could be obtained by the particle simulation method and it was also cleared that the intrinsic viscosity decreased as occurring of recoverable deformation of the rodlike particle induced by flow.

  7. Notes on the Mesh Handler and Mesh Data Conversion

    International Nuclear Information System (INIS)

    Lee, Sang Yong; Park, Chan Eok

    2009-01-01

    At the outset of the development of the thermal-hydraulic code (THC), efforts have been made to utilize the recent technology of the computational fluid dynamics. Among many of them, the unstructured mesh approach was adopted to alleviate the restriction of the grid handling system. As a natural consequence, a mesh handler (MH) has been developed to manipulate the complex mesh data from the mesh generator. The mesh generator, Gambit, was chosen at the beginning of the development of the code. But a new mesh generator, Pointwise, was introduced to get more flexible mesh generation capability. An open source code, Paraview, was chosen as a post processor, which can handle unstructured as well as structured mesh data. Overall data processing system for THC is shown in Figure-1. There are various file formats to save the mesh data in the permanent storage media. A couple of dozen of file formats are found even in the above mentioned programs. A competent mesh handler should have the capability to import or export mesh data as many as possible formats. But, in reality, there are two aspects that make it difficult to achieve the competence. The first aspect to consider is the time and efforts to program the interface code. And the second aspect, which is even more difficult one, is the fact that many mesh data file formats are proprietary information. In this paper, some experience of the development of the format conversion programs will be presented. File formats involved are Gambit neutral format, Ansys-CFX grid file format, VTK legacy file format, Nastran format and CGNS

  8. Macroscale implicit electromagnetic particle simulation of magnetized plasmas

    International Nuclear Information System (INIS)

    Tanaka, Motohiko.

    1988-01-01

    An electromagnetic and multi-dimensional macroscale particle simulation code (MACROS) is presented which enables us to make a large time and spatial scale kinetic simulation of magnetized plasmas. Particle ions, finite mass electrons with the guiding-center approximation and a complete set of Maxwell equations are employed. Implicit field-particle coupled equations are derived in which a time-decentered (slightly backward) finite differential scheme is used to achieve stability for large time and spatial scales. It is shown analytically that the present simulation scheme suppresses high frequency electromagnetic waves and that it accurately reproduces low frequency waves in the plasma. These properties are verified by numerical examination of eigenmodes in a 2-D thermal equilibrium plasma and by that of the kinetic Alfven wave. (author)

  9. Simulations of Shock Wave Interaction with a Particle Cloud

    Science.gov (United States)

    Koneru, Rahul; Rollin, Bertrand; Ouellet, Frederick; Annamalai, Subramanian; Balachandar, S.'Bala'

    2016-11-01

    Simulations of a shock wave interacting with a cloud of particles are performed in an attempt to understand similar phenomena observed in dispersal of solid particles under such extreme environment as an explosion. We conduct numerical experiments in which a particle curtain fills only 87% of the shock tube from bottom to top. As such, the particle curtain upon interaction with the shock wave is expected to experience Kelvin-Helmholtz (KH) and Richtmyer-Meshkov (RM) instabilities. In this study, the initial volume fraction profile matches with that of Sandia Multiphase Shock Tube experiments, and the shock Mach number is limited to M =1.66. In these simulations we use a Eulerian-Lagrangian approach along with state-of-the-art point-particle force and heat transfer models. Measurements of particle dispersion are made at different initial volume fractions of the particle cloud. A detailed analysis of the evolution of the particle curtain with respect to the initial conditions is presented. This work was supported by the U.S. Department of Energy, National Nuclear Security Administration, Advanced Simulation and Computing Program, as a Cooperative Agreement under the Predictive Science Academic Alliance Program, Contract No. DE-NA0002378.

  10. Kinetic-Monte-Carlo-Based Parallel Evolution Simulation Algorithm of Dust Particles

    Directory of Open Access Journals (Sweden)

    Xiaomei Hu

    2014-01-01

    Full Text Available The evolution simulation of dust particles provides an important way to analyze the impact of dust on the environment. KMC-based parallel algorithm is proposed to simulate the evolution of dust particles. In the parallel evolution simulation algorithm of dust particles, data distribution way and communication optimizing strategy are raised to balance the load of every process and reduce the communication expense among processes. The experimental results show that the simulation of diffusion, sediment, and resuspension of dust particles in virtual campus is realized and the simulation time is shortened by parallel algorithm, which makes up for the shortage of serial computing and makes the simulation of large-scale virtual environment possible.

  11. Deformation and fracture behavior of simulated particle gels

    NARCIS (Netherlands)

    Rzepiela, A.A.

    2003-01-01

    In this PhD project rheological properties of model particle gels are investigated using Brownian Dynamics (BD) simulations. Particle gels are systems of colloidal particles that form weakly bonded percolating networks interpenetrated by a suspending fluid. They are characterized as

  12. Low-noise Collision Operators for Particle-in-cell Simulations

    International Nuclear Information System (INIS)

    Lewandowski, J.L.V.

    2005-01-01

    A new method to implement low-noise collision operators in particle-in-cell simulations is presented. The method is based on the fact that relevant collision operators can be included naturally in the Lagrangian formulation that exemplifies the particle-in-cell simulation method. Numerical simulations show that the momentum and energy conservation properties of the simulated plasma associated with the low-noise collision operator are improved as compared with standard collision algorithms based on random numbers

  13. An Immersed Boundary - Adaptive Mesh Refinement solver (IB-AMR) for high fidelity fully resolved wind turbine simulations

    Science.gov (United States)

    Angelidis, Dionysios; Sotiropoulos, Fotis

    2015-11-01

    The geometrical details of wind turbines determine the structure of the turbulence in the near and far wake and should be taken in account when performing high fidelity calculations. Multi-resolution simulations coupled with an immersed boundary method constitutes a powerful framework for high-fidelity calculations past wind farms located over complex terrains. We develop a 3D Immersed-Boundary Adaptive Mesh Refinement flow solver (IB-AMR) which enables turbine-resolving LES of wind turbines. The idea of using a hybrid staggered/non-staggered grid layout adopted in the Curvilinear Immersed Boundary Method (CURVIB) has been successfully incorporated on unstructured meshes and the fractional step method has been employed. The overall performance and robustness of the second order accurate, parallel, unstructured solver is evaluated by comparing the numerical simulations against conforming grid calculations and experimental measurements of laminar and turbulent flows over complex geometries. We also present turbine-resolving multi-scale LES considering all the details affecting the induced flow field; including the geometry of the tower, the nacelle and especially the rotor blades of a wind tunnel scale turbine. This material is based upon work supported by the Department of Energy under Award Number DE-EE0005482 and the Sandia National Laboratories.

  14. Comparing AMR and SPH Cosmological Simulations. I. Dark Matter and Adiabatic Simulations

    Science.gov (United States)

    O'Shea, Brian W.; Nagamine, Kentaro; Springel, Volker; Hernquist, Lars; Norman, Michael L.

    2005-09-01

    We compare two cosmological hydrodynamic simulation codes in the context of hierarchical galaxy formation: the Lagrangian smoothed particle hydrodynamics (SPH) code GADGET, and the Eulerian adaptive mesh refinement (AMR) code Enzo. Both codes represent dark matter with the N-body method but use different gravity solvers and fundamentally different approaches for baryonic hydrodynamics. The SPH method in GADGET uses a recently developed ``entropy conserving'' formulation of SPH, while for the mesh-based Enzo two different formulations of Eulerian hydrodynamics are employed: the piecewise parabolic method (PPM) extended with a dual energy formulation for cosmology, and the artificial viscosity-based scheme used in the magnetohydrodynamics code ZEUS. In this paper we focus on a comparison of cosmological simulations that follow either only dark matter, or also a nonradiative (``adiabatic'') hydrodynamic gaseous component. We perform multiple simulations using both codes with varying spatial and mass resolution with identical initial conditions. The dark matter-only runs agree generally quite well provided Enzo is run with a comparatively fine root grid and a low overdensity threshold for mesh refinement, otherwise the abundance of low-mass halos is suppressed. This can be readily understood as a consequence of the hierarchical particle-mesh algorithm used by Enzo to compute gravitational forces, which tends to deliver lower force resolution than the tree-algorithm of GADGET at early times before any adaptive mesh refinement takes place. At comparable force resolution we find that the latter offers substantially better performance and lower memory consumption than the present gravity solver in Enzo. In simulations that include adiabatic gasdynamics we find general agreement in the distribution functions of temperature, entropy, and density for gas of moderate to high overdensity, as found inside dark matter halos. However, there are also some significant differences in

  15. Large scale particle simulations in a virtual memory computer

    International Nuclear Information System (INIS)

    Gray, P.C.; Million, R.; Wagner, J.S.; Tajima, T.

    1983-01-01

    Virtual memory computers are capable of executing large-scale particle simulations even when the memory requirements exceeds the computer core size. The required address space is automatically mapped onto slow disc memory the the operating system. When the simulation size is very large, frequent random accesses to slow memory occur during the charge accumulation and particle pushing processes. Assesses to slow memory significantly reduce the excecution rate of the simulation. We demonstrate in this paper that with the proper choice of sorting algorithm, a nominal amount of sorting to keep physically adjacent particles near particles with neighboring array indices can reduce random access to slow memory, increase the efficiency of the I/O system, and hence, reduce the required computing time. (orig.)

  16. Large-scale particle simulations in a virtual-memory computer

    International Nuclear Information System (INIS)

    Gray, P.C.; Wagner, J.S.; Tajima, T.; Million, R.

    1982-08-01

    Virtual memory computers are capable of executing large-scale particle simulations even when the memory requirements exceed the computer core size. The required address space is automatically mapped onto slow disc memory by the operating system. When the simulation size is very large, frequent random accesses to slow memory occur during the charge accumulation and particle pushing processes. Accesses to slow memory significantly reduce the execution rate of the simulation. We demonstrate in this paper that with the proper choice of sorting algorithm, a nominal amount of sorting to keep physically adjacent particles near particles with neighboring array indices can reduce random access to slow memory, increase the efficiency of the I/O system, and hence, reduce the required computing time

  17. Direct numerical simulation of bubbles with parallelized adaptive mesh refinement

    International Nuclear Information System (INIS)

    Talpaert, A.

    2015-01-01

    The study of two-phase Thermal-Hydraulics is a major topic for Nuclear Engineering for both security and efficiency of nuclear facilities. In addition to experiments, numerical modeling helps to knowing precisely where bubbles appear and how they behave, in the core as well as in the steam generators. This work presents the finest scale of representation of two-phase flows, Direct Numerical Simulation of bubbles. We use the 'Di-phasic Low Mach Number' equation model. It is particularly adapted to low-Mach number flows, that is to say flows which velocity is much slower than the speed of sound; this is very typical of nuclear thermal-hydraulics conditions. Because we study bubbles, we capture the front between vapor and liquid phases thanks to a downward flux limiting numerical scheme. The specific discrete analysis technique this work introduces is well-balanced parallel Adaptive Mesh Refinement (AMR). With AMR, we refined the coarse grid on a batch of patches in order to locally increase precision in areas which matter more, and capture fine changes in the front location and its topology. We show that patch-based AMR is very adapted for parallel computing. We use a variety of physical examples: forced advection, heat transfer, phase changes represented by a Stefan model, as well as the combination of all those models. We will present the results of those numerical simulations, as well as the speed up compared to equivalent non-AMR simulation and to serial computation of the same problems. This document is made up of an abstract and the slides of the presentation. (author)

  18. Loading relativistic Maxwell distributions in particle simulations

    Science.gov (United States)

    Zenitani, S.

    2015-12-01

    In order to study energetic plasma phenomena by using particle-in-cell (PIC) and Monte-Carlo simulations, we need to deal with relativistic velocity distributions in these simulations. However, numerical algorithms to deal with relativistic distributions are not well known. In this contribution, we overview basic algorithms to load relativistic Maxwell distributions in PIC and Monte-Carlo simulations. For stationary relativistic Maxwellian, the inverse transform method and the Sobol algorithm are reviewed. To boost particles to obtain relativistic shifted-Maxwellian, two rejection methods are newly proposed in a physically transparent manner. Their acceptance efficiencies are 􏰅50% for generic cases and 100% for symmetric distributions. They can be combined with arbitrary base algorithms.

  19. Numerical techniques for large cosmological N-body simulations

    International Nuclear Information System (INIS)

    Efstathiou, G.; Davis, M.; Frenk, C.S.; White, S.D.M.

    1985-01-01

    We describe and compare techniques for carrying out large N-body simulations of the gravitational evolution of clustering in the fundamental cube of an infinite periodic universe. In particular, we consider both particle mesh (PM) codes and P 3 M codes in which a higher resolution force is obtained by direct summation of contributions from neighboring particles. We discuss the mesh-induced anisotropies in the forces calculated by these schemes, and the extent to which they can model the desired 1/r 2 particle-particle interaction. We also consider how transformation of the time variable can improve the efficiency with which the equations of motion are integrated. We present tests of the accuracy with which the resulting schemes conserve energy and are able to follow individual particle trajectories. We have implemented an algorithm which allows initial conditions to be set up to model any desired spectrum of linear growing mode density fluctuations. A number of tests demonstrate the power of this algorithm and delineate the conditions under which it is effective. We carry out several test simulations using a variety of techniques in order to show how the results are affected by dynamic range limitations in the force calculations, by boundary effects, by residual artificialities in the initial conditions, and by the number of particles employed. For most purposes cosmological simulations are limited by the resolution of their force calculation rather than by the number of particles they can employ. For this reason, while PM codes are quite adequate to study the evolution of structure on large scale, P 3 M methods are to be preferred, in spite of their greater cost and complexity, whenever the evolution of small-scale structure is important

  20. Particle Number Dependence of the N-body Simulations of Moon Formation

    Science.gov (United States)

    Sasaki, Takanori; Hosono, Natsuki

    2018-04-01

    The formation of the Moon from the circumterrestrial disk has been investigated by using N-body simulations with the number N of particles limited from 104 to 105. We develop an N-body simulation code on multiple Pezy-SC processors and deploy Framework for Developing Particle Simulators to deal with large number of particles. We execute several high- and extra-high-resolution N-body simulations of lunar accretion from a circumterrestrial disk of debris generated by a giant impact on Earth. The number of particles is up to 107, in which 1 particle corresponds to a 10 km sized satellitesimal. We find that the spiral structures inside the Roche limit radius differ between low-resolution simulations (N ≤ 105) and high-resolution simulations (N ≥ 106). According to this difference, angular momentum fluxes, which determine the accretion timescale of the Moon also depend on the numerical resolution.

  1. Adaptive Mesh Iteration Method for Trajectory Optimization Based on Hermite-Pseudospectral Direct Transcription

    Directory of Open Access Journals (Sweden)

    Humin Lei

    2017-01-01

    Full Text Available An adaptive mesh iteration method based on Hermite-Pseudospectral is described for trajectory optimization. The method uses the Legendre-Gauss-Lobatto points as interpolation points; then the state equations are approximated by Hermite interpolating polynomials. The method allows for changes in both number of mesh points and the number of mesh intervals and produces significantly smaller mesh sizes with a higher accuracy tolerance solution. The derived relative error estimate is then used to trade the number of mesh points with the number of mesh intervals. The adaptive mesh iteration method is applied successfully to the examples of trajectory optimization of Maneuverable Reentry Research Vehicle, and the simulation experiment results show that the adaptive mesh iteration method has many advantages.

  2. Mesh-based parallel code coupling interface

    Energy Technology Data Exchange (ETDEWEB)

    Wolf, K.; Steckel, B. (eds.) [GMD - Forschungszentrum Informationstechnik GmbH, St. Augustin (DE). Inst. fuer Algorithmen und Wissenschaftliches Rechnen (SCAI)

    2001-04-01

    MpCCI (mesh-based parallel code coupling interface) is an interface for multidisciplinary simulations. It provides industrial end-users as well as commercial code-owners with the facility to combine different simulation tools in one environment. Thereby new solutions for multidisciplinary problems will be created. This opens new application dimensions for existent simulation tools. This Book of Abstracts gives a short overview about ongoing activities in industry and research - all presented at the 2{sup nd} MpCCI User Forum in February 2001 at GMD Sankt Augustin. (orig.) [German] MpCCI (mesh-based parallel code coupling interface) definiert eine Schnittstelle fuer multidisziplinaere Simulationsanwendungen. Sowohl industriellen Anwender als auch kommerziellen Softwarehersteller wird mit MpCCI die Moeglichkeit gegeben, Simulationswerkzeuge unterschiedlicher Disziplinen miteinander zu koppeln. Dadurch entstehen neue Loesungen fuer multidisziplinaere Problemstellungen und fuer etablierte Simulationswerkzeuge ergeben sich neue Anwendungsfelder. Dieses Book of Abstracts bietet einen Ueberblick ueber zur Zeit laufende Arbeiten in der Industrie und in der Forschung, praesentiert auf dem 2{sup nd} MpCCI User Forum im Februar 2001 an der GMD Sankt Augustin. (orig.)

  3. Poisson solvers for self-consistent multi-particle simulations

    International Nuclear Information System (INIS)

    Qiang, J; Paret, S

    2014-01-01

    Self-consistent multi-particle simulation plays an important role in studying beam-beam effects and space charge effects in high-intensity beams. The Poisson equation has to be solved at each time-step based on the particle density distribution in the multi-particle simulation. In this paper, we review a number of numerical methods that can be used to solve the Poisson equation efficiently. The computational complexity of those numerical methods will be O(N log(N)) or O(N) instead of O(N2), where N is the total number of grid points used to solve the Poisson equation

  4. Computational performance of Free Mesh Method applied to continuum mechanics problems

    Science.gov (United States)

    YAGAWA, Genki

    2011-01-01

    The free mesh method (FMM) is a kind of the meshless methods intended for particle-like finite element analysis of problems that are difficult to handle using global mesh generation, or a node-based finite element method that employs a local mesh generation technique and a node-by-node algorithm. The aim of the present paper is to review some unique numerical solutions of fluid and solid mechanics by employing FMM as well as the Enriched Free Mesh Method (EFMM), which is a new version of FMM, including compressible flow and sounding mechanism in air-reed instruments as applications to fluid mechanics, and automatic remeshing for slow crack growth, dynamic behavior of solid as well as large-scale Eigen-frequency of engine block as applications to solid mechanics. PMID:21558753

  5. A Reconfigurable Mesh-Ring Topology for Bluetooth Sensor Networks

    Directory of Open Access Journals (Sweden)

    Ben-Yi Wang

    2018-05-01

    Full Text Available In this paper, a Reconfigurable Mesh-Ring (RMR algorithm is proposed for Bluetooth sensor networks. The algorithm is designed in three stages to determine the optimal configuration of the mesh-ring network. Firstly, a designated root advertises and discovers its neighboring nodes. Secondly, a scatternet criterion is built to compute the minimum number of piconets and distributes the connection information for piconet and scatternet. Finally, a peak-search method is designed to determine the optimal mesh-ring configuration for various sizes of networks. To maximize the network capacity, the research problem is formulated by determining the best connectivity of available mesh links. During the formation and maintenance phases, three possible configurations (including piconet, scatternet, and hybrid are examined to determine the optimal placement of mesh links. The peak-search method is a systematic approach, and is implemented by three functional blocks: the topology formation block generates the mesh-ring topology, the routing efficiency block computes the routing performance, and the optimum decision block introduces a decision-making criterion to determine the optimum number of mesh links. Simulation results demonstrate that the optimal mesh-ring configuration can be determined and that the scatternet case achieves better overall performance than the other two configurations. The RMR topology also outperforms the conventional ring-based and cluster-based mesh methods in terms of throughput performance for Bluetooth configurable networks.

  6. Radiological dispersal device outdoor simulation test: Cesium chloride particle characteristics

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Sang Don, E-mail: lee.sangdon@epa.gov [U.S. Environmental Protection Agency, Research Triangle Park, NC 27711 (United States); Snyder, Emily G.; Willis, Robert [U.S. Environmental Protection Agency, Research Triangle Park, NC 27711 (United States); Fischer, Robert; Gates-Anderson, Dianne; Sutton, Mark [Lawrence Livermore National Laboratory, Livermore, CA 94550 (United States); Viani, Brian [Simbol Mining Corp., Pleasanton, CA 94566 (United States); Drake, John [U.S. Environmental Protection Agency, Cincinnati, OH 45268 (United States); MacKinney, John [U.S. Department of Homeland Security, Washington, DC 20528 (United States)

    2010-04-15

    Particles were generated from the detonation of simulated radiological dispersal devices (RDDs) using non-radioactive CsCl powder and explosive C4. The physical and chemical properties of the resulting particles were characterized. Two RDD simulation tests were conducted at Lawrence Livermore National Laboratory: one of the simulated RDDs was positioned 1 m above a steel plate and the other was partially buried in soil. Particles were collected with filters at a distance of 150 m from the origin of the RDD device, and particle mass concentrations were monitored to identify the particle plume intensity using real time particle samplers. Particles collected on filters were analyzed via computer-controlled scanning electron microscopy coupled with energy dispersive X-ray spectrometry (CCSEM/EDX) to determine their size distribution, morphology, and chemical constituents. This analysis showed that particles generated by the detonation of explosives can be associated with other materials (e.g., soil) that are in close proximity to the RDD device and that the morphology and chemical makeup of the particles change depending on the interactions of the RDD device with the surrounding materials.

  7. Radiological dispersal device outdoor simulation test: Cesium chloride particle characteristics

    International Nuclear Information System (INIS)

    Lee, Sang Don; Snyder, Emily G.; Willis, Robert; Fischer, Robert; Gates-Anderson, Dianne; Sutton, Mark; Viani, Brian; Drake, John; MacKinney, John

    2010-01-01

    Particles were generated from the detonation of simulated radiological dispersal devices (RDDs) using non-radioactive CsCl powder and explosive C4. The physical and chemical properties of the resulting particles were characterized. Two RDD simulation tests were conducted at Lawrence Livermore National Laboratory: one of the simulated RDDs was positioned 1 m above a steel plate and the other was partially buried in soil. Particles were collected with filters at a distance of 150 m from the origin of the RDD device, and particle mass concentrations were monitored to identify the particle plume intensity using real time particle samplers. Particles collected on filters were analyzed via computer-controlled scanning electron microscopy coupled with energy dispersive X-ray spectrometry (CCSEM/EDX) to determine their size distribution, morphology, and chemical constituents. This analysis showed that particles generated by the detonation of explosives can be associated with other materials (e.g., soil) that are in close proximity to the RDD device and that the morphology and chemical makeup of the particles change depending on the interactions of the RDD device with the surrounding materials.

  8. Simulating the Agulhas system in global ocean models - nesting vs. multi-resolution unstructured meshes

    Science.gov (United States)

    Biastoch, Arne; Sein, Dmitry; Durgadoo, Jonathan V.; Wang, Qiang; Danilov, Sergey

    2018-01-01

    Many questions in ocean and climate modelling require the combined use of high resolution, global coverage and multi-decadal integration length. For this combination, even modern resources limit the use of traditional structured-mesh grids. Here we compare two approaches: A high-resolution grid nested into a global model at coarser resolution (NEMO with AGRIF) and an unstructured-mesh grid (FESOM) which allows to variably enhance resolution where desired. The Agulhas system around South Africa is used as a testcase, providing an energetic interplay of a strong western boundary current and mesoscale dynamics. Its open setting into the horizontal and global overturning circulations also requires global coverage. Both model configurations simulate a reasonable large-scale circulation. Distribution and temporal variability of the wind-driven circulation are quite comparable due to the same atmospheric forcing. However, the overturning circulation differs, owing each model's ability to represent formation and spreading of deep water masses. In terms of regional, high-resolution dynamics, all elements of the Agulhas system are well represented. Owing to the strong nonlinearity in the system, Agulhas Current transports of both configurations and in comparison with observations differ in strength and temporal variability. Similar decadal trends in Agulhas Current transport and Agulhas leakage are linked to the trends in wind forcing.

  9. Numerical simulations on a high-temperature particle moving in coolant

    International Nuclear Information System (INIS)

    Li Xiaoyan; Shang Zhi; Xu Jijun

    2006-01-01

    This study considers the coupling effect between film boiling heat transfer and evaporation drag around a hot-particle in cold liquid. Taking momentum and energy equations of the vapor film into account, a transient single particle model under FCI conditions has been established. The numerical simulations on a high-temperature particle moving in coolant have been performed using Gear algorithm. Adaptive dynamic boundary method is adopted during simulating to matching the dynamic boundary that is caused by vapor film changing. Based on the method presented above, the transient process of high-temperature particles moving in coolant can be simulated. The experimental results prove the validity of the HPMC model. (authors)

  10. Numerical Investigation of Corrugated Wire Mesh Laminate

    Directory of Open Access Journals (Sweden)

    Jeongho Choi

    2013-01-01

    Full Text Available The aim of this work is to develop a numerical model of Corrugated Wire Mesh Laminate (CWML capturing all its complexities such as nonlinear material properties, nonlinear geometry and large deformation behaviour, and frictional behaviour. Development of such a model will facilitate numerical simulation of the mechanical behaviour of the wire mesh structure under various types of loading as well as the variation of the CWML configuration parameters to tailor its mechanical properties to suit the intended application. Starting with a single strand truss model consisting of four waves with a bilinear stress-strain model to represent the plastic behaviour of stainless steel, the finite element model is gradually built up to study single-layer structures with 18 strands of corrugated wire meshes consistency and double- and quadruple-layered laminates with alternating crossply orientations. The compressive behaviour of the CWML model is simulated using contact elements to model friction and is compared to the load-deflection behaviour determined experimentally in uniaxial compression tests. The numerical model of the CWML is then employed to conduct the aim of establishing the upper and lower bounds of stiffness and load capacity achievable by such structures.

  11. Modelling and simulation of particle-particle interaction in a magnetophoretic bio-separation chip

    Science.gov (United States)

    Alam, Manjurul; Golozar, Matin; Darabi, Jeff

    2018-04-01

    A Lagrangian particle trajectory model is developed to predict the interaction between cell-bead particle complexes and to track their trajectories in a magnetophoretic bio-separation chip. Magnetic flux gradients are simulated in the OpenFOAM CFD software and imported into MATLAB to obtain the trapping lengths and trajectories of the particles. A connector vector is introduced to calculate the interaction force between cell-bead complexes as they flow through a microfluidic device. The interaction force calculations are performed for cases where the connector vector is parallel, perpendicular, and at an angle of 45° with the applied magnetic field. The trajectories of the particles are simulated by solving a system of eight ordinary differential equations using a fourth order Runge-Kutta method. The model is then used to study the effects of geometric positions and angles of the connector vector between the particles as well as the cell size, number of beads per cell, and flow rate on the interaction force and trajectories of the particles. The results show that the interaction forces may be attractive or repulsive, depending on the orientation of the connector vector distance between the particle complexes and the applied magnetic field. When the interaction force is attractive, the particles are observed to merge and trap sooner than a single particle, whereas a repulsive interaction force has little or no effect on the trapping length.

  12. Recent progress of hybrid simulation for energetic particles and MHD

    International Nuclear Information System (INIS)

    Todo, Y.

    2013-01-01

    Several hybrid simulation models have been constructed to study the evolution of Alfven eigenmodes destabilized by energetic particles. Recent hybrid simulation results of energetic particle driven instabilities are presented in this paper. (J.P.N.)

  13. Shock Wave Speed and Transient Response of PE Pipe with Steel-Mesh Reinforcement

    Directory of Open Access Journals (Sweden)

    Wuyi Wan

    2016-01-01

    Full Text Available A steel mesh can improve the tensile strength and stability of a polyethylene (PE pipe in a water supply pipeline system. However, it can also cause more severe water hammer hazard due to increasing wave speed. In order to analyze the influence of the steel mesh on the shock wave speed and transient response processes, an improved wave speed formula is proposed by incorporating the equivalent elastic modulus. A field measurement validates the wave speed formula. Moreover, the transient wave propagation and extreme pressures are simulated and compared by the method of characteristics (MOC for reinforced PE pipes with various steel-mesh densities. Results show that a steel mesh can significantly increase the shock wave speed in a PE pipe and thus can cause severe peak pressure and hydraulic surges in a water supply pipeline system. The proposed wave speed formula can more reasonably evaluate the wave speed and improve the transient simulation of steel-mesh-reinforced PE pipes.

  14. High viscosity fluid simulation using particle-based method

    KAUST Repository

    Chang, Yuanzhang

    2011-03-01

    We present a new particle-based method for high viscosity fluid simulation. In the method, a new elastic stress term, which is derived from a modified form of the Hooke\\'s law, is included in the traditional Navier-Stokes equation to simulate the movements of the high viscosity fluids. Benefiting from the Lagrangian nature of Smoothed Particle Hydrodynamics method, large flow deformation can be well handled easily and naturally. In addition, in order to eliminate the particle deficiency problem near the boundary, ghost particles are employed to enforce the solid boundary condition. Compared with Finite Element Methods with complicated and time-consuming remeshing operations, our method is much more straightforward to implement. Moreover, our method doesn\\'t need to store and compare to an initial rest state. The experimental results show that the proposed method is effective and efficient to handle the movements of highly viscous flows, and a large variety of different kinds of fluid behaviors can be well simulated by adjusting just one parameter. © 2011 IEEE.

  15. Multiphase CFD simulation of a solid bowl centrifuge

    Energy Technology Data Exchange (ETDEWEB)

    Romani Fernandez, X.; Nirschl, H. [Universitaet Karlsruhe, Institut fuer MVM, Karlsruhe (Germany)

    2009-05-15

    This study presents some results from the numerical simulation of the flow in an industrial solid bowl centrifuge used for particle separation in industrial fluid processing. The computational fluid dynamics (CFD) software Fluent was used to simulate this multiphase flow. Simplified two-dimensional and three-dimensional geometries were built and meshed from the real centrifuge geometry. The CFD results show a boundary layer of axially fast moving fluid at the gas-liquid interface. Below this layer there is a thin recirculation. The obtained tangential velocity values are lower than the ones for the rigid-body motion. Also, the trajectories of the solid particles are evaluated. (Abstract Copyright [2009], Wiley Periodicals, Inc.)

  16. Simulation of deterministic energy-balance particle agglomeration in turbulent liquid-solid flows

    Science.gov (United States)

    Njobuenwu, Derrick O.; Fairweather, Michael

    2017-08-01

    An efficient technique to simulate turbulent particle-laden flow at high mass loadings within the four-way coupled simulation regime is presented. The technique implements large-eddy simulation, discrete particle simulation, a deterministic treatment of inter-particle collisions, and an energy-balanced particle agglomeration model. The algorithm to detect inter-particle collisions is such that the computational costs scale linearly with the number of particles present in the computational domain. On detection of a collision, particle agglomeration is tested based on the pre-collision kinetic energy, restitution coefficient, and van der Waals' interactions. The performance of the technique developed is tested by performing parametric studies on the influence of the restitution coefficient (en = 0.2, 0.4, 0.6, and 0.8), particle size (dp = 60, 120, 200, and 316 μm), Reynolds number (Reτ = 150, 300, and 590), and particle concentration (αp = 5.0 × 10-4, 1.0 × 10-3, and 5.0 × 10-3) on particle-particle interaction events (collision and agglomeration). The results demonstrate that the collision frequency shows a linear dependency on the restitution coefficient, while the agglomeration rate shows an inverse dependence. Collisions among smaller particles are more frequent and efficient in forming agglomerates than those of coarser particles. The particle-particle interaction events show a strong dependency on the shear Reynolds number Reτ, while increasing the particle concentration effectively enhances particle collision and agglomeration whilst having only a minor influence on the agglomeration rate. Overall, the sensitivity of the particle-particle interaction events to the selected simulation parameters is found to influence the population and distribution of the primary particles and agglomerates formed.

  17. Simulating Biomass Fast Pyrolysis at the Single Particle Scale

    Energy Technology Data Exchange (ETDEWEB)

    Ciesielski, Peter [National Renewable Energy Laboratory (NREL); Wiggins, Gavin [ORNL; Daw, C Stuart [ORNL; Jakes, Joseph E. [U.S. Forest Service, Forest Products Laboratory, Madison, Wisconsin, USA

    2017-07-01

    Simulating fast pyrolysis at the scale of single particles allows for the investigation of the impacts of feedstock-specific parameters such as particle size, shape, and species of origin. For this reason particle-scale modeling has emerged as an important tool for understanding how variations in feedstock properties affect the outcomes of pyrolysis processes. The origins of feedstock properties are largely dictated by the composition and hierarchical structure of biomass, from the microstructural porosity to the external morphology of milled particles. These properties may be accounted for in simulations of fast pyrolysis by several different computational approaches depending on the level of structural and chemical complexity included in the model. The predictive utility of particle-scale simulations of fast pyrolysis can still be enhanced substantially by advancements in several areas. Most notably, considerable progress would be facilitated by the development of pyrolysis kinetic schemes that are decoupled from transport phenomena, predict product evolution from whole-biomass with increased chemical speciation, and are still tractable with present-day computational resources.

  18. Cartesian anisotropic mesh adaptation for compressible flow

    International Nuclear Information System (INIS)

    Keats, W.A.; Lien, F.-S.

    2004-01-01

    Simulating transient compressible flows involving shock waves presents challenges to the CFD practitioner in terms of the mesh quality required to resolve discontinuities and prevent smearing. This paper discusses a novel two-dimensional Cartesian anisotropic mesh adaptation technique implemented for compressible flow. This technique, developed for laminar flow by Ham, Lien and Strong, is efficient because it refines and coarsens cells using criteria that consider the solution in each of the cardinal directions separately. In this paper the method will be applied to compressible flow. The procedure shows promise in its ability to deliver good quality solutions while achieving computational savings. The convection scheme used is the Advective Upstream Splitting Method (Plus), and the refinement/ coarsening criteria are based on work done by Ham et al. Transient shock wave diffraction over a backward step and shock reflection over a forward step are considered as test cases because they demonstrate that the quality of the solution can be maintained as the mesh is refined and coarsened in time. The data structure is explained in relation to the computational mesh, and the object-oriented design and implementation of the code is presented. Refinement and coarsening algorithms are outlined. Computational savings over uniform and isotropic mesh approaches are shown to be significant. (author)

  19. Low modulus and bioactive Ti/α-TCP/Ti-mesh composite prepared by spark plasma sintering.

    Science.gov (United States)

    Guo, Yu; Tan, Yanni; Liu, Yong; Liu, Shifeng; Zhou, Rui; Tang, Hanchun

    2017-11-01

    A titanium mesh scaffold composite filled with Ti/α-TCP particles was prepared by spark plasma sintering (SPS). The microstructures and interfacial reactions of the composites were investigated by scanning electron microscopy (SEM), Energy Dispersive Spectroscopy (EDS) and X-ray diffraction (XRD) analyses. The compressive strength and elastic modulus were also measured. In vitro bioactivity and biocompatibility was evaluated by using simulated body fluid and cells culture, respectively. After high temperature sintering, Ti oxides, Ti x P y and CaTiO 3 were formed. The formation of Ti oxides and Ti x P y were resulted from the diffusion of O and P elements from α-TCP to Ti. CaTiO 3 was the reaction product of Ti and α-TCP. The composite of 70Ti/α-TCP incorporated with Ti mesh showed a high compressive strength of 589MPa and a low compressive modulus of 30GPa. The bioactivity test showed the formation of a thick apatite layer on the composite and well-spread cells attachment. A good combination of mechanical properties and bioactivity indicated a high potential application of Ti/α-TCP/Ti-mesh composite for orthopedic implants. Copyright © 2017. Published by Elsevier B.V.

  20. Advanced computers and simulation

    International Nuclear Information System (INIS)

    Ryne, R.D.

    1993-01-01

    Accelerator physicists today have access to computers that are far more powerful than those available just 10 years ago. In the early 1980's, desktop workstations performed less one million floating point operations per second (Mflops), and the realized performance of vector supercomputers was at best a few hundred Mflops. Today vector processing is available on the desktop, providing researchers with performance approaching 100 Mflops at a price that is measured in thousands of dollars. Furthermore, advances in Massively Parallel Processors (MPP) have made performance of over 10 gigaflops a reality, and around mid-decade MPPs are expected to be capable of teraflops performance. Along with advances in MPP hardware, researchers have also made significant progress in developing algorithms and software for MPPS. These changes have had, and will continue to have, a significant impact on the work of computational accelerator physicists. Now, instead of running particle simulations with just a few thousand particles, we can perform desktop simulations with tens of thousands of simulation particles, and calculations with well over 1 million particles are being performed on MPPs. In the area of computational electromagnetics, simulations that used to be performed only on vector supercomputers now run in several hours on desktop workstations, and researchers are hoping to perform simulations with over one billion mesh points on future MPPs. In this paper we will discuss the latest advances, and what can be expected in the near future, in hardware, software and applications codes for advanced simulation of particle accelerators

  1. Multiscale Simulations Using Particles

    DEFF Research Database (Denmark)

    Walther, Jens Honore

    vortex methods for problems in continuum fluid dynamics, dissipative particle dynamics for flow at the meso scale, and atomistic molecular dynamics simulations of nanofluidic systems. We employ multiscale techniques to breach the atomistic and continuum scales to study fundamental problems in fluid...... dynamics. Recent work on the thermophoretic motion of water nanodroplets confined inside carbon nanotubes, and multiscale techniques for polar liquids will be discussed in detail at the symposium....

  2. Controlling the error on target motion through real-time mesh adaptation: Applications to deep brain stimulation.

    Science.gov (United States)

    Bui, Huu Phuoc; Tomar, Satyendra; Courtecuisse, Hadrien; Audette, Michel; Cotin, Stéphane; Bordas, Stéphane P A

    2018-05-01

    An error-controlled mesh refinement procedure for needle insertion simulations is presented. As an example, the procedure is applied for simulations of electrode implantation for deep brain stimulation. We take into account the brain shift phenomena occurring when a craniotomy is performed. We observe that the error in the computation of the displacement and stress fields is localised around the needle tip and the needle shaft during needle insertion simulation. By suitably and adaptively refining the mesh in this region, our approach enables to control, and thus to reduce, the error whilst maintaining a coarser mesh in other parts of the domain. Through academic and practical examples we demonstrate that our adaptive approach, as compared with a uniform coarse mesh, increases the accuracy of the displacement and stress fields around the needle shaft and, while for a given accuracy, saves computational time with respect to a uniform finer mesh. This facilitates real-time simulations. The proposed methodology has direct implications in increasing the accuracy, and controlling the computational expense of the simulation of percutaneous procedures such as biopsy, brachytherapy, regional anaesthesia, or cryotherapy. Moreover, the proposed approach can be helpful in the development of robotic surgeries because the simulation taking place in the control loop of a robot needs to be accurate, and to occur in real time. Copyright © 2018 John Wiley & Sons, Ltd.

  3. A mesh density study for application to large deformation rolling process evaluation

    International Nuclear Information System (INIS)

    Martin, J.A.

    1997-12-01

    When addressing large deformation through an elastic-plastic analysis the mesh density is paramount in determining the accuracy of the solution. However, given the nonlinear nature of the problem, a highly-refined mesh will generally require a prohibitive amount of computer resources. This paper addresses finite element mesh optimization studies considering accuracy of results and computer resource needs as applied to large deformation rolling processes. In particular, the simulation of the thread rolling manufacturing process is considered using the MARC software package and a Cray C90 supercomputer. Both mesh density and adaptive meshing on final results for both indentation of a rigid body to a specified depth and contact rolling along a predetermined length are evaluated

  4. A Generic Mesh Data Structure with Parallel Applications

    Science.gov (United States)

    Cochran, William Kenneth, Jr.

    2009-01-01

    High performance, massively-parallel multi-physics simulations are built on efficient mesh data structures. Most data structures are designed from the bottom up, focusing on the implementation of linear algebra routines. In this thesis, we explore a top-down approach to design, evaluating the various needs of many aspects of simulation, not just…

  5. Theory and simulation of epitaxial rotation. Light particles adsorbed on graphite

    DEFF Research Database (Denmark)

    Vives, E.; Lindgård, P.-A.

    1993-01-01

    We present a theory and Monte Carlo simulations of adsorbed particles on a corrugated substrate. We have focused on the case of rare gases and light molecules, H-2 and D2, adsorbed on graphite. The competition between the particle-particle and particle-substrate interactions gives rise to frustra...... found a modulated 4 x 4 structure. Energy, structure-factor intensities, peak positions, and epitaxial rotation angles as a function of temperature and coverage have been determined from the simulations. Good agreement with theory and experimental data is found.......We present a theory and Monte Carlo simulations of adsorbed particles on a corrugated substrate. We have focused on the case of rare gases and light molecules, H-2 and D2, adsorbed on graphite. The competition between the particle-particle and particle-substrate interactions gives rise...... between the commensurate and incommensurate phase for the adsorbed systems. From our simulations and our theory, we are, able to understand the gamma phase of D2 as an ordered phase stabilized by disorder. It can be described as a 2q-modulated structure. In agreement with the experiments, we have also...

  6. Electrostatics in biomolecular simulations : where are we now and where are we heading?

    NARCIS (Netherlands)

    Karttunen, M.E.J.; Rottler, J.; Vattulainen, I.; Sagui, C.

    2008-01-01

    Chapter 2. In this review, we discuss current methods and developments in the treatment of electrostatic interactions in biomolecular and soft matter simulations. We review the current ‘work horses’, namely, Ewald summation based methods such the Particle-Mesh Ewald, and others, and also newer

  7. Partitioning of unstructured meshes for load balancing

    International Nuclear Information System (INIS)

    Martin, O.C.; Otto, S.W.

    1994-01-01

    Many large-scale engineering and scientific calculations involve repeated updating of variables on an unstructured mesh. To do these types of computations on distributed memory parallel computers, it is necessary to partition the mesh among the processors so that the load balance is maximized and inter-processor communication time is minimized. This can be approximated by the problem, of partitioning a graph so as to obtain a minimum cut, a well-studied combinatorial optimization problem. Graph partitioning algorithms are discussed that give good but not necessarily optimum solutions. These algorithms include local search methods recursive spectral bisection, and more general purpose methods such as simulated annealing. It is shown that a general procedure enables to combine simulated annealing with Kernighan-Lin. The resulting algorithm is both very fast and extremely effective. (authors) 23 refs., 3 figs., 1 tab

  8. Are patient specific meshes required for EIT head imaging?

    Science.gov (United States)

    Jehl, Markus; Aristovich, Kirill; Faulkner, Mayo; Holder, David

    2016-06-01

    Head imaging with electrical impedance tomography (EIT) is usually done with time-differential measurements, to reduce time-invariant modelling errors. Previous research suggested that more accurate head models improved image quality, but no thorough analysis has been done on the required accuracy. We propose a novel pipeline for creation of precise head meshes from magnetic resonance imaging and computed tomography scans, which was applied to four different heads. Voltages were simulated on all four heads for perturbations of different magnitude, haemorrhage and ischaemia, in five different positions and for three levels of instrumentation noise. Statistical analysis showed that reconstructions on the correct mesh were on average 25% better than on the other meshes. However, the stroke detection rates were not improved. We conclude that a generic head mesh is sufficient for monitoring patients for secondary strokes following head trauma.

  9. Open boundaries for particle beams within fit-simulations

    International Nuclear Information System (INIS)

    Balk, M.C.; Schuhmann, R.; Weiland, T.

    2006-01-01

    A method is proposed to simulate open boundary conditions for charged particle beams with v< c in time domain or frequency domain within the Finite Integration Technique (FIT). Inside the calculation domain the moving charged particles are represented by a line current. Further, the simulated field components at the boundary of the calculation domain have to be modified for an undisturbed transmission of the space-charge field. This can be realised by a 'scattered field' formulation. The method is verified by several calculations

  10. Direct Numerical Simulations of Particle-Laden Turbulent Channel Flow

    Science.gov (United States)

    Jebakumar, Anand Samuel; Premnath, Kannan; Abraham, John

    2017-11-01

    In a recent experimental study, Lau and Nathan (2014) reported that the distribution of particles in a turbulent pipe flow is strongly influenced by the Stokes number (St). At St lower than 1, particles migrate toward the wall and at St greater than 10 they tend to migrate toward the axis. It was suggested that this preferential migration of particles is due to two forces, the Saffman lift force and the turbophoretic force. Saffman lift force represents a force acting on the particle as a result of a velocity gradient across the particle when it leads or lags the fluid flow. Turbophoretic force is induced by turbulence which tends to move the particle in the direction of decreasing turbulent kinetic energy. In this study, the Lattice Boltzmann Method (LBM) is employed to simulate a particle-laden turbulent channel flow through Direct Numerical Simulations (DNS). We find that the preferential migration is a function of particle size in addition to the St. We explain the effect of the particle size and St on the Saffman lift force and turbophoresis and present how this affects particle concentration at different conditions.

  11. Reactor physics verification of the MCNP6 unstructured mesh capability

    International Nuclear Information System (INIS)

    Burke, T. P.; Kiedrowski, B. C.; Martz, R. L.; Martin, W. R.

    2013-01-01

    The Monte Carlo software package MCNP6 has the ability to transport particles on unstructured meshes generated from the Computed-Aided Engineering software Abaqus. Verification is performed using benchmarks with features relevant to reactor physics - Big Ten and the C5G7 computational benchmark. Various meshing strategies are tested and results are compared to reference solutions. Computational performance results are also given. The conclusions show MCNP6 is capable of producing accurate calculations for reactor physics geometries and the computational requirements for small lattice benchmarks are reasonable on modern computing platforms. (authors)

  12. Reactor physics verification of the MCNP6 unstructured mesh capability

    Energy Technology Data Exchange (ETDEWEB)

    Burke, T. P. [Department of Nuclear Engineering and Radiological Sciences, University of Michigan, 2355 Bonisteel Boulevard, Ann Arbor, MI 48109 (United States); Kiedrowski, B. C.; Martz, R. L. [X-Computational Physics Division, Monte Carlo Codes Group, Los Alamos National Laboratory, P.O. Box 1663, Los Alamos, NM 87545 (United States); Martin, W. R. [Department of Nuclear Engineering and Radiological Sciences, University of Michigan, 2355 Bonisteel Boulevard, Ann Arbor, MI 48109 (United States)

    2013-07-01

    The Monte Carlo software package MCNP6 has the ability to transport particles on unstructured meshes generated from the Computed-Aided Engineering software Abaqus. Verification is performed using benchmarks with features relevant to reactor physics - Big Ten and the C5G7 computational benchmark. Various meshing strategies are tested and results are compared to reference solutions. Computational performance results are also given. The conclusions show MCNP6 is capable of producing accurate calculations for reactor physics geometries and the computational requirements for small lattice benchmarks are reasonable on modern computing platforms. (authors)

  13. Resolved-particle simulation by the Physalis method: Enhancements and new capabilities

    Energy Technology Data Exchange (ETDEWEB)

    Sierakowski, Adam J., E-mail: sierakowski@jhu.edu [Department of Mechanical Engineering, Johns Hopkins University, 3400 North Charles Street, Baltimore, MD 21218 (United States); Prosperetti, Andrea [Department of Mechanical Engineering, Johns Hopkins University, 3400 North Charles Street, Baltimore, MD 21218 (United States); Faculty of Science and Technology and J.M. Burgers Centre for Fluid Dynamics, University of Twente, P.O. Box 217, 7500 AE Enschede (Netherlands)

    2016-03-15

    We present enhancements and new capabilities of the Physalis method for simulating disperse multiphase flows using particle-resolved simulation. The current work enhances the previous method by incorporating a new type of pressure-Poisson solver that couples with a new Physalis particle pressure boundary condition scheme and a new particle interior treatment to significantly improve overall numerical efficiency. Further, we implement a more efficient method of calculating the Physalis scalar products and incorporate short-range particle interaction models. We provide validation and benchmarking for the Physalis method against experiments of a sedimenting particle and of normal wall collisions. We conclude with an illustrative simulation of 2048 particles sedimenting in a duct. In the appendix, we present a complete and self-consistent description of the analytical development and numerical methods.

  14. Hydrodynamics in adaptive resolution particle simulations: Multiparticle collision dynamics

    Energy Technology Data Exchange (ETDEWEB)

    Alekseeva, Uliana, E-mail: Alekseeva@itc.rwth-aachen.de [Jülich Supercomputing Centre (JSC), Institute for Advanced Simulation (IAS), Forschungszentrum Jülich, D-52425 Jülich (Germany); German Research School for Simulation Sciences (GRS), Forschungszentrum Jülich, D-52425 Jülich (Germany); Winkler, Roland G., E-mail: r.winkler@fz-juelich.de [Theoretical Soft Matter and Biophysics, Institute for Advanced Simulation (IAS), Forschungszentrum Jülich, D-52425 Jülich (Germany); Sutmann, Godehard, E-mail: g.sutmann@fz-juelich.de [Jülich Supercomputing Centre (JSC), Institute for Advanced Simulation (IAS), Forschungszentrum Jülich, D-52425 Jülich (Germany); ICAMS, Ruhr-University Bochum, D-44801 Bochum (Germany)

    2016-06-01

    A new adaptive resolution technique for particle-based multi-level simulations of fluids is presented. In the approach, the representation of fluid and solvent particles is changed on the fly between an atomistic and a coarse-grained description. The present approach is based on a hybrid coupling of the multiparticle collision dynamics (MPC) method and molecular dynamics (MD), thereby coupling stochastic and deterministic particle-based methods. Hydrodynamics is examined by calculating velocity and current correlation functions for various mixed and coupled systems. We demonstrate that hydrodynamic properties of the mixed fluid are conserved by a suitable coupling of the two particle methods, and that the simulation results agree well with theoretical expectations.

  15. Expected Transmission Energy Route Metric for Wireless Mesh Senor Networks

    Directory of Open Access Journals (Sweden)

    YanLiang Jin

    2011-01-01

    Full Text Available Mesh is a network topology that achieves high throughput and stable intercommunication. With great potential, it is expected to be the key architecture of future networks. Wireless sensor networks are an active research area with numerous workshops and conferences arranged each year. The overall performance of a WSN highly depends on the energy consumption of the network. This paper designs a new routing metric for wireless mesh sensor networks. Results from simulation experiments reveal that the new metric algorithm improves the energy balance of the whole network and extends the lifetime of wireless mesh sensor networks (WMSNs.

  16. Computational plasticity algorithm for particle dynamics simulations

    Science.gov (United States)

    Krabbenhoft, K.; Lyamin, A. V.; Vignes, C.

    2018-01-01

    The problem of particle dynamics simulation is interpreted in the framework of computational plasticity leading to an algorithm which is mathematically indistinguishable from the common implicit scheme widely used in the finite element analysis of elastoplastic boundary value problems. This algorithm provides somewhat of a unification of two particle methods, the discrete element method and the contact dynamics method, which usually are thought of as being quite disparate. In particular, it is shown that the former appears as the special case where the time stepping is explicit while the use of implicit time stepping leads to the kind of schemes usually labelled contact dynamics methods. The framing of particle dynamics simulation within computational plasticity paves the way for new approaches similar (or identical) to those frequently employed in nonlinear finite element analysis. These include mixed implicit-explicit time stepping, dynamic relaxation and domain decomposition schemes.

  17. Microscopic Simulation of Particle Detectors

    CERN Document Server

    Schindler, Heinrich

    Detailed computer simulations are indispensable tools for the development and optimization of modern particle detectors. The interaction of particles with the sensitive medium, giving rise to ionization or excitation of atoms, is stochastic by its nature. The transport of the resulting photons and charge carriers, which eventually generate the observed signal, is also subject to statistical fluctuations. Together with the readout electronics, these processes - which are ultimately governed by the atomic cross-sections for the respective interactions - pose a fundamental limit to the achievable detector performance. Conventional methods for calculating electron drift lines based on macroscopic transport coefficients used to provide an adequate description for traditional gas-based particle detectors such as wire chambers. However, they are not suitable for small-scale devices such as micropattern gas detectors, which have significantly gained importance in recent years. In this thesis, a novel approach, bas...

  18. Data simulation for the Associated Particle Imaging system

    International Nuclear Information System (INIS)

    Tunnell, L.N.

    1994-01-01

    A data simulation procedure for the Associated Particle Imaging (API) system has been developed by postprocessing output from the Monte Carlo Neutron Photon (MCNP) code. This paper compares the simulated results to our experimental data

  19. A New Approach to Adaptive Control of Multiple Scales in Plasma Simulations

    Science.gov (United States)

    Omelchenko, Yuri

    2007-04-01

    A new approach to temporal refinement of kinetic (Particle-in-Cell, Vlasov) and fluid (MHD, two-fluid) simulations of plasmas is presented: Discrete-Event Simulation (DES). DES adaptively distributes CPU resources in accordance with local time scales and enables asynchronous integration of inhomogeneous nonlinear systems with multiple time scales on meshes of arbitrary topologies. This removes computational penalties usually incurred in explicit codes due to the global Courant-Friedrich-Levy (CFL) restriction on a time-step size. DES stands apart from multiple time-stepping algorithms in that it requires neither selecting a global synchronization time step nor pre-determining a sequence of time-integration operations for individual parts of the system (local time increments need not bear any integer multiple relations). Instead, elements of a mesh-distributed solution self-adaptively predict and synchronize their temporal trajectories by directly enforcing local causality (accuracy) constraints, which are formulated in terms of incremental changes to the evolving solution. Together with flux-conservative propagation of information, this new paradigm ensures stable and fast asynchronous runs, where idle computation is automatically eliminated. DES is parallelized via a novel Preemptive Event Processing (PEP) technique, which automatically synchronizes elements with similar update rates. In this mode, events with close execution times are projected onto time levels, which are adaptively determined by the program. PEP allows reuse of standard message-passing algorithms on distributed architectures. For optimum accuracy, DES can be combined with adaptive mesh refinement (AMR) techniques for structured and unstructured meshes. Current examples of event-driven models range from electrostatic, hybrid particle-in-cell plasma systems to reactive fluid dynamics simulations. They demonstrate the superior performance of DES in terms of accuracy, speed and robustness.

  20. Granular dynamics, contact mechanics and particle system simulations a DEM study

    CERN Document Server

    Thornton, Colin

    2015-01-01

    This book is devoted to the Discrete Element Method (DEM) technique, a discontinuum modelling approach that takes into account the fact that granular materials are composed of discrete particles which interact with each other at the microscale level. This numerical simulation technique can be used both for dispersed systems in which the particle-particle interactions are collisional and compact systems of particles with multiple enduring contacts. The book provides an extensive and detailed explanation of the theoretical background of DEM. Contact mechanics theories for elastic, elastic-plastic, adhesive elastic and adhesive elastic-plastic particle-particle interactions are presented. Other contact force models are also discussed, including corrections to some of these models as described in the literature, and important areas of further research are identified. A key issue in DEM simulations is whether or not a code can reliably simulate the simplest of systems, namely the single particle oblique impact wit...

  1. Simulations of magnetic capturing of drug carriers in the brain vascular system

    Energy Technology Data Exchange (ETDEWEB)

    Kenjeres, S., E-mail: S.Kenjeres@tudelft.nl [Department of Multi-Scale Physics, Faculty of Applied Sciences, J.M. Burgerscentre for Fluid Dynamics, Delft University of Technology, Leeghwaterstraat 39, 2628 CB Delft (Netherlands); Righolt, B.W. [Department of Multi-Scale Physics, Faculty of Applied Sciences, J.M. Burgerscentre for Fluid Dynamics, Delft University of Technology, Leeghwaterstraat 39, 2628 CB Delft (Netherlands)

    2012-06-15

    Highlights: Black-Right-Pointing-Pointer Blood flow and magnetic particles distributions in the brain vascular system simulated. Black-Right-Pointing-Pointer Numerical mesh generated from raw MRI images. Black-Right-Pointing-Pointer Significant increase in local capturing of magnetic particles obtained. Black-Right-Pointing-Pointer Promising technique for localised non-invasive treatment of brain tumours. - Abstract: The present paper reports on numerical simulations of blood flow and magnetic drug carrier distributions in a complex brain vascular system. The blood is represented as a non-Newtonian fluid by the generalised power law. The Lagrangian tracking of the double-layer spherical particles is performed to estimate particle deposition under influence of imposed magnetic field gradients across arterial walls. Two situations are considered: neutral (magnetic field off) and active control (magnetic field on) case. The double-layer spherical particles that mimic a real medical drug are characterised by two characteristic diameters - the outer one and the inner one of the magnetic core. A numerical mesh of the brain vascular system consisting of multi-branching arteries is generated from raw MRI scan images of a patient. The blood is supplied through four main inlet arteries and the entire vascular system includes more than 30 outlets, which are modelled by Murray's law. The no-slip boundary condition is applied for velocity components along the smooth and rigid arterial walls. Numerical simulations revealed detailed insights into blood flow patterns, wall-shear-stress and local particle deposition efficiency along arterial walls. It is demonstrated that magnetically targeted drug delivery significantly increased the particle capturing efficiency in the pre-defined regions. This feature can be potentially useful for localised, non-invasive treatment of brain tumours.

  2. A constrained Delaunay discretization method for adaptively meshing highly discontinuous geological media

    Science.gov (United States)

    Wang, Yang; Ma, Guowei; Ren, Feng; Li, Tuo

    2017-12-01

    A constrained Delaunay discretization method is developed to generate high-quality doubly adaptive meshes of highly discontinuous geological media. Complex features such as three-dimensional discrete fracture networks (DFNs), tunnels, shafts, slopes, boreholes, water curtains, and drainage systems are taken into account in the mesh generation. The constrained Delaunay triangulation method is used to create adaptive triangular elements on planar fractures. Persson's algorithm (Persson, 2005), based on an analogy between triangular elements and spring networks, is enriched to automatically discretize a planar fracture into mesh points with varying density and smooth-quality gradient. The triangulated planar fractures are treated as planar straight-line graphs (PSLGs) to construct piecewise-linear complex (PLC) for constrained Delaunay tetrahedralization. This guarantees the doubly adaptive characteristic of the resulted mesh: the mesh is adaptive not only along fractures but also in space. The quality of elements is compared with the results from an existing method. It is verified that the present method can generate smoother elements and a better distribution of element aspect ratios. Two numerical simulations are implemented to demonstrate that the present method can be applied to various simulations of complex geological media that contain a large number of discontinuities.

  3. Filtration of submicrometer particles by pelagic tunicates.

    Science.gov (United States)

    Sutherland, Kelly R; Madin, Laurence P; Stocker, Roman

    2010-08-24

    Salps are common in oceanic waters and have higher per-individual filtration rates than any other zooplankton filter feeder. Although salps are centimeters in length, feeding via particle capture occurs on a fine, mucous mesh (fiber diameter d approximately 0.1 microm) at low velocity (U = 1.6 +/- 0.6 cmxs(-1), mean +/- SD) and is thus a low Reynolds-number (Re approximately 10(-3)) process. In contrast to the current view that particle encounter is dictated by simple sieving of particles larger than the mesh spacing, a low-Re mathematical model of encounter rates by the salp feeding apparatus for realistic oceanic particle-size distributions shows that submicron particles, due to their higher abundances, are encountered at higher rates (particles per time) than larger particles. Data from feeding experiments with 0.5-, 1-, and 3-microm diameter polystyrene spheres corroborate these findings. Although particles larger than 1 microm (e.g., flagellates, small diatoms) represent a larger carbon pool, smaller particles in the 0.1- to 1-microm range (e.g., bacteria, Prochlorococcus) may be more quickly digestible because they present more surface area, and we find that particles smaller than the mesh size (1.4 microm) can fully satisfy salp energetic needs. Furthermore, by packaging submicrometer particles into rapidly sinking fecal pellets, pelagic tunicates can substantially change particle-size spectra and increase downward fluxes in the ocean.

  4. The effects of particle recycling on the divertor plasma: A particle-in-cell with Monte Carlo collision simulation

    Science.gov (United States)

    Chang, Mingyu; Sang, Chaofeng; Sun, Zhenyue; Hu, Wanpeng; Wang, Dezhen

    2018-05-01

    A Particle-In-Cell (PIC) with Monte Carlo Collision (MCC) model is applied to study the effects of particle recycling on divertor plasma in the present work. The simulation domain is the scrape-off layer of the tokamak in one-dimension along the magnetic field line. At the divertor plate, the reflected deuterium atoms (D) and thermally released deuterium molecules (D2) are considered. The collisions between the plasma particles (e and D+) and recycled neutral particles (D and D2) are described by the MCC method. It is found that the recycled neutral particles have a great impact on divertor plasma. The effects of different collisions on the plasma are simulated and discussed. Moreover, the impacts of target materials on the plasma are simulated by comparing the divertor with Carbon (C) and Tungsten (W) targets. The simulation results show that the energy and momentum losses of the C target are larger than those of the W target in the divertor region even without considering the impurity particles, whereas the W target has a more remarkable influence on the core plasma.

  5. Open boundaries for particle beams within fit-simulations

    Energy Technology Data Exchange (ETDEWEB)

    Balk, M.C. [Technische Universitaet Darmstadt, Institut fuer Theorie Elektromagnetischer Felder, Schlossgartenstr. 8, 64289 Darmstadt (Germany)]. E-mail: balk@temf.tu-darmstadt.de; Schuhmann, R. [Technische Universitaet Darmstadt, Institut fuer Theorie Elektromagnetischer Felder, Schlossgartenstr. 8, 64289 Darmstadt (Germany); Weiland, T. [Technische Universitaet Darmstadt, Institut fuer Theorie Elektromagnetischer Felder, Schlossgartenstr. 8, 64289 Darmstadt (Germany)

    2006-03-01

    A method is proposed to simulate open boundary conditions for charged particle beams with vparticles are represented by a line current. Further, the simulated field components at the boundary of the calculation domain have to be modified for an undisturbed transmission of the space-charge field. This can be realised by a 'scattered field' formulation. The method is verified by several calculations.

  6. Tensile Behaviour of Welded Wire Mesh and Hexagonal Metal Mesh for Ferrocement Application

    Science.gov (United States)

    Tanawade, A. G.; Modhera, C. D.

    2017-08-01

    Tension tests were conducted on welded mesh and hexagonal Metal mesh. Welded Mesh is available in the market in different sizes. The two types are analysed viz. Ø 2.3 mm and Ø 2.7 mm welded mesh, having opening size 31.75 mm × 31.75 mm and 25.4 mm × 25.4 mm respectively. Tensile strength test was performed on samples of welded mesh in three different orientations namely 0°, 30° and 45° degrees with the loading axis and hexagonal Metal mesh of Ø 0.7 mm, having opening 19.05 × 19.05 mm. Experimental tests were conducted on samples of these meshes. The objective of this study was to investigate the behaviour of the welded mesh and hexagonal Metal mesh. The result shows that the tension load carrying capacity of welded mesh of Ø 2.7 mm of 0° orientation is good as compared to Ø2.3 mm mesh and ductility of hexagonal Metal mesh is good in behaviour.

  7. Validation of the coupling of mesh models to GEANT4 Monte Carlo code for simulation of internal sources of photons

    International Nuclear Information System (INIS)

    Caribe, Paulo Rauli Rafeson Vasconcelos; Cassola, Vagner Ferreira; Kramer, Richard; Khoury, Helen Jamil

    2013-01-01

    The use of three-dimensional models described by polygonal meshes in numerical dosimetry enables more accurate modeling of complex objects than the use of simple solid. The objectives of this work were validate the coupling of mesh models to the Monte Carlo code GEANT4 and evaluate the influence of the number of vertices in the simulations to obtain absorbed fractions of energy (AFEs). Validation of the coupling was performed to internal sources of photons with energies between 10 keV and 1 MeV for spherical geometries described by the GEANT4 and three-dimensional models with different number of vertices and triangular or quadrilateral faces modeled using Blender program. As a result it was found that there were no significant differences between AFEs for objects described by mesh models and objects described using solid volumes of GEANT4. Since that maintained the shape and the volume the decrease in the number of vertices to describe an object does not influence so meant dosimetric data, but significantly decreases the time required to achieve the dosimetric calculations, especially for energies less than 100 keV

  8. Competition between drag and Coulomb interactions in turbulent particle-laden flows using a coupled-fluid-Ewald-summation based approach

    Science.gov (United States)

    Yao, Yuan; Capecelatro, Jesse

    2018-03-01

    We present a numerical study on inertial electrically charged particles suspended in a turbulent carrier phase. Fluid-particle interactions are accounted for in an Eulerian-Lagrangian (EL) framework and coupled to a Fourier-based Ewald summation method, referred to as the particle-particle-particle-mesh (P3M ) method, to accurately capture short- and long-range electrostatic forces in a tractable manner. The EL P3M method is used to assess the competition between drag and Coulomb forces for a range of Stokes numbers and charge densities. Simulations of like- and oppositely charged particles suspended in a two-dimensional Taylor-Green vortex and three-dimensional homogeneous isotropic turbulence are reported. It is found that even in dilute suspensions, the short-range electric potential plays an important role in flows that admit preferential concentration. Suspensions of oppositely charged particles are observed to agglomerate in the form of chains and rings. Comparisons between the particle-mesh method typically employed in fluid-particle calculations and P3M are reported, in addition to one-point and two-point statistics to quantify the level of clustering as a function of Reynolds number, Stokes number, and nondimensional electric settling velocity.

  9. Seeking new surgical predictors of mesh exposure after transvaginal mesh repair.

    Science.gov (United States)

    Wu, Pei-Ying; Chang, Chih-Hung; Shen, Meng-Ru; Chou, Cheng-Yang; Yang, Yi-Ching; Huang, Yu-Fang

    2016-10-01

    The purpose of this study was to explore new preventable risk factors for mesh exposure. A retrospective review of 92 consecutive patients treated with transvaginal mesh (TVM) in the urogynecological unit of our university hospital. An analysis of perioperative predictors was conducted in patients after vaginal repairs using a type 1 mesh. Mesh complications were recorded according to International Urogynecological Association (IUGA) definitions. Mesh-exposure-free durations were calculated by using the Kaplan-Meier method and compared between different closure techniques using log-rank test. Hazard ratios (HR) of predictors for mesh exposure were estimated by univariate and multivariate analyses using Cox proportional hazards regression models. The median surveillance interval was 24.1 months. Two late occurrences were found beyond 1 year post operation. No statistically significant correlation was observed between mesh exposure and concomitant hysterectomy. Exposure risks were significantly higher in patients with interrupted whole-layer closure in univariate analysis. In the multivariate analysis, hematoma [HR 5.42, 95 % confidence interval (CI) 1.26-23.35, P = 0.024), Prolift mesh (HR 5.52, 95 % CI 1.15-26.53, P = 0.033), and interrupted whole-layer closure (HR 7.02, 95 % CI 1.62-30.53, P = 0.009) were the strongest predictors of mesh exposure. Findings indicate the risks of mesh exposure and reoperation may be prevented by avoiding hematoma, large amount of mesh, or interrupted whole-layer closure in TVM surgeries. If these risk factors are prevented, hysterectomy may not be a relative contraindication for TVM use. We also provide evidence regarding mesh exposure and the necessity for more than 1 year of follow-up and preoperative counselling.

  10. Progress on H5Part: A Portable High Performance Parallel Data Interface for Electromagnetics Simulations

    International Nuclear Information System (INIS)

    Adelmann, Andreas; Gsell, Achim; Oswald, Benedikt; Schietinger, Thomas; Bethel, Wes; Shalf, John; Siegerist, Cristina; Stockinger, Kurt

    2007-01-01

    Significant problems facing all experimental and computational sciences arise from growing data size and complexity. Common to all these problems is the need to perform efficient data I/O on diverse computer architectures. In our scientific application, the largest parallel particle simulations generate vast quantities of six-dimensional data. Such a simulation run produces data for an aggregate data size up to several TB per run. Motivated by the need to address data I/O and access challenges, we have implemented H5Part, an open source data I/O API that simplifies the use of the Hierarchical Data Format v5 library (HDF5). HDF5 is an industry standard for high performance, cross-platform data storage and retrieval that runs on all contemporary architectures from large parallel supercomputers to laptops. H5Part, which is oriented to the needs of the particle physics and cosmology communities, provides support for parallel storage and retrieval of particles, structured and in the future unstructured meshes. In this paper, we describe recent work focusing on I/O support for particles and structured meshes and provide data showing performance on modern supercomputer architectures like the IBM POWER 5

  11. Output power fluctuations due to different weights of macro particles used in particle-in-cell simulations of Cerenkov devices

    International Nuclear Information System (INIS)

    Bao, Rong; Li, Yongdong; Liu, Chunliang; Wang, Hongguang

    2016-01-01

    The output power fluctuations caused by weights of macro particles used in particle-in-cell (PIC) simulations of a backward wave oscillator and a travelling wave tube are statistically analyzed. It is found that the velocities of electrons passed a specific slow-wave structure form a specific electron velocity distribution. The electron velocity distribution obtained in PIC simulation with a relative small weight of macro particles is considered as an initial distribution. By analyzing this initial distribution with a statistical method, the estimations of the output power fluctuations caused by different weights of macro particles are obtained. The statistical method is verified by comparing the estimations with the simulation results. The fluctuations become stronger with increasing weight of macro particles, which can also be determined reversely from estimations of the output power fluctuations. With the weights of macro particles optimized by the statistical method, the output power fluctuations in PIC simulations are relatively small and acceptable.

  12. Three-dimensional simulation of the motion of a single particle under a simulated turbulent velocity field

    Science.gov (United States)

    Moreno-Casas, P. A.; Bombardelli, F. A.

    2015-12-01

    A 3D Lagrangian particle tracking model is coupled to a 3D channel velocity field to simulate the saltation motion of a single sediment particle moving in saltation mode. The turbulent field is a high-resolution three dimensional velocity field that reproduces a by-pass transition to turbulence on a flat plate due to free-stream turbulence passing above de plate. In order to reduce computational costs, a decoupled approached is used, i.e., the turbulent flow is simulated independently from the tracking model, and then used to feed the 3D Lagrangian particle model. The simulations are carried using the point-particle approach. The particle tracking model contains three sub-models, namely, particle free-flight, a post-collision velocity and bed representation sub-models. The free-flight sub-model considers the action of the following forces: submerged weight, non-linear drag, lift, virtual mass, Magnus and Basset forces. The model also includes the effect of particle angular velocity. The post-collision velocities are obtained by applying conservation of angular and linear momentum. The complete model was validated with experimental results from literature within the sand range. Results for particle velocity time series and distribution of particle turbulent intensities are presented.

  13. Trajectory Optimization Based on Multi-Interval Mesh Refinement Method

    Directory of Open Access Journals (Sweden)

    Ningbo Li

    2017-01-01

    Full Text Available In order to improve the optimization accuracy and convergence rate for trajectory optimization of the air-to-air missile, a multi-interval mesh refinement Radau pseudospectral method was introduced. This method made the mesh endpoints converge to the practical nonsmooth points and decreased the overall collocation points to improve convergence rate and computational efficiency. The trajectory was divided into four phases according to the working time of engine and handover of midcourse and terminal guidance, and then the optimization model was built. The multi-interval mesh refinement Radau pseudospectral method with different collocation points in each mesh interval was used to solve the trajectory optimization model. Moreover, this method was compared with traditional h method. Simulation results show that this method can decrease the dimensionality of nonlinear programming (NLP problem and therefore improve the efficiency of pseudospectral methods for solving trajectory optimization problems.

  14. A software platform for continuum modeling of ion channels based on unstructured mesh

    International Nuclear Information System (INIS)

    Tu, B; Bai, S Y; Xie, Y; Zhang, L B; Lu, B Z; Chen, M X

    2014-01-01

    Most traditional continuum molecular modeling adopted finite difference or finite volume methods which were based on a structured mesh (grid). Unstructured meshes were only occasionally used, but an increased number of applications emerge in molecular simulations. To facilitate the continuum modeling of biomolecular systems based on unstructured meshes, we are developing a software platform with tools which are particularly beneficial to those approaches. This work describes the software system specifically for the simulation of a typical, complex molecular procedure: ion transport through a three-dimensional channel system that consists of a protein and a membrane. The platform contains three parts: a meshing tool chain for ion channel systems, a parallel finite element solver for the Poisson–Nernst–Planck equations describing the electrodiffusion process of ion transport, and a visualization program for continuum molecular modeling. The meshing tool chain in the platform, which consists of a set of mesh generation tools, is able to generate high-quality surface and volume meshes for ion channel systems. The parallel finite element solver in our platform is based on the parallel adaptive finite element package PHG which wass developed by one of the authors [1]. As a featured component of the platform, a new visualization program, VCMM, has specifically been developed for continuum molecular modeling with an emphasis on providing useful facilities for unstructured mesh-based methods and for their output analysis and visualization. VCMM provides a graphic user interface and consists of three modules: a molecular module, a meshing module and a numerical module. A demonstration of the platform is provided with a study of two real proteins, the connexin 26 and hemolysin ion channels. (paper)

  15. Shadowfax: Moving mesh hydrodynamical integration code

    Science.gov (United States)

    Vandenbroucke, Bert

    2016-05-01

    Shadowfax simulates galaxy evolution. Written in object-oriented modular C++, it evolves a mixture of gas, subject to the laws of hydrodynamics and gravity, and any collisionless fluid only subject to gravity, such as cold dark matter or stars. For the hydrodynamical integration, it makes use of a (co-) moving Lagrangian mesh. The code has a 2D and 3D version, contains utility programs to generate initial conditions and visualize simulation snapshots, and its input/output is compatible with a number of other simulation codes, e.g. Gadget2 (ascl:0003.001) and GIZMO (ascl:1410.003).

  16. The Role of Chronic Mesh Infection in Delayed-Onset Vaginal Mesh Complications or Recurrent Urinary Tract Infections: Results From Explanted Mesh Cultures.

    Science.gov (United States)

    Mellano, Erin M; Nakamura, Leah Y; Choi, Judy M; Kang, Diana C; Grisales, Tamara; Raz, Shlomo; Rodriguez, Larissa V

    2016-01-01

    Vaginal mesh complications necessitating excision are increasingly prevalent. We aim to study whether subclinical chronically infected mesh contributes to the development of delayed-onset mesh complications or recurrent urinary tract infections (UTIs). Women undergoing mesh removal from August 2013 through May 2014 were identified by surgical code for vaginal mesh removal. Only women undergoing removal of anti-incontinence mesh were included. Exclusion criteria included any women undergoing simultaneous prolapse mesh removal. We abstracted preoperative and postoperative information from the medical record and compared mesh culture results from patients with and without mesh extrusion, de novo recurrent UTIs, and delayed-onset pain. One hundred seven women with only anti-incontinence mesh removed were included in the analysis. Onset of complications after mesh placement was within the first 6 months in 70 (65%) of 107 and delayed (≥6 months) in 37 (35%) of 107. A positive culture from the explanted mesh was obtained from 82 (77%) of 107 patients, and 40 (37%) of 107 were positive with potential pathogens. There were no significant differences in culture results when comparing patients with delayed-onset versus immediate pain, extrusion with no extrusion, and de novo recurrent UTIs with no infections. In this large cohort of patients with mesh removed for a diverse array of complications, cultures of the explanted vaginal mesh demonstrate frequent low-density bacterial colonization. We found no differences in culture results from women with delayed-onset pain versus acute pain, vaginal mesh extrusions versus no extrusions, or recurrent UTIs using standard culture methods. Chronic prosthetic infections in other areas of medicine are associated with bacterial biofilms, which are resistant to typical culture techniques. Further studies using culture-independent methods are needed to investigate the potential role of chronic bacterial infections in delayed vaginal mesh

  17. Direct numerical simulation of fluid-particle heat transfer in fixed random arrays of non-spherical particles

    NARCIS (Netherlands)

    Tavassoli Estahbanati, H.; Peters, E.A.J.F.; Kuipers, J.A.M.

    2015-01-01

    Direct numerical simulations are conducted to characterize the fluid-particle heat transfer coefficient in fixed random arrays of non-spherical particles. The objective of this study is to examine the applicability of well-known heat transfer correlations, that are proposed for spherical particles,

  18. Automatic mesh adaptivity for CADIS and FW-CADIS neutronics modeling of difficult shielding problems

    International Nuclear Information System (INIS)

    Ibrahim, A. M.; Peplow, D. E.; Mosher, S. W.; Wagner, J. C.; Evans, T. M.; Wilson, P. P.; Sawan, M. E.

    2013-01-01

    The CADIS and FW-CADIS hybrid Monte Carlo/deterministic techniques dramatically increase the efficiency of neutronics modeling, but their use in the accurate design analysis of very large and geometrically complex nuclear systems has been limited by the large number of processors and memory requirements for their preliminary deterministic calculations and final Monte Carlo calculation. Three mesh adaptivity algorithms were developed to reduce the memory requirements of CADIS and FW-CADIS without sacrificing their efficiency improvement. First, a macro-material approach enhances the fidelity of the deterministic models without changing the mesh. Second, a deterministic mesh refinement algorithm generates meshes that capture as much geometric detail as possible without exceeding a specified maximum number of mesh elements. Finally, a weight window coarsening algorithm de-couples the weight window mesh and energy bins from the mesh and energy group structure of the deterministic calculations in order to remove the memory constraint of the weight window map from the deterministic mesh resolution. The three algorithms were used to enhance an FW-CADIS calculation of the prompt dose rate throughout the ITER experimental facility. Using these algorithms resulted in a 23.3% increase in the number of mesh tally elements in which the dose rates were calculated in a 10-day Monte Carlo calculation and, additionally, increased the efficiency of the Monte Carlo simulation by a factor of at least 3.4. The three algorithms enabled this difficult calculation to be accurately solved using an FW-CADIS simulation on a regular computer cluster, obviating the need for a world-class super computer. (authors)

  19. Automatic mesh adaptivity for hybrid Monte Carlo/deterministic neutronics modeling of difficult shielding problems

    International Nuclear Information System (INIS)

    Ibrahim, Ahmad M.; Wilson, Paul P.H.; Sawan, Mohamed E.; Mosher, Scott W.; Peplow, Douglas E.; Wagner, John C.; Evans, Thomas M.; Grove, Robert E.

    2015-01-01

    The CADIS and FW-CADIS hybrid Monte Carlo/deterministic techniques dramatically increase the efficiency of neutronics modeling, but their use in the accurate design analysis of very large and geometrically complex nuclear systems has been limited by the large number of processors and memory requirements for their preliminary deterministic calculations and final Monte Carlo calculation. Three mesh adaptivity algorithms were developed to reduce the memory requirements of CADIS and FW-CADIS without sacrificing their efficiency improvement. First, a macromaterial approach enhances the fidelity of the deterministic models without changing the mesh. Second, a deterministic mesh refinement algorithm generates meshes that capture as much geometric detail as possible without exceeding a specified maximum number of mesh elements. Finally, a weight window coarsening algorithm decouples the weight window mesh and energy bins from the mesh and energy group structure of the deterministic calculations in order to remove the memory constraint of the weight window map from the deterministic mesh resolution. The three algorithms were used to enhance an FW-CADIS calculation of the prompt dose rate throughout the ITER experimental facility. Using these algorithms resulted in a 23.3% increase in the number of mesh tally elements in which the dose rates were calculated in a 10-day Monte Carlo calculation and, additionally, increased the efficiency of the Monte Carlo simulation by a factor of at least 3.4. The three algorithms enabled this difficult calculation to be accurately solved using an FW-CADIS simulation on a regular computer cluster, eliminating the need for a world-class super computer

  20. Deflating link buffers in a wireless mesh network

    KAUST Repository

    Jamshaid, Kamran; Shihada, Basem; Showail, Ahmad; Levis, Philip

    2014-01-01

    We analyze the problem of buffer sizing for backlogged TCP flows in 802.11-based wireless mesh networks. Our objective is to maintain high network utilization while providing low queueing delays. Unlike wired networks where a single link buffer feeds a bottleneck link, the radio spectral resource in a mesh network is shared among a set of contending mesh routers. We account for this by formulating the buffer size problem as sizing a collective buffer distributed over a set of interfering nodes. In this paper we propose mechanisms for sizing and distributing this collective buffer among the mesh nodes constituting the network bottleneck. Our mechanism factors in the network topology and wireless link rates, improving on pre-set buffer allocations that cannot optimally work across the range of configurations achievable with 802.11 radios. We evaluate our mechanisms using simulations as well as experiments on a testbed. Our results show that we can reduce the RTT of a flow by 6× or more, at the cost of less than 10% drop in end-to-end flow throughput.

  1. Deflating link buffers in a wireless mesh network

    KAUST Repository

    Jamshaid, Kamran

    2014-05-01

    We analyze the problem of buffer sizing for backlogged TCP flows in 802.11-based wireless mesh networks. Our objective is to maintain high network utilization while providing low queueing delays. Unlike wired networks where a single link buffer feeds a bottleneck link, the radio spectral resource in a mesh network is shared among a set of contending mesh routers. We account for this by formulating the buffer size problem as sizing a collective buffer distributed over a set of interfering nodes. In this paper we propose mechanisms for sizing and distributing this collective buffer among the mesh nodes constituting the network bottleneck. Our mechanism factors in the network topology and wireless link rates, improving on pre-set buffer allocations that cannot optimally work across the range of configurations achievable with 802.11 radios. We evaluate our mechanisms using simulations as well as experiments on a testbed. Our results show that we can reduce the RTT of a flow by 6× or more, at the cost of less than 10% drop in end-to-end flow throughput.

  2. Basic Algorithms for the Asynchronous Reconfigurable Mesh

    Directory of Open Access Journals (Sweden)

    Yosi Ben-Asher

    2002-01-01

    Full Text Available Many constant time algorithms for various problems have been developed for the reconfigurable mesh (RM in the past decade. All these algorithms are designed to work with synchronous execution, with no regard for the fact that large size RMs will probably be asynchronous. A similar observation about the PRAM model motivated many researchers to develop algorithms and complexity measures for the asynchronous PRAM (APRAM. In this work, we show how to define the asynchronous reconfigurable mesh (ARM and how to measure the complexity of asynchronous algorithms executed on it. We show that connecting all processors in a row of an n×n ARM (the analog of barrier synchronization in the APRAM model can be solved with complexity Θ(nlog⁡n. Intuitively, this is average work time for solving such a problem. Next, we describe general a technique for simulating T -step synchronous RM algorithms on the ARM with complexity of Θ(T⋅n2log⁡n. Finally, we consider the simulation of the classical synchronous algorithm for counting the number of non-zero bits in an n bits vector using (ksimulated, one can (at least in the case of counting improve upon the general simulation.

  3. Milestone Deliverable: FY18-Q1: Deploy production sliding mesh capability with linear solver benchmarking.

    Energy Technology Data Exchange (ETDEWEB)

    Domino, Stefan P. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2017-12-01

    This milestone was focused on deploying and verifying a “sliding-mesh interface,” and establishing baseline timings for blade-resolved simulations of a sub-MW-scale turbine. In the ExaWind project, we are developing both sliding-mesh and overset-mesh approaches for handling the rotating blades in an operating wind turbine. In the sliding-mesh approach, the turbine rotor and its immediate surrounding fluid are captured in a “disk” that is embedded in the larger fluid domain. The embedded fluid is simulated in a coordinate system that rotates with the rotor. It is important that the coupling algorithm (and its implementation) between the rotating and inertial discrete models maintains the accuracy of the numerical methods on either side of the interface, i.e., the interface is “design order.”

  4. A PIC-MCC code for simulation of streamer propagation in air

    DEFF Research Database (Denmark)

    Chanrion, Olivier Arnaud; Neubert, Torsten

    2008-01-01

    A particle code has been developed to study the distribution and acceleration of electrons in electric discharges in air. The code can follow the evolution of a discharge from the initial stage of a single free electron in a background electric field to the formation of an electron avalanche...... and its transition into a streamer. The code is in 2D axi-symmetric coordinates, allowing quasi 3D simulations during the initial stages of streamer formation. This is important for realistic simulations of problems where space charge fields are essential such as in streamer formation. The charged...... particles are followed in a Cartesian mesh and the electric field is updated with Poisson's equation from the charged particle densities. Collisional processes between electrons and air molecules are simulated with a Monte Carlo technique, according to cross section probabilities. The code also includes...

  5. Simulation of a pulsatile total artificial heart: Development of a partitioned Fluid Structure Interaction model

    Science.gov (United States)

    Sonntag, Simon J.; Kaufmann, Tim A. S.; Büsen, Martin R.; Laumen, Marco; Linde, Torsten; Schmitz-Rode, Thomas; Steinseifer, Ulrich

    2013-04-01

    Heart disease is one of the leading causes of death in the world. Due to a shortage in donor organs artificial hearts can be a bridge to transplantation or even serve as a destination therapy for patients with terminal heart insufficiency. A pusher plate driven pulsatile membrane pump, the Total Artificial Heart (TAH) ReinHeart, is currently under development at the Institute of Applied Medical Engineering of RWTH Aachen University.This paper presents the methodology of a fully coupled three-dimensional time-dependent Fluid Structure Interaction (FSI) simulation of the TAH using a commercial partitioned block-Gauss-Seidel coupling package. Partitioned coupling of the incompressible fluid with the slender flexible membrane as well as a high fluid/structure density ratio of about unity led inherently to a deterioration of the stability (‘artificial added mass instability’). The objective was to conduct a stable simulation with high accuracy of the pumping process. In order to achieve stability, a combined resistance and pressure outlet boundary condition as well as the interface artificial compressibility method was applied. An analysis of the contact algorithm and turbulence condition is presented. Independence tests are performed for the structural and the fluid mesh, the time step size and the number of pulse cycles. Because of the large deformation of the fluid domain, a variable mesh stiffness depending on certain mesh properties was specified for the fluid elements. Adaptive remeshing was avoided. Different approaches for the mesh stiffness function are compared with respect to convergence, preservation of mesh topology and mesh quality. The resulting mesh aspect ratios, mesh expansion factors and mesh orthogonalities are evaluated in detail. The membrane motion and flow distribution of the coupled simulations are compared with a top-view recording and stereo Particle Image Velocimetry (PIV) measurements, respectively, of the actual pump.

  6. Noiseless Vlasov–Poisson simulations with linearly transformed particles

    Energy Technology Data Exchange (ETDEWEB)

    Campos Pinto, Martin, E-mail: campos@ann.jussieu.fr [Lawrence Berkeley National Laboratory, Berkeley, CA 94720 (United States); CNRS, UMR 7598, Laboratoire Jacques-Louis Lions, F-75005, Paris (France); UPMC Univ. Paris 06, UMR 7598, Laboratoire Jacques-Louis Lions, F-75005, Paris (France); Sonnendrücker, Eric, E-mail: sonnen@math.unistra.fr [IRMA, UMR 7501, Université de Strasbourg and CNRS, 7 rue René Descartes, F-67084 Strasbourg Cedex (France); Project-team CALVI, INRIA Nancy Grand Est, 7 rue René Descartes, F-67084 Strasbourg Cedex (France); Friedman, Alex, E-mail: af@llnl.gov [Lawrence Livermore National Laboratory, Livermore, CA 94550 (United States); Lawrence Berkeley National Laboratory, Berkeley, CA 94720 (United States); Grote, David P., E-mail: grote1@llnl.gov [Lawrence Livermore National Laboratory, Livermore, CA 94550 (United States); Lawrence Berkeley National Laboratory, Berkeley, CA 94720 (United States); Lund, Steve M., E-mail: smlund@llnl.gov [Lawrence Livermore National Laboratory, Livermore, CA 94550 (United States); Lawrence Berkeley National Laboratory, Berkeley, CA 94720 (United States)

    2014-10-15

    We introduce a deterministic discrete-particle simulation approach, the Linearly-Transformed Particle-In-Cell (LTPIC) method, that employs linear deformations of the particles to reduce the noise traditionally associated with particle schemes. Formally, transforming the particles is justified by local first order expansions of the characteristic flow in phase space. In practice the method amounts of using deformation matrices within the particle shape functions; these matrices are updated via local evaluations of the forward numerical flow. Because it is necessary to periodically remap the particles on a regular grid to avoid excessively deforming their shapes, the method can be seen as a development of Denavit's Forward Semi-Lagrangian (FSL) scheme (Denavit, 1972 [8]). However, it has recently been established (Campos Pinto, 2012 [20]) that the underlying Linearly-Transformed Particle scheme converges for abstract transport problems, with no need to remap the particles; deforming the particles can thus be seen as a way to significantly lower the remapping frequency needed in the FSL schemes, and hence the associated numerical diffusion. To couple the method with electrostatic field solvers, two specific charge deposition schemes are examined, and their performance compared with that of the standard deposition method. Finally, numerical 1d1v simulations involving benchmark test cases and halo formation in an initially mismatched thermal sheet beam demonstrate some advantages of our LTPIC scheme over the classical PIC and FSL methods. Benchmarked test cases also indicate that, for numerical choices involving similar computational effort, the LTPIC method is capable of accuracy comparable to or exceeding that of state-of-the-art, high-resolution Vlasov schemes.

  7. Analysis of the Numerical Diffusion in Anisotropic Mediums: Benchmarks for Magnetic Field Aligned Meshes in Space Propulsion Simulations

    Directory of Open Access Journals (Sweden)

    Daniel Pérez-Grande

    2016-11-01

    Full Text Available This manuscript explores numerical errors in highly anisotropic diffusion problems. First, the paper addresses the use of regular structured meshes in numerical solutions versus meshes aligned with the preferential directions of the problem. Numerical diffusion in structured meshes is quantified by solving the classical anisotropic diffusion problem; the analysis is exemplified with the application to a numerical model of conducting fluids under magnetic confinement, where rates of transport in directions parallel and perpendicular to a magnetic field are quite different. Numerical diffusion errors in this problem promote the use of magnetic field aligned meshes (MFAM. The generation of this type of meshes presents some challenges; several meshing strategies are implemented and analyzed in order to provide insight into achieving acceptable mesh regularity. Second, Gradient Reconstruction methods for magnetically aligned meshes are addressed and numerical errors are compared for the structured and magnetically aligned meshes. It is concluded that using the latter provides a more correct and straightforward approach to solving problems where anisotropicity is present, especially, if the anisotropicity level is high or difficult to quantify. The conclusions of the study may be extrapolated to the study of anisotropic flows different from conducting fluids.

  8. Convergence study of global meshing on enamel-cement-bracket finite element model

    Science.gov (United States)

    Samshuri, S. F.; Daud, R.; Rojan, M. A.; Basaruddin, K. S.; Abdullah, A. B.; Ariffin, A. K.

    2017-09-01

    This paper presents on meshing convergence analysis of finite element (FE) model to simulate enamel-cement-bracket fracture. Three different materials used in this study involving interface fracture are concerned. Complex behavior ofinterface fracture due to stress concentration is the reason to have a well-constructed meshing strategy. In FE analysis, meshing size is a critical factor that influenced the accuracy and computational time of analysis. The convergence study meshing scheme involving critical area (CA) and non-critical area (NCA) to ensure an optimum meshing sizes are acquired for this FE model. For NCA meshing, the area of interest are at the back of enamel, bracket ligature groove and bracket wing. For CA meshing, area of interest are enamel area close to cement layer, the cement layer and bracket base. The value of constant NCA meshing tested are meshing size 1 and 0.4. The value constant CA meshing tested are 0.4 and 0.1. Manipulative variables are randomly selected and must abide the rule of NCA must be higher than CA. This study employed first principle stresses due to brittle failure nature of the materials used. Best meshing size are selected according to convergence error analysis. Results show that, constant CA are more stable compare to constant NCA meshing. Then, 0.05 constant CA meshing are tested to test the accuracy of smaller meshing. However, unpromising result obtained as the errors are increasing. Thus, constant CA 0.1 with NCA mesh of 0.15 until 0.3 are the most stable meshing as the error in this region are lowest. Convergence test was conducted on three selected coarse, medium and fine meshes at the range of NCA mesh of 0.15 until 3 and CA mesh area stay constant at 0.1. The result shows that, at coarse mesh 0.3, the error are 0.0003% compare to 3% acceptable error. Hence, the global meshing are converge as the meshing size at CA 0.1 and NCA 0.15 for this model.

  9. Topological patterns of mesh textures in serpentinites

    Science.gov (United States)

    Miyazawa, M.; Suzuki, A.; Shimizu, H.; Okamoto, A.; Hiraoka, Y.; Obayashi, I.; Tsuji, T.; Ito, T.

    2017-12-01

    Serpentinization is a hydration process that forms serpentine minerals and magnetite within the oceanic lithosphere. Microfractures crosscut these minerals during the reactions, and the structures look like mesh textures. It has been known that the patterns of microfractures and the system evolutions are affected by the hydration reaction and fluid transport in fractures and within matrices. This study aims at quantifying the topological patterns of the mesh textures and understanding possible conditions of fluid transport and reaction during serpentinization in the oceanic lithosphere. Two-dimensional simulation by the distinct element method (DEM) generates fracture patterns due to serpentinization. The microfracture patterns are evaluated by persistent homology, which measures features of connected components of a topological space and encodes multi-scale topological features in the persistence diagrams. The persistence diagrams of the different mesh textures are evaluated by principal component analysis to bring out the strong patterns of persistence diagrams. This approach help extract feature values of fracture patterns from high-dimensional and complex datasets.

  10. A Level-set based framework for viscous simulation of particle-laden supersonic flows

    Science.gov (United States)

    Das, Pratik; Sen, Oishik; Jacobs, Gustaaf; Udaykumar, H. S.

    2017-06-01

    Particle-laden supersonic flows are important in natural and industrial processes, such as, volcanic eruptions, explosions, pneumatic conveyance of particle in material processing etc. Numerical study of such high-speed particle laden flows at the mesoscale calls for a numerical framework which allows simulation of supersonic flow around multiple moving solid objects. Only a few efforts have been made toward development of numerical frameworks for viscous simulation of particle-fluid interaction in supersonic flow regime. The current work presents a Cartesian grid based sharp-interface method for viscous simulations of interaction between supersonic flow with moving rigid particles. The no-slip boundary condition is imposed at the solid-fluid interfaces using a modified ghost fluid method (GFM). The current method is validated against the similarity solution of compressible boundary layer over flat-plate and benchmark numerical solution for steady supersonic flow over cylinder. Further validation is carried out against benchmark numerical results for shock induced lift-off of a cylinder in a shock tube. 3D simulation of steady supersonic flow over sphere is performed to compare the numerically obtained drag co-efficient with experimental results. A particle-resolved viscous simulation of shock interaction with a cloud of particles is performed to demonstrate that the current method is suitable for large-scale particle resolved simulations of particle-laden supersonic flows.

  11. Direct numerical simulation of granular flows with fluid; Simulation numerique directe d'ecoulements granulaires en presence de fluide

    Energy Technology Data Exchange (ETDEWEB)

    Komiwes, V.

    1999-09-01

    Numerical models applied to simulation of granular flow with fluid are developed. The physical model selected to describe particles flow is a discrete approach. Particle trajectories are calculated by the Newton law and collision is describe by a soft-sphere approach. The fluid flow is modelled by Navier-Stokes equations. The modelling of the momentum transfer depends on the resolution scale: for a scale of the order of the particle diameter, it is modelled by a drag-law and for a scale smaller than the particle diameter, it is directly calculated by stress tensor computation around particles. The direct model is used to find representative elementary volume and prove the local character of the Ergun's law. This application shows the numerical (mesh size), physical (Reynolds number) and computational (CPU time and memory consumptions) limitations. The drag law model and the direct model are validated with analytical and empirical solutions and compared. For the two models, the CPU time and the memory consumptions are discussed. The drag law model is applied to the simulation of gas-solid dense fluidized-beds. In the case of uniform gas distribution, the fluidized-bed simulation heights are compared to experimental data for particle of group A and B of the Geldart classification. (author)

  12. Urogynecologic Surgical Mesh Implants

    Science.gov (United States)

    ... procedures performed to treat pelvic floor disorders with surgical mesh: Transvaginal mesh to treat POP Transabdominal mesh to treat ... address safety risks Final Order for Reclassification of Surgical Mesh for Transvaginal Pelvic Organ Prolapse Repair Final Order for Effective ...

  13. Hybrid mesh generation for the new generation of oil reservoir simulators: 3D extension; Generation de maillage hybride pour les simulateurs de reservoir petrolier de nouvelle generation: extension 3D

    Energy Technology Data Exchange (ETDEWEB)

    Flandrin, N.

    2005-09-15

    During the exploitation of an oil reservoir, it is important to predict the recovery of hydrocarbons and to optimize its production. A better comprehension of the physical phenomena requires to simulate 3D multiphase flows in increasingly complex geological structures. In this thesis, we are interested in this spatial discretization and we propose to extend in 3D the 2D hybrid model proposed by IFP in 1998 that allows to take directly into account in the geometry the radial characteristics of the flows. In these hybrid meshes, the wells and their drainage areas are described by structured radial circular meshes and the reservoirs are represented by structured meshes that can be a non uniform Cartesian grid or a Corner Point Geometry grids. In order to generate a global conforming mesh, unstructured transition meshes based on power diagrams and satisfying finite volume properties are used to connect the structured meshes together. Two methods have been implemented to generate these transition meshes: the first one is based on a Delaunay triangulation, the other one uses a frontal approach. Finally, some criteria are introduced to measure the quality of the transition meshes and optimization procedures are proposed to increase this quality under finite volume properties constraints. (author)

  14. SIMULATION OF TRANSLATIONAL - ROTATIONAL MOTION OF WOOD PARTICLES DURING THE PROCESS OF PARTICLE ORIENTATION

    Directory of Open Access Journals (Sweden)

    Sergey PLOTNIKOV

    2014-09-01

    Full Text Available The simulation from the motion of flat particle revealed that the fall depends on the height of the drop, the thickness and density of the particles and does not depend on its length and width. The drop in air is about 20% longer than in vacuum. During orientation from angular particles the velocity of rotating particles with a length of 150mm is reduced by 18%, for particles with a length of 75mm by 12%. This reduction increases linearly with decreasing density of particles. A velocity field acting on the particle in the fall and rotation was presented. The results of the study prove the possibility to reduce the scatter of the particles during the mat's formation, that in turns can increase the board’s bending strength.

  15. Intravesical midurethral sling mesh erosion secondary to transvaginal mesh reconstructive surgery

    Directory of Open Access Journals (Sweden)

    Sukanda Bin Jaili

    2015-05-01

    Conclusion: Repeated vaginal reconstructive surgery may jeopardize a primary mesh or sling, and pose a high risk of mesh erosion, which may be delayed for several years. Removal of the mesh erosion and bladder repair are feasible pervaginally with good outcome.

  16. Mesh versus non-mesh repair of ventral abdominal hernias

    International Nuclear Information System (INIS)

    Jawaid, M.A.; Talpur, A.H.

    2008-01-01

    To investigate the relative effectiveness of mesh and suture repair of ventral abdominal hernias in terms of clinical outcome, quality of life and rate of recurrence in both the techniques. This is a retrospective descriptive analysis of 236 patients with mesh and non-mesh repair of primary ventral hernias performed between January 2000 to December 2004 at Surgery Department, Liaquat University of Medical and Health Sciences, Jamshoro. The record sheets of the patients were analyzed and data retrieved to compare the results of both techniques for short-term and long-term results. The data retrieved is statistically analyzed on SPSS version 11. There were 43 (18.22%) males and 193 (81.77%) females with a mean age of 51.79 years and a range of 59 (81-22). Para-umbilical hernia was the commonest of ventral hernia and accounted for 49.8% (n=118) of the total study population followed by incisional hernia comprising 24% (n=57) of the total number. There was a significant difference in the recurrent rate at 3 years interval with 23/101 (22.77%) recurrences in suture-repaired subjects compared to 10/135 (7.40%) in mesh repair group. Chronic pain lasting up to 1-2 years was noted in 14 patients with suture repair. Wound infection is comparatively more common (8.14%) in mesh group. The other variables such as operative and postoperative complications, total hospital stay and quality of life is also discussed. Mesh repair of ventral hernia is much superior to non-mesh suture repair in terms of recurrence and overall outcome. (author)

  17. Auxiliary plasma heating and fueling models for use in particle simulation codes

    International Nuclear Information System (INIS)

    Procassini, R.J.; Cohen, B.I.

    1989-01-01

    Computational models of a radiofrequency (RF) heating system and neutral-beam injector are presented. These physics packages, when incorporated into a particle simulation code allow one to simulate the auxiliary heating and fueling of fusion plasmas. The RF-heating package is based upon a quasilinear diffusion equation which describes the slow evolution of the heated particle distribution. The neutral-beam injector package models the charge exchange and impact ionization processes which transfer energy and particles from the beam to the background plasma. Particle simulations of an RF-heated and a neutral-beam-heated simple-mirror plasma are presented. 8 refs., 5 figs

  18. Cobra-IE Evaluation by Simulation of the NUPEC BWR Full-Size Fine-Mesh Bundle Test (BFBT)

    International Nuclear Information System (INIS)

    Burns, C. J.; Aumiler, D.L.

    2006-01-01

    The COBRA-IE computer code is a thermal-hydraulic subchannel analysis program capable of simulating phenomena present in both PWRs and BWRs. As part of ongoing COBRA-IE assessment efforts, the code has been evaluated against experimental data from the NUPEC BWR Full-Size Fine-Mesh Bundle Tests (BFBT). The BFBT experiments utilized an 8 x 8 rod bundle to simulate BWR operating conditions and power profiles, providing an excellent database for investigation of the capabilities of the code. Benchmarks performed included steady-state and transient void distribution, single-phase and two-phase pressure drop, and steady-state and transient critical power measurements. COBRA-IE effectively captured the trends seen in the experimental data with acceptable prediction error. Future sensitivity studies are planned to investigate the effects of enabling and/or modifying optional code models dealing with void drift, turbulent mixing, rewetting, and CHF

  19. Synergy between NMR measurements and MD simulations of protein/RNA complexes: application to the RRMs, the most common RNA recognition motifs

    Czech Academy of Sciences Publication Activity Database

    Krepl, Miroslav; Clery, A.; Blatter, M.; Allain, F.H.T.; Šponer, Jiří

    2016-01-01

    Roč. 44, č. 13 (2016), s. 6452-6470 ISSN 0305-1048 Institutional support: RVO:68081707 Keywords : molecular- dynamics simulations * particle mesh ewald * pre-ribosomal-rna Subject RIV: BO - Biophysics Impact factor: 10.162, year: 2016

  20. Loading relativistic Maxwell distributions in particle simulations

    Energy Technology Data Exchange (ETDEWEB)

    Zenitani, Seiji, E-mail: seiji.zenitani@nao.ac.jp [National Astronomical Observatory of Japan, 2-21-1 Osawa, Mitaka, Tokyo 181-8588 (Japan)

    2015-04-15

    Numerical algorithms to load relativistic Maxwell distributions in particle-in-cell (PIC) and Monte-Carlo simulations are presented. For stationary relativistic Maxwellian, the inverse transform method and the Sobol algorithm are reviewed. To boost particles to obtain relativistic shifted-Maxwellian, two rejection methods are proposed in a physically transparent manner. Their acceptance efficiencies are ≈50% for generic cases and 100% for symmetric distributions. They can be combined with arbitrary base algorithms.

  1. Loading relativistic Maxwell distributions in particle simulations

    International Nuclear Information System (INIS)

    Zenitani, Seiji

    2015-01-01

    Numerical algorithms to load relativistic Maxwell distributions in particle-in-cell (PIC) and Monte-Carlo simulations are presented. For stationary relativistic Maxwellian, the inverse transform method and the Sobol algorithm are reviewed. To boost particles to obtain relativistic shifted-Maxwellian, two rejection methods are proposed in a physically transparent manner. Their acceptance efficiencies are ≈50% for generic cases and 100% for symmetric distributions. They can be combined with arbitrary base algorithms

  2. Quantitative cross-sectional measurement of solid concentration distribution in slurries using a wire-mesh sensor

    International Nuclear Information System (INIS)

    Dos Santos, Eduardo Nunes; Da Silva, Marco Jose; Schleicher, Eckhard; Reinecke, Sebastian; Hampel, Uwe

    2016-01-01

    Wire-mesh sensors have so far been widely applied in gas–liquid flows where resistance or capacitance distributions are measured and converted into gas or liquid holdup distributions. In this work we report on the qualification of the wire-mesh imaging technique for the measurement of cross-sectional solid concentrations in solid–liquid mixtures. As the dielectric constants of solid particles are different from those of gas, water or oil in the flow, measuring this property can be used as an indication of solid distribution. Experiments were performed in a stirred tank of 100 mm diameter equipped with a capacitance wire-mesh sensor. The wire-mesh sensor was operated at an acquisition speed of 4000 frames per second and has a spatial resolution of 6.25 mm. As solids we used silica sand particles (diameter ∼250 μm) which were suspended in water in a volume concentration range of 1% to 35% to form slurries. By varying the stirring speed, different solid concentration distributions were produced and investigated. In order to convert the measured relative permittivity distribution into a solid concentration distribution, an empirical approach was employed. (paper)

  3. Particle simulation of intense electron cyclotron heating and beat-wave current drive

    International Nuclear Information System (INIS)

    Cohen, B.I.

    1987-01-01

    High-power free-electron lasers make new methods possible for heating plasmas and driving current in toroidal plasmas with electromagnetic waves. We have undertaken particle simulation studies with one and two dimensional, relativistic particle simulation codes of intense pulsed electron cyclotron heating and beat-wave current drive. The particle simulation methods here are conventional: the algorithms are time-centered, second-order-accurate, explicit, leap-frog difference schemes. The use of conventional methods restricts the range of space and time scales to be relatively compact in the problems addressed. Nevertheless, experimentally relevant simulations have been performed. 10 refs., 2 figs

  4. A splitting integration scheme for the SPH simulation of concentrated particle suspensions

    Science.gov (United States)

    Bian, Xin; Ellero, Marco

    2014-01-01

    Simulating nearly contacting solid particles in suspension is a challenging task due to the diverging behavior of short-range lubrication forces, which pose a serious time-step limitation for explicit integration schemes. This general difficulty limits severely the total duration of simulations of concentrated suspensions. Inspired by the ideas developed in [S. Litvinov, M. Ellero, X.Y. Hu, N.A. Adams, J. Comput. Phys. 229 (2010) 5457-5464] for the simulation of highly dissipative fluids, we propose in this work a splitting integration scheme for the direct simulation of solid particles suspended in a Newtonian liquid. The scheme separates the contributions of different forces acting on the solid particles. In particular, intermediate- and long-range multi-body hydrodynamic forces, which are computed from the discretization of the Navier-Stokes equations using the smoothed particle hydrodynamics (SPH) method, are taken into account using an explicit integration; for short-range lubrication forces, velocities of pairwise interacting solid particles are updated implicitly by sweeping over all the neighboring pairs iteratively, until convergence in the solution is obtained. By using the splitting integration, simulations can be run stably and efficiently up to very large solid particle concentrations. Moreover, the proposed scheme is not limited to the SPH method presented here, but can be easily applied to other simulation techniques employed for particulate suspensions.

  5. Simulating immersed particle collisions: the Devil's in the details

    Science.gov (United States)

    Biegert, Edward; Vowinckel, Bernhard; Meiburg, Eckart

    2015-11-01

    Simulating densely-packed particle-laden flows with any degree of confidence requires accurate modeling of particle-particle collisions. To this end, we investigate a few collision models from the fluids and granular flow communities using sphere-wall collisions, which have been studied by a number of experimental groups. These collisions involve enough complexities--gravity, particle-wall lubrication forces, particle-wall contact stresses, particle-wake interactions--to challenge any collision model. Evaluating the successes and shortcomings of the collision models, we seek improvements in order to obtain more consistent results. We will highlight several implementation details that are crucial for obtaining accurate results.

  6. Mesh removal following transvaginal mesh placement: a case series of 104 operations.

    Science.gov (United States)

    Marcus-Braun, Naama; von Theobald, Peter

    2010-04-01

    The objective of the study was to reveal the way we treat vaginal mesh complications in a trained referral center. This is a retrospective review of all patients who underwent surgical removal of transvaginal mesh for mesh-related complications during a 5-year period. Eighty-three patients underwent 104 operations including 61 complete mesh removal, 14 partial excision, 15 section of sub-urethral sling, and five laparoscopies. Main indications were erosion, infection, granuloma, incomplete voiding, and pain. Fifty-eight removals occurred more than 2 years after the primary mesh placement. Mean operation time was 21 min, and there were two intraoperative and ten minor postoperative complications. Stress urinary incontinence (SUI) recurred in 38% and cystocele in 19% of patients. In a trained center, mesh removal was found to be a quick and safe procedure. Mesh-related complications may frequently occur more than 2 years after the primary operation. Recurrence was mostly associated with SUI and less with genital prolapse.

  7. Parallel paving: An algorithm for generating distributed, adaptive, all-quadrilateral meshes on parallel computers

    Energy Technology Data Exchange (ETDEWEB)

    Lober, R.R.; Tautges, T.J.; Vaughan, C.T.

    1997-03-01

    Paving is an automated mesh generation algorithm which produces all-quadrilateral elements. It can additionally generate these elements in varying sizes such that the resulting mesh adapts to a function distribution, such as an error function. While powerful, conventional paving is a very serial algorithm in its operation. Parallel paving is the extension of serial paving into parallel environments to perform the same meshing functions as conventional paving only on distributed, discretized models. This extension allows large, adaptive, parallel finite element simulations to take advantage of paving`s meshing capabilities for h-remap remeshing. A significantly modified version of the CUBIT mesh generation code has been developed to host the parallel paving algorithm and demonstrate its capabilities on both two dimensional and three dimensional surface geometries and compare the resulting parallel produced meshes to conventionally paved meshes for mesh quality and algorithm performance. Sandia`s {open_quotes}tiling{close_quotes} dynamic load balancing code has also been extended to work with the paving algorithm to retain parallel efficiency as subdomains undergo iterative mesh refinement.

  8. In vitro extracellular matrix model to evaluate stroma cell response to transvaginal mesh.

    Science.gov (United States)

    Wu, Ming-Ping; Huang, Kuan-Hui; Long, Cheng-Yu; Yang, Chau-Chen; Tong, Yat-Ching

    2014-04-01

    The use of surgical mesh for female pelvic floor reconstruction has increased in recent years. However, there is paucity of information about the biological responses of host stroma cells to different meshes. This study was aimed to establish an in vitro experimental model to study the micro-environment of extracellular matrix (ECM) with embedded mesh and the stroma cell behaviors to different synthetic meshes. Matrigel multi-cellular co-culture system with embedded mesh was used to evaluate the interaction of stroma cells and synthetic mesh in a simulated ECM environment. Human umbilical vein endothelial cells (HUVEC) and NIH3T3 fibroblasts were inoculated in the system. The established multi-cellular Matrigel co-culture system was used to detect stroma cell recruitment and tube formation ability for different synthetic meshes. HUVEC and NIH3T3 cells were recruited into the mesh interstices and organized into tube-like structures in type I mesh material from Perigee, Marlex and Prolift 24 hr after cell inoculation. On the contrary, there was little recruitment of HUVEC and NIH3T3 cells into the type III mesh of intra-vaginal sling (IVS). The Matrigel multi-cellular co-culture system with embedded mesh offers a useful in vitro model to study the biological behaviors of stroma cells in response to different types of synthetic meshes. The system can help to select ideal mesh candidates before actual implantation into the human body. © 2013 Wiley Periodicals, Inc.

  9. Hairpins participating in folding of human telomeric sequence quadruplexes studied by standard and T-REMD simulations

    Czech Academy of Sciences Publication Activity Database

    Stadlbauer, Petr; Kuehrova, P.; Banáš, P.; Koča, J.; Bussi, G.; Trantírek, L.; Otyepka, M.; Šponer, Jiří

    2016-01-01

    Roč. 43, č. 20 (2016), s. 9626-9644 ISSN 0305-1048 R&D Projects: GA ČR(CZ) GAP208/11/1822 Institutional support: RVO:68081707 Keywords : MOLECULAR-DYNAMICS SIMULATIONS * INTRAMOLECULAR DNA QUADRUPLEXES * PARTICLE MESH EWALD Subject RIV: BO - Biophysics Impact factor: 10.162, year: 2016

  10. Dynamic Simulation of Random Packing of Polydispersive Fine Particles

    Science.gov (United States)

    Ferraz, Carlos Handrey Araujo; Marques, Samuel Apolinário

    2018-02-01

    In this paper, we perform molecular dynamic (MD) simulations to study the two-dimensional packing process of both monosized and random size particles with radii ranging from 1.0 to 7.0 μm. The initial positions as well as the radii of five thousand fine particles were defined inside a rectangular box by using a random number generator. Both the translational and rotational movements of each particle were considered in the simulations. In order to deal with interacting fine particles, we take into account both the contact forces and the long-range dispersive forces. We account for normal and static/sliding tangential friction forces between particles and between particle and wall by means of a linear model approach, while the long-range dispersive forces are computed by using a Lennard-Jones-like potential. The packing processes were studied assuming different long-range interaction strengths. We carry out statistical calculations of the different quantities studied such as packing density, mean coordination number, kinetic energy, and radial distribution function as the system evolves over time. We find that the long-range dispersive forces can strongly influence the packing process dynamics as they might form large particle clusters, depending on the intensity of the long-range interaction strength.

  11. Short term post-operative morphing of sacrocolpopexy mesh measured by magnetic resonance imaging.

    Science.gov (United States)

    Sindhwani, Nikhil; Callewaert, Geertje; Deprest, Thomas; Housmans, Susanne; Van Beckevoort, Dirk; Deprest, Jan

    2018-04-01

    Sacrocolpopexy (SC) involves suspension of the vaginal vault or cervix to the sacrum using a mesh. Following insertion, the meshes have been observed to have undergone dimensional changes. To quantify dimensional changes of meshes following implantation and characterize their morphology in-vivo. 24 patients underwent SC using PolyVinyliDeneFluoride mesh loaded with Fe 3 O 4 particles. Tailored anterior and posterior mesh flaps were sutured to the respective vaginal walls, uniting at the apex. The posterior flap continued to the sacrum and was attached there. Meshes were visualized on magnetic resonance (MR) imaging at 12 [3-12] (median [range]) months postoperatively and 3D models of the mesh were generated. Dynamic MR sequences were acquired during valsalva to record mesh mobility. The area of the vagina effectively supported by the mesh (Effective Support Area (ESA)) was calculated. The 3D models' wall thickness map was analyzed to identify the locations of mesh folding. Intraclass correlation (ICC) was calculated to test the reliability of the methods. To measure the laxity and flatness of the mesh, the curvature and the ellipticity of the sacral flap were calculated. The ESA calculation methodology had ICC = 0.97. A reduction of 75.49 [61.55-78.67] % (median [IQR]) in area, 47.64 [38.07-59.81] % in anterior flap, and of 23.95 [10.96-27.21] % in the posterior flap was measured. The mesh appeared thicker near its attachment at the sacral promontory (n = 19) and near the vaginal apex (n = 22). The laxity of the mesh was 1.13 [1.10-1.16] and 60.55 [49.76-76.25] % of the sacral flap was flat. We could not reliably measure mesh mobility (ICC = 0.16). A methodology for complete 3D characterization of SC meshes using MR images was presented. After implantation, the supported area is much lower than what is prepared prior to implantation. We propose this happened during the surgery itself. Copyright © 2018 Elsevier Ltd. All rights reserved.

  12. Polygonal Prism Mesh in the Viscous Layers for the Polyhedral Mesh Generator, PolyGen

    International Nuclear Information System (INIS)

    Lee, Sang Yong; Park, Chan Eok; Kim, Shin Whan

    2015-01-01

    Polyhedral mesh has been known to have some benefits over the tetrahedral mesh. Efforts have been made to set up a polyhedral mesh generation system with open source programs SALOME and TetGen. The evaluation has shown that the polyhedral mesh generation system is promising. But it is necessary to extend the capability of the system to handle the viscous layers to be a generalized mesh generator. A brief review to the previous works on the mesh generation for the viscous layers will be made in section 2. Several challenging issues for the polygonal prism mesh generation will be discussed as well. The procedure to generate a polygonal prism mesh will be discussed in detail in section 3. Conclusion will be followed in section 4. A procedure to generate meshes in the viscous layers with PolyGen has been successfully designed. But more efforts have to be exercised to find the best way for the generating meshes for viscous layers. Using the extrusion direction of the STL data will the first of the trials in the near future

  13. Two-way coupling of magnetohydrodynamic simulations with embedded particle-in-cell simulations

    Science.gov (United States)

    Makwana, K. D.; Keppens, R.; Lapenta, G.

    2017-12-01

    We describe a method for coupling an embedded domain in a magnetohydrodynamic (MHD) simulation with a particle-in-cell (PIC) method. In this two-way coupling we follow the work of Daldorff et al. (2014) [19] in which the PIC domain receives its initial and boundary conditions from MHD variables (MHD to PIC coupling) while the MHD simulation is updated based on the PIC variables (PIC to MHD coupling). This method can be useful for simulating large plasma systems, where kinetic effects captured by particle-in-cell simulations are localized but affect global dynamics. We describe the numerical implementation of this coupling, its time-stepping algorithm, and its parallelization strategy, emphasizing the novel aspects of it. We test the stability and energy/momentum conservation of this method by simulating a steady-state plasma. We test the dynamics of this coupling by propagating plasma waves through the embedded PIC domain. Coupling with MHD shows satisfactory results for the fast magnetosonic wave, but significant distortion for the circularly polarized Alfvén wave. Coupling with Hall-MHD shows excellent coupling for the whistler wave. We also apply this methodology to simulate a Geospace Environmental Modeling (GEM) challenge type of reconnection with the diffusion region simulated by PIC coupled to larger scales with MHD and Hall-MHD. In both these cases we see the expected signatures of kinetic reconnection in the PIC domain, implying that this method can be used for reconnection studies.

  14. Impact of detector simulation in particle physics collider experiments

    Science.gov (United States)

    Daniel Elvira, V.

    2017-06-01

    Through the last three decades, accurate simulation of the interactions of particles with matter and modeling of detector geometries has proven to be of critical importance to the success of the international high-energy physics (HEP) experimental programs. For example, the detailed detector modeling and accurate physics of the Geant4-based simulation software of the CMS and ATLAS particle physics experiments at the European Center of Nuclear Research (CERN) Large Hadron Collider (LHC) was a determinant factor for these collaborations to deliver physics results of outstanding quality faster than any hadron collider experiment ever before. This review article highlights the impact of detector simulation on particle physics collider experiments. It presents numerous examples of the use of simulation, from detector design and optimization, through software and computing development and testing, to cases where the use of simulation samples made a difference in the precision of the physics results and publication turnaround, from data-taking to submission. It also presents estimates of the cost and economic impact of simulation in the CMS experiment. Future experiments will collect orders of magnitude more data with increasingly complex detectors, taxing heavily the performance of simulation and reconstruction software. Consequently, exploring solutions to speed up simulation and reconstruction software to satisfy the growing demand of computing resources in a time of flat budgets is a matter that deserves immediate attention. The article ends with a short discussion on the potential solutions that are being considered, based on leveraging core count growth in multicore machines, using new generation coprocessors, and re-engineering HEP code for concurrency and parallel computing.

  15. Evaluating the performance of the particle finite element method in parallel architectures

    Science.gov (United States)

    Gimenez, Juan M.; Nigro, Norberto M.; Idelsohn, Sergio R.

    2014-05-01

    This paper presents a high performance implementation for the particle-mesh based method called particle finite element method two (PFEM-2). It consists of a material derivative based formulation of the equations with a hybrid spatial discretization which uses an Eulerian mesh and Lagrangian particles. The main aim of PFEM-2 is to solve transport equations as fast as possible keeping some level of accuracy. The method was found to be competitive with classical Eulerian alternatives for these targets, even in their range of optimal application. To evaluate the goodness of the method with large simulations, it is imperative to use of parallel environments. Parallel strategies for Finite Element Method have been widely studied and many libraries can be used to solve Eulerian stages of PFEM-2. However, Lagrangian stages, such as streamline integration, must be developed considering the parallel strategy selected. The main drawback of PFEM-2 is the large amount of memory needed, which limits its application to large problems with only one computer. Therefore, a distributed-memory implementation is urgently needed. Unlike a shared-memory approach, using domain decomposition the memory is automatically isolated, thus avoiding race conditions; however new issues appear due to data distribution over the processes. Thus, a domain decomposition strategy for both particle and mesh is adopted, which minimizes the communication between processes. Finally, performance analysis running over multicore and multinode architectures are presented. The Courant-Friedrichs-Lewy number used influences the efficiency of the parallelization and, in some cases, a weighted partitioning can be used to improve the speed-up. However the total cputime for cases presented is lower than that obtained when using classical Eulerian strategies.

  16. Novel methods in the Particle-In-Cell accelerator Code-Framework Warp

    Energy Technology Data Exchange (ETDEWEB)

    Vay, J-L [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Grote, D. P. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Cohen, R. H. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Friedman, A. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2012-12-26

    The Particle-In-Cell (PIC) Code-Framework Warp is being developed by the Heavy Ion Fusion Science Virtual National Laboratory (HIFS-VNL) to guide the development of accelerators that can deliver beams suitable for high-energy density experiments and implosion of inertial fusion capsules. It is also applied in various areas outside the Heavy Ion Fusion program to the study and design of existing and next-generation high-energy accelerators, including the study of electron cloud effects and laser wakefield acceleration for example. This study presents an overview of Warp's capabilities, summarizing recent original numerical methods that were developed by the HIFS-VNL (including PIC with adaptive mesh refinement, a large-timestep 'drift-Lorentz' mover for arbitrarily magnetized species, a relativistic Lorentz invariant leapfrog particle pusher, simulations in Lorentz-boosted frames, an electromagnetic solver with tunable numerical dispersion and efficient stride-based digital filtering), with special emphasis on the description of the mesh refinement capability. In addition, selected examples of the applications of the methods to the abovementioned fields are given.

  17. Direct numerical simulation of 3D particle motion in an evaporating liquid film

    International Nuclear Information System (INIS)

    Hwang, Ho Chan; Son, Gi Hun

    2016-01-01

    A direct numerical simulation method is developed for 3D particle motion in liquid film evaporation. The liquid-gas and fluid-solid interfaces are tracked by a sharp-interface Level-set (LS) method, which includes the effects of evaporation, contact line and solid particles. The LS method is validated through simulation of the interaction between two particles falling in a single-phase fluid. The LS based DNS method is applied to computation of the particle motion in liquid film evaporation to investigate the particle-interface and particle-particle interactions

  18. Ef: Software for Nonrelativistic Beam Simulation by Particle-in-Cell Algorithm

    Science.gov (United States)

    Boytsov, A. Yu.; Bulychev, A. A.

    2018-04-01

    Understanding of particle dynamics is crucial in construction of electron guns, ion sources and other types of nonrelativistic beam devices. Apart from external guiding and focusing systems, a prominent role in evolution of such low-energy beams is played by particle-particle interaction. Numerical simulations taking into account these effects are typically accomplished by a well-known particle-in-cell method. In practice, for convenient work a simulation program should not only implement this method, but also support parallelization, provide integration with CAD systems and allow access to details of the simulation algorithm. To address the formulated requirements, development of a new open source code - Ef - has been started. It's current features and main functionality are presented. Comparison with several analytical models demonstrates good agreement between the numerical results and the theory. Further development plans are discussed.

  19. Ef: Software for Nonrelativistic Beam Simulation by Particle-in-Cell Algorithm

    Directory of Open Access Journals (Sweden)

    Boytsov A. Yu.

    2018-01-01

    Full Text Available Understanding of particle dynamics is crucial in construction of electron guns, ion sources and other types of nonrelativistic beam devices. Apart from external guiding and focusing systems, a prominent role in evolution of such low-energy beams is played by particle-particle interaction. Numerical simulations taking into account these effects are typically accomplished by a well-known particle-in-cell method. In practice, for convenient work a simulation program should not only implement this method, but also support parallelization, provide integration with CAD systems and allow access to details of the simulation algorithm. To address the formulated requirements, development of a new open source code - Ef - has been started. It's current features and main functionality are presented. Comparison with several analytical models demonstrates good agreement between the numerical results and the theory. Further development plans are discussed.

  20. Parallel-vector algorithms for particle simulations on shared-memory multiprocessors

    International Nuclear Information System (INIS)

    Nishiura, Daisuke; Sakaguchi, Hide

    2011-01-01

    Over the last few decades, the computational demands of massive particle-based simulations for both scientific and industrial purposes have been continuously increasing. Hence, considerable efforts are being made to develop parallel computing techniques on various platforms. In such simulations, particles freely move within a given space, and so on a distributed-memory system, load balancing, i.e., assigning an equal number of particles to each processor, is not guaranteed. However, shared-memory systems achieve better load balancing for particle models, but suffer from the intrinsic drawback of memory access competition, particularly during (1) paring of contact candidates from among neighboring particles and (2) force summation for each particle. Here, novel algorithms are proposed to overcome these two problems. For the first problem, the key is a pre-conditioning process during which particle labels are sorted by a cell label in the domain to which the particles belong. Then, a list of contact candidates is constructed by pairing the sorted particle labels. For the latter problem, a table comprising the list indexes of the contact candidate pairs is created and used to sum the contact forces acting on each particle for all contacts according to Newton's third law. With just these methods, memory access competition is avoided without additional redundant procedures. The parallel efficiency and compatibility of these two algorithms were evaluated in discrete element method (DEM) simulations on four types of shared-memory parallel computers: a multicore multiprocessor computer, scalar supercomputer, vector supercomputer, and graphics processing unit. The computational efficiency of a DEM code was found to be drastically improved with our algorithms on all but the scalar supercomputer. Thus, the developed parallel algorithms are useful on shared-memory parallel computers with sufficient memory bandwidth.

  1. Solar energetic particles: observational studies and magnetohydrodynamic simulation

    International Nuclear Information System (INIS)

    Masson, S.

    2010-10-01

    Solar activity manifests itself through highly dynamical events, such as flares and coronal mass ejections, which result in energy release by magnetic reconnection. This thesis focuses on two manifestations of this energy release: solar energetic particles and dynamics of magnetic reconnection. The first part of my work consists in the detailed temporal analysis of several electromagnetic signatures, produced by energetic particles in the solar atmosphere, with respect to the energetic particle flux at Earth. Using multi-instrument observations, I highlighted that particles can be accelerated by the flare to relativistic energies during a specific episode of acceleration in the impulsive phase. This showed that particles traveled a longer path length than the theoretical length generally assumed. Using in-situ measurements of magnetic field and plasma, I identified the interplanetary magnetic field for 10 particle events, and performing a velocity dispersion analysis I obtained the interplanetary length traveled by particles. I showed that the magnetic structure of the interplanetary medium play a crucial role in the association of the particle flux at Earth and the acceleration signatures of particles at the Sun. The second part of my work focuses on the dynamics of magnetic reconnection. Observationally, the best evidence for magnetic reconnection is the appearance of brightnesses at the solar surface. Performing the first data-driven 3 dimensional magneto-hydrodynamic (MHD) simulation of an observed event, I discovered that the evolution of brightnesses can be explained by the succession of two different reconnection regimes, induced by a new topological association where null-point separatrix lines are embedded in quasi-separatrix layers. This new topological association induces a change of field line connectivity, but also a continuous reconnection process, leading to an apparent slipping motion of reconnected field lines. From a MHD simulation I showed that

  2. RadSim: a program to simulate individual particle interactions for educational purposes

    International Nuclear Information System (INIS)

    Verhaegen, Frank; Palefsky, Steven; DeBlois, Francois

    2006-01-01

    A program was developed, RadSim, which can be used to simulate certain individual interactions of photons, electrons, positrons and alpha particles with a single atom for educational purposes. The program can be run in two modes: manual and simulated. In the manual mode, an individual particle undergoing a specified interaction with a target atom can be simulated, which essentially comes down to a graphical evaluation of kinematic equations. In the simulated mode, a preset number of identical particles are allowed to undergo a specified interaction type with a target atom. The exit channel of the interaction is sampled from probability distributions using Monte Carlo methods. The incoming and outgoing particles are visualized and the frequency distribution of the kinematic variables of the exit channel is displayed graphically. It has to be emphasized that RadSim was mainly developed for educational purposes. (note)

  3. A Nonlinear Dynamic Model and Free Vibration Analysis of Deployable Mesh Reflectors

    Science.gov (United States)

    Shi, H.; Yang, B.; Thomson, M.; Fang, H.

    2011-01-01

    This paper presents a dynamic model of deployable mesh reflectors, in which geometric and material nonlinearities of such a space structure are fully described. Then, by linearization around an equilibrium configuration of the reflector structure, a linearized model is obtained. With this linearized model, the natural frequencies and mode shapes of a reflector can be computed. The nonlinear dynamic model of deployable mesh reflectors is verified by using commercial finite element software in numerical simulation. As shall be seen, the proposed nonlinear model is useful for shape (surface) control of deployable mesh reflectors under thermal loads.

  4. Implicit particle simulation of electromagnetic plasma phenomena

    International Nuclear Information System (INIS)

    Kamimura, T.; Montalvo, E.; Barnes, D.C.; Leboeuf, J.N.; Tajima, T.

    1986-11-01

    A direct method for the implicit particle simulation of electromagnetic phenomena in magnetized, multi-dimensional plasmas is developed. The method is second-order accurate for ωΔt < 1, with ω a characteristic frequency and time step Δt. Direct time integration of the implicit equations with simplified space differencing allows the consistent inclusion of finite particle size. Decentered time differencing of the Lorentz force permits the efficient simulation of strongly magnetized plasmas. A Fourier-space iterative technique for solving the implicit field corrector equation, based on the separation of plasma responses perpendicular and parallel to the magnetic field and longitudinal and transverse to the wavevector, is described. Wave propagation properties in a uniform plasma are in excellent agreement with theoretical expectations. Applications to collisionless tearing and coalescence instabilities further demonstrate the usefulness of the algorithm. (author)

  5. Verification of Gyrokinetic Particle of Turbulent Simulation of Device Size Scaling Transport

    Institute of Scientific and Technical Information of China (English)

    LIN Zhihong; S. ETHIER; T. S. HAHM; W. M. TANG

    2012-01-01

    Verification and historical perspective are presented on the gyrokinetic particle simulations that discovered the device size scaling of turbulent transport and indentified the geometry model as the source of the long-standing disagreement between gyrokinetic particle and continuum simulations.

  6. Simulation of Particle Fluxes at the DESY-II Test Beam Facility

    International Nuclear Information System (INIS)

    Schuetz, Anne

    2015-05-01

    In the course of this Master's thesis ''Simulation of Particle Fluxes at the DESY-II Test Beam Facility'' the test beam generation for the DESY test beam line was studied in detail and simulated with the simulation software SLIC. SLIC uses the Geant4 toolkit for realistic Monte Carlo simulations of particles passing through detector material.After discussing the physics processes relevant for the test beam generation and the principles of the beam generation itself, the software used is introduced together with a description of the functionality of the Geant4 Monte Carlo simulation. The simulation of the test beam line follows the sequence of the test beam generation. Therefore, it starts with the simulation of the beam bunch of the synchrotron accelerator DESY-II, and proceeds step by step with the single test beam line components. An additional benefit of this thesis is the provision of particle flux and trajectory maps, which make fluxes directly visible by following the particle tracks through the simulated beam line. These maps allow us to see each of the test beam line components, because flux rates and directions change rapidly at these points. They will also guide the decision for placements of future test beam line components and measurement equipment.In the end, the beam energy and its spread, and the beam rate of the final test beam in the test beam area were studied in the simulation, so that the results can be compared to the measured beam parameters. The test beam simulation of this Master's thesis will serve as a key input for future test beam line improvements.

  7. Mesh optimization for microbial fuel cell cathodes constructed around stainless steel mesh current collectors

    KAUST Repository

    Zhang, Fang

    2011-02-01

    Mesh current collectors made of stainless steel (SS) can be integrated into microbial fuel cell (MFC) cathodes constructed of a reactive carbon black and Pt catalyst mixture and a poly(dimethylsiloxane) (PDMS) diffusion layer. It is shown here that the mesh properties of these cathodes can significantly affect performance. Cathodes made from the coarsest mesh (30-mesh) achieved the highest maximum power of 1616 ± 25 mW m-2 (normalized to cathode projected surface area; 47.1 ± 0.7 W m-3 based on liquid volume), while the finest mesh (120-mesh) had the lowest power density (599 ± 57 mW m-2). Electrochemical impedance spectroscopy showed that charge transfer and diffusion resistances decreased with increasing mesh opening size. In MFC tests, the cathode performance was primarily limited by reaction kinetics, and not mass transfer. Oxygen permeability increased with mesh opening size, accounting for the decreased diffusion resistance. At higher current densities, diffusion became a limiting factor, especially for fine mesh with low oxygen transfer coefficients. These results demonstrate the critical nature of the mesh size used for constructing MFC cathodes. © 2010 Elsevier B.V. All rights reserved.

  8. A Numerical Study of Mesh Adaptivity in Multiphase Flows with Non-Newtonian Fluids

    Science.gov (United States)

    Percival, James; Pavlidis, Dimitrios; Xie, Zhihua; Alberini, Federico; Simmons, Mark; Pain, Christopher; Matar, Omar

    2014-11-01

    We present an investigation into the computational efficiency benefits of dynamic mesh adaptivity in the numerical simulation of transient multiphase fluid flow problems involving Non-Newtonian fluids. Such fluids appear in a range of industrial applications, from printing inks to toothpastes and introduce new challenges for mesh adaptivity due to the additional ``memory'' of viscoelastic fluids. Nevertheless, the multiscale nature of these flows implies huge potential benefits for a successful implementation. The study is performed using the open source package Fluidity, which couples an unstructured mesh control volume finite element solver for the multiphase Navier-Stokes equations to a dynamic anisotropic mesh adaptivity algorithm, based on estimated solution interpolation error criteria, and conservative mesh-to-mesh interpolation routine. The code is applied to problems involving rheologies ranging from simple Newtonian to shear-thinning to viscoelastic materials and verified against experimental data for various industrial and microfluidic flows. This work was undertaken as part of the EPSRC MEMPHIS programme grant EP/K003976/1.

  9. Computer Models Simulate Fine Particle Dispersion

    Science.gov (United States)

    2010-01-01

    Through a NASA Seed Fund partnership with DEM Solutions Inc., of Lebanon, New Hampshire, scientists at Kennedy Space Center refined existing software to study the electrostatic phenomena of granular and bulk materials as they apply to planetary surfaces. The software, EDEM, allows users to import particles and obtain accurate representations of their shapes for modeling purposes, such as simulating bulk solids behavior, and was enhanced to be able to more accurately model fine, abrasive, cohesive particles. These new EDEM capabilities can be applied in many industries unrelated to space exploration and have been adopted by several prominent U.S. companies, including John Deere, Pfizer, and Procter & Gamble.

  10. Comparison of GPU-Based Numerous Particles Simulation and Experiment

    International Nuclear Information System (INIS)

    Park, Sang Wook; Jun, Chul Woong; Sohn, Jeong Hyun; Lee, Jae Wook

    2014-01-01

    The dynamic behavior of numerous grains interacting with each other can be easily observed. In this study, this dynamic behavior was analyzed based on the contact between numerous grains. The discrete element method was used for analyzing the dynamic behavior of each particle and the neighboring-cell algorithm was employed for detecting their contact. The Hertzian and tangential sliding friction contact models were used for calculating the contact force acting between the particles. A GPU-based parallel program was developed for conducting the computer simulation and calculating the numerous contacts. The dam break experiment was performed to verify the simulation results. The reliability of the program was verified by comparing the results of the simulation with those of the experiment

  11. Electrokinetic Particle Transport in Micro-Nanofluidics Direct Numerical Simulation Analysis

    CERN Document Server

    Qian, Shizhi

    2012-01-01

    Numerous applications of micro-/nanofluidics are related to particle transport in micro-/nanoscale channels, and electrokinetics has proved to be one of the most promising tools to manipulate particles in micro/nanofluidics. Therefore, a comprehensive understanding of electrokinetic particle transport in micro-/nanoscale channels is crucial to the development of micro/nano-fluidic devices. Electrokinetic Particle Transport in Micro-/Nanofluidics: Direct Numerical Simulation Analysis provides a fundamental understanding of electrokinetic particle transport in micro-/nanofluidics involving elect

  12. Wire-mesh capped deposition sensors: Novel passive tool for coarse fraction flux estimation of radon thoron progeny in indoor environments

    International Nuclear Information System (INIS)

    Mayya, Y.S.; Mishra, Rosaline; Prajith, Rama; Sapra, B.K.; Kushwaha, H.S.

    2010-01-01

    Deposition-based 222 Rn and 220 Rn progeny sensors act as unique, passive tools for determining the long time-averaged progeny deposition fluxes in the environment. The use of these deposition sensors as progeny concentration monitors was demonstrated in typical indoor environments as conceptually superior alternatives to gas-based indirect monitoring methods. In the present work, the dependency of these deposition monitors on various environmental parameters is minimized by capping the deposition sensor with a suitable wire mesh. These wire-mesh capped deposition sensors measure the coarse fraction deposition flux, which is less dependent on the change in environmental parameters like ventilation rate and turbulence. The calibration of these wire-mesh capped coarse fraction progeny sensors was carried out by laboratory controlled experiments. These sensors were deployed both in indoor and in occupational environments having widely different ventilation rates. The obtained coarse fraction deposition velocities were fairly constant in these environments, which further confirmed that the signal on the wire-mesh capped sensors show the least dependency on the change in environmental parameters. This technique has the potential to serve as a passive particle sizer in the general context of nanoparticles using progeny species as surrogates. On the whole, there exists a strong case for developing a passive system that responds only to coarse fraction for providing alternative tools for dosimetry and environmental fine particle research. - Research highlights: → Wire-mesh capped deposition sensor measures the coarse fraction deposition flux → Coarse fraction deposition flux less dependent on environmental conditions → Wire-mesh capped deposition sensor as passive particle sizer

  13. Thermal Fluctuations in Smooth Dissipative Particle Dynamics simulation of mesoscopic thermal systems

    Science.gov (United States)

    Gatsonis, Nikolaos; Yang, Jun

    2013-11-01

    The SDPD-DV is implemented in our work for arbitrary 3D wall bounded geometries. The particle position and momentum equations are integrated with a velocity-Verlet algorithm and the entropy equation is integrated with a Runge-Kutta algorithm. Simulations of nitrogen gas are performed to evaluate the effects of timestep and particle scale on temperature, self-diffusion coefficient and shear viscosity. The hydrodynamic fluctuations in temperature, density, pressure and velocity from the SDPD-DV simulations are evaluated and compared with theoretical predictions. Steady planar thermal Couette flows are simulated and compared with analytical solutions. Simulations cover the hydrodynamic and mesocopic regime and show thermal fluctuations and their dependence on particle size.

  14. A cellular automaton - finite volume method for the simulation of dendritic and eutectic growth in binary alloys using an adaptive mesh refinement

    Science.gov (United States)

    Dobravec, Tadej; Mavrič, Boštjan; Šarler, Božidar

    2017-11-01

    A two-dimensional model to simulate the dendritic and eutectic growth in binary alloys is developed. A cellular automaton method is adopted to track the movement of the solid-liquid interface. The diffusion equation is solved in the solid and liquid phases by using an explicit finite volume method. The computational domain is divided into square cells that can be hierarchically refined or coarsened using an adaptive mesh based on the quadtree algorithm. Such a mesh refines the regions of the domain near the solid-liquid interface, where the highest concentration gradients are observed. In the regions where the lowest concentration gradients are observed the cells are coarsened. The originality of the work is in the novel, adaptive approach to the efficient and accurate solution of the posed multiscale problem. The model is verified and assessed by comparison with the analytical results of the Lipton-Glicksman-Kurz model for the steady growth of a dendrite tip and the Jackson-Hunt model for regular eutectic growth. Several examples of typical microstructures are simulated and the features of the method as well as further developments are discussed.

  15. Classical trajectory Monte Carlo simulations of particle confinement using dual levitated coils

    Directory of Open Access Journals (Sweden)

    R. A. Lane

    2014-07-01

    Full Text Available The particle confinement properties of plasma confinement systems that employ dual levitated magnetic coils are investigated using classical trajectory Monte Carlo simulations. Two model systems are examined. In one, two identical current-carrying loops are coaxial and separated axially. In the second, two concentric and coplanar loops have different radii and carry equal currents. In both systems, a magnetic null circle is present between the current loops. Simulations are carried out for seven current loop separations for each system and at numerous values of magnetic field strength. Particle confinement is investigated at three locations between the loops at different distances from the magnetic null circle. Each simulated particle that did not escape the system exhibited one of four modes of confinement. Reduced results are given for both systems as the lowest magnetic field strength that exhibits complete confinement of all simulated particles for a particular loop separation.

  16. Analytical technical of lightning surges induced on grounding mesh of PWR nuclear power plant

    International Nuclear Information System (INIS)

    Ikeda, I.; Tani, M.; Yonezawa, T.

    1990-01-01

    An analytical lightning surge technique is needed to make a qualitative and predictive evaluation of transient voltages induced on local grounding meshes and instrumentation cables by a lightning strike on a lightning rod in a PWR plant. This paper discusses an experiment with lightning surge impulses in a PWR plant which was setup to observe lightning caused transient voltages. Experimental data when compared with EMTP simulation results improved the simulation method. The improved method provides a good estimation of induced voltages on grounding meshes and instrumentation cables

  17. Modelling of pedestrian level wind environment on a high-quality mesh: A case study for the HKPolyU campus

    DEFF Research Database (Denmark)

    Du, Yaxing; Mak, Cheuk Ming; Ai, Zhengtao

    2018-01-01

    Quality and efficiency of computational fluid dynamics (CFD) simulation of pedestrian level wind environment in a complex urban area are often compromised by many influencing factors, particularly mesh quality. This paper first proposes a systematic and efficient mesh generation method and then p......Quality and efficiency of computational fluid dynamics (CFD) simulation of pedestrian level wind environment in a complex urban area are often compromised by many influencing factors, particularly mesh quality. This paper first proposes a systematic and efficient mesh generation method...... and then performs detailed sensitivity analysis of some important computational parameters. The geometrically complex Hong Kong Polytechnic University (HKPolyU) campus is taken as a case study. Based on the high-quality mesh system, the influences of three important computational parameters, namely, turbulence...... model, near-wall mesh density and computational domain size, on the CFD predicted results of pedestrian level wind environment are quantitatively evaluated. Validation of CFD models is conducted against wind tunnel experimental data, where a good agreement is achieved. It is found that the proposed mesh...

  18. Parallel treatment of simulation particles in particle-in-cell codes on SUPRENUM

    International Nuclear Information System (INIS)

    Seldner, D.

    1990-02-01

    This report contains the program documentation and description of the program package 2D-PLAS, which has been developed at the Nuclear Research Center Karlsruhe in the Institute for Data Processing in Technology (IDT) under the auspices of the BMFT. 2D-PLAS is a parallel program version of the treatment of the simulation particles of the two-dimensional stationary particle-in-cell code BFCPIC which has been developed at the Nuclear Research Center Karlsruhe. This parallel version has been designed for the parallel computer SUPRENUM. (orig.) [de

  19. Constraining Diameters of Ash Particles in Io's Pele Plume by DSMC Simulation

    Science.gov (United States)

    McDoniel, William; Goldstein, D. B.; Varghese, P. L.; Trafton, L. M.

    2013-10-01

    The black “butterfly wings” seen at Pele are produced by silicate ash which is to some extent entrained in the gas flow from very low altitudes. These particles are key to understanding the volcanism at Pele. However, the Pele plume is not nearly as dusty as Prometheus, and these are not the only particles in the plume, as the SO2 in the plume will also condense as it cools. It is therefore difficult to estimate a size distribution for the ash particles by observation, and the drag on ash particles from the plume flow is significant enough that ballistic models are also of limited use. Using Direct Simulation Monte Carlo, we can simulate a gas plume at Pele which demonstrates very good agreement with observations. By extending this model down to nearly the surface of the lava lake, ash particles can be included in the simulation by assuming that they are initially entrained in the very dense (for Io) gas immediately above the magma. Particles are seen to fall to the ground to the east and west of the vent, agreeing with the orientation of the “butterfly wings”, and particles with larger diameters fall to the ground closer to the lava lake. We present a model for mapping simulated deposition density to the coloration of the surface and we use it to estimate the size distribution of ash particles in the plume.

  20. Assessment of fusion facility dose rate map using mesh adaptivity enhancements of hybrid Monte Carlo/deterministic techniques

    International Nuclear Information System (INIS)

    Ibrahim, Ahmad M.; Wilson, Paul P.; Sawan, Mohamed E.; Mosher, Scott W.; Peplow, Douglas E.; Grove, Robert E.

    2014-01-01

    Highlights: •Calculate the prompt dose rate everywhere throughout the entire fusion energy facility. •Utilize FW-CADIS to accurately perform difficult neutronics calculations for fusion energy systems. •Develop three mesh adaptivity algorithms to enhance FW-CADIS efficiency in fusion-neutronics calculations. -- Abstract: Three mesh adaptivity algorithms were developed to facilitate and expedite the use of the CADIS and FW-CADIS hybrid Monte Carlo/deterministic techniques in accurate full-scale neutronics simulations of fusion energy systems with immense sizes and complicated geometries. First, a macromaterial approach enhances the fidelity of the deterministic models without changing the mesh. Second, a deterministic mesh refinement algorithm generates meshes that capture as much geometric detail as possible without exceeding a specified maximum number of mesh elements. Finally, a weight window coarsening algorithm decouples the weight window mesh and energy bins from the mesh and energy group structure of the deterministic calculations in order to remove the memory constraint of the weight window map from the deterministic mesh resolution. The three algorithms were used to enhance an FW-CADIS calculation of the prompt dose rate throughout the ITER experimental facility and resulted in a 23.3% increase in the number of mesh tally elements in which the dose rates were calculated in a 10-day Monte Carlo calculation. Additionally, because of the significant increase in the efficiency of FW-CADIS simulations, the three algorithms enabled this difficult calculation to be accurately solved on a regular computer cluster, eliminating the need for a world-class super computer

  1. MeSH Now: automatic MeSH indexing at PubMed scale via learning to rank.

    Science.gov (United States)

    Mao, Yuqing; Lu, Zhiyong

    2017-04-17

    MeSH indexing is the task of assigning relevant MeSH terms based on a manual reading of scholarly publications by human indexers. The task is highly important for improving literature retrieval and many other scientific investigations in biomedical research. Unfortunately, given its manual nature, the process of MeSH indexing is both time-consuming (new articles are not immediately indexed until 2 or 3 months later) and costly (approximately ten dollars per article). In response, automatic indexing by computers has been previously proposed and attempted but remains challenging. In order to advance the state of the art in automatic MeSH indexing, a community-wide shared task called BioASQ was recently organized. We propose MeSH Now, an integrated approach that first uses multiple strategies to generate a combined list of candidate MeSH terms for a target article. Through a novel learning-to-rank framework, MeSH Now then ranks the list of candidate terms based on their relevance to the target article. Finally, MeSH Now selects the highest-ranked MeSH terms via a post-processing module. We assessed MeSH Now on two separate benchmarking datasets using traditional precision, recall and F 1 -score metrics. In both evaluations, MeSH Now consistently achieved over 0.60 in F-score, ranging from 0.610 to 0.612. Furthermore, additional experiments show that MeSH Now can be optimized by parallel computing in order to process MEDLINE documents on a large scale. We conclude that MeSH Now is a robust approach with state-of-the-art performance for automatic MeSH indexing and that MeSH Now is capable of processing PubMed scale documents within a reasonable time frame. http://www.ncbi.nlm.nih.gov/CBBresearch/Lu/Demo/MeSHNow/ .

  2. Multi-grid Particle-in-cell Simulations of Plasma Microturbulence

    International Nuclear Information System (INIS)

    Lewandowski, J.L.V.

    2003-01-01

    A new scheme to accurately retain kinetic electron effects in particle-in-cell (PIC) simulations for the case of electrostatic drift waves is presented. The splitting scheme, which is based on exact separation between adiabatic and on adiabatic electron responses, is shown to yield more accurate linear growth rates than the standard df scheme. The linear and nonlinear elliptic problems that arise in the splitting scheme are solved using a multi-grid solver. The multi-grid particle-in-cell approach offers an attractive path, both from the physics and numerical points of view, to simulate kinetic electron dynamics in global toroidal plasmas

  3. Particle beam dynamics simulations using the POOMA framework

    International Nuclear Information System (INIS)

    Humphrey, W.; Ryne, R.; Cleland, T.; Cummings, J.; Habib, S.; Mark, G.; Ji Qiang

    1998-01-01

    A program for simulation of the dynamics of high intensity charged particle beams in linear particle accelerators has been developed in C++ using the POOMA Framework, for use on serial and parallel architectures. The code models the trajectories of charged particles through a sequence of different accelerator beamline elements such as drift chambers, quadrupole magnets, or RF cavities. An FFT-based particle-in-cell algorithm is used to solve the Poisson equation that models the Coulomb interactions of the particles. The code employs an object-oriented design with software abstractions for the particle beam, accelerator beamline, and beamline elements, using C++ templates to efficiently support both 2D and 3D capabilities in the same code base. The POOMA Framework, which encapsulates much of the effort required for parallel execution, provides particle and field classes, particle-field interaction capabilities, and parallel FFT algorithms. The performance of this application running serially and in parallel is compared to an existing HPF implementation, with the POOMA version seen to run four times faster than the HPF code

  4. Memory-efficient optimization of Gyrokinetic particle-to-grid interpolation for multicore processors

    Energy Technology Data Exchange (ETDEWEB)

    Madduri, Kamesh [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Williams, Samuel [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Ethier, Stephane [Princeton Plasma Physics Lab. (PPPL), Princeton, NJ (United States); Oliker, Leonid [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Shalf, John [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Strohmaier, Erich [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Yelicky, Katherine [Univ. of California, Berkeley, CA (United States)

    2009-01-01

    We present multicore parallelization strategies for the particle-to-grid interpolation step in the Gyrokinetic Toroidal Code (GTC), a 3D particle-in-cell (PIC) application to study turbulent transport in magnetic-confinement fusion devices. Particle-grid interpolation is a known performance bottleneck in several PIC applications. In GTC, this step involves particles depositing charges to a 3D toroidal mesh, and multiple particles may contribute to the charge at a grid point. We design new parallel algorithms for the GTC charge deposition kernel, and analyze their performance on three leading multicore platforms. We implement thirteen different variants for this kernel and identify the best-performing ones given typical PIC parameters such as the grid size, number of particles per cell, and the GTC-specific particle Larmor radius variation. We find that our best strategies can be 2x faster than the reference optimized MPI implementation, and our analysis provides insight into desirable architectural features for high-performance PIC simulation codes.

  5. Simulation study of effects of initial particle size distribution on dissolution

    International Nuclear Information System (INIS)

    Wang, G.; Xu, D.S.; Ma, N.; Zhou, N.; Payton, E.J.; Yang, R.; Mills, M.J.; Wang, Y.

    2009-01-01

    Dissolution kinetics of γ' particles in binary Ni-Al alloys with different initial particle size distributions (PSD) is studied using a three-dimensional (3D) quantitative phase field model. By linking model inputs directly to thermodynamic and atomic mobility databases, microstructural evolution during dissolution is simulated in real time and length scales. The model is first validated against analytical solution for dissolution of a single γ' particle in 1D and numerical solution in 3D before it is applied to investigate the effects of initial PSD on dissolution kinetics. Four different types of PSD, uniform, normal, log-normal and bimodal, are considered. The simulation results show that the volume fraction of γ' particles decreases exponentially with time, while the temporal evolution of average particle size depends strongly on the initial PSD

  6. Self-ordering of small-diameter metal nanoparticles by dewetting on hexagonal mesh templates.

    Science.gov (United States)

    Meshot, Eric R; Zhao, Zhouzhou; Lu, Wei; Hart, A John

    2014-09-07

    Arrays of small-diameter nanoparticles with high spatial order are useful for chemical and biological sensors, data storage, synthesis of nanowires and nanotubes, and many other applications. We show that self-ordered metal nanoparticle arrays can be formed by dewetting of thin films on hexagonal mesh substrates made of anodic aluminum oxide (AAO). Upon heating, the metal (Fe) film dewets onto the interstitial sites (i.e., the node points) between pores on the top surface of the AAO. We investigated the particle morphology and dynamics of dewetting using a combination of atomic force microscopy (AFM), grazing-incidence small-angle X-ray scattering (GISAXS), and numerical simulations. Templated metal particles are more monodisperse and have higher local order than those formed by the same dewetting process on flat, nonporous alumina. The degree of order depends on the initial film thickness, and for the optimal thickness tested (nominally 2 nm), we achieved uniform coverage and high order of the particles, comparable to that of the AAO template itself. Computational modeling of dewetting on templates with various pore order and size shows that the order of AAO pores is primarily influential in determining particle position and spacing, while the variance in pore size is less impactful. Potential uses of these ordered nanoparticle arrays on porous materials include plasmonic sensors and spatially controlled catalysts.

  7. Verification of Eulerian-Eulerian and Eulerian-Lagrangian simulations for fluid-particle flows

    Science.gov (United States)

    Kong, Bo; Patel, Ravi G.; Capecelatro, Jesse; Desjardins, Olivier; Fox, Rodney O.

    2017-11-01

    In this work, we study the performance of three simulation techniques for fluid-particle flows: (1) a volume-filtered Euler-Lagrange approach (EL), (2) a quadrature-based moment method using the anisotropic Gaussian closure (AG), and (3) a traditional two-fluid model. By simulating two problems: particles in frozen homogeneous isotropic turbulence (HIT), and cluster-induced turbulence (CIT), the convergence of the methods under grid refinement is found to depend on the simulation method and the specific problem, with CIT simulations facing fewer difficulties than HIT. Although EL converges under refinement for both HIT and CIT, its statistical results exhibit dependence on the techniques used to extract statistics for the particle phase. For HIT, converging both EE methods (TFM and AG) poses challenges, while for CIT, AG and EL produce similar results. Overall, all three methods face challenges when trying to extract converged, parameter-independent statistics due to the presence of shocks in the particle phase. National Science Foundation and National Energy Technology Laboratory.

  8. Evaluating simulant materials for understanding cranial backspatter from a ballistic projectile.

    Science.gov (United States)

    Das, Raj; Collins, Alistair; Verma, Anurag; Fernandez, Justin; Taylor, Michael

    2015-05-01

    In cranial wounds resulting from a gunshot, the study of backspatter patterns can provide information about the actual incidents by linking material to surrounding objects. This study investigates the physics of backspatter from a high-speed projectile impact and evaluates a range of simulant materials using impact tests. Next, we evaluate a mesh-free method called smoothed particle hydrodynamics (SPH) to model the splashing mechanism during backspatter. The study has shown that a projectile impact causes fragmentation at the impact site, while transferring momentum to fragmented particles. The particles travel along the path of least resistance, leading to partial material movement in the reverse direction of the projectile motion causing backspatter. Medium-density fiberboard is a better simulant for a human skull than polycarbonate, and lorica leather is a better simulant for a human skin than natural rubber. SPH is an effective numerical method for modeling the high-speed impact fracture and fragmentations. © 2015 American Academy of Forensic Sciences.

  9. Monte Carlo simulation of particle-induced bit upsets

    Science.gov (United States)

    Wrobel, Frédéric; Touboul, Antoine; Vaillé, Jean-Roch; Boch, Jérôme; Saigné, Frédéric

    2017-09-01

    We investigate the issue of radiation-induced failures in electronic devices by developing a Monte Carlo tool called MC-Oracle. It is able to transport the particles in device, to calculate the energy deposited in the sensitive region of the device and to calculate the transient current induced by the primary particle and the secondary particles produced during nuclear reactions. We compare our simulation results with SRAM experiments irradiated with neutrons, protons and ions. The agreement is very good and shows that it is possible to predict the soft error rate (SER) for a given device in a given environment.

  10. Monte Carlo simulation of particle-induced bit upsets

    Directory of Open Access Journals (Sweden)

    Wrobel Frédéric

    2017-01-01

    Full Text Available We investigate the issue of radiation-induced failures in electronic devices by developing a Monte Carlo tool called MC-Oracle. It is able to transport the particles in device, to calculate the energy deposited in the sensitive region of the device and to calculate the transient current induced by the primary particle and the secondary particles produced during nuclear reactions. We compare our simulation results with SRAM experiments irradiated with neutrons, protons and ions. The agreement is very good and shows that it is possible to predict the soft error rate (SER for a given device in a given environment.

  11. Mesh-morphing algorithms for specimen-specific finite element modeling.

    Science.gov (United States)

    Sigal, Ian A; Hardisty, Michael R; Whyne, Cari M

    2008-01-01

    Despite recent advances in software for meshing specimen-specific geometries, considerable effort is still often required to produce and analyze specimen-specific models suitable for biomechanical analysis through finite element modeling. We hypothesize that it is possible to obtain accurate models by adapting a pre-existing geometry to represent a target specimen using morphing techniques. Here we present two algorithms for morphing, automated wrapping (AW) and manual landmarks (ML), and demonstrate their use to prepare specimen-specific models of caudal rat vertebrae. We evaluate the algorithms by measuring the distance between target and morphed geometries and by comparing response to axial loading simulated with finite element (FE) methods. First a traditional reconstruction process based on microCT was used to obtain two natural specimen-specific FE models. Next, the two morphing algorithms were used to compute mappings from the surface of one model, the source, to the other, the target, and to use this mapping to morph the source mesh to produce a target mesh. The microCT images were then used to assign element-specific material properties. In AW the mappings were obtained by wrapping the source and target surfaces with an auxiliary triangulated surface. In ML, landmarks were manually placed on corresponding locations on the surfaces of both source and target. Both morphing algorithms were successful in reproducing the shape of the target vertebra with a median distance between natural and morphed models of 18.8 and 32.2 microm, respectively, for AW and ML. Whereas AW-morphing produced a surface more closely resembling that of the target, ML guaranteed correspondence of the landmark locations between source and target. Morphing preserved the quality of the mesh producing models suitable for FE simulation. Moreover, there were only minor differences between natural and morphed models in predictions of deformation, strain and stress. We therefore conclude that

  12. Natural tracer test simulation by stochastic particle tracking method

    International Nuclear Information System (INIS)

    Ackerer, P.; Mose, R.; Semra, K.

    1990-01-01

    Stochastic particle tracking methods are well adapted to 3D transport simulations where discretization requirements of other methods usually cannot be satisfied. They do need a very accurate approximation of the velocity field. The described code is based on the mixed hybrid finite element method (MHFEM) to calculated the piezometric and velocity field. The random-walk method is used to simulate mass transport. The main advantages of the MHFEM over FD or FE are the simultaneous calculation of pressure and velocity, which are considered as unknowns; the possibility of interpolating velocities everywhere; and the continuity of the normal component of the velocity vector from one element to another. For these reasons, the MHFEM is well adapted for particle tracking methods. After a general description of the numerical methods, the model is used to simulate the observations made during the Twin Lake Tracer Test in 1983. A good match is found between observed and simulated heads and concentrations. (Author) (12 refs., 4 figs.)

  13. 3D finite element modelling of force transmission and particle fracture of sand

    Energy Technology Data Exchange (ETDEWEB)

    Imseeh, Wadi H.; Alshibli, Khalid A. (Tennessee-K)

    2018-02-01

    Global compressive loading of granular media causes rearrangements of particles into a denser configuration. Under 1D compression, researchers observed that particles initially translate and rotate which lead to more contacts between particles and the development of force chains to resist applied loads. Particles within force chains resist most of the applied loads while neighbor particles provide lateral support to prevent particles within force chains from buckling. Several experimental and numerical models have been proposed in the literature to characterize force chains within granular materials. This paper presents a 3D finite element (FE) model that simulates 1D compression experiment on F-75 Ottawa sand. The FE mesh of particles closely matched 3D physical shape of sand particles that were acquired using 3D synchrotron micro-computed tomography (SMT) technique. The paper presents a quantitative assessment of the model, in which evolution of force chains, fracture modes, and stress-strain relationships showed an excellent agreement with experimental measurements reported by Cil et al. Alshibli (2017).

  14. Particle-in-cell Simulations with Kinetic Electrons

    International Nuclear Information System (INIS)

    Lewandowski, J.L.V.

    2004-01-01

    A new scheme, based on an exact separation between adiabatic and nonadiabatic electron responses, for particle-in-cell (PIC) simulations of drift-type modes is presented. The (linear and nonlinear) elliptic equations for the scalar fields are solved using a multi-grid solver. The new scheme yields linear growth rates in excellent agreement with theory and it is shown to conserve energy well into the nonlinear regime. It is also demonstrated that simulations with few electrons are reliable and accurate, suggesting that large-scale, PIC simulations with electron dynamics in toroidal geometry (e.g., tokamaks and stellarators plasmas) are within reach of present-day massively parallel supercomputers

  15. Laparoscopic appendicectomy for suspected mesh-induced appendicitis after laparoscopic transabdominal preperitoneal polypropylene mesh inguinal herniorraphy

    Directory of Open Access Journals (Sweden)

    Jennings Jason

    2010-01-01

    Full Text Available Laparoscopic inguinal herniorraphy via a transabdominal preperitoneal (TAPP approach using Polypropylene Mesh (Mesh and staples is an accepted technique. Mesh induces a localised inflammatory response that may extend to, and involve, adjacent abdominal and pelvic viscera such as the appendix. We present an interesting case of suspected Mesh-induced appendicitis treated successfully with laparoscopic appendicectomy, without Mesh removal, in an elderly gentleman who presented with symptoms and signs of acute appendicitis 18 months after laparoscopic inguinal hernia repair. Possible mechanisms for Mesh-induced appendicitis are briefly discussed.

  16. Gyrokinetic Electron and Fully Kinetic Ion Particle Simulation of Collisionless Plasma Dynamics

    Energy Technology Data Exchange (ETDEWEB)

    Yu Lin; Xueyi Wang; Liu Chen; Zhihong Lin

    2009-08-11

    Fully kinetic-particle simulations and hybrid simulations have been utilized for decades to investigate various fundamental plasma processes, such as magnetic reconnection, fast compressional waves, and wave-particle interaction. Nevertheless, due to disparate temporal and spatial scales between electrons and ions, existing fully kinetic-particle codes have to employ either unrealistically high electron-to-ion mass ratio, me/mi, or simulation domain limited to a few or a few ten's of the ion Larmor radii, or/and time much less than the global Alfven time scale in order to accommodate available computing resources. On the other hand, in the hybrid simulation, the ions are treated as fully kinetic particles but the electrons are treated as a massless fluid. The electron kinetic effects, e.g., wave-particle resonances and finite electron Larmor radius effects, are completely missing. Important physics, such as the electron transit time damping of fast compressional waves or the triggering mechanism of magnetic reconnection in collisionless plasmas is absent in the hybrid codes. Motivated by these considerations and noting that dynamics of interest to us has frequencies lower than the electron gyrofrequency, we planned to develop an innovative particle simulation model, gyrokinetic (GK) electrons and fully kinetic (FK) ions. In the GK-electron and FK-ion (GKe/FKi) particle simulation model, the rapid electron cyclotron motion is removed, while keeping finite electron Larmor radii, realistic me/mi ratio, wave-particle interactions, and off-diagonal components of electron pressure tensor. The computation power can thus be significantly improved over that of the full-particle codes. As planned in the project DE-FG02-05ER54826, we have finished the development of the new GK-electron and FK-ion scheme, finished its benchmark for a uniform plasma in 1-D, 2-D, and 3-D systems against linear waves obtained from analytical theories, and carried out a further convergence

  17. Gyrokinetic Electron and Fully Kinetic Ion Particle Simulation of Collisionless Plasma Dynamics

    International Nuclear Information System (INIS)

    Lin, Yu; Wang, Xueyi; Chen, Liu; Lin, Zhihong

    2009-01-01

    Fully kinetic-particle simulations and hybrid simulations have been utilized for decades to investigate various fundamental plasma processes, such as magnetic reconnection, fast compressional waves, and wave-particle interaction. Nevertheless, due to disparate temporal and spatial scales between electrons and ions, existing fully kinetic-particle codes have to employ either unrealistically high electron-to-ion mass ratio, me/mi, or simulation domain limited to a few or a few ten's of the ion Larmor radii, or/and time much less than the global Alfven time scale in order to accommodate available computing resources. On the other hand, in the hybrid simulation, the ions are treated as fully kinetic particles but the electrons are treated as a massless fluid. The electron kinetic effects, e.g., wave-particle resonances and finite electron Larmor radius effects, are completely missing. Important physics, such as the electron transit time damping of fast compressional waves or the triggering mechanism of magnetic reconnection in collisionless plasmas is absent in the hybrid codes. Motivated by these considerations and noting that dynamics of interest to us has frequencies lower than the electron gyrofrequency, we planned to develop an innovative particle simulation model, gyrokinetic (GK) electrons and fully kinetic (FK) ions. In the GK-electron and FK-ion (GKe/FKi) particle simulation model, the rapid electron cyclotron motion is removed, while keeping finite electron Larmor radii, realistic me/mi ratio, wave-particle interactions, and off-diagonal components of electron pressure tensor. The computation power can thus be significantly improved over that of the full-particle codes. As planned in the project DE-FG02-05ER54826, we have finished the development of the new GK-electron and FK-ion scheme, finished its benchmark for a uniform plasma in 1-D, 2-D, and 3-D systems against linear waves obtained from analytical theories, and carried out a further convergence test

  18. Explicit simulation of ice particle habits in a Numerical Weather Prediction Model

    Science.gov (United States)

    Hashino, Tempei

    2007-05-01

    This study developed a scheme for explicit simulation of ice particle habits in Numerical Weather Prediction (NWP) Models. The scheme is called Spectral Ice Habit Prediction System (SHIPS), and the goal is to retain growth history of ice particles in the Eulerian dynamics framework. It diagnoses characteristics of ice particles based on a series of particle property variables (PPVs) that reflect history of microphysieal processes and the transport between mass bins and air parcels in space. Therefore, categorization of ice particles typically used in bulk microphysical parameterization and traditional bin models is not necessary, so that errors that stem from the categorization can be avoided. SHIPS predicts polycrystals as well as hexagonal monocrystals based on empirically derived habit frequency and growth rate, and simulates the habit-dependent aggregation and riming processes by use of the stochastic collection equation with predicted PPVs. Idealized two dimensional simulations were performed with SHIPS in a NWP model. The predicted spatial distribution of ice particle habits and types, and evolution of particle size distributions showed good quantitative agreement with observation This comprehensive model of ice particle properties, distributions, and evolution in clouds can be used to better understand problems facing wide range of research disciplines, including microphysics processes, radiative transfer in a cloudy atmosphere, data assimilation, and weather modification.

  19. The Microwave Properties of Simulated Melting Precipitation Particles: Sensitivity to Initial Melting

    Science.gov (United States)

    Johnson, B. T.; Olson, W. S.; Skofronick-Jackson, G.

    2016-01-01

    A simplified approach is presented for assessing the microwave response to the initial melting of realistically shaped ice particles. This paper is divided into two parts: (1) a description of the Single Particle Melting Model (SPMM), a heuristic melting simulation for ice-phase precipitation particles of any shape or size (SPMM is applied to two simulated aggregate snow particles, simulating melting up to 0.15 melt fraction by mass), and (2) the computation of the single-particle microwave scattering and extinction properties of these hydrometeors, using the discrete dipole approximation (via DDSCAT), at the following selected frequencies: 13.4, 35.6, and 94.0GHz for radar applications and 89, 165.0, and 183.31GHz for radiometer applications. These selected frequencies are consistent with current microwave remote-sensing platforms, such as CloudSat and the Global Precipitation Measurement (GPM) mission. Comparisons with calculations using variable-density spheres indicate significant deviations in scattering and extinction properties throughout the initial range of melting (liquid volume fractions less than 0.15). Integration of the single-particle properties over an exponential particle size distribution provides additional insight into idealized radar reflectivity and passive microwave brightness temperature sensitivity to variations in size/mass, shape, melt fraction, and particle orientation.

  20. STUDY ON SIMULATION METHOD OF AVALANCHE : FLOW ANALYSIS OF AVALANCHE USING PARTICLE METHOD

    OpenAIRE

    塩澤, 孝哉

    2015-01-01

    In this paper, modeling for the simulation of the avalanche by a particle method is discussed. There are two kinds of the snow avalanches, one is the surface avalanche which shows a smoke-like flow, and another is the total-layer avalanche which shows a flow like Bingham fluid. In the simulation of the surface avalanche, the particle method in consideration of a rotation resistance model is used. The particle method by Bingham fluid is used in the simulation of the total-layer avalanche. At t...

  1. Boltzmann Solver with Adaptive Mesh in Velocity Space

    International Nuclear Information System (INIS)

    Kolobov, Vladimir I.; Arslanbekov, Robert R.; Frolova, Anna A.

    2011-01-01

    We describe the implementation of direct Boltzmann solver with Adaptive Mesh in Velocity Space (AMVS) using quad/octree data structure. The benefits of the AMVS technique are demonstrated for the charged particle transport in weakly ionized plasmas where the collision integral is linear. We also describe the implementation of AMVS for the nonlinear Boltzmann collision integral. Test computations demonstrate both advantages and deficiencies of the current method for calculations of narrow-kernel distributions.

  2. Persistent pelvic pain following transvaginal mesh surgery: a cause for mesh removal.

    Science.gov (United States)

    Marcus-Braun, Naama; Bourret, Antoine; von Theobald, Peter

    2012-06-01

    Persistent pelvic pain after vaginal mesh surgery is an uncommon but serious complication that greatly affects women's quality of life. Our aim was to evaluate various procedures for mesh removal performed at a tertiary referral center in cases of persistent pelvic pain, and to evaluate the ensuing complications and outcomes. A retrospective study was conducted at the University Hospital of Caen, France, including all patients treated for removal or section of vaginal mesh due to pelvic pain as a primary cause, between January 2004 and September 2009. Ten patients met the inclusion criteria. Patients were diagnosed between 10 months and 3 years after their primary operation. Eight cases followed suburethral sling procedures and two followed mesh surgery for pelvic organ prolapse. Patients presented with obturator neuralgia (6), pudendal neuralgia (2), dyspareunia (1), and non-specific pain (1). The surgical treatment to release the mesh included: three cases of extra-peritoneal laparoscopy, four cases of complete vaginal mesh removal, one case of partial mesh removal and two cases of section of the suburethral sling. In all patients with obturator neuralgia, symptoms were resolved or improved, whereas in both cases of pudendal neuralgia the symptoms continued. There were no intra-operative complications. Post-operative Retzius hematoma was observed in one patient after laparoscopy. Mesh removal in a tertiary center is a safe procedure, necessary in some cases of persistent pelvic pain. Obturator neuralgia seems to be easier to treat than pudendal neuralgia. Early diagnosis is the key to success in prevention of chronic disease. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.

  3. PAM: Particle automata model in simulation of Fusarium graminearum pathogen expansion.

    Science.gov (United States)

    Wcisło, Rafał; Miller, S Shea; Dzwinel, Witold

    2016-01-21

    The multi-scale nature and inherent complexity of biological systems are a great challenge for computer modeling and classical modeling paradigms. We present a novel particle automata modeling metaphor in the context of developing a 3D model of Fusarium graminearum infection in wheat. The system consisting of the host plant and Fusarium pathogen cells can be represented by an ensemble of discrete particles defined by a set of attributes. The cells-particles can interact with each other mimicking mechanical resistance of the cell walls and cell coalescence. The particles can move, while some of their attributes can be changed according to prescribed rules. The rules can represent cellular scales of a complex system, while the integrated particle automata model (PAM) simulates its overall multi-scale behavior. We show that due to the ability of mimicking mechanical interactions of Fusarium tip cells with the host tissue, the model is able to simulate realistic penetration properties of the colonization process reproducing both vertical and lateral Fusarium invasion scenarios. The comparison of simulation results with micrographs from laboratory experiments shows encouraging qualitative agreement between the two. Copyright © 2015 Elsevier Ltd. All rights reserved.

  4. Parallel Adaptive Mesh Refinement for High-Order Finite-Volume Schemes in Computational Fluid Dynamics

    Science.gov (United States)

    Schwing, Alan Michael

    For computational fluid dynamics, the governing equations are solved on a discretized domain of nodes, faces, and cells. The quality of the grid or mesh can be a driving source for error in the results. While refinement studies can help guide the creation of a mesh, grid quality is largely determined by user expertise and understanding of the flow physics. Adaptive mesh refinement is a technique for enriching the mesh during a simulation based on metrics for error, impact on important parameters, or location of important flow features. This can offload from the user some of the difficult and ambiguous decisions necessary when discretizing the domain. This work explores the implementation of adaptive mesh refinement in an implicit, unstructured, finite-volume solver. Consideration is made for applying modern computational techniques in the presence of hanging nodes and refined cells. The approach is developed to be independent of the flow solver in order to provide a path for augmenting existing codes. It is designed to be applicable for unsteady simulations and refinement and coarsening of the grid does not impact the conservatism of the underlying numerics. The effect on high-order numerical fluxes of fourth- and sixth-order are explored. Provided the criteria for refinement is appropriately selected, solutions obtained using adapted meshes have no additional error when compared to results obtained on traditional, unadapted meshes. In order to leverage large-scale computational resources common today, the methods are parallelized using MPI. Parallel performance is considered for several test problems in order to assess scalability of both adapted and unadapted grids. Dynamic repartitioning of the mesh during refinement is crucial for load balancing an evolving grid. Development of the methods outlined here depend on a dual-memory approach that is described in detail. Validation of the solver developed here against a number of motivating problems shows favorable

  5. MPPhys—A many-particle simulation package for computational physics education

    Science.gov (United States)

    Müller, Thomas

    2014-03-01

    In a first course to classical mechanics elementary physical processes like elastic two-body collisions, the mass-spring model, or the gravitational two-body problem are discussed in detail. The continuation to many-body systems, however, is deferred to graduate courses although the underlying equations of motion are essentially the same and although there is a strong motivation for high-school students in particular because of the use of particle systems in computer games. The missing link between the simple and the more complex problem is a basic introduction to solve the equations of motion numerically which could be illustrated, however, by means of the Euler method. The many-particle physics simulation package MPPhys offers a platform to experiment with simple particle simulations. The aim is to give a principle idea how to implement many-particle simulations and how simulation and visualization can be combined for interactive visual explorations. Catalogue identifier: AERR_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AERR_v1_0.html Program obtainable from: CPC Program Library, Queen’s University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 111327 No. of bytes in distributed program, including test data, etc.: 608411 Distribution format: tar.gz Programming language: C++, OpenGL, GLSL, OpenCL. Computer: Linux and Windows platforms with OpenGL support. Operating system: Linux and Windows. RAM: Source Code 4.5 MB Complete package 242 MB Classification: 14, 16.9. External routines: OpenGL, OpenCL Nature of problem: Integrate N-body simulations, mass-spring models Solution method: Numerical integration of N-body-simulations, 3D-Rendering via OpenGL. Running time: Problem dependent

  6. Adaptive upscaling with the dual mesh method

    Energy Technology Data Exchange (ETDEWEB)

    Guerillot, D.; Verdiere, S.

    1997-08-01

    The objective of this paper is to demonstrate that upscaling should be calculated during the flow simulation instead of trying to enhance the a priori upscaling methods. Hence, counter-examples are given to motivate our approach, the so-called Dual Mesh Method. The main steps of this numerical algorithm are recalled. Applications illustrate the necessity to consider different average relative permeability values depending on the direction in space. Moreover, these values could be different for the same average saturation. This proves that an a priori upscaling cannot be the answer even in homogeneous cases because of the {open_quotes}dynamical heterogeneity{close_quotes} created by the saturation profile. Other examples show the efficiency of the Dual Mesh Method applied to heterogeneous medium and to an actual field case in South America.

  7. Lagrangian Particle Tracking Simulation for Warm-Rain Processes in Quasi-One-Dimensional Domain

    Science.gov (United States)

    Kunishima, Y.; Onishi, R.

    2017-12-01

    Conventional cloud simulations are based on the Euler method and compute each microphysics process in a stochastic way assuming infinite numbers of particles within each numerical grid. They therefore cannot provide the Lagrangian statistics of individual particles in cloud microphysics (i.e., aerosol particles, cloud particles, and rain drops) nor discuss the statistical fluctuations due to finite number of particles. We here simulate the entire precipitation process of warm-rain, with tracking individual particles. We use the Lagrangian Cloud Simulator (LCS), which is based on the Euler-Lagrangian framework. In that framework, flow motion and scalar transportation are computed with the Euler method, and particle motion with the Lagrangian one. The LCS tracks particle motions and collision events individually with considering the hydrodynamic interaction between approaching particles with a superposition method, that is, it can directly represent the collisional growth of cloud particles. It is essential for trustworthy collision detection to take account of the hydrodynamic interaction. In this study, we newly developed a stochastic model based on the Twomey cloud condensation nuclei (CCN) activation for the Lagrangian tracking simulation and integrated it into the LCS. Coupling with the Euler computation for water vapour and temperature fields, the initiation and condensational growth of water droplets were computed in the Lagrangian way. We applied the integrated LCS for a kinematic simulation of warm-rain processes in a vertically-elongated domain of, at largest, 0.03×0.03×3000 (m3) with horizontal periodicity. Aerosol particles with a realistic number density, 5×107 (m3), were evenly distributed over the domain at the initial state. Prescribed updraft at the early stage initiated development of a precipitating cloud. We have confirmed that the obtained bulk statistics fairly agree with those from a conventional spectral-bin scheme for a vertical column

  8. Automated detection and analysis of particle beams in laser-plasma accelerator simulations

    International Nuclear Information System (INIS)

    Ushizima, Daniela Mayumi; Geddes, C.G.; Cormier-Michel, E.; Bethel, E. Wes; Jacobsen, J.; Prabhat; Ruebel, O.; Weber, G.; Hamann, B.

    2010-01-01

    Numerical simulations of laser-plasma wakefield (particle) accelerators model the acceleration of electrons trapped in plasma oscillations (wakes) left behind when an intense laser pulse propagates through the plasma. The goal of these simulations is to better understand the process involved in plasma wake generation and how electrons are trapped and accelerated by the wake. Understanding of such accelerators, and their development, offer high accelerating gradients, potentially reducing size and cost of new accelerators. One operating regime of interest is where a trapped subset of electrons loads the wake and forms an isolated group of accelerated particles with low spread in momentum and position, desirable characteristics for many applications. The electrons trapped in the wake may be accelerated to high energies, the plasma gradient in the wake reaching up to a gigaelectronvolt per centimeter. High-energy electron accelerators power intense X-ray radiation to terahertz sources, and are used in many applications including medical radiotherapy and imaging. To extract information from the simulation about the quality of the beam, a typical approach is to examine plots of the entire dataset, visually determining the adequate parameters necessary to select a subset of particles, which is then further analyzed. This procedure requires laborious examination of massive data sets over many time steps using several plots, a routine that is unfeasible for large data collections. Demand for automated analysis is growing along with the volume and size of simulations. Current 2D LWFA simulation datasets are typically between 1GB and 100GB in size, but simulations in 3D are of the order of TBs. The increase in the number of datasets and dataset sizes leads to a need for automatic routines to recognize particle patterns as particle bunches (beam of electrons) for subsequent analysis. Because of the growth in dataset size, the application of machine learning techniques for

  9. Parallel Block Structured Adaptive Mesh Refinement on Graphics Processing Units

    Energy Technology Data Exchange (ETDEWEB)

    Beckingsale, D. A. [Atomic Weapons Establishment (AWE), Aldermaston (United Kingdom); Gaudin, W. P. [Atomic Weapons Establishment (AWE), Aldermaston (United Kingdom); Hornung, R. D. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Gunney, B. T. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Gamblin, T. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Herdman, J. A. [Atomic Weapons Establishment (AWE), Aldermaston (United Kingdom); Jarvis, S. A. [Atomic Weapons Establishment (AWE), Aldermaston (United Kingdom)

    2014-11-17

    Block-structured adaptive mesh refinement is a technique that can be used when solving partial differential equations to reduce the number of zones necessary to achieve the required accuracy in areas of interest. These areas (shock fronts, material interfaces, etc.) are recursively covered with finer mesh patches that are grouped into a hierarchy of refinement levels. Despite the potential for large savings in computational requirements and memory usage without a corresponding reduction in accuracy, AMR adds overhead in managing the mesh hierarchy, adding complex communication and data movement requirements to a simulation. In this paper, we describe the design and implementation of a native GPU-based AMR library, including: the classes used to manage data on a mesh patch, the routines used for transferring data between GPUs on different nodes, and the data-parallel operators developed to coarsen and refine mesh data. We validate the performance and accuracy of our implementation using three test problems and two architectures: an eight-node cluster, and over four thousand nodes of Oak Ridge National Laboratory’s Titan supercomputer. Our GPU-based AMR hydrodynamics code performs up to 4.87× faster than the CPU-based implementation, and has been scaled to over four thousand GPUs using a combination of MPI and CUDA.

  10. Thermal-chemical Mantle Convection Models With Adaptive Mesh Refinement

    Science.gov (United States)

    Leng, W.; Zhong, S.

    2008-12-01

    In numerical modeling of mantle convection, resolution is often crucial for resolving small-scale features. New techniques, adaptive mesh refinement (AMR), allow local mesh refinement wherever high resolution is needed, while leaving other regions with relatively low resolution. Both computational efficiency for large- scale simulation and accuracy for small-scale features can thus be achieved with AMR. Based on the octree data structure [Tu et al. 2005], we implement the AMR techniques into the 2-D mantle convection models. For pure thermal convection models, benchmark tests show that our code can achieve high accuracy with relatively small number of elements both for isoviscous cases (i.e. 7492 AMR elements v.s. 65536 uniform elements) and for temperature-dependent viscosity cases (i.e. 14620 AMR elements v.s. 65536 uniform elements). We further implement tracer-method into the models for simulating thermal-chemical convection. By appropriately adding and removing tracers according to the refinement of the meshes, our code successfully reproduces the benchmark results in van Keken et al. [1997] with much fewer elements and tracers compared with uniform-mesh models (i.e. 7552 AMR elements v.s. 16384 uniform elements, and ~83000 tracers v.s. ~410000 tracers). The boundaries of the chemical piles in our AMR code can be easily refined to the scales of a few kilometers for the Earth's mantle and the tracers are concentrated near the chemical boundaries to precisely trace the evolvement of the boundaries. It is thus very suitable for our AMR code to study the thermal-chemical convection problems which need high resolution to resolve the evolvement of chemical boundaries, such as the entrainment problems [Sleep, 1988].

  11. Radial particle distributions in PARMILA simulation beams

    International Nuclear Information System (INIS)

    Boicourt, G.P.

    1984-03-01

    The estimation of beam spill in particle accelerators is becoming of greater importance as higher current designs are being funded. To the present, no numerical method for predicting beam-spill has been available. In this paper, we present an approach to the loss-estimation problem that uses probability distributions fitted to particle-simulation beams. The properties of the PARMILA code's radial particle distribution are discussed, and a broad class of probability distributions are examined to check their ability to fit it. The possibility that the PARMILA distribution is a mixture is discussed, and a fitting distribution consisting of a mixture of two generalized gamma distributions is found. An efficient algorithm to accomplish the fit is presented. Examples of the relative prediction of beam spill are given. 26 references, 18 figures, 1 table

  12. Simulation of capillary bridges between nanoscale particles.

    Science.gov (United States)

    Dörmann, Michael; Schmid, Hans-Joachim

    2014-02-04

    Capillary forces are very important as they exceed in general other adhesion forces. But at the same time the exact calculation of these forces is very complex, so often assumptions and approximations are used. Previous research was done with regard to micrometer sized particles, but the behavior of nanoscale particles is different. Hence, the results for micrometer sized particles cannot be directly transferred when considering nanoscale particles. Therefore, a simulation method was developed to calculate numerically the shape of a rotationally symmetrical capillary bridge between two spherical particles or a particle and a plate. The capillary bridge in the gap between the particles is formed due to capillary condensation and is in thermodynamic equilibrium with the gas phase. Hence the Kelvin equation and the Young-Laplace equation can be used to calculate the profile of the capillary bridge, depending on the relative humidity of the surrounding air. The bridge profile consists of several elements that are determined consecutively and interpolated linearly. After the shape is determined, the volume and force, divided into capillary pressure force and surface tension force, can be calculated. The validation of this numerical model will be shown by comparison with several different analytical calculations for micrometer-sized particles. Furthermore, it is demonstrated that two often used approximations, (1) the toroidal approximation and (2) the use of an effective radius, cannot be used for nanoscale particles without remarkable mistake. It will be discussed how the capillary force and its components depend on different parameters, like particle size, relative humidity, contact angle, and distance, respectively. The rupture of a capillary bridge due to particle separation will also be presented.

  13. Numerical simulation of DPF filter for selected regimes with deposited soot particles

    Science.gov (United States)

    Lávička, David; Kovařík, Petr

    2012-04-01

    For the purpose of accumulation of particulate matter from Diesel engine exhaust gas, particle filters are used (referred to as DPF or FAP filters in the automotive industry). However, the cost of these filters is quite high. As the emission limits become stricter, the requirements for PM collection are rising accordingly. Particulate matters are very dangerous for human health and these are not invisible for human eye. They can often cause various diseases of the respiratory tract, even what can cause lung cancer. Performed numerical simulations were used to analyze particle filter behavior under various operating modes. The simulations were especially focused on selected critical states of particle filter, when engine is switched to emergency regime. The aim was to prevent and avoid critical situations due the filter behavior understanding. The numerical simulations were based on experimental analysis of used diesel particle filters.

  14. Fortran interface layer of the framework for developing particle simulator FDPS

    Science.gov (United States)

    Namekata, Daisuke; Iwasawa, Masaki; Nitadori, Keigo; Tanikawa, Ataru; Muranushi, Takayuki; Wang, Long; Hosono, Natsuki; Nomura, Kentaro; Makino, Junichiro

    2018-06-01

    Numerical simulations based on particle methods have been widely used in various fields including astrophysics. To date, various versions of simulation software have been developed by individual researchers or research groups in each field, through a huge amount of time and effort, even though the numerical algorithms used are very similar. To improve the situation, we have developed a framework, called FDPS (Framework for Developing Particle Simulators), which enables researchers to develop massively parallel particle simulation codes for arbitrary particle methods easily. Until version 3.0, FDPS provided an API (application programming interface) for the C++ programming language only. This limitation comes from the fact that FDPS is developed using the template feature in C++, which is essential to support arbitrary data types of particle. However, there are many researchers who use Fortran to develop their codes. Thus, the previous versions of FDPS require such people to invest much time to learn C++. This is inefficient. To cope with this problem, we developed a Fortran interface layer in FDPS, which provides API for Fortran. In order to support arbitrary data types of particle in Fortran, we design the Fortran interface layer as follows. Based on a given derived data type in Fortran representing particle, a PYTHON script provided by us automatically generates a library that manipulates the C++ core part of FDPS. This library is seen as a Fortran module providing an API of FDPS from the Fortran side and uses C programs internally to interoperate Fortran with C++. In this way, we have overcome several technical issues when emulating a `template' in Fortran. Using the Fortran interface, users can develop all parts of their codes in Fortran. We show that the overhead of the Fortran interface part is sufficiently small and a code written in Fortran shows a performance practically identical to the one written in C++.

  15. Scalar energy fluctuations in Large-Eddy Simulation of turbulent flames: Statistical budgets and mesh quality criterion

    Energy Technology Data Exchange (ETDEWEB)

    Vervisch, Luc; Domingo, Pascale; Lodato, Guido [CORIA - CNRS and INSA de Rouen, Technopole du Madrillet, BP 8, 76801 Saint-Etienne-du-Rouvray (France); Veynante, Denis [EM2C - CNRS and Ecole Centrale Paris, Grande Voie des Vignes, 92295 Chatenay-Malabry (France)

    2010-04-15

    Large-Eddy Simulation (LES) provides space-filtered quantities to compare with measurements, which usually have been obtained using a different filtering operation; hence, numerical and experimental results can be examined side-by-side in a statistical sense only. Instantaneous, space-filtered and statistically time-averaged signals feature different characteristic length-scales, which can be combined in dimensionless ratios. From two canonical manufactured turbulent solutions, a turbulent flame and a passive scalar turbulent mixing layer, the critical values of these ratios under which measured and computed variances (resolved plus sub-grid scale) can be compared without resorting to additional residual terms are first determined. It is shown that actual Direct Numerical Simulation can hardly accommodate a sufficiently large range of length-scales to perform statistical studies of LES filtered reactive scalar-fields energy budget based on sub-grid scale variances; an estimation of the minimum Reynolds number allowing for such DNS studies is given. From these developments, a reliability mesh criterion emerges for scalar LES and scaling for scalar sub-grid scale energy is discussed. (author)

  16. Early experience with mesh excision for adverse outcomes after transvaginal mesh placement using prolapse kits.

    Science.gov (United States)

    Ridgeway, Beri; Walters, Mark D; Paraiso, Marie Fidela R; Barber, Matthew D; McAchran, Sarah E; Goldman, Howard B; Jelovsek, J Eric

    2008-12-01

    The purpose of this study was to determine the complications, treatments, and outcomes in patients choosing to undergo removal of mesh previously placed with a mesh procedural kit. This was a retrospective review of all patients who underwent surgical removal of transvaginal mesh for mesh-related complications during a 3-year period at Cleveland Clinic. At last follow-up, patients reported degree of pain, level of improvement, sexual activity, and continued symptoms. Nineteen patients underwent removal of mesh during the study period. Indications for removal included chronic pain (6/19), dyspareunia (6/19), recurrent pelvic organ prolapse (8/19), mesh erosion (12/19), and vesicovaginal fistula (3/19), with most patients (16/19) citing more than 1 reason. There were few complications related to the mesh removal. Most patients reported significant relief of symptoms. Mesh removal can be technically difficult but appears to be safe with few complications and high relief of symptoms, although some symptoms can persist.

  17. A direct simulation method for flows with suspended paramagnetic particles

    NARCIS (Netherlands)

    Kang, T.G.; Hulsen, M.A.; Toonder, den J.M.J.; Anderson, P.D.; Meijer, H.E.H.

    2008-01-01

    A direct numerical simulation method based on the Maxwell stress tensor and a fictitious domain method has been developed to solve flows with suspended paramagnetic particles. The numerical scheme enables us to take into account both hydrodynamic and magnetic interactions between particles in a

  18. Mesh erosion after abdominal sacrocolpopexy.

    Science.gov (United States)

    Kohli, N; Walsh, P M; Roat, T W; Karram, M M

    1998-12-01

    To report our experience with erosion of permanent suture or mesh material after abdominal sacrocolpopexy. A retrospective chart review was performed to identify patients who underwent sacrocolpopexy by the same surgeon over 8 years. Demographic data, operative notes, hospital records, and office charts were reviewed after sacrocolpopexy. Patients with erosion of either suture or mesh were treated initially with conservative therapy followed by surgical intervention as required. Fifty-seven patients underwent sacrocolpopexy using synthetic mesh during the study period. The mean (range) postoperative follow-up was 19.9 (1.3-50) months. Seven patients (12%) had erosions after abdominal sacrocolpopexy with two suture erosions and five mesh erosions. Patients with suture erosion were asymptomatic compared with patients with mesh erosion, who presented with vaginal bleeding or discharge. The mean (+/-standard deviation) time to erosion was 14.0+/-7.7 (range 4-24) months. Both patients with suture erosion were treated conservatively with estrogen cream. All five patients with mesh erosion required transvaginal removal of the mesh. Mesh erosion can follow abdominal sacrocolpopexy over a long time, and usually presents as vaginal bleeding or discharge. Although patients with suture erosion can be managed successfully with conservative treatment, patients with mesh erosion require surgical intervention. Transvaginal removal of the mesh with vaginal advancement appears to be an effective treatment in patients failing conservative management.

  19. Surgical management of lower urinary mesh perforation after mid-urethral polypropylene mesh sling: mesh excision, urinary tract reconstruction and concomitant pubovaginal sling with autologous rectus fascia.

    Science.gov (United States)

    Shah, Ketul; Nikolavsky, Dmitriy; Gilsdorf, Daniel; Flynn, Brian J

    2013-12-01

    We present our management of lower urinary tract (LUT) mesh perforation after mid-urethral polypropylene mesh sling using a novel combination of surgical techniques including total or near total mesh excision, urinary tract reconstruction, and concomitant pubovaginal sling with autologous rectus fascia in a single operation. We retrospectively reviewed the medical records of 189 patients undergoing transvaginal removal of polypropylene mesh from the lower urinary tract or vagina. The focus of this study is 21 patients with LUT mesh perforation after mid-urethral polypropylene mesh sling. We excluded patients with LUT mesh perforation from prolapse kits (n = 4) or sutures (n = 11), or mesh that was removed because of isolated vaginal wall exposure without concomitant LUT perforation (n = 164). Twenty-one patients underwent surgical removal of mesh through a transvaginal approach or combined transvaginal/abdominal approaches. The location of the perforation was the urethra in 14 and the bladder in 7. The mean follow-up was 22 months. There were no major intraoperative complications. All patients had complete resolution of the mesh complication and the primary symptom. Of the patients with urethral perforation, continence was achieved in 10 out of 14 (71.5 %). Of the patients with bladder perforation, continence was achieved in all 7. Total or near total removal of lower urinary tract (LUT) mesh perforation after mid-urethral polypropylene mesh sling can completely resolve LUT mesh perforation in a single operation. A concomitant pubovaginal sling can be safely performed in efforts to treat existing SUI or avoid future surgery for SUI.

  20. Self-Assembly of DNA-Coated Particles: Experiment, Simulation and Theory

    Science.gov (United States)

    Song, Minseok

    The bottom-up assembly of material architectures with tunable complexity, function, composition, and structure is a long sought goal in rational materials design. One promising approach aims to harnesses the programmability and specificity of DNA hybridization in order to direct the assembly of oligonucleotide-functionalized nano- and micro-particles by tailoring, in part, interparticle interactions. DNA-programmable assembly into three-dimensionally ordered structures has attracted extensive research interest owing to emergent applications in photonics, plasmonics and catalysis and potentially many other areas. Progress on the rational design of DNA-mediated interactions to create useful two-dimensional structures (e.g., structured films), on the other hand, has been rather slow. In this thesis, we establish strategies to engineer a diversity of 2D crystalline arrangements by designing and exploiting DNA-programmable interparticle interactions. We employ a combination of simulation, theory and experiments to predict and confirm accessibility of 2D structural diversity in an effort to establish a rational approach to 2D DNA-mediated particle assembly. We start with the experimental realization of 2D DNA-mediated assembly by decorating micron-sized silica particles with covalently attached single-stranded DNA through a two-step reaction. Subsequently, we elucidate sensitivity and ultimate controllability of DNA-mediated assembly---specifically the melting transition from dispersed singlet particles to aggregated or assembled structures---through control of the concentration of commonly employed nonionic surfactants. We relate the observed tunability to an apparent coupling with the critical micelle temperature in these systems. Also, both square and hexagonal 2D ordered particle arrangements are shown to evolve from disordered aggregates under appropriate annealing conditions defined based upon pre-established melting profiles. Subsequently, the controlled mixing of

  1. Sensitivity of electrospray molecular dynamics simulations to long-range Coulomb interaction models.

    Science.gov (United States)

    Mehta, Neil A; Levin, Deborah A

    2018-03-01

    Molecular dynamics (MD) electrospray simulations of 1-ethyl-3-methylimidazolium tetrafluoroborate (EMIM-BF_{4}) ion liquid were performed with the goal of evaluating the influence of long-range Coulomb models on ion emission characteristics. The direct Coulomb (DC), shifted force Coulomb sum (SFCS), and particle-particle particle-mesh (PPPM) long-range Coulomb models were considered in this work. The DC method with a sufficiently large cutoff radius was found to be the most accurate approach for modeling electrosprays, but, it is computationally expensive. The Coulomb potential energy modeled by the DC method in combination with the radial electric fields were found to be necessary to generate the Taylor cone. The differences observed between the SFCS and the DC in terms of predicting the total ion emission suggest that the former should not be used in MD electrospray simulations. Furthermore, the common assumption of domain periodicity was observed to be detrimental to the accuracy of the capillary-based electrospray simulations.

  2. Target Lagrangian kinematic simulation for particle-laden flows.

    Science.gov (United States)

    Murray, S; Lightstone, M F; Tullis, S

    2016-09-01

    The target Lagrangian kinematic simulation method was motivated as a stochastic Lagrangian particle model that better synthesizes turbulence structure, relative to stochastic separated flow models. By this method, the trajectories of particles are constructed according to synthetic turbulent-like fields, which conform to a target Lagrangian integral timescale. In addition to recovering the expected Lagrangian properties of fluid tracers, this method is shown to reproduce the crossing trajectories and continuity effects, in agreement with an experimental benchmark.

  3. Analytical solutions and particle simulations of cross-field plasma sheaths

    International Nuclear Information System (INIS)

    Gerver, M.J.; Parker, S.E.; Theilhaber, K.

    1989-01-01

    Particles simulations have been made of an infinite plasma slab, bounded by absorbing conducting walls, with a magnetic field parallel to the walls. The simulations have been either 1-D, or 2-D, with the magnetic field normal to the simulation plane. Initially, the plasma has a uniform density between the walls, and there is a uniform source of ions and electrons to replace particles lost to the walls. In the 1-D case, there is no diffusion of the particle guiding centers, and the plasma remains uniform in density and potential over most of the slab, with sheaths about a Debye length wide where the potential rises to the wall potential. In the 2-D case, the density profile becomes parabolic, going almost to zero at the walls, and there is a quasineutral presheath in the bulk of the plasma, in addition to sheaths near the walls. Analytic expressions are found for the density and potential profiles in both cases, including, in the 2-D case, the magnetic presheath due to finite ion Larmor radius, and the effects of the guiding center diffusion rate being either much less than or much grater than the energy diffusion rate. These analytic expressions are shown to agree with the simulations. A 1-D simulation with Monte Carlo guiding center diffusion included gives results that are good agreement with the much more expensive 2-D simulation. 17 refs., 10 figs

  4. Computer simulation of backscattered alpha particles

    International Nuclear Information System (INIS)

    Sanchez, A. Martin; Bland, C.J.; Timon, A. Fernandez

    2000-01-01

    Alpha-particle spectrometry forms an important aspect of radionuclide metrology. Accurate measurements require corrections to be made for factors such as self-absorption within the source and backscattering from the backing material. The theory of the latter phenomenon has only received limited attention. Furthermore the experimental verification of these theoretical results requires adequate counting statistics for a variety of sources with different activities. These problems could be resolved by computer simulations of the various interactions which occur as alpha-particles move through different materials. The pioneering work of Ziegler and his coworkers over several years, has provided the sophisticated software (SRIM) which has enabled us to obtain the results presented here. These results are compared with theoretical and experimental values obtained previously

  5. Dissipative particle dynamics simulation of fluid motion through an unsaturated fracture and fracture junction

    International Nuclear Information System (INIS)

    Liu Moubin; Meakin, Paul; Huang Hai

    2007-01-01

    Multiphase fluid motion in unsaturated fractures and fracture networks involves complicated fluid dynamics, which is difficult to model using grid-based continuum methods. In this paper, the application of dissipative particle dynamics (DPD), a relatively new mesoscale method to simulate fluid motion in unsaturated fractures is described. Unlike the conventional DPD method that employs a purely repulsive conservative (non-dissipative) particle-particle interaction to simulate the behavior of gases, we used conservative particle-particle interactions that combine short-range repulsive and long-range attractive interactions. This new conservative particle-particle interaction allows the behavior of multiphase systems consisting of gases, liquids and solids to be simulated. Our simulation results demonstrate that, for a fracture with flat parallel walls, the DPD method with the new interaction potential function is able to reproduce the hydrodynamic behavior of fully saturated flow, and various unsaturated flow modes including thin film flow, wetting and non-wetting flow. During simulations of flow through a fracture junction, the fracture junction can be fully or partially saturated depending on the wetting property of the fluid, the injection rate and the geometry of the fracture junction. Flow mode switching from a fully saturated flow to a thin film flow can also be observed in the fracture junction

  6. Launch Environment Water Flow Simulations Using Smoothed Particle Hydrodynamics

    Science.gov (United States)

    Vu, Bruce T.; Berg, Jared J.; Harris, Michael F.; Crespo, Alejandro C.

    2015-01-01

    This paper describes the use of Smoothed Particle Hydrodynamics (SPH) to simulate the water flow from the rainbird nozzle system used in the sound suppression system during pad abort and nominal launch. The simulations help determine if water from rainbird nozzles will impinge on the rocket nozzles and other sensitive ground support elements.

  7. A dynamic mesh refinement technique for Lattice Boltzmann simulations on octree-like grids

    KAUST Repository

    Neumann, Philipp

    2012-04-27

    In this contribution, we present our new adaptive Lattice Boltzmann implementation within the Peano framework, with special focus on nanoscale particle transport problems. With the continuum hypothesis not holding anymore on these small scales, new physical effects - such as Brownian fluctuations - need to be incorporated. We explain the overall layout of the application, including memory layout and access, and shortly review the adaptive algorithm. The scheme is validated by different benchmark computations in two and three dimensions. An extension to dynamically changing grids and a spatially adaptive approach to fluctuating hydrodynamics, allowing for the thermalisation of the fluid in particular regions of interest, is proposed. Both dynamic adaptivity and adaptive fluctuating hydrodynamics are validated separately in simulations of particle transport problems. The application of this scheme to an oscillating particle in a nanopore illustrates the importance of Brownian fluctuations in such setups. © 2012 Springer-Verlag.

  8. Meshing Force of Misaligned Spline Coupling and the Influence on Rotor System

    Directory of Open Access Journals (Sweden)

    Guang Zhao

    2008-01-01

    Full Text Available Meshing force of misaligned spline coupling is derived, dynamic equation of rotor-spline coupling system is established based on finite element analysis, the influence of meshing force on rotor-spline coupling system is simulated by numerical integral method. According to the theoretical analysis, meshing force of spline coupling is related to coupling parameters, misalignment, transmitting torque, static misalignment, dynamic vibration displacement, and so on. The meshing force increases nonlinearly with increasing the spline thickness and static misalignment or decreasing alignment meshing distance (AMD. Stiffness of coupling relates to dynamic vibration displacement, and static misalignment is not a constant. Dynamic behaviors of rotor-spline coupling system reveal the following: 1X-rotating speed is the main response frequency of system when there is no misalignment; while 2X-rotating speed appears when misalignment is present. Moreover, when misalignment increases, vibration of the system gets intricate; shaft orbit departs from origin, and magnitudes of all frequencies increase. Research results can provide important criterions on both optimization design of spline coupling and trouble shooting of rotor systems.

  9. Numerical simulation of flow fields and particle trajectories

    DEFF Research Database (Denmark)

    Mayer, Stefan

    2000-01-01

    . The time-dependent flow is approximated with a continuous sequence of steady state creeping flow fields, where metachronously beating ciliary bands are modelled by linear combinations of singularity solutions to the Stokes equations. Generally, the computed flow fields can be divided into an unsteady......A model describing the ciliary driven flow and motion of suspended particles in downstream suspension feeders is developed. The quasi-steady Stokes equations for creeping flow are solved numerically in an unbounded fluid domain around cylindrical bodies using a boundary integral formulation...... in the simulated unsteady ciliary driven flow. A fraction of particles appear to follow trajectories, that resemble experimentally observed particle capture events in the downstream feeding system of the polycheate Sabella penicillus, indicating that particles can be captured by ciliary systems without mechanical...

  10. Optimization of the particle pusher in a diode simulation code

    International Nuclear Information System (INIS)

    Theimer, M.M.; Quintenz, J.P.

    1979-09-01

    The particle pusher in Sandia's particle-in-cell diode simulation code has been rewritten to reduce the required run time of a typical simulation. The resulting new version of the code has been found to run up to three times as fast as the original with comparable accuracy. The cost of this optimization was an increase in storage requirements of about 15%. The new version has also been written to run efficiently on a CRAY-1 computing system. Steps taken to affect this reduced run time are described. Various test cases are detailed

  11. Context-Based Topology Control for Wireless Mesh Networks

    Directory of Open Access Journals (Sweden)

    Pragasen Mudali

    2016-01-01

    Full Text Available Topology Control has been shown to provide several benefits to wireless ad hoc and mesh networks. However these benefits have largely been demonstrated using simulation-based evaluations. In this paper, we demonstrate the negative impact that the PlainTC Topology Control prototype has on topology stability. This instability is found to be caused by the large number of transceiver power adjustments undertaken by the prototype. A context-based solution is offered to reduce the number of transceiver power adjustments undertaken without sacrificing the cumulative transceiver power savings and spatial reuse advantages gained from employing Topology Control in an infrastructure wireless mesh network. We propose the context-based PlainTC+ prototype and show that incorporating context information in the transceiver power adjustment process significantly reduces topology instability. In addition, improvements to network performance arising from the improved topology stability are also observed. Future plans to add real-time context-awareness to PlainTC+ will have the scheme being prototyped in a software-defined wireless mesh network test-bed being planned.

  12. Load management strategy for Particle-In-Cell simulations in high energy particle acceleration

    Energy Technology Data Exchange (ETDEWEB)

    Beck, A., E-mail: beck@llr.in2p3.fr [Laboratoire Leprince-Ringuet, École polytechnique, CNRS-IN2P3, Palaiseau 91128 (France); Frederiksen, J.T. [Niels Bohr Institute, University of Copenhagen, Blegdamsvej 17, 2100 København Ø (Denmark); Dérouillat, J. [CEA, Maison de La Simulation, 91400 Saclay (France)

    2016-09-01

    In the wake of the intense effort made for the experimental CILEX project, numerical simulation campaigns have been carried out in order to finalize the design of the facility and to identify optimal laser and plasma parameters. These simulations bring, of course, important insight into the fundamental physics at play. As a by-product, they also characterize the quality of our theoretical and numerical models. In this paper, we compare the results given by different codes and point out algorithmic limitations both in terms of physical accuracy and computational performances. These limitations are illustrated in the context of electron laser wakefield acceleration (LWFA). The main limitation we identify in state-of-the-art Particle-In-Cell (PIC) codes is computational load imbalance. We propose an innovative algorithm to deal with this specific issue as well as milestones towards a modern, accurate high-performance PIC code for high energy particle acceleration.

  13. Self-Adaptive Event-Driven Simulation of Multi-Scale Plasma Systems

    Science.gov (United States)

    Omelchenko, Yuri; Karimabadi, Homayoun

    2005-10-01

    Multi-scale plasmas pose a formidable computational challenge. The explicit time-stepping models suffer from the global CFL restriction. Efficient application of adaptive mesh refinement (AMR) to systems with irregular dynamics (e.g. turbulence, diffusion-convection-reaction, particle acceleration etc.) may be problematic. To address these issues, we developed an alternative approach to time stepping: self-adaptive discrete-event simulation (DES). DES has origin in operations research, war games and telecommunications. We combine finite-difference and particle-in-cell techniques with this methodology by assuming two caveats: (1) a local time increment, dt for a discrete quantity f can be expressed in terms of a physically meaningful quantum value, df; (2) f is considered to be modified only when its change exceeds df. Event-driven time integration is self-adaptive as it makes use of causality rules rather than parametric time dependencies. This technique enables asynchronous flux-conservative update of solution in accordance with local temporal scales, removes the curse of the global CFL condition, eliminates unnecessary computation in inactive spatial regions and results in robust and fast parallelizable codes. It can be naturally combined with various mesh refinement techniques. We discuss applications of this novel technology to diffusion-convection-reaction systems and hybrid simulations of magnetosonic shocks.

  14. Surface meshing with curvature convergence

    KAUST Repository

    Li, Huibin; Zeng, Wei; Morvan, Jean-Marie; Chen, Liming; Gu, Xianfengdavid

    2014-01-01

    Surface meshing plays a fundamental role in graphics and visualization. Many geometric processing tasks involve solving geometric PDEs on meshes. The numerical stability, convergence rates and approximation errors are largely determined by the mesh qualities. In practice, Delaunay refinement algorithms offer satisfactory solutions to high quality mesh generations. The theoretical proofs for volume based and surface based Delaunay refinement algorithms have been established, but those for conformal parameterization based ones remain wide open. This work focuses on the curvature measure convergence for the conformal parameterization based Delaunay refinement algorithms. Given a metric surface, the proposed approach triangulates its conformal uniformization domain by the planar Delaunay refinement algorithms, and produces a high quality mesh. We give explicit estimates for the Hausdorff distance, the normal deviation, and the differences in curvature measures between the surface and the mesh. In contrast to the conventional results based on volumetric Delaunay refinement, our stronger estimates are independent of the mesh structure and directly guarantee the convergence of curvature measures. Meanwhile, our result on Gaussian curvature measure is intrinsic to the Riemannian metric and independent of the embedding. In practice, our meshing algorithm is much easier to implement and much more efficient. The experimental results verified our theoretical results and demonstrated the efficiency of the meshing algorithm. © 2014 IEEE.

  15. Surface meshing with curvature convergence

    KAUST Repository

    Li, Huibin

    2014-06-01

    Surface meshing plays a fundamental role in graphics and visualization. Many geometric processing tasks involve solving geometric PDEs on meshes. The numerical stability, convergence rates and approximation errors are largely determined by the mesh qualities. In practice, Delaunay refinement algorithms offer satisfactory solutions to high quality mesh generations. The theoretical proofs for volume based and surface based Delaunay refinement algorithms have been established, but those for conformal parameterization based ones remain wide open. This work focuses on the curvature measure convergence for the conformal parameterization based Delaunay refinement algorithms. Given a metric surface, the proposed approach triangulates its conformal uniformization domain by the planar Delaunay refinement algorithms, and produces a high quality mesh. We give explicit estimates for the Hausdorff distance, the normal deviation, and the differences in curvature measures between the surface and the mesh. In contrast to the conventional results based on volumetric Delaunay refinement, our stronger estimates are independent of the mesh structure and directly guarantee the convergence of curvature measures. Meanwhile, our result on Gaussian curvature measure is intrinsic to the Riemannian metric and independent of the embedding. In practice, our meshing algorithm is much easier to implement and much more efficient. The experimental results verified our theoretical results and demonstrated the efficiency of the meshing algorithm. © 2014 IEEE.

  16. The Monte Carlo photoionization and moving-mesh radiation hydrodynamics code CMACIONIZE

    Science.gov (United States)

    Vandenbroucke, B.; Wood, K.

    2018-04-01

    We present the public Monte Carlo photoionization and moving-mesh radiation hydrodynamics code CMACIONIZE, which can be used to simulate the self-consistent evolution of HII regions surrounding young O and B stars, or other sources of ionizing radiation. The code combines a Monte Carlo photoionization algorithm that uses a complex mix of hydrogen, helium and several coolants in order to self-consistently solve for the ionization and temperature balance at any given type, with a standard first order hydrodynamics scheme. The code can be run as a post-processing tool to get the line emission from an existing simulation snapshot, but can also be used to run full radiation hydrodynamical simulations. Both the radiation transfer and the hydrodynamics are implemented in a general way that is independent of the grid structure that is used to discretize the system, allowing it to be run both as a standard fixed grid code, but also as a moving-mesh code.

  17. Particle simulation in curvilinear coordinate systems

    International Nuclear Information System (INIS)

    LeBrun, M.J.; Tajima, T.

    1989-01-01

    We present methods for particle simulation of plasmas in a nearly arbitrary coordinate metric and describe a toroidal electrostatic simulation code that evolved from this effort. A Mercier-type coordinate system is used, with a nonuniform radial grid for improved cross-field resolution. A fast iterative method for solving the Poisson equation is employed, and the interpolation/filtering technique shown to be momentum and energy conserving in the continuum limit. Lorentz ion and drift electron species are used. The code has been thoroughly tested for its reproduction of linear and nonlinear physics, and has been applied to the toroidal drift wave problem and its impact on anomalous transport in tokamaks. 40 refs., 10 figs., 1 tab

  18. Numerical simulation of turbulent combustion: Scientific challenges

    Science.gov (United States)

    Ren, ZhuYin; Lu, Zhen; Hou, LingYun; Lu, LiuYan

    2014-08-01

    Predictive simulation of engine combustion is key to understanding the underlying complicated physicochemical processes, improving engine performance, and reducing pollutant emissions. Critical issues as turbulence modeling, turbulence-chemistry interaction, and accommodation of detailed chemical kinetics in complex flows remain challenging and essential for high-fidelity combustion simulation. This paper reviews the current status of the state-of-the-art large eddy simulation (LES)/prob-ability density function (PDF)/detailed chemistry approach that can address the three challenging modelling issues. PDF as a subgrid model for LES is formulated and the hybrid mesh-particle method for LES/PDF simulations is described. Then the development need in micro-mixing models for the PDF simulations of turbulent premixed combustion is identified. Finally the different acceleration methods for detailed chemistry are reviewed and a combined strategy is proposed for further development.

  19. Analysis of dynamic meshing characteristic of planetary gear transmission in wind power increasing gearbox

    Directory of Open Access Journals (Sweden)

    Wang Jungang

    2017-01-01

    Full Text Available Dynamic behavior of planetary gear’s tooth contact surface in the different location can better conform operation condition comparing to the general gear pair. Nonlinear finite element algorithm was derived according to the basic control equation of contact dynamics. A finite element model of planetary gear transmission in wind power increasing gearbox was proposed considering different meshing locations based on nonlinear finite element solution. The characteristics of stress distribution at different meshing positions were analyzed. A simulation of the meshing process was conducted using finite element analysis. It was shown that node stresses of external meshing planetary gear varied significantly at different position. The analysis provides some useful insights into the performance of planetary gear’s tooth contact surface.

  20. Mesh size in Lichtenstein repair: a systematic review and meta-analysis to determine the importance of mesh size.

    Science.gov (United States)

    Seker, D; Oztuna, D; Kulacoglu, H; Genc, Y; Akcil, M

    2013-04-01

    Small mesh size has been recognized as one of the factors responsible for recurrence after Lichtenstein hernia repair due to insufficient coverage or mesh shrinkage. The Lichtenstein Hernia Institute recommends a 7 × 15 cm mesh that can be trimmed up to 2 cm from the lateral side. We performed a systematic review to determine surgeons' mesh size preference for the Lichtenstein hernia repair and made a meta-analysis to determine the effect of mesh size, mesh type, and length of follow-up time on recurrence. Two medical databases, PubMed and ISI Web of Science, were systematically searched using the key word "Lichtenstein repair." All full text papers were selected. Publications mentioning mesh size were brought for further analysis. A mesh surface area of 90 cm(2) was accepted as the threshold for defining the mesh as small or large. Also, a subgroup analysis for recurrence pooled proportion according to the mesh size, mesh type, and follow-up period was done. In total, 514 papers were obtained. There were no prospective or retrospective clinical studies comparing mesh size and clinical outcome. A total of 141 papers were duplicated in both databases. As a result, 373 papers were obtained. The full text was available in over 95 % of papers. Only 41 (11.2 %) papers discussed mesh size. In 29 studies, a mesh larger than 90 cm(2) was used. The most frequently preferred commercial mesh size was 7.5 × 15 cm. No papers mentioned the size of the mesh after trimming. There was no information about the relationship between mesh size and patient BMI. The pooled proportion in recurrence for small meshes was 0.0019 (95 % confidence interval: 0.007-0.0036), favoring large meshes to decrease the chance of recurrence. Recurrence becomes more marked when follow-up period is longer than 1 year (p < 0.001). Heavy meshes also decreased recurrence (p = 0.015). This systematic review demonstrates that the size of the mesh used in Lichtenstein hernia repair is rarely

  1. Large-eddy simulations of 3D Taylor-Green vortex: comparison of Smoothed Particle Hydrodynamics, Lattice Boltzmann and Finite Volume methods

    International Nuclear Information System (INIS)

    Kajzer, A; Pozorski, J; Szewc, K

    2014-01-01

    In the paper we present Large-eddy simulation (LES) results of 3D Taylor- Green vortex obtained by the three different computational approaches: Smoothed Particle Hydrodynamics (SPH), Lattice Boltzmann Method (LBM) and Finite Volume Method (FVM). The Smagorinsky model was chosen as a subgrid-scale closure in LES for all considered methods and a selection of spatial resolutions have been investigated. The SPH and LBM computations have been carried out with the use of the in-house codes executed on GPU and compared, for validation purposes, with the FVM results obtained using the open-source CFD software OpenFOAM. A comparative study in terms of one-point statistics and turbulent energy spectra shows a good agreement of LES results for all methods. An analysis of the GPU code efficiency and implementation difficulties has been made. It is shown that both SPH and LBM may offer a significant advantage over mesh-based CFD methods.

  2. Automatic mesh refinement and parallel load balancing for Fokker-Planck-DSMC algorithm

    Science.gov (United States)

    Küchlin, Stephan; Jenny, Patrick

    2018-06-01

    Recently, a parallel Fokker-Planck-DSMC algorithm for rarefied gas flow simulation in complex domains at all Knudsen numbers was developed by the authors. Fokker-Planck-DSMC (FP-DSMC) is an augmentation of the classical DSMC algorithm, which mitigates the near-continuum deficiencies in terms of computational cost of pure DSMC. At each time step, based on a local Knudsen number criterion, the discrete DSMC collision operator is dynamically switched to the Fokker-Planck operator, which is based on the integration of continuous stochastic processes in time, and has fixed computational cost per particle, rather than per collision. In this contribution, we present an extension of the previous implementation with automatic local mesh refinement and parallel load-balancing. In particular, we show how the properties of discrete approximations to space-filling curves enable an efficient implementation. Exemplary numerical studies highlight the capabilities of the new code.

  3. Comparing semi-analytic particle tagging and hydrodynamical simulations of the Milky Way's stellar halo

    Science.gov (United States)

    Cooper, Andrew P.; Cole, Shaun; Frenk, Carlos S.; Le Bret, Theo; Pontzen, Andrew

    2017-08-01

    Particle tagging is an efficient, but approximate, technique for using cosmological N-body simulations to model the phase-space evolution of the stellar populations predicted, for example, by a semi-analytic model of galaxy formation. We test the technique developed by Cooper et al. (which we call stings here) by comparing particle tags with stars in a smooth particle hydrodynamic (SPH) simulation. We focus on the spherically averaged density profile of stars accreted from satellite galaxies in a Milky Way (MW)-like system. The stellar profile in the SPH simulation can be recovered accurately by tagging dark matter (DM) particles in the same simulation according to a prescription based on the rank order of particle binding energy. Applying the same prescription to an N-body version of this simulation produces a density profile differing from that of the SPH simulation by ≲10 per cent on average between 1 and 200 kpc. This confirms that particle tagging can provide a faithful and robust approximation to a self-consistent hydrodynamical simulation in this regime (in contradiction to previous claims in the literature). We find only one systematic effect, likely due to the collisionless approximation, namely that massive satellites in the SPH simulation are disrupted somewhat earlier than their collisionless counterparts. In most cases, this makes remarkably little difference to the spherically averaged distribution of their stellar debris. We conclude that, for galaxy formation models that do not predict strong baryonic effects on the present-day DM distribution of MW-like galaxies or their satellites, differences in stellar halo predictions associated with the treatment of star formation and feedback are much more important than those associated with the dynamical limitations of collisionless particle tagging.

  4. Program BETA for simulation of particle decays and reactions

    International Nuclear Information System (INIS)

    Takhtamyshev, G.G.; Merkulova, T.A.

    1997-01-01

    Program BETA is designed for simulation of particle decays and reactions. The program also produces integration over the phase space and decay rate or the reaction cross section are calculated as a result of such integration. At the simulation process the adaptive random number generator SMART may be used, what is found to be useful for some difficult cases

  5. Augmenting Sand Simulation Environments through Subdivision and Particle Refinement

    Science.gov (United States)

    Clothier, M.; Bailey, M.

    2012-12-01

    Recent advances in computer graphics and parallel processing hardware have provided disciplines with new methods to evaluate and visualize data. These advances have proven useful for earth and planetary scientists as many researchers are using this hardware to process large amounts of data for analysis. As such, this has provided opportunities for collaboration between computer graphics and the earth sciences. Through collaboration with the Oregon Space Grant and IGERT Ecosystem Informatics programs, we are investigating techniques for simulating the behavior of sand. We are also collaborating with the Jet Propulsion Laboratory's (JPL) DARTS Lab to exchange ideas and gain feedback on our research. The DARTS Lab specializes in simulation of planetary vehicles, such as the Mars rovers. Their simulations utilize a virtual "sand box" to test how a planetary vehicle responds to different environments. Our research builds upon this idea to create a sand simulation framework so that planetary environments, such as the harsh, sandy regions on Mars, are more fully realized. More specifically, we are focusing our research on the interaction between a planetary vehicle, such as a rover, and the sand beneath it, providing further insight into its performance. Unfortunately, this can be a computationally complex problem, especially if trying to represent the enormous quantities of sand particles interacting with each other. However, through the use of high-performance computing, we have developed a technique to subdivide areas of actively participating sand regions across a large landscape. Similar to a Level of Detail (LOD) technique, we only subdivide regions of a landscape where sand particles are actively participating with another object. While the sand is within this subdivision window and moves closer to the surface of the interacting object, the sand region subdivides into smaller regions until individual sand particles are left at the surface. As an example, let's say

  6. Fully resolved simulations of expansion waves propagating into particle beds

    Science.gov (United States)

    Marjanovic, Goran; Hackl, Jason; Annamalai, Subramanian; Jackson, Thomas; Balachandar, S.

    2017-11-01

    There is a tremendous amount of research that has been done on compression waves and shock waves moving over particles but very little concerning expansion waves. Using 3-D direct numerical simulations, this study will explore expansion waves propagating into fully resolved particle beds of varying volume fractions and geometric arrangements. The objectives of these simulations are as follows: 1) To fully resolve all (1-way coupled) forces on the particles in a time varying flow and 2) to verify state-of-the-art drag models for such complex flows. We will explore a range of volume fractions, from very low ones that are similar to single particle flows, to higher ones where nozzling effects are observed between neighboring particles. Further, we will explore two geometric arrangements: body centered cubic and face centered cubic. We will quantify the effects that volume fraction and geometric arrangement plays on the drag forces and flow fields experienced by the particles. These results will then be compared to theoretical predictions from a model based on the generalized Faxen's theorem. This work was supported in part by the U.S. Department of Energy under the Predictive Science Academic Alliance Program, under Contract No. DE-NA0002378.

  7. Vectorization of a particle simulation method for hypersonic rarefied flow

    Science.gov (United States)

    Mcdonald, Jeffrey D.; Baganoff, Donald

    1988-01-01

    An efficient particle simulation technique for hypersonic rarefied flows is presented at an algorithmic and implementation level. The implementation is for a vector computer architecture, specifically the Cray-2. The method models an ideal diatomic Maxwell molecule with three translational and two rotational degrees of freedom. Algorithms are designed specifically for compatibility with fine grain parallelism by reducing the number of data dependencies in the computation. By insisting on this compatibility, the method is capable of performing simulation on a much larger scale than previously possible. A two-dimensional simulation of supersonic flow over a wedge is carried out for the near-continuum limit where the gas is in equilibrium and the ideal solution can be used as a check on the accuracy of the gas model employed in the method. Also, a three-dimensional, Mach 8, rarefied flow about a finite-span flat plate at a 45 degree angle of attack was simulated. It utilized over 10 to the 7th particles carried through 400 discrete time steps in less than one hour of Cray-2 CPU time. This problem was chosen to exhibit the capability of the method in handling a large number of particles and a true three-dimensional geometry.

  8. Vectorization of a particle simulation method for hypersonic rarefied flow

    International Nuclear Information System (INIS)

    Mcdonald, J.D.; Baganoff, D.

    1988-01-01

    An efficient particle simulation technique for hypersonic rarefied flows is presented at an algorithmic and implementation level. The implementation is for a vector computer architecture, specifically the Cray-2. The method models an ideal diatomic Maxwell molecule with three translational and two rotational degrees of freedom. Algorithms are designed specifically for compatibility with fine grain parallelism by reducing the number of data dependencies in the computation. By insisting on this compatibility, the method is capable of performing simulation on a much larger scale than previously possible. A two-dimensional simulation of supersonic flow over a wedge is carried out for the near-continuum limit where the gas is in equilibrium and the ideal solution can be used as a check on the accuracy of the gas model employed in the method. Also, a three-dimensional, Mach 8, rarefied flow about a finite-span flat plate at a 45 degree angle of attack was simulated. It utilized over 10 to the 7th particles carried through 400 discrete time steps in less than one hour of Cray-2 CPU time. This problem was chosen to exhibit the capability of the method in handling a large number of particles and a true three-dimensional geometry. 14 references

  9. A mixed finite element method for particle simulation in lasertron

    International Nuclear Information System (INIS)

    Le Meur, G.

    1987-03-01

    A particle simulation code is being developed with the aim to treat the motion of charged particles in electromagnetic devices, such as Lasertron. The paper describes the use of mixed finite element methods in computing the field components, without derivating them from scalar or vector potentials. Graphical results are shown

  10. A mixed finite element method for particle simulation in Lasertron

    International Nuclear Information System (INIS)

    Le Meur, G.

    1987-01-01

    A particle simulation code is being developed with the aim to treat the motion of charged particles in electromagnetic devices, such as Lasertron. The paper describes the use of mixed finite element methods in computing the field components, without derivating them from scalar or vector potentials. Graphical results are shown

  11. Parallel 3D Mortar Element Method for Adaptive Nonconforming Meshes

    Science.gov (United States)

    Feng, Huiyu; Mavriplis, Catherine; VanderWijngaart, Rob; Biswas, Rupak

    2004-01-01

    High order methods are frequently used in computational simulation for their high accuracy. An efficient way to avoid unnecessary computation in smooth regions of the solution is to use adaptive meshes which employ fine grids only in areas where they are needed. Nonconforming spectral elements allow the grid to be flexibly adjusted to satisfy the computational accuracy requirements. The method is suitable for computational simulations of unsteady problems with very disparate length scales or unsteady moving features, such as heat transfer, fluid dynamics or flame combustion. In this work, we select the Mark Element Method (MEM) to handle the non-conforming interfaces between elements. A new technique is introduced to efficiently implement MEM in 3-D nonconforming meshes. By introducing an "intermediate mortar", the proposed method decomposes the projection between 3-D elements and mortars into two steps. In each step, projection matrices derived in 2-D are used. The two-step method avoids explicitly forming/deriving large projection matrices for 3-D meshes, and also helps to simplify the implementation. This new technique can be used for both h- and p-type adaptation. This method is applied to an unsteady 3-D moving heat source problem. With our new MEM implementation, mesh adaptation is able to efficiently refine the grid near the heat source and coarsen the grid once the heat source passes. The savings in computational work resulting from the dynamic mesh adaptation is demonstrated by the reduction of the the number of elements used and CPU time spent. MEM and mesh adaptation, respectively, bring irregularity and dynamics to the computer memory access pattern. Hence, they provide a good way to gauge the performance of computer systems when running scientific applications whose memory access patterns are irregular and unpredictable. We select a 3-D moving heat source problem as the Unstructured Adaptive (UA) grid benchmark, a new component of the NAS Parallel

  12. Numerical simulation of the motion of charged suspended particle in multi-phase flow

    Energy Technology Data Exchange (ETDEWEB)

    Abd Elkhalek, M M [Nuclear Research Center-Atomic Energy Authority, Cairo (Egypt)

    1997-12-31

    A method for computing numerical simulation of the motion of charged suspended particle in multi-phase flow between two-long parallel plates is described in detail. The equation of motion of a suspended particle was suggested by closkin. The equations of motion are reduced to ordinary differential equations by similarity transformations and solved numerically by using Runge-Kutta method. The trajectories of particles are calculated by integrating the equation of motion of a single particle. Numerical solutions of the resulting ordinary differential equations provide velocity distributions for both fluid and solid phases and density distributions for the solid. The present simulation requires some empirical parameters concerning the collision of the particles with the wall. Some typical results for both fluid and particle phases and density distributions of the particles are presented graphically. 4 figs.

  13. Numerical Simulation of the Motion of Charged Suspended Particle in Multi-Phase Flow

    International Nuclear Information System (INIS)

    Abd-El Khalek, M.M.

    1998-01-01

    A method for computing Numerical simulation of the motion of charged suspended particle in multi-phase flow between two-long parallel plates is described in detail. The equation of motion of a suspended particle was suggested by Closkin. The equations of motion are reduced to ordinary differential equations by similarity transformations and solved numerically by using the Runge-Kutta method. The trajectories of particles are calculated by integrating the equation of motion of a single particle. Numerical solutions of the resulting ordinary differential equations provide velocity distributions for both fluid and solid phases and density distributions for the solid. The present simulation requires some empirical parameters concerning the collision of the particles with the wall. Some typical results for both fluid and particle phases and density distributions of the particles are presented graphically

  14. Transonic Airfoil Flow Simulation. Part I: Mesh Generation and Inviscid Method

    Directory of Open Access Journals (Sweden)

    Vladimir CARDOS

    2010-06-01

    Full Text Available A calculation method for the subsonic and transonic viscous flow over airfoil using thedisplacement surface concept is described. Part I presents a mesh generation method forcomputational grid and a finite volume method for the time-dependent Euler equations. The inviscidsolution is used for the inviscid-viscous coupling procedure presented in the Part II.

  15. Mesh Generation and Adaption for High Reynolds Number RANS Computations, Phase I

    Data.gov (United States)

    National Aeronautics and Space Administration — This proposal offers to provide NASA with an automatic mesh generator for the simulation of aerodynamic flows using Reynolds-Averages Navier-Stokes (RANS) models....

  16. Mesh Generation and Adaption for High Reynolds Number RANS Computations, Phase II

    Data.gov (United States)

    National Aeronautics and Space Administration — This proposal offers to provide NASA with an automatic mesh generator for the simulation of aerodynamic flows using Reynolds-Averages Navier-Stokes (RANS) models....

  17. Lagrangian Particle Tracking in a Discontinuous Galerkin Method for Hypersonic Reentry Flows in Dusty Environments

    Science.gov (United States)

    Ching, Eric; Lv, Yu; Ihme, Matthias

    2017-11-01

    Recent interest in human-scale missions to Mars has sparked active research into high-fidelity simulations of reentry flows. A key feature of the Mars atmosphere is the high levels of suspended dust particles, which can not only enhance erosion of thermal protection systems but also transfer energy and momentum to the shock layer, increasing surface heat fluxes. Second-order finite-volume schemes are typically employed for hypersonic flow simulations, but such schemes suffer from a number of limitations. An attractive alternative is discontinuous Galerkin methods, which benefit from arbitrarily high spatial order of accuracy, geometric flexibility, and other advantages. As such, a Lagrangian particle method is developed in a discontinuous Galerkin framework to enable the computation of particle-laden hypersonic flows. Two-way coupling between the carrier and disperse phases is considered, and an efficient particle search algorithm compatible with unstructured curved meshes is proposed. In addition, variable thermodynamic properties are considered to accommodate high-temperature gases. The performance of the particle method is demonstrated in several test cases, with focus on the accurate prediction of particle trajectories and heating augmentation. Financial support from a Stanford Graduate Fellowship and the NASA Early Career Faculty program are gratefully acknowledged.

  18. Voltammetry at micro-mesh electrodes

    Directory of Open Access Journals (Sweden)

    Wadhawan Jay D.

    2003-01-01

    Full Text Available The voltammetry at three micro-mesh electrodes is explored. It is found that at sufficiently short experimental durations, the micro-mesh working electrode first behaves as an ensemble of microband electrodes, then follows the behaviour anticipated for an array of diffusion-independent micro-ring electrodes of the same perimeter as individual grid-squares within the mesh. During prolonged electrolysis, the micro-mesh electrode follows that behaviour anticipated theoretically for a cubically-packed partially-blocked electrode. Application of the micro-mesh electrode for the electrochemical determination of carbon dioxide in DMSO electrolyte solutions is further illustrated.

  19. Dissolution of uranium and plutonium particles: simulations using the Mercer equation

    International Nuclear Information System (INIS)

    Cowan, C.E.; Jenne, E.A.

    1983-10-01

    There is a need to be able to predict the amount of plutonium that will be in solution at a given time from dissolution of particles in order to better predict the environmental behavior and possible adverse effects of plutonium spills. The equation developed by Mercer (1967) to simulate the dissolution of particles in lungs was parameterized and used to simulate the dissolution of a population of plutonium or uranium particles in the soil. Parameter values for the size distribution of particles in soil, and the density of the particles were found; however, values for the shape factors, and the dissolution rate were virtually non-existent. The calculated mass dissolved was most sensitive to the median diameter of the population of particles and least sensitive to the geometric standard deviation. A given percent change in the shape parameter and the dissolution rate resulted in approximately an equal percent change in the mass dissolved. Provided that the population of particles follows a log-normal distribution, the particles are homogeneous in composition and the dissolution can be represented by first-order kinetics, this equation can probably be applied with slight modification to estimate the mass dissolved at a given time. 66 references, 7 figures, 4 tables

  20. Streaming simplification of tetrahedral meshes.

    Science.gov (United States)

    Vo, Huy T; Callahan, Steven P; Lindstrom, Peter; Pascucci, Valerio; Silva, Cláudio T

    2007-01-01

    Unstructured tetrahedral meshes are commonly used in scientific computing to represent scalar, vector, and tensor fields in three dimensions. Visualization of these meshes can be difficult to perform interactively due to their size and complexity. By reducing the size of the data, we can accomplish real-time visualization necessary for scientific analysis. We propose a two-step approach for streaming simplification of large tetrahedral meshes. Our algorithm arranges the data on disk in a streaming, I/O-efficient format that allows coherent access to the tetrahedral cells. A quadric-based simplification is sequentially performed on small portions of the mesh in-core. Our output is a coherent streaming mesh which facilitates future processing. Our technique is fast, produces high quality approximations, and operates out-of-core to process meshes too large for main memory.

  1. Atlas-Based Automatic Generation of Subject-Specific Finite Element Tongue Meshes.

    Science.gov (United States)

    Bijar, Ahmad; Rohan, Pierre-Yves; Perrier, Pascal; Payan, Yohan

    2016-01-01

    Generation of subject-specific 3D finite element (FE) models requires the processing of numerous medical images in order to precisely extract geometrical information about subject-specific anatomy. This processing remains extremely challenging. To overcome this difficulty, we present an automatic atlas-based method that generates subject-specific FE meshes via a 3D registration guided by Magnetic Resonance images. The method extracts a 3D transformation by registering the atlas' volume image to the subject's one, and establishes a one-to-one correspondence between the two volumes. The 3D transformation field deforms the atlas' mesh to generate the subject-specific FE mesh. To preserve the quality of the subject-specific mesh, a diffeomorphic non-rigid registration based on B-spline free-form deformations is used, which guarantees a non-folding and one-to-one transformation. Two evaluations of the method are provided. First, a publicly available CT-database is used to assess the capability to accurately capture the complexity of each subject-specific Lung's geometry. Second, FE tongue meshes are generated for two healthy volunteers and two patients suffering from tongue cancer using MR images. It is shown that the method generates an appropriate representation of the subject-specific geometry while preserving the quality of the FE meshes for subsequent FE analysis. To demonstrate the importance of our method in a clinical context, a subject-specific mesh is used to simulate tongue's biomechanical response to the activation of an important tongue muscle, before and after cancer surgery.

  2. Mathematical simulation of cascade-probabilistic functions for charged particles

    International Nuclear Information System (INIS)

    Kupchishin, A.A.; Kupchishin, A.I.; Smygaleva, T.A.

    1998-01-01

    Analytical expressions for cascade-probabilistic functions (CPF) for electrons, protons, α-particles and ions with taking into account energy losses are received. Mathematical analysis of these functions is carried out and main properties of function are determined. Algorithms of CPF are developed and their computer calculation were conducted. Regularities in behavior of function in dependence on initial particles energy, atomic number and registration depth are established. Book is intended to specialists on mathematical simulation of radiation defects, solid state physics, elementary particle physics and applied mathematics. There are 3 chapters in the book: 1. Cascade-probabilistic functions for electrons; 2. CPF for protons and α-particles; 3. CPF with taking unto account energy losses of ions. (author)

  3. Particle Simulation of Fractional Diffusion Equations

    KAUST Repository

    Allouch, Samer

    2017-07-12

    This work explores different particle-based approaches to the simulation of one-dimensional fractional subdiffusion equations in unbounded domains. We rely on smooth particle approximations, and consider four methods for estimating the fractional diffusion term. The first method is based on direct differentiation of the particle representation, it follows the Riesz definition of the fractional derivative and results in a non-conservative scheme. The other three methods follow the particle strength exchange (PSE) methodology and are by construction conservative, in the sense that the total particle strength is time invariant. The first PSE algorithm is based on using direct differentiation to estimate the fractional diffusion flux, and exploiting the resulting estimates in an integral representation of the divergence operator. Meanwhile, the second one relies on the regularized Riesz representation of the fractional diffusion term to derive a suitable interaction formula acting directly on the particle representation of the diffusing field. A third PSE construction is considered that exploits the Green\\'s function of the fractional diffusion equation. The performance of all four approaches is assessed for the case of a one-dimensional diffusion equation with constant diffusivity. This enables us to take advantage of known analytical solutions, and consequently conduct a detailed analysis of the performance of the methods. This includes a quantitative study of the various sources of error, namely filtering, quadrature, domain truncation, and time integration, as well as a space and time self-convergence analysis. These analyses are conducted for different values of the order of the fractional derivatives, and computational experiences are used to gain insight that can be used for generalization of the present constructions.

  4. Particle Simulation of Fractional Diffusion Equations

    KAUST Repository

    Allouch, Samer; Lucchesi, Marco; Maî tre, O. P. Le; Mustapha, K. A.; Knio, Omar

    2017-01-01

    This work explores different particle-based approaches to the simulation of one-dimensional fractional subdiffusion equations in unbounded domains. We rely on smooth particle approximations, and consider four methods for estimating the fractional diffusion term. The first method is based on direct differentiation of the particle representation, it follows the Riesz definition of the fractional derivative and results in a non-conservative scheme. The other three methods follow the particle strength exchange (PSE) methodology and are by construction conservative, in the sense that the total particle strength is time invariant. The first PSE algorithm is based on using direct differentiation to estimate the fractional diffusion flux, and exploiting the resulting estimates in an integral representation of the divergence operator. Meanwhile, the second one relies on the regularized Riesz representation of the fractional diffusion term to derive a suitable interaction formula acting directly on the particle representation of the diffusing field. A third PSE construction is considered that exploits the Green's function of the fractional diffusion equation. The performance of all four approaches is assessed for the case of a one-dimensional diffusion equation with constant diffusivity. This enables us to take advantage of known analytical solutions, and consequently conduct a detailed analysis of the performance of the methods. This includes a quantitative study of the various sources of error, namely filtering, quadrature, domain truncation, and time integration, as well as a space and time self-convergence analysis. These analyses are conducted for different values of the order of the fractional derivatives, and computational experiences are used to gain insight that can be used for generalization of the present constructions.

  5. PEPT: An invaluable tool for 3-D particle tracking and CFD simulation verification in hydrocyclone studies

    Directory of Open Access Journals (Sweden)

    Hoffmann Alex C.

    2013-05-01

    Full Text Available Particle tracks in a hydrocyclone generated both experimentally by positron emission particle tracking (PEPT and numerically with Eulerian-Lagranian CFD have been studied and compared. A hydrocyclone with a cylinder-on-cone design was used in this study, the geometries used in the CFD simulations and in the experiments being identical. It is shown that it is possible to track a fast-moving particle in a hydrocyclone using PEPT with high temporal and spatial resolutions. The numerical 3-D particle trajectories were generated using the Large Eddy Simulation (LES turbulence model for the fluid and Lagrangian particle tracking for the particles. The behaviors of the particles were analyzed in detail and were found to be consistent between experiments and CFD simulations. The tracks of the particles are discussed and related to the fluid flow field visualized in the CFD simulations using the cross-sectional static pressure distribution.

  6. 3D simulation studies of tokamak plasmas using MHD and extended-MHD models

    International Nuclear Information System (INIS)

    Park, W.; Chang, Z.; Fredrickson, E.; Fu, G.Y.

    1996-01-01

    The M3D (Multi-level 3D) tokamak simulation project aims at the simulation of tokamak plasmas using a multi-level tokamak code package. Several current applications using MHD and Extended-MHD models are presented; high-β disruption studies in reversed shear plasmas using the MHD level MH3D code, ω *i stabilization and nonlinear island saturation of TAE mode using the hybrid particle/MHD level MH3D-K code, and unstructured mesh MH3D ++ code studies. In particular, three internal mode disruption mechanisms are identified from simulation results which agree which agree well with experimental data

  7. Influence of reinforcement mesh configuration for improvement of concrete durability

    Science.gov (United States)

    Pan, Chong-gen; Jin, Wei-liang; Mao, Jiang-hong; Zhang, Hua; Sun, Li-hao; Wei, Dong

    2017-10-01

    Steel bar in concrete structures under harsh environmental conditions, such as chlorine corrosion, seriously affects its service life. Bidirectional electromigration rehabilitation (BIEM) is a new method of repair technology for reinforced concrete structures in such chloride corrosion environments. By applying the BIEM, chloride ions can be removed from the concrete and the migrating corrosion inhibit can be moved to the steel surface. In conventional engineering, the concrete structure is often configured with a multi-layer steel mesh. However, the effect of the BIEM in such structures has not yet been investigated. In this paper, the relevant simulation test is carried out to study the migration law of chloride ions and the migrating corrosion inhibitor in a concrete specimen with complex steel mesh under different energizing modes. The results show that the efficiency of the BIEM increases 50% in both the monolayer steel mesh and the double-layer steel mesh. By using the single-sided BIEM, 87% of the chloride ions are removed from the steel surface. The different step modes can affect the chloride ion removal. The chloride ions within the range of the reinforcement protective cover are easier to be removed than those in the concrete between the two layers of steel mesh. However, the amount of migrating corrosion inhibitor is larger in the latter circumstances.

  8. Particle Tracking and Simulation on the .NET Framework

    International Nuclear Information System (INIS)

    Nishimura, Hiroshi; Scarvie, Tom

    2006-01-01

    Particle tracking and simulation studies are becoming increasingly complex. In addition to the use of more sophisticated graphics, interactive scripting is becoming popular. Compatibility with different control systems requires network and database capabilities. It is not a trivial task to fulfill all the various requirements without sacrificing runtime performance. We evaluated the effectiveness of the .NET framework by converting a C++ simulation code to C. The portability to other platforms is mentioned in terms of Mono

  9. Developments of DPF systems with mesh laminated structures. Performances of DPF systems which consist of the metal-mesh laminated filter combustion with the alumina-fiber mesh, and the combustion device of trapped diesel particles; Mesh taso kozo no DPF no kaihatsu. Kinzokusen to arumina sen`i mesh ni yoru fukugo filter to filter heiyo heater ni yoru DPF no seino

    Energy Technology Data Exchange (ETDEWEB)

    Kojima, T; Tange, A; Matsuda, K [NHK Spring Co. Ltd., Yokohama (Japan)

    1997-10-01

    For the purpose of continuous run without any maintenance, new DPF (diesel particulate filter)systems laminated by both metal-wire mesh and alumina-fiber mesh alternately, are under the developments. The perfect combustion of trapped diesel particulate can be achieved by a couple of the resistance heating devices inserted into the filter. 5 refs., 7 figs., 3 tabs.

  10. Energy Efficient Monitoring for Intrusion Detection in Battery-Powered Wireless Mesh Networks

    KAUST Repository

    Hassanzadeh, Amin

    2011-07-18

    Wireless Mesh Networks (WMN) are easy-to-deploy, low cost solutions for providing networking and internet services in environments with no network infrastructure, e.g., disaster areas and battlefields. Since electric power is not readily available in such environments battery-powered mesh routers, operating in an energy efficient manner, are required. To the best of our knowledge, the impact of energy efficient solutions, e.g., involving duty-cycling, on WMN intrusion detection systems, which require continuous monitoring, remains an open research problem. In this paper we propose that carefully chosen monitoring mesh nodes ensure continuous and complete detection coverage, while allowing non-monitoring mesh nodes to save energy through duty-cycling. We formulate the monitoring node selection problem as an optimization problem and propose distributed and centralized solutions for it, with different tradeoffs. Through extensive simulations and a proof-of-concept hardware/software implementation we demonstrate that our solutions extend the WMN lifetime by 8%, while ensuring, at the minimum, a 97% intrusion detection rate.

  11. Lattice Boltzmann Simulation of Collision between 2D Circular Particles Suspension in Couette Flow

    Directory of Open Access Journals (Sweden)

    Li-Zhong Huang

    2013-01-01

    Full Text Available Collision between 2D circular particles suspension in Couette flow is simulated by using multiple-relaxation-time based lattice Boltzmann and direct forcing/fictitious domain method in this paper. The patterns of particle collisions are simulated and analyzed in detail by changing the velocity of top and bottom walls in the Couette flow. It can be seen from the simulation results that, while the velocity is large enough, the number of collisions between particles will change little as this velocity varies.

  12. Progress of laser-plasma interaction simulations with the particle-in-cell code

    International Nuclear Information System (INIS)

    Sakagami, Hitoshi; Kishimoto, Yasuaki; Sentoku, Yasuhiko; Taguchi, Toshihiro

    2005-01-01

    As the laser-plasma interaction is a non-equilibrium, non-linear and relativistic phenomenon, we must introduce a microscopic method, namely, the relativistic electromagnetic PIC (Particle-In-Cell) simulation code. The PIC code requires a huge number of particles to validate simulation results, and its task is very computation-intensive. Thus simulation researches by the PIC code have been progressing along with advances in computer technology. Recently, parallel computers with tremendous computational power have become available, and thus we can perform three-dimensional PIC simulations for the laser-plasma interaction to investigate laser fusion. Some simulation results are shown with figures. We discuss a recent trend of large-scale PIC simulations that enable direct comparison between experimental facts and computational results. We also discharge/lightning simulations by the extended PIC code, which include various atomic and relaxation processes. (author)

  13. Laparoscopic mesh explantation and drainage of sacral abscess remote from transvaginal excision of exposed sacral colpopexy mesh.

    Science.gov (United States)

    Roth, Ted M; Reight, Ian

    2012-07-01

    Sacral colpopexy may be complicated by mesh exposure, and the surgical treatment of mesh exposure typically results in minor postoperative morbidity and few delayed complications. A 75-year-old woman presented 7 years after a laparoscopic sacral colpopexy, with Mersilene mesh, with an apical mesh exposure. She underwent an uncomplicated transvaginal excision and was asymptomatic until 8 months later when she presented with vaginal drainage and a sacral abscess. This was successfully treated with laparoscopic enterolysis, drainage of the abscess, and explantation of the remaining mesh. Incomplete excision of exposed colpopexy mesh can lead to ascending infection and sacral abscess. Laparoscopic drainage and mesh removal may be considered in these patients.

  14. Direct Lagrangian tracking simulations of particles in vertically-developing atmospheric clouds

    Science.gov (United States)

    Onishi, Ryo; Kunishima, Yuichi

    2017-11-01

    We have been developing the Lagrangian Cloud Simulator (LCS), which follows the so-called Euler-Lagrangian framework, where flow motion and scalar transportations (i.e., temperature and humidity) are computed with the Euler method and particle motion with the Lagrangian method. The LCS simulation considers the hydrodynamic interaction between approaching particles for robust collision detection. This leads to reliable simulations of collision growth of cloud droplets. Recently the activation process, in which aerosol particles become tiny liquid droplets, has been implemented in the LCS. The present LCS can therefore consider the whole warm-rain precipitation processes -activation, condensation, collision and drop precipitation. In this talk, after briefly introducing the LCS, we will show kinematic simulations using the LCS for quasi-one dimensional domain, i.e., vertically elongated 3D domain. They are compared with one-dimensional kinematic simulations using a spectral-bin cloud microphysics scheme, which is based on the Euler method. The comparisons show fairly good agreement with small discrepancies, the source of which will be presented. The Lagrangian statistics, obtained for the first time for the vertical domain, will be the center of discussion. This research was supported by MEXT as ``Exploratory Challenge on Post-K computer'' (Frontiers of Basic Science: Challenging the Limits).

  15. Application of State Quantization-Based Methods in HEP Particle Transport Simulation

    Science.gov (United States)

    Santi, Lucio; Ponieman, Nicolás; Jun, Soon Yung; Genser, Krzysztof; Elvira, Daniel; Castro, Rodrigo

    2017-10-01

    Simulation of particle-matter interactions in complex geometries is one of the main tasks in high energy physics (HEP) research. An essential aspect of it is an accurate and efficient particle transportation in a non-uniform magnetic field, which includes the handling of volume crossings within a predefined 3D geometry. Quantized State Systems (QSS) is a family of numerical methods that provides attractive features for particle transportation processes, such as dense output (sequences of polynomial segments changing only according to accuracy-driven discrete events) and lightweight detection and handling of volume crossings (based on simple root-finding of polynomial functions). In this work we present a proof-of-concept performance comparison between a QSS-based standalone numerical solver and an application based on the Geant4 simulation toolkit, with its default Runge-Kutta based adaptive step method. In a case study with a charged particle circulating in a vacuum (with interactions with matter turned off), in a uniform magnetic field, and crossing up to 200 volume boundaries twice per turn, simulation results showed speedups of up to 6 times in favor of QSS while it being 10 times slower in the case with zero volume boundaries.

  16. Numerical simulation of nucleate boiling and heat transfer using MPL-MAFL

    Energy Technology Data Exchange (ETDEWEB)

    Han Young Yoon, Hee Cheol Kim [Korea Atomic Energy Research Inst., Taejon (Korea, Republic of); Koshizuka, Seiichi; Oka, Yoshiaki [Tokyo Univ., Tokai, Ibaraki (Japan). Nuclear Engineering Research Lab

    2000-10-01

    A mesh-free numerical method is presented for direct calculation of bubble growth. It is a combination of particle and gridless methods where the terms, 'particle' and 'gridless', refer to Lagrangian and Eulerian schemes respectively. Thus, an arbitrary-Lagrangian-Eulerian calculation is possible, in this method, with a cloud of computing points that are equivalent to the computing cells in mesh-based methods. The moving interface is traced through the Lagrangian motion of the computing points using a particle method and, at the fixed computing points, convection is calculated using a gridless method. The particle interaction model of the moving-particle semi-implicit (MPS) method is applied to the differential operators and the meshless-advection using a flow-directional local-grid (MAFL) scheme is utilized for the gridless method. A complex moving interface problems can be effectively analyzed by MPS-MAFL since the mesh is no longer used. The present method is applied to the calculation of gas-liquid two-phase flow with and without the phase change in two dimensions. The pressure and temperature gradients are ignored for the vapor region and the phase interface is treated as a free boundary. As an isothermal flow, a gas bubble rising in viscous liquids is simulated numerically and the results are compared with the empirical correlation. The energy equation is coupled with the equation of motion for the calculation of nucleate pool boiling. The numerical results are provided for the bubble growth rate, departure radius, and the heat transfer rate, which show good agreement with the experimental observations. The heat transfer mechanism associated with nucleate pool boiling is quantitatively evaluated and discussed with previous empirical studies. (author)

  17. Large shear deformation of particle gels studied by Brownian Dynamics simulations

    NARCIS (Netherlands)

    Rzepiela, A.A.; Opheusden, van J.H.J.; Vliet, van T.

    2004-01-01

    Brownian Dynamics (BD) simulations have been performed to study structure and rheology of particle gels under large shear deformation. The model incorporates soft spherical particles, and reversible flexible bond formation. Two different methods of shear deformation are discussed, namely affine and

  18. Sensitivity of Particle Size in Discrete Element Method to Particle Gas Method (DEM_PGM) Coupling in Underbody Blast Simulations

    Science.gov (United States)

    2016-06-12

    Particle Size in Discrete Element Method to Particle Gas Method (DEM_PGM) Coupling in Underbody Blast Simulations Venkatesh Babu, Kumar Kulkarni, Sanjay...buried in soil viz., (1) coupled discrete element & particle gas methods (DEM-PGM) and (2) Arbitrary Lagrangian-Eulerian (ALE), are investigated. The...DEM_PGM and identify the limitations/strengths compared to the ALE method. Discrete Element Method (DEM) can model individual particle directly, and

  19. Mesh optimization for microbial fuel cell cathodes constructed around stainless steel mesh current collectors

    KAUST Repository

    Zhang, Fang; Merrill, Matthew D.; Tokash, Justin C.; Saito, Tomonori; Cheng, Shaoan; Hickner, Michael A.; Logan, Bruce E.

    2011-01-01

    that the mesh properties of these cathodes can significantly affect performance. Cathodes made from the coarsest mesh (30-mesh) achieved the highest maximum power of 1616 ± 25 mW m-2 (normalized to cathode projected surface area; 47.1 ± 0.7 W m-3 based on liquid

  20. SUPERIMPOSED MESH PLOTTING IN MCNP

    Energy Technology Data Exchange (ETDEWEB)

    J. HENDRICKS

    2001-02-01

    The capability to plot superimposed meshes has been added to MCNP{trademark}. MCNP4C featured a superimposed mesh weight window generator which enabled users to set up geometries without having to subdivide geometric cells for variance reduction. The variance reduction was performed with weight windows on a rectangular or cylindrical mesh superimposed over the physical geometry. Experience with the new capability was favorable but also indicated that a number of enhancements would be very beneficial, particularly a means of visualizing the mesh and its values. The mathematics for plotting the mesh and its values is described here along with a description of other upgrades.

  1. GPUs, a new tool of acceleration in CFD: efficiency and reliability on smoothed particle hydrodynamics methods.

    Directory of Open Access Journals (Sweden)

    Alejandro C Crespo

    Full Text Available Smoothed Particle Hydrodynamics (SPH is a numerical method commonly used in Computational Fluid Dynamics (CFD to simulate complex free-surface flows. Simulations with this mesh-free particle method far exceed the capacity of a single processor. In this paper, as part of a dual-functioning code for either central processing units (CPUs or Graphics Processor Units (GPUs, a parallelisation using GPUs is presented. The GPU parallelisation technique uses the Compute Unified Device Architecture (CUDA of nVidia devices. Simulations with more than one million particles on a single GPU card exhibit speedups of up to two orders of magnitude over using a single-core CPU. It is demonstrated that the code achieves different speedups with different CUDA-enabled GPUs. The numerical behaviour of the SPH code is validated with a standard benchmark test case of dam break flow impacting on an obstacle where good agreement with the experimental results is observed. Both the achieved speed-ups and the quantitative agreement with experiments suggest that CUDA-based GPU programming can be used in SPH methods with efficiency and reliability.

  2. Surface mesh to voxel data registration for patient-specific anatomical modeling

    Science.gov (United States)

    de Oliveira, Júlia E. E.; Giessler, Paul; Keszei, András.; Herrler, Andreas; Deserno, Thomas M.

    2016-03-01

    Virtual Physiological Human (VPH) models are frequently used for training, planning, and performing medical procedures. The Regional Anaesthesia Simulator and Assistant (RASimAs) project has the goal of increasing the application and effectiveness of regional anesthesia (RA) by combining a simulator of ultrasound-guided and electrical nerve-stimulated RA procedures and a subject-specific assistance system through an integration of image processing, physiological models, subject-specific data, and virtual reality. Individualized models enrich the virtual training tools for learning and improving regional anaesthesia (RA) skills. Therefore, we suggest patient-specific VPH models that are composed by registering the general mesh-based models with patient voxel data-based recordings. Specifically, the pelvis region has been focused for the support of the femoral nerve block. The processing pipeline is composed of different freely available toolboxes such as MatLab, the open Simulation framework (SOFA), and MeshLab. The approach of Gilles is applied for mesh-to-voxel registration. Personalized VPH models include anatomical as well as mechanical properties of the tissues. Two commercial VPH models (Zygote and Anatomium) were used together with 34 MRI data sets. Results are presented for the skin surface and pelvic bones. Future work will extend the registration procedure to cope with all model tissue (i.e., skin, muscle, bone, vessel, nerve, fascia) in a one-step procedure and extrapolating the personalized models to body regions actually being out of the captured field of view.

  3. Simulations and measurements of coupling impedance for modern particle accelerator devices

    CERN Document Server

    AUTHOR|(CDS)2158523; Biancacci, Nicolò; Mostacci, Andrea

    In this document it has been treated the study of the coupling impedance in modern devices, already installed or not, in different particle accelerators. In the specific case: • For a device in-phase of project, several simulations for impedance calculation have been done. • For a component already realized and used, measurements of coupling impedance value have been done. Simulations are used to determine the impact of the interconnect between to magnets, designed for the future particle accelerator FCC, on the overall impedance of the machine which is about 100 km long. In particular has been done a check between theory, simulations and measurements of components already built, allowing a better and deeper study of the component we have analysed. Controls that probably will be helpful to have a clear guideline in future works. The measurements instead concern in an existing component that was already used in LHC, the longest particle accelerator ever realised on the planet, 27 km long. The coupling impe...

  4. Optimized Loading for Particle-in-cell Gyrokinetic Simulations

    International Nuclear Information System (INIS)

    Lewandowski, J.L.V.

    2004-01-01

    The problem of particle loading in particle-in-cell gyrokinetic simulations is addressed using a quadratic optimization algorithm. Optimized loading in configuration space dramatically reduces the short wavelength modes in the electrostatic potential that are partly responsible for the non-conservation of total energy; further, the long wavelength modes are resolved with good accuracy. As a result, the conservation of energy for the optimized loading is much better that the conservation of energy for the random loading. The method is valid for any geometry and can be coupled to optimization algorithms in velocity space

  5. Cache-Oblivious Mesh Layouts

    International Nuclear Information System (INIS)

    Yoon, S; Lindstrom, P; Pascucci, V; Manocha, D

    2005-01-01

    We present a novel method for computing cache-oblivious layouts of large meshes that improve the performance of interactive visualization and geometric processing algorithms. Given that the mesh is accessed in a reasonably coherent manner, we assume no particular data access patterns or cache parameters of the memory hierarchy involved in the computation. Furthermore, our formulation extends directly to computing layouts of multi-resolution and bounding volume hierarchies of large meshes. We develop a simple and practical cache-oblivious metric for estimating cache misses. Computing a coherent mesh layout is reduced to a combinatorial optimization problem. We designed and implemented an out-of-core multilevel minimization algorithm and tested its performance on unstructured meshes composed of tens to hundreds of millions of triangles. Our layouts can significantly reduce the number of cache misses. We have observed 2-20 times speedups in view-dependent rendering, collision detection, and isocontour extraction without any modification of the algorithms or runtime applications

  6. Assessment of finite element and smoothed particles hydrodynamics methods for modeling serrated chip formation in hardened steel

    Directory of Open Access Journals (Sweden)

    Usama Umer

    2016-05-01

    Full Text Available This study aims to perform comparative analyses in modeling serrated chip morphologies using traditional finite element and smoothed particles hydrodynamics methods. Although finite element models are being employed in predicting machining performance variables for the last two decades, many drawbacks and limitations exist with the current finite element models. The problems like excessive mesh distortions, high numerical cost of adaptive meshing techniques, and need of geometric chip separation criteria hinder its practical implementation in metal cutting industries. In this study, a mesh free method, namely, smoothed particles hydrodynamics, is implemented for modeling serrated chip morphology while machining AISI H13 hardened tool steel. The smoothed particles hydrodynamics models are compared with the traditional finite element models, and it has been found that the smoothed particles hydrodynamics models have good capabilities in handling large distortions and do not need any geometric or mesh-based chip separation criterion.

  7. An arbitrary curvilinear-coordinate method for particle-in-cell modeling

    International Nuclear Information System (INIS)

    Fichtl, C A; Finn, J M; Cartwright, K L

    2012-01-01

    A new approach to kinetic simulation of plasmas in complex geometries, based on the particle-in-cell (PIC) simulation method, is explored. In the two-dimensional (2D) electrostatic version of our method, called the arbitrary curvilinear-coordinate PIC method, all essential PIC operations are carried out in 2D on a uniform grid on the unit square logical domain, and mapped to a nonuniform boundary-fitted grid on the physical domain. As the resulting logical grid equations of motion are not separable, we have developed an extension of the semi-implicit modified leapfrog integration technique to preserve the symplectic nature of the logical grid particle mover. A generalized, curvilinear-coordinate formulation of Poisson's equations to solve for the electrostatic fields on the uniform logical grid is also developed. By our formulation, we compute the plasma charge density on the logical grid based on the particles' positions on the logical domain. That is, the plasma particles are weighted to the uniform logical grid and the self-consistent mean electrostatic fields obtained from the solution of the logical grid Poisson equation are interpolated to the particle positions on the logical grid. This process eliminates the complexity associated with the weighting and interpolation processes on the nonuniform physical grid and allows us to run the PIC method on arbitrary boundary-fitted meshes. (paper)

  8. Implementing particle-in-cell plasma simulation code on the BBN TC2000

    International Nuclear Information System (INIS)

    Sturtevant, J.E.; Maccabe, A.B.

    1990-01-01

    The BBN TC2000 is a multiple instruction, multiple data (MIMD) machine that combines a physically distributed memory with a logically shared memory programming environment using the unique Butterfly switch. Particle-In-Cell (PIC) plasma simulations model the interaction of charged particles with electric and magnetic fields. This paper describes the implementation of both a 1-D electrostatic and a 2 1/2-D electromagnetic PIC (particle-in-cell) plasma simulation code on a BBN TC2000. Performance is compared to implementations of the same code on the shared memory Sequent Balance and distributed memory Intel iPSC hypercube

  9. Simulating water hammer with corrective smoothed particle method

    NARCIS (Netherlands)

    Hou, Q.; Kruisbrink, A.C.H.; Tijsseling, A.S.; Keramat, A.

    2012-01-01

    The corrective smoothed particle method (CSPM) is used to simulate water hammer. The spatial derivatives in the water-hammer equations are approximated by a corrective kernel estimate. For the temporal derivatives, the Euler-forward time integration algorithm is employed. The CSPM results are in

  10. Pt deposition on carbon paper and Ti mesh substrates by surface limited redox replacement

    CSIR Research Space (South Africa)

    Modibedi, M

    2011-12-01

    Full Text Available , the material used as gas diffusion layer for proton exchange membrane fuel cells (PEMFCs) and Ti-mesh. The deposition uses multiple redox replacement of underpotentially deposited Cu used as a sacrificial metal. The morphology and particle size of the deposited...

  11. Development of modifications to the material point method for the simulation of thin membranes, compressible fluids, and their interactions

    Energy Technology Data Exchange (ETDEWEB)

    York, A.R. II [Sandia National Labs., Albuquerque, NM (United States). Engineering and Process Dept.

    1997-07-01

    The material point method (MPM) is an evolution of the particle in cell method where Lagrangian particles or material points are used to discretize the volume of a material. The particles carry properties such as mass, velocity, stress, and strain and move through a Eulerian or spatial mesh. The momentum equation is solved on the Eulerian mesh. Modifications to the material point method are developed that allow the simulation of thin membranes, compressible fluids, and their dynamic interactions. A single layer of material points through the thickness is used to represent a membrane. The constitutive equation for the membrane is applied in the local coordinate system of each material point. Validation problems are presented and numerical convergence is demonstrated. Fluid simulation is achieved by implementing a constitutive equation for a compressible, viscous, Newtonian fluid and by solution of the energy equation. The fluid formulation is validated by simulating a traveling shock wave in a compressible fluid. Interactions of the fluid and membrane are handled naturally with the method. The fluid and membrane communicate through the Eulerian grid on which forces are calculated due to the fluid and membrane stress states. Validation problems include simulating a projectile impacting an inflated airbag. In some impact simulations with the MPM, bodies may tend to stick together when separating. Several algorithms are proposed and tested that allow bodies to separate from each other after impact. In addition, several methods are investigated to determine the local coordinate system of a membrane material point without relying upon connectivity data.

  12. User Manual for the PROTEUS Mesh Tools

    Energy Technology Data Exchange (ETDEWEB)

    Smith, Micheal A. [Argonne National Lab. (ANL), Argonne, IL (United States); Shemon, Emily R. [Argonne National Lab. (ANL), Argonne, IL (United States)

    2015-06-01

    This report describes the various mesh tools that are provided with the PROTEUS code giving both descriptions of the input and output. In many cases the examples are provided with a regression test of the mesh tools. The most important mesh tools for any user to consider using are the MT_MeshToMesh.x and the MT_RadialLattice.x codes. The former allows the conversion between most mesh types handled by PROTEUS while the second allows the merging of multiple (assembly) meshes into a radial structured grid. Note that the mesh generation process is recursive in nature and that each input specific for a given mesh tool (such as .axial or .merge) can be used as “mesh” input for any of the mesh tools discussed in this manual.

  13. Anisotropic evaluation of synthetic surgical meshes.

    Science.gov (United States)

    Saberski, E R; Orenstein, S B; Novitsky, Y W

    2011-02-01

    The material properties of meshes used in hernia repair contribute to the overall mechanical behavior of the repair. The anisotropic potential of synthetic meshes, representing a difference in material properties (e.g., elasticity) in different material axes, is not well defined to date. Haphazard orientation of anisotropic mesh material can contribute to inconsistent surgical outcomes. We aimed to characterize and compare anisotropic properties of commonly used synthetic meshes. Six different polypropylene (Trelex(®), ProLite™, Ultrapro™), polyester (Parietex™), and PTFE-based (Dualmesh(®), Infinit) synthetic meshes were selected. Longitudinal and transverse axes were defined for each mesh, and samples were cut in each axis orientation. Samples underwent uniaxial tensile testing, from which the elastic modulus (E) in each axis was determined. The degree of anisotropy (λ) was calculated as a logarithmic expression of the ratio between the elastic modulus in each axis. Five of six meshes displayed significant anisotropic behavior. Ultrapro™ and Infinit exhibited approximately 12- and 20-fold differences between perpendicular axes, respectively. Trelex(®), ProLite™, and Parietex™ were 2.3-2.4 times. Dualmesh(®) was the least anisotropic mesh, without marked difference between the axes. Anisotropy of synthetic meshes has been underappreciated. In this study, we found striking differences between elastic properties of perpendicular axes for most commonly used synthetic meshes. Indiscriminate orientation of anisotropic mesh may adversely affect hernia repairs. Proper labeling of all implants by manufacturers should be mandatory. Understanding the specific anisotropic behavior of synthetic meshes should allow surgeons to employ rational implant orientation to maximize outcomes of hernia repair.

  14. HIGH-RESOLUTION SIMULATIONS OF CONVECTION PRECEDING IGNITION IN TYPE Ia SUPERNOVAE USING ADAPTIVE MESH REFINEMENT

    International Nuclear Information System (INIS)

    Nonaka, A.; Aspden, A. J.; Almgren, A. S.; Bell, J. B.; Zingale, M.; Woosley, S. E.

    2012-01-01

    We extend our previous three-dimensional, full-star simulations of the final hours of convection preceding ignition in Type Ia supernovae to higher resolution using the adaptive mesh refinement capability of our low Mach number code, MAESTRO. We report the statistics of the ignition of the first flame at an effective 4.34 km resolution and general flow field properties at an effective 2.17 km resolution. We find that off-center ignition is likely, with radius of 50 km most favored and a likely range of 40-75 km. This is consistent with our previous coarser (8.68 km resolution) simulations, implying that we have achieved sufficient resolution in our determination of likely ignition radii. The dynamics of the last few hot spots preceding ignition suggest that a multiple ignition scenario is not likely. With improved resolution, we can more clearly see the general flow pattern in the convective region, characterized by a strong outward plume with a lower speed recirculation. We show that the convective core is turbulent with a Kolmogorov spectrum and has a lower turbulent intensity and larger integral length scale than previously thought (on the order of 16 km s –1 and 200 km, respectively), and we discuss the potential consequences for the first flames.

  15. Practical implementation of tetrahedral mesh reconstruction in emission tomography

    Science.gov (United States)

    Boutchko, R.; Sitek, A.; Gullberg, G. T.

    2013-05-01

    This paper presents a practical implementation of image reconstruction on tetrahedral meshes optimized for emission computed tomography with parallel beam geometry. Tetrahedral mesh built on a point cloud is a convenient image representation method, intrinsically three-dimensional and with a multi-level resolution property. Image intensities are defined at the mesh nodes and linearly interpolated inside each tetrahedron. For the given mesh geometry, the intensities can be computed directly from tomographic projections using iterative reconstruction algorithms with a system matrix calculated using an exact analytical formula. The mesh geometry is optimized for a specific patient using a two stage process. First, a noisy image is reconstructed on a finely-spaced uniform cloud. Then, the geometry of the representation is adaptively transformed through boundary-preserving node motion and elimination. Nodes are removed in constant intensity regions, merged along the boundaries, and moved in the direction of the mean local intensity gradient in order to provide higher node density in the boundary regions. Attenuation correction and detector geometric response are included in the system matrix. Once the mesh geometry is optimized, it is used to generate the final system matrix for ML-EM reconstruction of node intensities and for visualization of the reconstructed images. In dynamic PET or SPECT imaging, the system matrix generation procedure is performed using a quasi-static sinogram, generated by summing projection data from multiple time frames. This system matrix is then used to reconstruct the individual time frame projections. Performance of the new method is evaluated by reconstructing simulated projections of the NCAT phantom and the method is then applied to dynamic SPECT phantom and patient studies and to a dynamic microPET rat study. Tetrahedral mesh-based images are compared to the standard voxel-based reconstruction for both high and low signal-to-noise ratio

  16. Practical implementation of tetrahedral mesh reconstruction in emission tomography

    International Nuclear Information System (INIS)

    Boutchko, R; Gullberg, G T; Sitek, A

    2013-01-01

    This paper presents a practical implementation of image reconstruction on tetrahedral meshes optimized for emission computed tomography with parallel beam geometry. Tetrahedral mesh built on a point cloud is a convenient image representation method, intrinsically three-dimensional and with a multi-level resolution property. Image intensities are defined at the mesh nodes and linearly interpolated inside each tetrahedron. For the given mesh geometry, the intensities can be computed directly from tomographic projections using iterative reconstruction algorithms with a system matrix calculated using an exact analytical formula. The mesh geometry is optimized for a specific patient using a two stage process. First, a noisy image is reconstructed on a finely-spaced uniform cloud. Then, the geometry of the representation is adaptively transformed through boundary-preserving node motion and elimination. Nodes are removed in constant intensity regions, merged along the boundaries, and moved in the direction of the mean local intensity gradient in order to provide higher node density in the boundary regions. Attenuation correction and detector geometric response are included in the system matrix. Once the mesh geometry is optimized, it is used to generate the final system matrix for ML-EM reconstruction of node intensities and for visualization of the reconstructed images. In dynamic PET or SPECT imaging, the system matrix generation procedure is performed using a quasi-static sinogram, generated by summing projection data from multiple time frames. This system matrix is then used to reconstruct the individual time frame projections. Performance of the new method is evaluated by reconstructing simulated projections of the NCAT phantom and the method is then applied to dynamic SPECT phantom and patient studies and to a dynamic microPET rat study. Tetrahedral mesh-based images are compared to the standard voxel-based reconstruction for both high and low signal-to-noise ratio

  17. To mesh or not to mesh: a review of pelvic organ reconstructive surgery

    Directory of Open Access Journals (Sweden)

    Dällenbach P

    2015-04-01

    Full Text Available Patrick Dällenbach Department of Gynecology and Obstetrics, Division of Gynecology, Urogynecology Unit, Geneva University Hospitals, Geneva, Switzerland Abstract: Pelvic organ prolapse (POP is a major health issue with a lifetime risk of undergoing at least one surgical intervention estimated at close to 10%. In the 1990s, the risk of reoperation after primary standard vaginal procedure was estimated to be as high as 30% to 50%. In order to reduce the risk of relapse, gynecological surgeons started to use mesh implants in pelvic organ reconstructive surgery with the emergence of new complications. Recent studies have nevertheless shown that the risk of POP recurrence requiring reoperation is lower than previously estimated, being closer to 10% rather than 30%. The development of mesh surgery – actively promoted by the marketing industry – was tremendous during the past decade, and preceded any studies supporting its benefit for our patients. Randomized trials comparing the use of mesh to native tissue repair in POP surgery have now shown better anatomical but similar functional outcomes, and meshes are associated with more complications, in particular for transvaginal mesh implants. POP is not a life-threatening condition, but a functional problem that impairs quality of life for women. The old adage “primum non nocere” is particularly appropriate when dealing with this condition which requires no treatment when asymptomatic. It is currently admitted that a certain degree of POP is physiological with aging when situated above the landmark of the hymen. Treatment should be individualized and the use of mesh needs to be selective and appropriate. Mesh implants are probably an important tool in pelvic reconstructive surgery, but the ideal implant has yet to be found. The indications for its use still require caution and discernment. This review explores the reasons behind the introduction of mesh augmentation in POP surgery, and aims to

  18. A general concurrent algorithm for plasma particle-in-cell simulation codes

    International Nuclear Information System (INIS)

    Liewer, P.C.; Decyk, V.K.

    1989-01-01

    We have developed a new algorithm for implementing plasma particle-in-cell (PIC) simulation codes on concurrent processors with distributed memory. This algorithm, named the general concurrent PIC algorithm (GCPIC), has been used to implement an electrostatic PIC code on the 33-node JPL Mark III Hypercube parallel computer. To decompose at PIC code using the GCPIC algorithm, the physical domain of the particle simulation is divided into sub-domains, equal in number to the number of processors, such that all sub-domains have roughly equal numbers of particles. For problems with non-uniform particle densities, these sub-domains will be of unequal physical size. Each processor is assigned a sub-domain and is responsible for updating the particles in its sub-domain. This algorithm has led to a a very efficient parallel implementation of a well-benchmarked 1-dimensional PIC code. The dominant portion of the code, updating the particle positions and velocities, is nearly 100% efficient when the number of particles is increased linearly with the number of hypercube processors used so that the number of particles per processor is constant. For example, the increase in time spent updating particles in going from a problem with 11,264 particles run on 1 processor to 360,448 particles on 32 processors was only 3% (parallel efficiency of 97%). Although implemented on a hypercube concurrent computer, this algorithm should also be efficient for PIC codes on other parallel architectures and for large PIC codes on sequential computers where part of the data must reside on external disks. copyright 1989 Academic Press, Inc

  19. Simulating Sand Behavior through Terrain Subdivision and Particle Refinement

    Science.gov (United States)

    Clothier, M.

    2013-12-01

    Advances in computer graphics, GPUs, and parallel processing hardware have provided researchers with new methods to visualize scientific data. In fact, these advances have spurred new research opportunities between computer graphics and other disciplines, such as Earth sciences. Through collaboration, Earth and planetary scientists have benefited by using these advances in hardware technology to process large amounts of data for visualization and analysis. At Oregon State University, we are collaborating with the Oregon Space Grant and IGERT Ecosystem Informatics programs to investigate techniques for simulating the behavior of sand. In addition, we have also been collaborating with the Jet Propulsion Laboratory's DARTS Lab to exchange ideas on our research. The DARTS Lab specializes in the simulation of planetary vehicles, such as the Mars rovers. One aspect of their work is testing these vehicles in a virtual "sand box" to test their performance in different environments. Our research builds upon this idea to create a sand simulation framework to allow for more complex and diverse environments. As a basis for our framework, we have focused on planetary environments, such as the harsh, sandy regions on Mars. To evaluate our framework, we have used simulated planetary vehicles, such as a rover, to gain insight into the performance and interaction between the surface sand and the vehicle. Unfortunately, simulating the vast number of individual sand particles and their interaction with each other has been a computationally complex problem in the past. However, through the use of high-performance computing, we have developed a technique to subdivide physically active terrain regions across a large landscape. To achieve this, we only subdivide terrain regions where sand particles are actively participating with another object or force, such as a rover wheel. This is similar to a Level of Detail (LOD) technique, except that the density of subdivisions are determined by

  20. 6th International Meshing Roundtable '97

    Energy Technology Data Exchange (ETDEWEB)

    White, D.

    1997-09-01

    The goal of the 6th International Meshing Roundtable is to bring together researchers and developers from industry, academia, and government labs in a stimulating, open environment for the exchange of technical information related to the meshing process. In the pas~ the Roundtable has enjoyed significant participation born each of these groups from a wide variety of countries. The Roundtable will consist of technical presentations from contributed papers and abstracts, two invited speakers, and two invited panels of experts discussing topics related to the development and use of automatic mesh generation tools. In addition, this year we will feature a "Bring Your Best Mesh" competition and poster session to encourage discussion and participation from a wide variety of mesh generation tool users. The schedule and evening social events are designed to provide numerous opportunities for informal dialog. A proceedings will be published by Sandia National Laboratories and distributed at the Roundtable. In addition, papers of exceptionally high quaIity will be submitted to a special issue of the International Journal of Computational Geometry and Applications. Papers and one page abstracts were sought that present original results on the meshing process. Potential topics include but are got limited to: Unstructured triangular and tetrahedral mesh generation Unstructured quadrilateral and hexahedral mesh generation Automated blocking and structured mesh generation Mixed element meshing Surface mesh generation Geometry decomposition and clean-up techniques Geometry modification techniques related to meshing Adaptive mesh refinement and mesh quality control Mesh visualization Special purpose meshing algorithms for particular applications Theoretical or novel ideas with practical potential Technical presentations from industrial researchers.

  1. User Manual for the PROTEUS Mesh Tools

    Energy Technology Data Exchange (ETDEWEB)

    Smith, Micheal A. [Argonne National Lab. (ANL), Argonne, IL (United States); Shemon, Emily R [Argonne National Lab. (ANL), Argonne, IL (United States)

    2016-09-19

    PROTEUS is built around a finite element representation of the geometry for visualization. In addition, the PROTEUS-SN solver was built to solve the even-parity transport equation on a finite element mesh provided as input. Similarly, PROTEUS-MOC and PROTEUS-NEMO were built to apply the method of characteristics on unstructured finite element meshes. Given the complexity of real world problems, experience has shown that using commercial mesh generator to create rather simple input geometries is overly complex and slow. As a consequence, significant effort has been put into place to create multiple codes that help assist in the mesh generation and manipulation. There are three input means to create a mesh in PROTEUS: UFMESH, GRID, and NEMESH. At present, the UFMESH is a simple way to generate two-dimensional Cartesian and hexagonal fuel assembly geometries. The UFmesh input allows for simple assembly mesh generation while the GRID input allows the generation of Cartesian, hexagonal, and regular triangular structured grid geometry options. The NEMESH is a way for the user to create their own mesh or convert another mesh file format into a PROTEUS input format. Given that one has an input mesh format acceptable for PROTEUS, we have constructed several tools which allow further mesh and geometry construction (i.e. mesh extrusion and merging). This report describes the various mesh tools that are provided with the PROTEUS code giving both descriptions of the input and output. In many cases the examples are provided with a regression test of the mesh tools. The most important mesh tools for any user to consider using are the MT_MeshToMesh.x and the MT_RadialLattice.x codes. The former allows the conversion between most mesh types handled by PROTEUS while the second allows the merging of multiple (assembly) meshes into a radial structured grid. Note that the mesh generation process is recursive in nature and that each input specific for a given mesh tool (such as .axial

  2. Particle-in-Cell Laser-Plasma Simulation on Xeon Phi Coprocessors

    OpenAIRE

    Surmin, I. A.; Bastrakov, S. I.; Efimenko, E. S.; Gonoskov, A. A.; Korzhimanov, A. V.; Meyerov, I. B.

    2015-01-01

    This paper concerns development of a high-performance implementation of the Particle-in-Cell method for plasma simulation on Intel Xeon Phi coprocessors. We discuss suitability of the method for Xeon Phi architecture and present our experience of porting and optimization of the existing parallel Particle-in-Cell code PICADOR. Direct porting with no code modification gives performance on Xeon Phi close to 8-core CPU on a benchmark problem with 50 particles per cell. We demonstrate step-by-step...

  3. Particle-in-cell simulations on spontaneous thermal magnetic field fluctuations

    Energy Technology Data Exchange (ETDEWEB)

    Simões, F. J. R. Jr.; Pavan, J. [Instituto de Física e Matemática, UFPel, Pelotas, RS (Brazil); Gaelzer, R.; Ziebell, L. F. [Instituto de Física, UFRGS, Porto Alegre, RS (Brazil); Yoon, P. H. [Institute for Physical Science and Technology, University of Maryland, College Park, Maryland 20742 (United States)

    2013-10-15

    In this paper an electromagnetic particle code is used to investigate the spontaneous thermal emission. Specifically we perform particle-in-cell simulations employing a non-relativistic isotropic Maxwellian particle distribution to show that thermal fluctuations are related to the origin of spontaneous magnetic field fluctuation. These thermal fluctuations can become seed for further amplification mechanisms and thus be considered at the origin of the cosmological magnetic field, at microgauss levels. Our numerical results are in accordance with theoretical results presented in the literature.

  4. A highly efficient sharp-interface immersed boundary method with adaptive mesh refinement for bio-inspired flow simulations

    Science.gov (United States)

    Deng, Xiaolong; Dong, Haibo

    2017-11-01

    Developing a high-fidelity, high-efficiency numerical method for bio-inspired flow problems with flow-structure interaction is important for understanding related physics and developing many bio-inspired technologies. To simulate a fast-swimming big fish with multiple finlets or fish schooling, we need fine grids and/or a big computational domain, which are big challenges for 3-D simulations. In current work, based on the 3-D finite-difference sharp-interface immersed boundary method for incompressible flows (Mittal et al., JCP 2008), we developed an octree-like Adaptive Mesh Refinement (AMR) technique to enhance the computational ability and increase the computational efficiency. The AMR is coupled with a multigrid acceleration technique and a MPI +OpenMP hybrid parallelization. In this work, different AMR layers are treated separately and the synchronization is performed in the buffer regions and iterations are performed for the convergence of solution. Each big region is calculated by a MPI process which then uses multiple OpenMP threads for further acceleration, so that the communication cost is reduced. With these acceleration techniques, various canonical and bio-inspired flow problems with complex boundaries can be simulated accurately and efficiently. This work is supported by the MURI Grant Number N00014-14-1-0533 and NSF Grant CBET-1605434.

  5. Simulation and scaling analysis of a spherical particle-laden blast wave

    Science.gov (United States)

    Ling, Y.; Balachandar, S.

    2018-05-01

    A spherical particle-laden blast wave, generated by a sudden release of a sphere of compressed gas-particle mixture, is investigated by numerical simulation. The present problem is a multiphase extension of the classic finite-source spherical blast-wave problem. The gas-particle flow can be fully determined by the initial radius of the spherical mixture and the properties of gas and particles. In many applications, the key dimensionless parameters, such as the initial pressure and density ratios between the compressed gas and the ambient air, can vary over a wide range. Parametric studies are thus performed to investigate the effects of these parameters on the characteristic time and spatial scales of the particle-laden blast wave, such as the maximum radius the contact discontinuity can reach and the time when the particle front crosses the contact discontinuity. A scaling analysis is conducted to establish a scaling relation between the characteristic scales and the controlling parameters. A length scale that incorporates the initial pressure ratio is proposed, which is able to approximately collapse the simulation results for the gas flow for a wide range of initial pressure ratios. This indicates that an approximate similarity solution for a spherical blast wave exists, which is independent of the initial pressure ratio. The approximate scaling is also valid for the particle front if the particles are small and closely follow the surrounding gas.

  6. Simulation and scaling analysis of a spherical particle-laden blast wave

    Science.gov (United States)

    Ling, Y.; Balachandar, S.

    2018-02-01

    A spherical particle-laden blast wave, generated by a sudden release of a sphere of compressed gas-particle mixture, is investigated by numerical simulation. The present problem is a multiphase extension of the classic finite-source spherical blast-wave problem. The gas-particle flow can be fully determined by the initial radius of the spherical mixture and the properties of gas and particles. In many applications, the key dimensionless parameters, such as the initial pressure and density ratios between the compressed gas and the ambient air, can vary over a wide range. Parametric studies are thus performed to investigate the effects of these parameters on the characteristic time and spatial scales of the particle-laden blast wave, such as the maximum radius the contact discontinuity can reach and the time when the particle front crosses the contact discontinuity. A scaling analysis is conducted to establish a scaling relation between the characteristic scales and the controlling parameters. A length scale that incorporates the initial pressure ratio is proposed, which is able to approximately collapse the simulation results for the gas flow for a wide range of initial pressure ratios. This indicates that an approximate similarity solution for a spherical blast wave exists, which is independent of the initial pressure ratio. The approximate scaling is also valid for the particle front if the particles are small and closely follow the surrounding gas.

  7. Analisis Kinerja EIGRP dan OSPF pada Topologi Ring dan Mesh

    Directory of Open Access Journals (Sweden)

    DWI ARYANTA

    2014-01-01

    Full Text Available ABSTRAK EIGRP (Enhanced Interior Gateway Routing Protocol dan OSPF (Open Shortest Path Fisrt adalah routing protokol yang banyak digunakan pada suatu jaringan komputer. EIGRP hanya dapat digunakan pada perangkat Merk CISCO, sedangkan OSPF dapat digunakan pada semua merk jaringan. Pada penelitian ini dibandingkan delay dan rute dari kedua routing protokol yang diimplementasikan pada topologi Ring dan Mesh. Cisco Packet Tracer 5.3 digunakan untuk mensimulasikan kedua routing protokol ini. Skenario pertama adalah perancangan jaringan kemudian dilakukan pengujian waktu delay 100 kali dalam 5 kasus. Skenario kedua dilakukan pengujian trace route untuk mengetahui jalur yang dilewati paket data lalu memutus link utama. Pada skenario kedua juga dilakukan perbandingan nilai metric dan cost hasil simulasi dengan perhitungan rumus. Skenario ketiga dilakukan pengujian waktu konvergensi untuk setiap routing protokol pada setiap topologi. Hasilnya EIGRP lebih cepat 386 µs daripada OSPF untuk topologi Ring sedangkan OSPF lebih cepat 453 µs daripada EIGRP untuk topologi Mesh. Hasil trace route menunjukan rute yang dipilih oleh routing protokol yaitu nilai metric dan cost yang terkecil. Waktu konvergensi rata-rata topologi Ring pada EIGRP sebesar 12,75 detik dan 34,5 detik pada OSPF sedangkan topologi Mesh di EIGRP sebesar 13 detik dan 35,25 detik di OSPF. Kata Kunci: EIGRP, OSPF, Packet Tracer 5.3, Ring, Mesh, Konvergensi ABSTRACT EIGRP (Enhanced Interior Gateway Routing Protocol and OSPF (Open Shortest Path Fisrt is the routing protocol that is widely used in a computer network. EIGRP can only be used on devices Brand CISCO, while OSPF can be used on all brands of network. In this study comparison of both the delay and the routing protocol implemented on Ring and Mesh topology. Cisco Packet Tracer 5.3 is used to simulate both the routing protocol. The first scenario is the design of the network and then do the test of time delay 100 times in 5 cases. The

  8. Analisis Kinerja EIGRP dan OSPF pada Topologi Ring dan Mesh

    Directory of Open Access Journals (Sweden)

    DWI ARYANTA

    2016-02-01

    Full Text Available ABSTRAK EIGRP (Enhanced Interior Gateway Routing Protocol dan OSPF (Open Shortest Path Fisrt adalah routing protokol yang banyak digunakan pada suatu jaringan komputer. EIGRP hanya dapat digunakan pada perangkat Merk CISCO, sedangkan OSPF dapat digunakan pada semua merk jaringan. Pada penelitian ini dibandingkan delay dan rute dari kedua routing protokol yang diimplementasikan pada topologi Ring dan Mesh. Cisco Packet Tracer 5.3 digunakan untuk mensimulasikan kedua routing protokol ini. Skenario pertama adalah perancangan jaringan kemudian dilakukan pengujian waktu delay 100 kali dalam 5 kasus. Skenario kedua dilakukan pengujian trace route untuk mengetahui jalur yang dilewati paket data lalu memutus link utama. Pada skenario kedua juga dilakukan perbandingan nilai metric dan cost hasil simulasi dengan perhitungan rumus. Skenario ketiga dilakukan pengujian waktu konvergensi untuk setiap routing protokol pada setiap topologi. Hasilnya EIGRP lebih cepat 386 µs daripada OSPF untuk topologi Ring sedangkan OSPF lebih cepat 453 µs daripada EIGRP untuk topologi Mesh. Hasil trace route menunjukan rute yang dipilih oleh routing protokol yaitu nilai metric dan cost yang terkecil. Waktu konvergensi rata-rata topologi Ring pada EIGRP sebesar 12,75 detik dan 34,5 detik pada OSPF sedangkan topologi Mesh di EIGRP sebesar 13 detik dan 35,25 detik di OSPF. Kata Kunci : EIGRP, OSPF, Packet Tracer 5.3, Ring, Mesh, Konvergensi ABSTRACT EIGRP (Enhanced Interior Gateway Routing Protocol and OSPF (Open Shortest Path Fisrt is the routing protocol that is widely used in a computer network. EIGRP can only be used on devices Brand CISCO, while OSPF can be used on all brands of network. In this study comparison of both the delay and the routing protocol implemented on Ring and Mesh topology. Cisco Packet Tracer 5.3 is used to simulate both the routing protocol. The first scenario is the design of the network and then do the test of time delay 100 times in 5 cases. The

  9. Management of complications of mesh surgery.

    Science.gov (United States)

    Lee, Dominic; Zimmern, Philippe E

    2015-07-01

    Transvaginal placements of synthetic mid-urethral slings and vaginal meshes have largely superseded traditional tissue repairs in the current era because of presumed efficacy and ease of implant with device 'kits'. The use of synthetic material has generated novel complications including mesh extrusion, pelvic and vaginal pain and mesh contraction. In this review, our aim is to discuss the management, surgical techniques and outcomes associated with mesh removal. Recent publications have seen an increase in presentation of these mesh-related complications, and reports from multiple tertiary centers have suggested that not all patients benefit from surgical intervention. Although the true incidence of mesh complications is unknown, recent publications can serve to guide physicians and inform patients of the surgical outcomes from mesh-related complications. In addition, the literature highlights the growing need for a registry to account for a more accurate reporting of these events and to counsel patients on the risk and benefits before proceeding with mesh surgeries.

  10. User Manual for the PROTEUS Mesh Tools

    International Nuclear Information System (INIS)

    Smith, Micheal A.; Shemon, Emily R.

    2015-01-01

    This report describes the various mesh tools that are provided with the PROTEUS code giving both descriptions of the input and output. In many cases the examples are provided with a regression test of the mesh tools. The most important mesh tools for any user to consider using are the MT M eshToMesh.x and the MT R adialLattice.x codes. The former allows the conversion between most mesh types handled by PROTEUS while the second allows the merging of multiple (assembly) meshes into a radial structured grid. Note that the mesh generation process is recursive in nature and that each input specific for a given mesh tool (such as .axial or .merge) can be used as ''mesh'' input for any of the mesh tools discussed in this manual.

  11. To mesh or not to mesh: a review of pelvic organ reconstructive surgery

    Science.gov (United States)

    Dällenbach, Patrick

    2015-01-01

    Pelvic organ prolapse (POP) is a major health issue with a lifetime risk of undergoing at least one surgical intervention estimated at close to 10%. In the 1990s, the risk of reoperation after primary standard vaginal procedure was estimated to be as high as 30% to 50%. In order to reduce the risk of relapse, gynecological surgeons started to use mesh implants in pelvic organ reconstructive surgery with the emergence of new complications. Recent studies have nevertheless shown that the risk of POP recurrence requiring reoperation is lower than previously estimated, being closer to 10% rather than 30%. The development of mesh surgery – actively promoted by the marketing industry – was tremendous during the past decade, and preceded any studies supporting its benefit for our patients. Randomized trials comparing the use of mesh to native tissue repair in POP surgery have now shown better anatomical but similar functional outcomes, and meshes are associated with more complications, in particular for transvaginal mesh implants. POP is not a life-threatening condition, but a functional problem that impairs quality of life for women. The old adage “primum non nocere” is particularly appropriate when dealing with this condition which requires no treatment when asymptomatic. It is currently admitted that a certain degree of POP is physiological with aging when situated above the landmark of the hymen. Treatment should be individualized and the use of mesh needs to be selective and appropriate. Mesh implants are probably an important tool in pelvic reconstructive surgery, but the ideal implant has yet to be found. The indications for its use still require caution and discernment. This review explores the reasons behind the introduction of mesh augmentation in POP surgery, and aims to clarify the risks, benefits, and the recognized indications for its use. PMID:25848324

  12. N-MODY: A Code for Collisionless N-body Simulations in Modified Newtonian Dynamics

    Science.gov (United States)

    Londrillo, Pasquale; Nipoti, Carlo

    2011-02-01

    N-MODY is a parallel particle-mesh code for collisionless N-body simulations in modified Newtonian dynamics (MOND). N-MODY is based on a numerical potential solver in spherical coordinates that solves the non-linear MOND field equation, and is ideally suited to simulate isolated stellar systems. N-MODY can be used also to compute the MOND potential of arbitrary static density distributions. A few applications of N-MODY indicate that some astrophysically relevant dynamical processes are profoundly different in MOND and in Newtonian gravity with dark matter.

  13. Streaming Compression of Hexahedral Meshes

    Energy Technology Data Exchange (ETDEWEB)

    Isenburg, M; Courbet, C

    2010-02-03

    We describe a method for streaming compression of hexahedral meshes. Given an interleaved stream of vertices and hexahedral our coder incrementally compresses the mesh in the presented order. Our coder is extremely memory efficient when the input stream documents when vertices are referenced for the last time (i.e. when it contains topological finalization tags). Our coder then continuously releases and reuses data structures that no longer contribute to compressing the remainder of the stream. This means in practice that our coder has only a small fraction of the whole mesh in memory at any time. We can therefore compress very large meshes - even meshes that do not file in memory. Compared to traditional, non-streaming approaches that load the entire mesh and globally reorder it during compression, our algorithm trades a less compact compressed representation for significant gains in speed, memory, and I/O efficiency. For example, on the 456k hexahedra 'blade' mesh, our coder is twice as fast and uses 88 times less memory (only 3.1 MB) with the compressed file increasing about 3% in size. We also present the first scheme for predictive compression of properties associated with hexahedral cells.

  14. Simulating variable source problems via post processing of individual particle tallies

    International Nuclear Information System (INIS)

    Bleuel, D.L.; Donahue, R.J.; Ludewigt, B.A.; Vujic, J.

    2000-01-01

    Monte Carlo is an extremely powerful method of simulating complex, three dimensional environments without excessive problem simplification. However, it is often time consuming to simulate models in which the source can be highly varied. Similarly difficult are optimization studies involving sources in which many input parameters are variable, such as particle energy, angle, and spatial distribution. Such studies are often approached using brute force methods or intelligent guesswork. One field in which these problems are often encountered is accelerator-driven Boron Neutron Capture Therapy (BNCT) for the treatment of cancers. Solving the reverse problem of determining the best neutron source for optimal BNCT treatment can be accomplished by separating the time-consuming particle-tracking process of a full Monte Carlo simulation from the calculation of the source weighting factors which is typically performed at the beginning of a Monte Carlo simulation. By post-processing these weighting factors on a recorded file of individual particle tally information, the effect of changing source variables can be realized in a matter of seconds, instead of requiring hours or days for additional complete simulations. By intelligent source biasing, any number of different source distributions can be calculated quickly from a single Monte Carlo simulation. The source description can be treated as variable and the effect of changing multiple interdependent source variables on the problem's solution can be determined. Though the focus of this study is on BNCT applications, this procedure may be applicable to any problem that involves a variable source

  15. COMPLEX OPTICAL CHARACTERIZATION OF MESH IMPLANTS AND ENCAPSULATION AREA

    Directory of Open Access Journals (Sweden)

    VALERIY P. ZAKHAROV

    2013-04-01

    Full Text Available Complex investigation of mesh implants was performed involving laser confocal microscopy, backscattered probing and OCT imaging methods. The growth of endomysium and fat tissue with microcirculation vessels was observed in the mesh encapsulation region. Confocal microscopy analysis shows that such pathologies complications such as necrosis formation and microcavities were localized in the area near implant fibers with the size compatible with fiber diameter. And the number of such formations increase with the increase of the size, number and density of microdefects on the implant surface. Results of numerical simulations show that it is possible to control implant installation up to the depth to 4 mm with a help of backscattering probing. The applicability of OCT imaging for mesh implant control was demonstrated. Special two-stage OCT image noise-reduction algorithm, including empirical mode decomposition, was proposed for contrast increase and better abnormalities visualization by halving the signal-to-noise ratio. Joint usage of backscattered probing and OCT allows to accurately ascertain implant and surrounding tissue conditions, which reduces the risk of relapse probability.

  16. RGG: Reactor geometry (and mesh) generator

    International Nuclear Information System (INIS)

    Jain, R.; Tautges, T.

    2012-01-01

    The reactor geometry (and mesh) generator RGG takes advantage of information about repeated structures in both assembly and core lattices to simplify the creation of geometry and mesh. It is released as open source software as a part of the MeshKit mesh generation library. The methodology operates in three stages. First, assembly geometry models of various types are generated by a tool called AssyGen. Next, the assembly model or models are meshed by using MeshKit tools or the CUBIT mesh generation tool-kit, optionally based on a journal file output by AssyGen. After one or more assembly model meshes have been constructed, a tool called CoreGen uses a copy/move/merge process to arrange the model meshes into a core model. In this paper, we present the current state of tools and new features in RGG. We also discuss the parallel-enabled CoreGen, which in several cases achieves super-linear speedups since the problems fit in available RAM at higher processor counts. Several RGG applications - 1/6 VHTR model, 1/4 PWR reactor core, and a full-core model for Monju - are reported. (authors)

  17. Numerical simulation of microstructure formation of suspended particles in magnetorheological fluids

    International Nuclear Information System (INIS)

    Ido, Y; Inagaki, T; Yamaguchi, T

    2010-01-01

    Microstructure formation of magnetic particles and nonmagnetic particles in magnetorheological (MR) fluids is investigated using the particle method simulation based on simplified Stokesian dynamics. Spherical nonmagnetic particles are rearranged in the field direction due to the formation of magnetic particles in chain-like clusters. Cluster formation of spherocylindrical magnetic particles forces spherical nonmagnetic particles to arrange in the direction of the field. In contrast, the spherocylindrical nonmagnetic particles, with an aspect ratio of two or three, are not sufficiently rearranged in the field direction by cluster formation of spherical magnetic particles. Even after cluster formation in the presence of a magnetic field, the uniformity of distribution of particles on the plane perpendicular to the field direction shows very little change. However, the deviation of uniformity in particle distribution is reduced when the volume fraction of magnetic particles is the same as that of nonmagnetic particles.

  18. Properties of meshes used in hernia repair: a comprehensive review of synthetic and biologic meshes.

    Science.gov (United States)

    Ibrahim, Ahmed M S; Vargas, Christina R; Colakoglu, Salih; Nguyen, John T; Lin, Samuel J; Lee, Bernard T

    2015-02-01

    Data on the mechanical properties of the adult human abdominal wall have been difficult to obtain rendering manufacture of the ideal mesh for ventral hernia repair a challenge. An ideal mesh would need to exhibit greater biomechanical strength and elasticity than that of the abdominal wall. The aim of this study is to quantitatively compare the biomechanical properties of the most commonly used synthetic and biologic meshes in ventral hernia repair and presents a comprehensive literature review. A narrative review of the literature was performed using the PubMed database spanning articles from 1982 to 2012 including a review of company Web sites to identify all available information relating to the biomechanical properties of various synthetic and biologic meshes used in ventral hernia repair. There exist differences in the mechanical properties and the chemical nature of different meshes. In general, most synthetic materials have greater stiffness and elasticity than what is required for abdominal wall reconstruction; however, each exhibits unique properties that may be beneficial for clinical use. On the contrary, biologic meshes are more elastic but less stiff and with a lower tensile strength than their synthetic counterparts. The current standard of practice for the treatment of ventral hernias is the use of permanent synthetic mesh material. Recently, biologic meshes have become more frequently used. Most meshes exhibit biomechanical properties over the known abdominal wall thresholds. Augmenting strength requires increasing amounts of material contributing to more stiffness and foreign body reaction, which is not necessarily an advantage. Thieme Medical Publishers 333 Seventh Avenue, New York, NY 10001, USA.

  19. Parallel pic plasma simulation through particle decomposition techniques

    International Nuclear Information System (INIS)

    Briguglio, S.; Vlad, G.; Di Martino, B.; Naples, Univ. 'Federico II'

    1998-02-01

    Particle-in-cell (PIC) codes are among the major candidates to yield a satisfactory description of the detail of kinetic effects, such as the resonant wave-particle interaction, relevant in determining the transport mechanism in magnetically confined plasmas. A significant improvement of the simulation performance of such codes con be expected from parallelization, e.g., by distributing the particle population among several parallel processors. Parallelization of a hybrid magnetohydrodynamic-gyrokinetic code has been accomplished within the High Performance Fortran (HPF) framework, and tested on the IBM SP2 parallel system, using a 'particle decomposition' technique. The adopted technique requires a moderate effort in porting the code in parallel form and results in intrinsic load balancing and modest inter processor communication. The performance tests obtained confirm the hypothesis of high effectiveness of the strategy, if targeted towards moderately parallel architectures. Optimal use of resources is also discussed with reference to a specific physics problem [it

  20. Computational Particle Dynamic Simulations on Multicore Processors (CPDMu) Final Report Phase I

    Energy Technology Data Exchange (ETDEWEB)

    Schmalz, Mark S

    2011-07-24

    Statement of Problem - Department of Energy has many legacy codes for simulation of computational particle dynamics and computational fluid dynamics applications that are designed to run on sequential processors and are not easily parallelized. Emerging high-performance computing architectures employ massively parallel multicore architectures (e.g., graphics processing units) to increase throughput. Parallelization of legacy simulation codes is a high priority, to achieve compatibility, efficiency, accuracy, and extensibility. General Statement of Solution - A legacy simulation application designed for implementation on mainly-sequential processors has been represented as a graph G. Mathematical transformations, applied to G, produce a graph representation {und G} for a high-performance architecture. Key computational and data movement kernels of the application were analyzed/optimized for parallel execution using the mapping G {yields} {und G}, which can be performed semi-automatically. This approach is widely applicable to many types of high-performance computing systems, such as graphics processing units or clusters comprised of nodes that contain one or more such units. Phase I Accomplishments - Phase I research decomposed/profiled computational particle dynamics simulation code for rocket fuel combustion into low and high computational cost regions (respectively, mainly sequential and mainly parallel kernels), with analysis of space and time complexity. Using the research team's expertise in algorithm-to-architecture mappings, the high-cost kernels were transformed, parallelized, and implemented on Nvidia Fermi GPUs. Measured speedups (GPU with respect to single-core CPU) were approximately 20-32X for realistic model parameters, without final optimization. Error analysis showed no loss of computational accuracy. Commercial Applications and Other Benefits - The proposed research will constitute a breakthrough in solution of problems related to efficient