Parallel Adaptive Mesh Refinement
Diachin, L; Hornung, R; Plassmann, P; WIssink, A
2005-03-04
As large-scale, parallel computers have become more widely available and numerical models and algorithms have advanced, the range of physical phenomena that can be simulated has expanded dramatically. Many important science and engineering problems exhibit solutions with localized behavior where highly-detailed salient features or large gradients appear in certain regions which are separated by much larger regions where the solution is smooth. Examples include chemically-reacting flows with radiative heat transfer, high Reynolds number flows interacting with solid objects, and combustion problems where the flame front is essentially a two-dimensional sheet occupying a small part of a three-dimensional domain. Modeling such problems numerically requires approximating the governing partial differential equations on a discrete domain, or grid. Grid spacing is an important factor in determining the accuracy and cost of a computation. A fine grid may be needed to resolve key local features while a much coarser grid may suffice elsewhere. Employing a fine grid everywhere may be inefficient at best and, at worst, may make an adequately resolved simulation impractical. Moreover, the location and resolution of fine grid required for an accurate solution is a dynamic property of a problem's transient features and may not be known a priori. Adaptive mesh refinement (AMR) is a technique that can be used with both structured and unstructured meshes to adjust local grid spacing dynamically to capture solution features with an appropriate degree of resolution. Thus, computational resources can be focused where and when they are needed most to efficiently achieve an accurate solution without incurring the cost of a globally-fine grid. Figure 1.1 shows two example computations using AMR; on the left is a structured mesh calculation of a impulsively-sheared contact surface and on the right is the fuselage and volume discretization of an RAH-66 Comanche helicopter [35]. Note the
Adaptive mesh refinement in titanium
Colella, Phillip; Wen, Tong
2005-01-21
In this paper, we evaluate Titanium's usability as a high-level parallel programming language through a case study, where we implement a subset of Chombo's functionality in Titanium. Chombo is a software package applying the Adaptive Mesh Refinement methodology to numerical Partial Differential Equations at the production level. In Chombo, the library approach is used to parallel programming (C++ and Fortran, with MPI), whereas Titanium is a Java dialect designed for high-performance scientific computing. The performance of our implementation is studied and compared with that of Chombo in solving Poisson's equation based on two grid configurations from a real application. Also provided are the counts of lines of code from both sides.
Adaptive mesh refinement for storm surge
Mandli, Kyle T.
2014-03-01
An approach to utilizing adaptive mesh refinement algorithms for storm surge modeling is proposed. Currently numerical models exist that can resolve the details of coastal regions but are often too costly to be run in an ensemble forecasting framework without significant computing resources. The application of adaptive mesh refinement algorithms substantially lowers the computational cost of a storm surge model run while retaining much of the desired coastal resolution. The approach presented is implemented in the GeoClaw framework and compared to ADCIRC for Hurricane Ike along with observed tide gauge data and the computational cost of each model run. © 2014 Elsevier Ltd.
Adaptive Mesh Refinement for Storm Surge
Mandli, Kyle T
2014-01-01
An approach to utilizing adaptive mesh refinement algorithms for storm surge modeling is proposed. Currently numerical models exist that can resolve the details of coastal regions but are often too costly to be run in an ensemble forecasting framework without significant computing resources. The application of adaptive mesh refinement algorithms substantially lowers the computational cost of a storm surge model run while retaining much of the desired coastal resolution. The approach presented is implemented in the \\geoclaw framework and compared to \\adcirc for Hurricane Ike along with observed tide gauge data and the computational cost of each model run.
Relativistic MHD with Adaptive Mesh Refinement
Anderson, M; Liebling, S L; Neilsen, D; Anderson, Matthew; Hirschmann, Eric; Liebling, Steven L.; Neilsen, David
2006-01-01
We solve the relativistic magnetohydrodynamics (MHD) equations using a finite difference Convex ENO method (CENO) in 3+1 dimensions within a distributed parallel adaptive mesh refinement (AMR) infrastructure. In flat space we examine a Balsara blast wave problem along with a spherical blast wave and a relativistic rotor test both with unigrid and AMR simulations. The AMR simulations substantially improve performance while reproducing the resolution equivalent unigrid simulation results. We also investigate the impact of hyperbolic divergence cleaning for the spherical blast wave and relativistic rotor. We include unigrid and mesh refinement parallel performance measurements for the spherical blast wave.
Parallel object-oriented adaptive mesh refinement
Balsara, D.; Quinlan, D.J.
1997-04-01
In this paper we study adaptive mesh refinement (AMR) for elliptic and hyperbolic systems. We use the Asynchronous Fast Adaptive Composite Grid Method (AFACX), a parallel algorithm based upon the of Fast Adaptive Composite Grid Method (FAC) as a test case of an adaptive elliptic solver. For our hyperbolic system example we use TVD and ENO schemes for solving the Euler and MHD equations. We use the structured grid load balancer MLB as a tool for obtaining a load balanced distribution in a parallel environment. Parallel adaptive mesh refinement poses difficulties in expressing both the basic single grid solver, whether elliptic or hyperbolic, in a fashion that parallelizes seamlessly. It also requires that these basic solvers work together within the adaptive mesh refinement algorithm which uses the single grid solvers as one part of its adaptive solution process. We show that use of AMR++, an object-oriented library within the OVERTURE Framework, simplifies the development of AMR applications. Parallel support is provided and abstracted through the use of the P++ parallel array class.
GRChombo: Numerical relativity with adaptive mesh refinement
Clough, Katy; Figueras, Pau; Finkel, Hal; Kunesch, Markus; Lim, Eugene A.; Tunyasuvunakool, Saran
2015-12-01
In this work, we introduce {\\mathtt{GRChombo}}: a new numerical relativity code which incorporates full adaptive mesh refinement (AMR) using block structured Berger-Rigoutsos grid generation. The code supports non-trivial ‘many-boxes-in-many-boxes’ mesh hierarchies and massive parallelism through the message passing interface. {\\mathtt{GRChombo}} evolves the Einstein equation using the standard BSSN formalism, with an option to turn on CCZ4 constraint damping if required. The AMR capability permits the study of a range of new physics which has previously been computationally infeasible in a full 3 + 1 setting, while also significantly simplifying the process of setting up the mesh for these problems. We show that {\\mathtt{GRChombo}} can stably and accurately evolve standard spacetimes such as binary black hole mergers and scalar collapses into black holes, demonstrate the performance characteristics of our code, and discuss various physics problems which stand to benefit from the AMR technique.
Relativistic MHD with adaptive mesh refinement
Anderson, Matthew [Department of Physics and Astronomy, Louisiana State University, Baton Rouge, LA 70803-4001 (United States); Hirschmann, Eric W [Department of Physics and Astronomy, Brigham Young University, Provo, UT 84602 (United States); Liebling, Steven L [Department of Physics, Long Island University-C W Post Campus, Brookville, NY 11548 (United States); Neilsen, David [Department of Physics and Astronomy, Brigham Young University, Provo, UT 84602 (United States)
2006-11-22
This paper presents a new computer code to solve the general relativistic magnetohydrodynamics (GRMHD) equations using distributed parallel adaptive mesh refinement (AMR). The fluid equations are solved using a finite difference convex ENO method (CENO) in 3 + 1 dimensions, and the AMR is Berger-Oliger. Hyperbolic divergence cleaning is used to control the {nabla} . B = 0 constraint. We present results from three flat space tests, and examine the accretion of a fluid onto a Schwarzschild black hole, reproducing the Michel solution. The AMR simulations substantially improve performance while reproducing the resolution equivalent unigrid simulation results. Finally, we discuss strong scaling results for parallel unigrid and AMR runs.
Adaptive Mesh Refinement for Characteristic Grids
Thornburg, Jonathan
2009-01-01
I consider techniques for Berger-Oliger adaptive mesh refinement (AMR) when numerically solving partial differential equations with wave-like solutions, using characteristic (double-null) grids. Such AMR algorithms are naturally recursive, and the best-known past Berger-Oliger characteristic AMR algorithm, that of Pretorius & Lehner (J. Comp. Phys. 198 (2004), 10), recurses on individual "diamond" characteristic grid cells. This leads to the use of fine-grained memory management, with individual grid cells kept in 2-dimensional linked lists at each refinement level. This complicates the implementation and adds overhead in both space and time. Here I describe a Berger-Oliger characteristic AMR algorithm which instead recurses on null \\emph{slices}. This algorithm is very similar to the usual Cauchy Berger-Oliger algorithm, and uses relatively coarse-grained memory management, allowing entire null slices to be stored stored in contiguous arrays in memory. The algorithm is very efficient in both space and ti...
GRChombo : Numerical Relativity with Adaptive Mesh Refinement
Clough, Katy; Finkel, Hal; Kunesch, Markus; Lim, Eugene A; Tunyasuvunakool, Saran
2015-01-01
Numerical relativity has undergone a revolution in the past decade. With a well-understood mathematical formalism, and full control over the gauge modes, it is now entering an era in which the science can be properly explored. In this work, we introduce GRChombo, a new numerical relativity code written to take full advantage of modern parallel computing techniques. GRChombo's features include full adaptive mesh refinement with block structured Berger-Rigoutsos grid generation which supports non-trivial "many-boxes-in-many-boxes" meshing hierarchies, and massive parallelism through the Message Passing Interface (MPI). GRChombo evolves the Einstein equation with the standard BSSN formalism, with an option to turn on CCZ4 constraint damping if required. We show that GRChombo passes all the standard "Apples-to-Apples" code comparison tests. We also show that it can stably and accurately evolve vacuum black hole spacetimes such as binary black hole mergers, and non-vacuum spacetimes such as scalar collapses into b...
Elliptic Solvers for Adaptive Mesh Refinement Grids
Quinlan, D.J.; Dendy, J.E., Jr.; Shapira, Y.
1999-06-03
We are developing multigrid methods that will efficiently solve elliptic problems with anisotropic and discontinuous coefficients on adaptive grids. The final product will be a library that provides for the simplified solution of such problems. This library will directly benefit the efforts of other Laboratory groups. The focus of this work is research on serial and parallel elliptic algorithms and the inclusion of our black-box multigrid techniques into this new setting. The approach applies the Los Alamos object-oriented class libraries that greatly simplify the development of serial and parallel adaptive mesh refinement applications. In the final year of this LDRD, we focused on putting the software together; in particular we completed the final AMR++ library, we wrote tutorials and manuals, and we built example applications. We implemented the Fast Adaptive Composite Grid method as the principal elliptic solver. We presented results at the Overset Grid Conference and other more AMR specific conferences. We worked on optimization of serial and parallel performance and published several papers on the details of this work. Performance remains an important issue and is the subject of continuing research work.
Carpet: Adaptive Mesh Refinement for the Cactus Framework
Schnetter, Erik; Hawley, Scott; Hawke, Ian
2016-11-01
Carpet is an adaptive mesh refinement and multi-patch driver for the Cactus Framework (ascl:1102.013). Cactus is a software framework for solving time-dependent partial differential equations on block-structured grids, and Carpet acts as driver layer providing adaptive mesh refinement, multi-patch capability, as well as parallelization and efficient I/O.
Adaptive mesh refinement for stochastic reaction-diffusion processes
Bayati, Basil; Chatelain, Philippe; Koumoutsakos, Petros
2011-01-01
We present an algorithm for adaptive mesh refinement applied to mesoscopic stochastic simulations of spatially evolving reaction-diffusion processes. The transition rates for the diffusion process are derived on adaptive, locally refined structured meshes. Convergence of the diffusion process is presented and the fluctuations of the stochastic process are verified. Furthermore, a refinement criterion is proposed for the evolution of the adaptive mesh. The method is validated in simulations of reaction-diffusion processes as described by the Fisher-Kolmogorov and Gray-Scott equations.
Adjoint Methods for Guiding Adaptive Mesh Refinement in Tsunami Modeling
Davis, B. N.; LeVeque, R. J.
2016-12-01
One difficulty in developing numerical methods for tsunami modeling is the fact that solutions contain time-varying regions where much higher resolution is required than elsewhere in the domain, particularly when tracking a tsunami propagating across the ocean. The open source GeoClaw software deals with this issue by using block-structured adaptive mesh refinement to selectively refine around propagating waves. For problems where only a target area of the total solution is of interest (e.g., one coastal community), a method that allows identifying and refining the grid only in regions that influence this target area would significantly reduce the computational cost of finding a solution. In this work, we show that solving the time-dependent adjoint equation and using a suitable inner product with the forward solution allows more precise refinement of the relevant waves. We present the adjoint methodology first in one space dimension for illustration and in a broad context since it could also be used in other adaptive software, and potentially for other tsunami applications beyond adaptive refinement. We then show how this adjoint method has been integrated into the adaptive mesh refinement strategy of the open source GeoClaw software and present tsunami modeling results showing that the accuracy of the solution is maintained and the computational time required is significantly reduced through the integration of the adjoint method into adaptive mesh refinement.
Adaptive mesh refinement for shocks and material interfaces
Dai, William Wenlong [Los Alamos National Laboratory
2010-01-01
There are three kinds of adaptive mesh refinement (AMR) in structured meshes. Block-based AMR sometimes over refines meshes. Cell-based AMR treats cells cell by cell and thus loses the advantage of the nature of structured meshes. Patch-based AMR is intended to combine advantages of block- and cell-based AMR, i.e., the nature of structured meshes and sharp regions of refinement. But, patch-based AMR has its own difficulties. For example, patch-based AMR typically cannot preserve symmetries of physics problems. In this paper, we will present an approach for a patch-based AMR for hydrodynamics simulations. The approach consists of clustering, symmetry preserving, mesh continuity, flux correction, communications, management of patches, and load balance. The special features of this patch-based AMR include symmetry preserving, efficiency of refinement across shock fronts and material interfaces, special implementation of flux correction, and patch management in parallel computing environments. To demonstrate the capability of the AMR framework, we will show both two- and three-dimensional hydrodynamics simulations with many levels of refinement.
Elliptic Solvers with Adaptive Mesh Refinement on Complex Geometries
Phillip, B.
2000-07-24
Adaptive Mesh Refinement (AMR) is a numerical technique for locally tailoring the resolution computational grids. Multilevel algorithms for solving elliptic problems on adaptive grids include the Fast Adaptive Composite grid method (FAC) and its parallel variants (AFAC and AFACx). Theory that confirms the independence of the convergence rates of FAC and AFAC on the number of refinement levels exists under certain ellipticity and approximation property conditions. Similar theory needs to be developed for AFACx. The effectiveness of multigrid-based elliptic solvers such as FAC, AFAC, and AFACx on adaptively refined overlapping grids is not clearly understood. Finally, a non-trivial eye model problem will be solved by combining the power of using overlapping grids for complex moving geometries, AMR, and multilevel elliptic solvers.
Parallel Block Structured Adaptive Mesh Refinement on Graphics Processing Units
Beckingsale, D. A. [Atomic Weapons Establishment (AWE), Aldermaston (United Kingdom); Gaudin, W. P. [Atomic Weapons Establishment (AWE), Aldermaston (United Kingdom); Hornung, R. D. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Gunney, B. T. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Gamblin, T. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Herdman, J. A. [Atomic Weapons Establishment (AWE), Aldermaston (United Kingdom); Jarvis, S. A. [Atomic Weapons Establishment (AWE), Aldermaston (United Kingdom)
2014-11-17
Block-structured adaptive mesh refinement is a technique that can be used when solving partial differential equations to reduce the number of zones necessary to achieve the required accuracy in areas of interest. These areas (shock fronts, material interfaces, etc.) are recursively covered with finer mesh patches that are grouped into a hierarchy of refinement levels. Despite the potential for large savings in computational requirements and memory usage without a corresponding reduction in accuracy, AMR adds overhead in managing the mesh hierarchy, adding complex communication and data movement requirements to a simulation. In this paper, we describe the design and implementation of a native GPU-based AMR library, including: the classes used to manage data on a mesh patch, the routines used for transferring data between GPUs on different nodes, and the data-parallel operators developed to coarsen and refine mesh data. We validate the performance and accuracy of our implementation using three test problems and two architectures: an eight-node cluster, and over four thousand nodes of Oak Ridge National Laboratory’s Titan supercomputer. Our GPU-based AMR hydrodynamics code performs up to 4.87× faster than the CPU-based implementation, and has been scaled to over four thousand GPUs using a combination of MPI and CUDA.
Toward parallel, adaptive mesh refinement for chemically reacting flow simulations
Devine, K.D.; Shadid, J.N.; Salinger, A.G. Hutchinson, S.A. [Sandia National Labs., Albuquerque, NM (United States); Hennigan, G.L. [New Mexico State Univ., Las Cruces, NM (United States)
1997-12-01
Adaptive numerical methods offer greater efficiency than traditional numerical methods by concentrating computational effort in regions of the problem domain where the solution is difficult to obtain. In this paper, the authors describe progress toward adding mesh refinement to MPSalsa, a computer program developed at Sandia National laboratories to solve coupled three-dimensional fluid flow and detailed reaction chemistry systems for modeling chemically reacting flow on large-scale parallel computers. Data structures that support refinement and dynamic load-balancing are discussed. Results using uniform refinement with mesh sequencing to improve convergence to steady-state solutions are also presented. Three examples are presented: a lid driven cavity, a thermal convection flow, and a tilted chemical vapor deposition reactor.
Enzo: An Adaptive Mesh Refinement Code for Astrophysics
The Enzo Collaboration; Bryan, Greg L.; Norman, Michael L.; O'Shea, Brian W.; Abel, Tom; Wise, John H.; Turk, Matthew J.; Reynolds, Daniel R.; Collins, David C.; Wang, Peng; Skillman, Samuel W.; Smith, Britton; Harkness, Robert P.; Bordner, James; Kim, Ji-Hoon
2013-01-01
This paper describes the open-source code Enzo, which uses block-structured adaptive mesh refinement to provide high spatial and temporal resolution for modeling astrophysical fluid flows. The code is Cartesian, can be run in 1, 2, and 3 dimensions, and supports a wide variety of physics including hydrodynamics, ideal and non-ideal magnetohydrodynamics, N-body dynamics (and, more broadly, self-gravity of fluids and particles), primordial gas chemistry, optically-thin radiative cooling of prim...
The Nonlinear Sigma Model With Distributed Adaptive Mesh Refinement
Liebling, S L
2004-01-01
An adaptive mesh refinement (AMR) scheme is implemented in a distributed environment using Message Passing Interface (MPI) to find solutions to the nonlinear sigma model. Previous work studied behavior similar to black hole critical phenomena at the threshold for singularity formation in this flat space model. This work is a follow-up describing extensions to distribute the grid hierarchy and presenting tests showing the correctness of the model.
Boxlib with tiling: an adaptive mesh refinement software framework
Unat, Didem; Zhang, W.; Almgren, A.; Day, M.; Nguyen, T.; Shalf, J.
2016-01-01
In this paper we introduce a block-structured adaptive mesh refinement software framework that incorporates tiling, a well-known loop transformation. Because the multiscale, multiphysics codes built in boxlib are designed to solve complex systems at high resolution, performance on current and next generation architectures is essential. With the expectation of many more cores per node on next generation architectures, the ability to effectively utilize threads within a node is essential, and t...
Fully implicit adaptive mesh refinement algorithm for reduced MHD
Philip, Bobby; Pernice, Michael; Chacon, Luis
2006-10-01
In the macroscopic simulation of plasmas, the numerical modeler is faced with the challenge of dealing with multiple time and length scales. Traditional approaches based on explicit time integration techniques and fixed meshes are not suitable for this challenge, as such approaches prevent the modeler from using realistic plasma parameters to keep the computation feasible. We propose here a novel approach, based on implicit methods and structured adaptive mesh refinement (SAMR). Our emphasis is on both accuracy and scalability with the number of degrees of freedom. As a proof-of-principle, we focus on the reduced resistive MHD model as a basic MHD model paradigm, which is truly multiscale. The approach taken here is to adapt mature physics-based technology to AMR grids, and employ AMR-aware multilevel techniques (such as fast adaptive composite grid --FAC-- algorithms) for scalability. We demonstrate that the concept is indeed feasible, featuring near-optimal scalability under grid refinement. Results of fully-implicit, dynamically-adaptive AMR simulations in challenging dissipation regimes will be presented on a variety of problems that benefit from this capability, including tearing modes, the island coalescence instability, and the tilt mode instability. L. Chac'on et al., J. Comput. Phys. 178 (1), 15- 36 (2002) B. Philip, M. Pernice, and L. Chac'on, Lecture Notes in Computational Science and Engineering, accepted (2006)
Block-Structured Adaptive Mesh Refinement Algorithms for Vlasov Simulation
Hittinger, J A F
2012-01-01
Direct discretization of continuum kinetic equations, like the Vlasov equation, are under-utilized because the distribution function generally exists in a high-dimensional (>3D) space and computational cost increases geometrically with dimension. We propose to use high-order finite-volume techniques with block-structured adaptive mesh refinement (AMR) to reduce the computational cost. The primary complication comes from a solution state comprised of variables of different dimensions. We develop the algorithms required to extend standard single-dimension block structured AMR to the multi-dimension case. Specifically, algorithms for reduction and injection operations that transfer data between mesh hierarchies of different dimensions are explained in detail. In addition, modifications to the basic AMR algorithm that enable the use of high-order spatial and temporal discretizations are discussed. Preliminary results for a standard 1D+1V Vlasov-Poisson test problem are presented. Results indicate that there is po...
Block-structured Adaptive Mesh Refinement - Theory, Implementation and Application
Deiterding Ralf
2011-12-01
Full Text Available Structured adaptive mesh refinement (SAMR techniques can enable cutting-edge simulations of problems governed by conservation laws. Focusing on the strictly hyperbolic case, these notes explain all algorithmic and mathematical details of a technically relevant implementation tailored for distributed memory computers. An overview of the background of commonly used finite volume discretizations for gas dynamics is included and typical benchmarks to quantify accuracy and performance of the dynamically adaptive code are discussed. Large-scale simulations of shock-induced realistic combustion in non-Cartesian geometry and shock-driven fluid-structure interaction with fully coupled dynamic boundary motion demonstrate the applicability of the discussed techniques for complex scenarios.
ENZO: AN ADAPTIVE MESH REFINEMENT CODE FOR ASTROPHYSICS
Bryan, Greg L.; Turk, Matthew J. [Columbia University, Department of Astronomy, New York, NY 10025 (United States); Norman, Michael L.; Bordner, James; Xu, Hao; Kritsuk, Alexei G. [CASS, University of California, San Diego, 9500 Gilman Drive La Jolla, CA 92093-0424 (United States); O' Shea, Brian W.; Smith, Britton [Department of Physics and Astronomy, Michigan State University, East Lansing, MI 48824 (United States); Abel, Tom; Wang, Peng; Skillman, Samuel W. [Kavli Institute for Particle Astrophysics and Cosmology, Stanford University, Menlo Park, CA 94025 (United States); Wise, John H. [Center for Relativistic Astrophysics, School of Physics, Georgia Institute of Technology, 837 State Street, Atlanta, GA (United States); Reynolds, Daniel R. [Department of Mathematics, Southern Methodist University, Box 750156, Dallas, TX 75205-0156 (United States); Collins, David C. [Department of Physics, Florida State University, Tallahassee, FL 32306 (United States); Harkness, Robert P. [NICS, Oak Ridge National Laboratory, 1 Bethel Valley Rd, Oak Ridge, TN 37831 (United States); Kim, Ji-hoon [Department of Astronomy and Astrophysics, University of California, 1156 High Street, Santa Cruz, CA 95064 (United States); Kuhlen, Michael [Theoretical Astrophysics Center, University of California Berkeley, Hearst Field Annex, Berkeley, CA 94720 (United States); Goldbaum, Nathan [Institute for Astronomy, University of Edinburgh, Edinburgh EH9 3HJ (United Kingdom); Hummels, Cameron [Department of Astronomy/Steward Observatory, University of Arizona, 933 N. Cherry Ave., Tucson, AZ 85721 (United States); Tasker, Elizabeth [Physics Department, Faculty of Science, Hokkaido University, Kita-10 Nishi 8, Kita-ku, Sapporo 060-0810 (Japan); Collaboration: Enzo Collaboration; and others
2014-04-01
This paper describes the open-source code Enzo, which uses block-structured adaptive mesh refinement to provide high spatial and temporal resolution for modeling astrophysical fluid flows. The code is Cartesian, can be run in one, two, and three dimensions, and supports a wide variety of physics including hydrodynamics, ideal and non-ideal magnetohydrodynamics, N-body dynamics (and, more broadly, self-gravity of fluids and particles), primordial gas chemistry, optically thin radiative cooling of primordial and metal-enriched plasmas (as well as some optically-thick cooling models), radiation transport, cosmological expansion, and models for star formation and feedback in a cosmological context. In addition to explaining the algorithms implemented, we present solutions for a wide range of test problems, demonstrate the code's parallel performance, and discuss the Enzo collaboration's code development methodology.
Enzo: An Adaptive Mesh Refinement Code for Astrophysics
Bryan, Greg L; O'Shea, Brian W; Abel, Tom; Wise, John H; Turk, Matthew J; Reynolds, Daniel R; Collins, David C; Wang, Peng; Skillman, Samuel W; Smith, Britton; Harkness, Robert P; Bordner, James; Kim, Ji-hoon; Kuhlen, Michael; Xu, Hao; Goldbaum, Nathan; Hummels, Cameron; Kritsuk, Alexei G; Tasker, Elizabeth; Skory, Stephen; Simpson, Christine M; Hahn, Oliver; Oishi, Jeffrey S; So, Geoffrey C; Zhao, Fen; Cen, Renyue; Li, Yuan
2013-01-01
This paper describes the open-source code Enzo, which uses block-structured adaptive mesh refinement to provide high spatial and temporal resolution for modeling astrophysical fluid flows. The code is Cartesian, can be run in 1, 2, and 3 dimensions, and supports a wide variety of physics including hydrodynamics, ideal and non-ideal magnetohydrodynamics, N-body dynamics (and, more broadly, self-gravity of fluids and particles), primordial gas chemistry, optically-thin radiative cooling of primordial and metal-enriched plasmas (as well as some optically-thick cooling models), radiation transport, cosmological expansion, and models for star formation and feedback in a cosmological context. In addition to explaining the algorithms implemented, we present solutions for a wide range of test problems, demonstrate the code's parallel performance, and discuss the Enzo collaboration's code development methodology.
Hydrodynamical Adaptive Mesh Refinement Simulations of Disk Galaxies
Gibson, Brad K; Sanchez-Blazquez, Patricia; Teyssier, Romain; House, Elisa L; Brook, Chris B; Kawata, Daisuke
2008-01-01
To date, fully cosmological hydrodynamic disk simulations to redshift zero have only been undertaken with particle-based codes, such as GADGET, Gasoline, or GCD+. In light of the (supposed) limitations of traditional implementations of smoothed particle hydrodynamics (SPH), or at the very least, their respective idiosyncrasies, it is important to explore complementary approaches to the SPH paradigm to galaxy formation. We present the first high-resolution cosmological disk simulations to redshift zero using an adaptive mesh refinement (AMR)-based hydrodynamical code, in this case, RAMSES. We analyse the temporal and spatial evolution of the simulated stellar disks' vertical heating, velocity ellipsoids, stellar populations, vertical and radial abundance gradients (gas and stars), assembly/infall histories, warps/lopsideness, disk edges/truncations (gas and stars), ISM physics implementations, and compare and contrast these properties with our sample of cosmological SPH disks, generated with GCD+. These prelim...
Production-quality Tools for Adaptive Mesh RefinementVisualization
Weber, Gunther H.; Childs, Hank; Bonnell, Kathleen; Meredith,Jeremy; Miller, Mark; Whitlock, Brad; Bethel, E. Wes
2007-10-25
Adaptive Mesh Refinement (AMR) is a highly effectivesimulation method for spanning a large range of spatiotemporal scales,such as astrophysical simulations that must accommodate ranges frominterstellar to sub-planetary. Most mainstream visualization tools stilllack support for AMR as a first class data type and AMR code teams usecustom built applications for AMR visualization. The Department ofEnergy's (DOE's) Science Discovery through Advanced Computing (SciDAC)Visualization and Analytics Center for Enabling Technologies (VACET) isextending and deploying VisIt, an open source visualization tool thataccommodates AMR as a first-class data type, for use asproduction-quality, parallel-capable AMR visual data analysisinfrastructure. This effort will help science teams that use AMR-basedsimulations and who develop their own AMR visual data analysis softwareto realize cost and labor savings.
Adaptive Mesh Refinement in Reactive Transport Modeling of Subsurface Environments
Molins, S.; Day, M.; Trebotich, D.; Graves, D. T.
2015-12-01
Adaptive mesh refinement (AMR) is a numerical technique for locally adjusting the resolution of computational grids. AMR makes it possible to superimpose levels of finer grids on the global computational grid in an adaptive manner allowing for more accurate calculations locally. AMR codes rely on the fundamental concept that the solution can be computed in different regions of the domain with different spatial resolutions. AMR codes have been applied to a wide range of problem including (but not limited to): fully compressible hydrodynamics, astrophysical flows, cosmological applications, combustion, blood flow, heat transfer in nuclear reactors, and land ice and atmospheric models for climate. In subsurface applications, in particular, reactive transport modeling, AMR may be particularly useful in accurately capturing concentration gradients (hence, reaction rates) that develop in localized areas of the simulation domain. Accurate evaluation of reaction rates is critical in many subsurface applications. In this contribution, we will discuss recent applications that bring to bear AMR capabilities on reactive transport problems from the pore scale to the flood plain scale.
Parallel adaptive mesh refinement techniques for plasticity problems
Barry, W.J. [Carnegie Mellon Univ., Pittsburgh, PA (United States). Dept. of Civil and Environmental Engineering; Jones, M.T. [Virginia Polytechnic Institute, Blacksburg, VA (United States). Dept. of Electrical and Computer Engineering]|[State Univ., Blacksburg, VA (United States); Plassmann, P.E. [Argonne National Lab., IL (United States)
1997-12-31
The accurate modeling of the nonlinear properties of materials can be computationally expensive. Parallel computing offers an attractive way for solving such problems; however, the efficient use of these systems requires the vertical integration of a number of very different software components, we explore the solution of two- and three-dimensional, small-strain plasticity problems. We consider a finite-element formulation of the problem with adaptive refinement of an unstructured mesh to accurately model plastic transition zones. We present a framework for the parallel implementation of such complex algorithms. This framework, using libraries from the SUMAA3d project, allows a user to build a parallel finite-element application without writing any parallel code. To demonstrate the effectiveness of this approach on widely varying parallel architectures, we present experimental results from an IBM SP parallel computer and an ATM-connected network of Sun UltraSparc workstations. The results detail the parallel performance of the computational phases of the application during the process while the material is incrementally loaded.
3D level set methods for evolving fronts on tetrahedral meshes with adaptive mesh refinement
Morgan, Nathaniel R.; Waltz, Jacob I.
2017-05-01
The level set method is commonly used to model dynamically evolving fronts and interfaces. In this work, we present new methods for evolving fronts with a specified velocity field or in the surface normal direction on 3D unstructured tetrahedral meshes with adaptive mesh refinement (AMR). The level set field is located at the nodes of the tetrahedral cells and is evolved using new upwind discretizations of Hamilton-Jacobi equations combined with a Runge-Kutta method for temporal integration. The level set field is periodically reinitialized to a signed distance function using an iterative approach with a new upwind gradient. The details of these level set and reinitialization methods are discussed. Results from a range of numerical test problems are presented.
RAM: a Relativistic Adaptive Mesh Refinement Hydrodynamics Code
Zhang, Wei-Qun; /KIPAC, Menlo Park; MacFadyen, Andrew I.; /Princeton, Inst. Advanced Study
2005-06-06
The authors have developed a new computer code, RAM, to solve the conservative equations of special relativistic hydrodynamics (SRHD) using adaptive mesh refinement (AMR) on parallel computers. They have implemented a characteristic-wise, finite difference, weighted essentially non-oscillatory (WENO) scheme using the full characteristic decomposition of the SRHD equations to achieve fifth-order accuracy in space. For time integration they use the method of lines with a third-order total variation diminishing (TVD) Runge-Kutta scheme. They have also implemented fourth and fifth order Runge-Kutta time integration schemes for comparison. The implementation of AMR and parallelization is based on the FLASH code. RAM is modular and includes the capability to easily swap hydrodynamics solvers, reconstruction methods and physics modules. In addition to WENO they have implemented a finite volume module with the piecewise parabolic method (PPM) for reconstruction and the modified Marquina approximate Riemann solver to work with TVD Runge-Kutta time integration. They examine the difficulty of accurately simulating shear flows in numerical relativistic hydrodynamics codes. They show that under-resolved simulations of simple test problems with transverse velocity components produce incorrect results and demonstrate the ability of RAM to correctly solve these problems. RAM has been tested in one, two and three dimensions and in Cartesian, cylindrical and spherical coordinates. they have demonstrated fifth-order accuracy for WENO in one and two dimensions and performed detailed comparison with other schemes for which they show significantly lower convergence rates. Extensive testing is presented demonstrating the ability of RAM to address challenging open questions in relativistic astrophysics.
A Parallel Algorithm for Adaptive Local Refinement of Tetrahedral Meshes Using Bisection
LinBo Zhang
2009-01-01
Local mesh refinement is one of the key steps in the implementations of adaptive finite element methods. This paper presents a parallel algorithm for distributed memory parallel computers for adaptive local refinement of tetrahedral meshes using bisection. This algorithm is used in PHG, Parallel Hierarchical Grid (http: //lsec. cc. ac. cn/phg/J, a toolbox under active development for parallel adaptive finite element solutions of partial differential equations. The algorithm proposed is characterized by allowing simultaneous refinement of submeshes to arbitrary levels before synchronization between submeshes and without the need of a central coordinator process for managing new vertices. Using the concept of canonical refinement, a simple proof of the independence of the resulting mesh on the mesh partitioning is given, which is useful in better understanding the behaviour of the bisectioning refinement procedure.AMS subject classifications: 65Y05, 65N50
Schwing, Alan Michael
For computational fluid dynamics, the governing equations are solved on a discretized domain of nodes, faces, and cells. The quality of the grid or mesh can be a driving source for error in the results. While refinement studies can help guide the creation of a mesh, grid quality is largely determined by user expertise and understanding of the flow physics. Adaptive mesh refinement is a technique for enriching the mesh during a simulation based on metrics for error, impact on important parameters, or location of important flow features. This can offload from the user some of the difficult and ambiguous decisions necessary when discretizing the domain. This work explores the implementation of adaptive mesh refinement in an implicit, unstructured, finite-volume solver. Consideration is made for applying modern computational techniques in the presence of hanging nodes and refined cells. The approach is developed to be independent of the flow solver in order to provide a path for augmenting existing codes. It is designed to be applicable for unsteady simulations and refinement and coarsening of the grid does not impact the conservatism of the underlying numerics. The effect on high-order numerical fluxes of fourth- and sixth-order are explored. Provided the criteria for refinement is appropriately selected, solutions obtained using adapted meshes have no additional error when compared to results obtained on traditional, unadapted meshes. In order to leverage large-scale computational resources common today, the methods are parallelized using MPI. Parallel performance is considered for several test problems in order to assess scalability of both adapted and unadapted grids. Dynamic repartitioning of the mesh during refinement is crucial for load balancing an evolving grid. Development of the methods outlined here depend on a dual-memory approach that is described in detail. Validation of the solver developed here against a number of motivating problems shows favorable
Multilevel Error Estimation and Adaptive h-Refinement for Cartesian Meshes with Embedded Boundaries
Aftosmis, M. J.; Berger, M. J.; Kwak, Dochan (Technical Monitor)
2002-01-01
This paper presents the development of a mesh adaptation module for a multilevel Cartesian solver. While the module allows mesh refinement to be driven by a variety of different refinement parameters, a central feature in its design is the incorporation of a multilevel error estimator based upon direct estimates of the local truncation error using tau-extrapolation. This error indicator exploits the fact that in regions of uniform Cartesian mesh, the spatial operator is exactly the same on the fine and coarse grids, and local truncation error estimates can be constructed by evaluating the residual on the coarse grid of the restricted solution from the fine grid. A new strategy for adaptive h-refinement is also developed to prevent errors in smooth regions of the flow from being masked by shocks and other discontinuous features. For certain classes of error histograms, this strategy is optimal for achieving equidistribution of the refinement parameters on hierarchical meshes, and therefore ensures grid converged solutions will be achieved for appropriately chosen refinement parameters. The robustness and accuracy of the adaptation module is demonstrated using both simple model problems and complex three dimensional examples using meshes with from 10(exp 6), to 10(exp 7) cells.
Soghrati, Soheil; Xiao, Fei; Nagarajan, Anand
2016-12-01
A Conforming to Interface Structured Adaptive Mesh Refinement (CISAMR) technique is introduced for the automated transformation of a structured grid into a conforming mesh with appropriate element aspect ratios. The CISAMR algorithm is composed of three main phases: (i) Structured Adaptive Mesh Refinement (SAMR) of the background grid; (ii) r-adaptivity of the nodes of elements cut by the crack; (iii) sub-triangulation of the elements deformed during the r-adaptivity process and those with hanging nodes generated during the SAMR process. The required considerations for the treatment of crack tips and branching cracks are also discussed in this manuscript. Regardless of the complexity of the problem geometry and without using iterative smoothing or optimization techniques, CISAMR ensures that aspect ratios of conforming elements are lower than three. Multiple numerical examples are presented to demonstrate the application of CISAMR for modeling linear elastic fracture problems with intricate morphologies.
Enzo+Moray: Radiation Hydrodynamics Adaptive Mesh Refinement Simulations with Adaptive Ray Tracing
Wise, John H
2010-01-01
We describe a photon-conserving radiative transfer algorithm, using a spatially-adaptive ray tracing scheme, and its parallel implementation into the adaptive mesh refinement (AMR) cosmological hydrodynamics code, Enzo. By coupling the solver with the energy equation and non-equilibrium chemistry network, our radiation hydrodynamics framework can be utilised to study a broad range of astrophysical problems, such as stellar and black hole (BH) feedback. Inaccuracies can arise from large timesteps and poor sampling, therefore we devised an adaptive time-stepping scheme and a fast approximation of the optically-thin radiation field with multiple sources. We test the method with several radiative transfer and radiation hydrodynamics tests that are given in Iliev et al. (2006, 2009). We further test our method with more dynamical situations, for example, the propagation of an ionisation front through a Rayleigh-Taylor instability, time-varying luminosities, and collimated radiation. The test suite also includes an...
Improvement of neutronic calculations on a Masurca core using adaptive mesh refinement capabilities
Fournier, D.; Archier, P.; Le Tellier, R.; Suteau, C., E-mail: damien.fournier@cea.fr, E-mail: pascal.archier@cea.fr, E-mail: romain.le-tellier@cea.fr, E-mail: christophe.suteau@cea.fr [CEA, DEN, DER/SPRC/LEPh, Cadarache, Saint Paul-lez-Durance (France)
2011-07-01
The simulation of 3D cores with homogenized assemblies in transport theory remains time and memory consuming for production calculations. With a multigroup discretization for the energy variable and a discrete ordinate method for the angle, a system of about 10{sup 4} coupled hyperbolic transport equations has to be solved. For these equations, we intend to optimize the spatial discretization. In the framework of the SNATCH solver used in this study, the spatial problem is dealt with by using a structured hexahedral mesh and applying a Discontinuous Galerkin Finite Element Method (DGFEM). This paper shows the improvements due to the development of Adaptive Mesh Refinement (AMR) methods. As the SNATCH solver uses a hierarchical polynomial basis, p−refinement is possible but also h−refinement thanks to non conforming capabilities. Besides, as the flux spatial behavior is highly dependent on the energy, we propose to adapt differently the spatial discretization according to the energy group. To avoid dealing with too many meshes, some energy groups are joined and share the same mesh. The different energy-dependent AMR strategies are compared to each other but also with the classical approach of a conforming and highly refined spatial mesh. This comparison is carried out on different quantities such as the multiplication factor, the flux or the current. The gain in time and memory is shown for 2D and 3D benchmarks coming from the ZONA2B experimental core configuration of the MASURCA mock-up at CEA Cadarache. (author)
Hornung, R.D. [Duke Univ., Durham, NC (United States)
1996-12-31
An adaptive local mesh refinement (AMR) algorithm originally developed for unsteady gas dynamics is extended to multi-phase flow in porous media. Within the AMR framework, we combine specialized numerical methods to treat the different aspects of the partial differential equations. Multi-level iteration and domain decomposition techniques are incorporated to accommodate elliptic/parabolic behavior. High-resolution shock capturing schemes are used in the time integration of the hyperbolic mass conservation equations. When combined with AMR, these numerical schemes provide high resolution locally in a more efficient manner than if they were applied on a uniformly fine computational mesh. We will discuss the interplay of physical, mathematical, and numerical concerns in the application of adaptive mesh refinement to flow in porous media problems of practical interest.
1995-01-01
The generation of suitable fine mesh divisions is essential to obtan two -dimensional electric field analysis solutions with desired accuracy. This process , however, requires considerable technical knowledge and experience. To solve this kind of problem, adaptive methods prove effective. In electric field problems, for example, researches are usually interested in the values of electric field intensity and its distributions. In this paper, we have developed an h-adaptive refinement procedure...
Adaptive mesh refinement with spectral accuracy for magnetohydrodynamics in two space dimensions
Rosenberg, D; Pouquet, A
2007-01-01
We examine the effect of accuracy of high-order spectral element methods, with or without adaptive mesh refinement (AMR), in the context of a classical configuration of magnetic reconnection in two space dimensions, the so-called Orszag-Tang vortex made up of a magnetic X-point centered on a stagnation point of the velocity. A recently developed spectral-element adaptive refinement incompressible magnetohydrodynamic (MHD) code is applied to simulate this problem. The MHD solver is explicit, and uses the Elsasser formulation on high-order elements. It automatically takes advantage of the adaptive grid mechanics that have been described elsewhere in the fluid context [Rosenberg, Fournier, Fischer, Pouquet, J. Comp. Phys. 215, 59-80 (2006)]; the code allows both statically refined and dynamically refined grids. Tests of the algorithm using analytic solutions are described, and comparisons of the Orszag-Tang solutions with pseudo-spectral computations are performed. We demonstrate for moderate Reynolds numbers th...
Relativistic Vlasov-Maxwell modelling using finite volumes and adaptive mesh refinement
Wettervik, Benjamin Svedung; Siminos, Evangelos; Fülöp, Tünde
2016-01-01
The dynamics of collisionless plasmas can be modelled by the Vlasov-Maxwell system of equations. An Eulerian approach is needed to accurately describe processes that are governed by high energy tails in the distribution function, but is of limited efficiency for high dimensional problems. The use of an adaptive mesh can reduce the scaling of the computational cost with the dimension of the problem. Here, we present a relativistic Eulerian Vlasov-Maxwell solver with block-structured adaptive mesh refinement in one spatial and one momentum dimension. The discretization of the Vlasov equation is based on a high-order finite volume method. A flux corrected transport algorithm is applied to limit spurious oscillations and ensure the physical character of the distribution function. We demonstrate a speed-up by a factor of five, because of the use of an adaptive mesh, in a typical scenario involving laser-plasma interaction in the self-induced transparency regime.
Ray, Jaideep; Lefantzi, Sophia; Najm, Habib N.; Kennedy, Christopher A.
2006-01-01
Block-structured adaptively refined meshes (SAMR) strive for efficient resolution of partial differential equations (PDEs) solved on large computational domains by clustering mesh points only where required by large gradients. Previous work has indicated that fourth-order convergence can be achieved on such meshes by using a suitable combination of high-order discretizations, interpolations, and filters and can deliver significant computational savings over conventional second-order methods at engineering error tolerances. In this paper, we explore the interactions between the errors introduced by discretizations, interpolations and filters. We develop general expressions for high-order discretizations, interpolations, and filters, in multiple dimensions, using a Fourier approach, facilitating the high-order SAMR implementation. We derive a formulation for the necessary interpolation order for given discretization and derivative orders. We also illustrate this order relationship empirically using one and two-dimensional model problems on refined meshes. We study the observed increase in accuracy with increasing interpolation order. We also examine the empirically observed order of convergence, as the effective resolution of the mesh is increased by successively adding levels of refinement, with different orders of discretization, interpolation, or filtering.
A Field-length based refinement criterion for adaptive mesh simulations of the interstellar medium
Gressel, Oliver
2009-01-01
Adequate modelling of the multiphase interstellar medium requires optically thin radiative cooling, comprising an inherent thermal instability. The size of the occurring condensation and evaporation interfaces is determined by the so-called Field-length, which gives the dimension at which the instability is significantly damped by thermal conduction. Our aim is to study the relevance of conduction scale effects in the numerical modelling of a bistable medium and check the applicability of conventional and alternative adaptive mesh techniques. The low physical value of the thermal conduction within the ISM defines a multiscale problem, hence promoting the use of adaptive meshes. We here introduce a new refinement strategy that applies the Field condition by Koyama & Inutsuka as a refinement criterion. The described method is very similar to the Jeans criterion for gravitational instability by Truelove and efficiently allows to trace the unstable gas situated at the thermal interfaces. We present test compu...
ENZO+MORAY: radiation hydrodynamics adaptive mesh refinement simulations with adaptive ray tracing
Wise, John H.; Abel, Tom
2011-07-01
We describe a photon-conserving radiative transfer algorithm, using a spatially-adaptive ray-tracing scheme, and its parallel implementation into the adaptive mesh refinement cosmological hydrodynamics code ENZO. By coupling the solver with the energy equation and non-equilibrium chemistry network, our radiation hydrodynamics framework can be utilized to study a broad range of astrophysical problems, such as stellar and black hole feedback. Inaccuracies can arise from large time-steps and poor sampling; therefore, we devised an adaptive time-stepping scheme and a fast approximation of the optically-thin radiation field with multiple sources. We test the method with several radiative transfer and radiation hydrodynamics tests that are given in Iliev et al. We further test our method with more dynamical situations, for example, the propagation of an ionization front through a Rayleigh-Taylor instability, time-varying luminosities and collimated radiation. The test suite also includes an expanding H II region in a magnetized medium, utilizing the newly implemented magnetohydrodynamics module in ENZO. This method linearly scales with the number of point sources and number of grid cells. Our implementation is scalable to 512 processors on distributed memory machines and can include the radiation pressure and secondary ionizations from X-ray radiation. It is included in the newest public release of ENZO.
Leng, Wei; Zhong, Shijie
2011-01-01
Numerical modeling of mantle convection is challenging. Owing to the multiscale nature of mantle dynamics, high resolution is often required in localized regions, with coarser resolution being sufficient elsewhere. When investigating thermochemical mantle convection, high resolution is required to resolve sharp and often discontinuous boundaries between distinct chemical components. In this paper, we present a 2-D finite element code with adaptive mesh refinement techniques for si...
A GPU implementation of adaptive mesh refinement to simulate tsunamis generated by landslides
de la Asunción, Marc; Castro, Manuel J.
2016-04-01
In this work we propose a CUDA implementation for the simulation of landslide-generated tsunamis using a two-layer Savage-Hutter type model and adaptive mesh refinement (AMR). The AMR method consists of dynamically increasing the spatial resolution of the regions of interest of the domain while keeping the rest of the domain at low resolution, thus obtaining better runtimes and similar results compared to increasing the spatial resolution of the entire domain. Our AMR implementation uses a patch-based approach, it supports up to three levels, power-of-two ratios of refinement, different refinement criteria and also several user parameters to control the refinement and clustering behaviour. A strategy based on the variation of the cell values during the simulation is used to interpolate and propagate the values of the fine cells. Several numerical experiments using artificial and realistic scenarios are presented.
Greg L. Bryan
2002-01-01
Full Text Available As an entry for the 2001 Gordon Bell Award in the "special" category, we describe our 3-d, hybrid, adaptive mesh refinement (AMR code Enzo designed for high-resolution, multiphysics, cosmological structure formation simulations. Our parallel implementation places no limit on the depth or complexity of the adaptive grid hierarchy, allowing us to achieve unprecedented spatial and temporal dynamic range. We report on a simulation of primordial star formation which develops over 8000 subgrids at 34 levels of refinement to achieve a local refinement of a factor of 1012 in space and time. This allows us to resolve the properties of the first stars which form in the universe assuming standard physics and a standard cosmological model. Achieving extreme resolution requires the use of 128-bit extended precision arithmetic (EPA to accurately specify the subgrid positions. We describe our EPA AMR implementation on the IBM SP2 Blue Horizon system at the San Diego Supercomputer Center.
Nicholas, Paul; Stasiuk, David; Nørgaard, Esben
2015-01-01
and material. Adaptive mesh refinement is used to support localised variance in resolution and information flow across these scales. The adaptation of mesh resolution is linked to structural analysis, panelisation, local geometric formation, connectivity, and the calculation of forming strains and material...
Error estimation and adaptive mesh refinement for parallel analysis of shell structures
Keating, Scott C.; Felippa, Carlos A.; Park, K. C.
1994-01-01
The formulation and application of element-level, element-independent error indicators is investigated. This research culminates in the development of an error indicator formulation which is derived based on the projection of element deformation onto the intrinsic element displacement modes. The qualifier 'element-level' means that no information from adjacent elements is used for error estimation. This property is ideally suited for obtaining error values and driving adaptive mesh refinements on parallel computers where access to neighboring elements residing on different processors may incur significant overhead. In addition such estimators are insensitive to the presence of physical interfaces and junctures. An error indicator qualifies as 'element-independent' when only visible quantities such as element stiffness and nodal displacements are used to quantify error. Error evaluation at the element level and element independence for the error indicator are highly desired properties for computing error in production-level finite element codes. Four element-level error indicators have been constructed. Two of the indicators are based on variational formulation of the element stiffness and are element-dependent. Their derivations are retained for developmental purposes. The second two indicators mimic and exceed the first two in performance but require no special formulation of the element stiffness mesh refinement which we demonstrate for two dimensional plane stress problems. The parallelizing of substructures and adaptive mesh refinement is discussed and the final error indicator using two-dimensional plane-stress and three-dimensional shell problems is demonstrated.
ADER-WENO finite volume schemes with space-time adaptive mesh refinement
Dumbser, Michael; Zanotti, Olindo; Hidalgo, Arturo; Balsara, Dinshaw S.
2013-09-01
We present the first high order one-step ADER-WENO finite volume scheme with adaptive mesh refinement (AMR) in multiple space dimensions. High order spatial accuracy is obtained through a WENO reconstruction, while a high order one-step time discretization is achieved using a local space-time discontinuous Galerkin predictor method. Due to the one-step nature of the underlying scheme, the resulting algorithm is particularly well suited for an AMR strategy on space-time adaptive meshes, i.e. with time-accurate local time stepping. The AMR property has been implemented 'cell-by-cell', with a standard tree-type algorithm, while the scheme has been parallelized via the message passing interface (MPI) paradigm. The new scheme has been tested over a wide range of examples for nonlinear systems of hyperbolic conservation laws, including the classical Euler equations of compressible gas dynamics and the equations of magnetohydrodynamics (MHD). High order in space and time have been confirmed via a numerical convergence study and a detailed analysis of the computational speed-up with respect to highly refined uniform meshes is also presented. We also show test problems where the presented high order AMR scheme behaves clearly better than traditional second order AMR methods. The proposed scheme that combines for the first time high order ADER methods with space-time adaptive grids in two and three space dimensions is likely to become a useful tool in several fields of computational physics, applied mathematics and mechanics.
Benoit, Commercon; Romain, Teyssier
2014-01-01
Implicit solvers present strong limitations when used on supercomputing facilities and in particular for adaptive mesh-refinement codes. We present a new method for implicit adaptive time-stepping on adaptive mesh refinement-grids and implementing it in the radiation hydrodynamics solver we designed for the RAMSES code for astrophysical purposes and, more particularly, for protostellar collapse. We briefly recall the radiation hydrodynamics equations and the adaptive time-stepping methodology used for hydrodynamical solvers. We then introduce the different types of boundary conditions (Dirichlet, Neumann, and Robin) that are used at the interface between levels and present our implementation of the new method in the RAMSES code. The method is tested against classical diffusion and radiation hydrodynamics tests, after which we present an application for protostellar collapse. We show that using Dirichlet boundary conditions at level interfaces is a good compromise between robustness and accuracy and that it ca...
Adaptive mesh refinement with spectral accuracy for magnetohydrodynamics in two space dimensions
Rosenberg, D.; Pouquet, A.; Mininni, P. D.
2007-08-01
We examine the effect of accuracy of high-order spectral element methods, with or without adaptive mesh refinement (AMR), in the context of a classical configuration of magnetic reconnection in two space dimensions, the so-called Orszag-Tang (OT) vortex made up of a magnetic X-point centred on a stagnation point of the velocity. A recently developed spectral-element adaptive refinement incompressible magnetohydrodynamic (MHD) code is applied to simulate this problem. The MHD solver is explicit, and uses the Elsässer formulation on high-order elements. It automatically takes advantage of the adaptive grid mechanics that have been described elsewhere in the fluid context (Rosenberg et al 2006 J. Comput. Phys. 215 59-80) the code allows both statically refined and dynamically refined grids. Tests of the algorithm using analytic solutions are described, and comparisons of the OT solutions with pseudo-spectral computations are performed. We demonstrate for moderate Reynolds numbers that the algorithms using both static and refined grids reproduce the pseudo-spectral solutions quite well. We show that low-order truncation—even with a comparable number of global degrees of freedom—fails to correctly model some strong (sup-norm) quantities in this problem, even though it satisfies adequately the weak (integrated) balance diagnostics.
A Parallel Ocean Model With Adaptive Mesh Refinement Capability For Global Ocean Prediction
Herrnstein, Aaron R. [Univ. of California, Davis, CA (United States)
2005-12-01
An ocean model with adaptive mesh refinement (AMR) capability is presented for simulating ocean circulation on decade time scales. The model closely resembles the LLNL ocean general circulation model with some components incorporated from other well known ocean models when appropriate. Spatial components are discretized using finite differences on a staggered grid where tracer and pressure variables are defined at cell centers and velocities at cell vertices (B-grid). Horizontal motion is modeled explicitly with leapfrog and Euler forward-backward time integration, and vertical motion is modeled semi-implicitly. New AMR strategies are presented for horizontal refinement on a B-grid, leapfrog time integration, and time integration of coupled systems with unequal time steps. These AMR capabilities are added to the LLNL software package SAMRAI (Structured Adaptive Mesh Refinement Application Infrastructure) and validated with standard benchmark tests. The ocean model is built on top of the amended SAMRAI library. The resulting model has the capability to dynamically increase resolution in localized areas of the domain. Limited basin tests are conducted using various refinement criteria and produce convergence trends in the model solution as refinement is increased. Carbon sequestration simulations are performed on decade time scales in domains the size of the North Atlantic and the global ocean. A suggestion is given for refinement criteria in such simulations. AMR predicts maximum pH changes and increases in CO_{2} concentration near the injection sites that are virtually unattainable with a uniform high resolution due to extremely long run times. Fine scale details near the injection sites are achieved by AMR with shorter run times than the finest uniform resolution tested despite the need for enhanced parallel performance. The North Atlantic simulations show a reduction in passive tracer errors when AMR is applied instead of a uniform coarse resolution. No
Conformal refinement of unstructured quadrilateral meshes
Garmella, Rao [Los Alamos National Laboratory
2009-01-01
We present a multilevel adaptive refinement technique for unstructured quadrilateral meshes in which the mesh is kept conformal at all times. This means that the refined mesh, like the original, is formed of only quadrilateral elements that intersect strictly along edges or at vertices, i.e., vertices of one quadrilateral element do not lie in an edge of another quadrilateral. Elements are refined using templates based on 1:3 refinement of edges. We demonstrate that by careful design of the refinement and coarsening strategy, we can maintain high quality elements in the refined mesh. We demonstrate the method on a number of examples with dynamically changing refinement regions.
Single-Pass GPU-Raycasting for Structured Adaptive Mesh Refinement Data
Kaehler, Ralf
2012-01-01
Structured Adaptive Mesh Refinement (SAMR) is a popular numerical technique to study processes with high spatial and temporal dynamic range. It reduces computational requirements by adapting the lattice on which the underlying differential equations are solved to most efficiently represent the solution. Particularly in astrophysics and cosmology such simulations now can capture spatial scales ten orders of magnitude apart and more. The irregular locations and extensions of the refined regions in the SAMR scheme and the fact that different resolution levels partially overlap, poses a challenge for GPU-based direct volume rendering methods. kD-trees have proven to be advantageous to subdivide the data domain into non-overlapping blocks of equally sized cells, optimal for the texture units of current graphics hardware, but previous GPU-supported raycasting approaches for SAMR data using this data structure required a separate rendering pass for each node, preventing the application of many advanced lighting sche...
A high order special relativistic hydrodynamic code with space-time adaptive mesh refinement
Zanotti, Olindo
2013-01-01
We present a high order one-step ADER-WENO finite volume scheme with space-time adaptive mesh refinement (AMR) for the solution of the special relativistic hydrodynamics equations. By adopting a local discontinuous Galerkin predictor method, a high order one-step time discretization is obtained, with no need for Runge-Kutta sub-steps. This turns out to be particularly advantageous in combination with space-time adaptive mesh refinement, which has been implemented following a "cell-by-cell" approach. As in existing second order AMR methods, also the present higher order AMR algorithm features time-accurate local time stepping (LTS), where grids on different spatial refinement levels are allowed to use different time steps. We also compare two different Riemann solvers for the computation of the numerical fluxes at the cell interfaces. The new scheme has been validated over a sample of numerical test problems in one, two and three spatial dimensions, exploring its ability in resolving the propagation of relativ...
A New MHD Code with Adaptive Mesh Refinement and Parallelization for Astrophysics
Jiang, R L; Chen, P F
2012-01-01
A new code, named MAP, is written in Fortran language for magnetohydrodynamics (MHD) calculation with the adaptive mesh refinement (AMR) and Message Passing Interface (MPI) parallelization. There are several optional numerical schemes for computing the MHD part, namely, modified Mac Cormack Scheme (MMC), Lax-Friedrichs scheme (LF) and weighted essentially non-oscillatory (WENO) scheme. All of them are second order, two-step, component-wise schemes for hyperbolic conservative equations. The total variation diminishing (TVD) limiters and approximate Riemann solvers are also equipped. A high resolution can be achieved by the hierarchical block-structured AMR mesh. We use the extended generalized Lagrange multiplier (EGLM) MHD equations to reduce the non-divergence free error produced by the scheme in the magnetic induction equation. The numerical algorithms for the non-ideal terms, e.g., the resistivity and the thermal conduction, are also equipped in the MAP code. The details of the AMR and MPI algorithms are d...
Constrained-Transport Magnetohydrodynamics with Adaptive-Mesh-Refinement in CHARM
Miniati, Francesco
2011-01-01
We present the implementation of a three-dimensional, second order accurate Godunov-type algorithm for magneto-hydrodynamic (MHD), in the adaptive-mesh-refinement (AMR) cosmological code {\\tt CHARM}. The algorithm is based on the full 12-solve spatially unsplit Corner-Transport-Upwind (CTU) scheme. The fluid quantities are cell-centered and are updated using the Piecewise-Parabolic-Method (PPM), while the magnetic field variables are face-centered and are evolved through application of the Stokes theorem on cell edges via a Constrained-Transport (CT) method. The multidimensional MHD source terms required in the predictor step for high-order accuracy are applied in a simplified form which reduces their complexity in three dimensions without loss of accuracy or robustness. The algorithm is implemented on an AMR framework which requires specific synchronization steps across refinement levels. These include face-centered restriction and prolongation operations and a {\\it reflux-curl} operation, which maintains a ...
Fromang, S; Teyssier, R
2006-01-01
In this paper, we present a new method to perform numerical simulations of astrophysical MHD flows using the Adaptive Mesh Refinement framework and Constrained Transport. The algorithm is based on a previous work in which the MUSCL--Hancock scheme was used to evolve the induction equation. In this paper, we detail the extension of this scheme to the full MHD equations and discuss its properties. Through a series of test problems, we illustrate the performances of this new code using two different MHD Riemann solvers (Lax-Friedrich and Roe) and the need of the Adaptive Mesh Refinement capabilities in some cases. Finally, we show its versatility by applying it to two completely different astrophysical situations well studied in the past years: the growth of the magnetorotational instability in the shearing box and the collapse of magnetized cloud cores. We have implemented this new Godunov scheme to solve the ideal MHD equations in the AMR code RAMSES. It results in a powerful tool that can be applied to a grea...
Cell-based Adaptive Mesh Refinement on the GPU with Applications to Exascale Supercomputing
Trujillo, Dennis; Robey, Robert; Davis, Neal; Nicholaeff, David
2011-10-01
We present an OpenCL implementation of a cell-based adaptive mesh refinement (AMR) scheme for the shallow water equations. The challenges associated with ensuring the locality of algorithm architecture to fully exploit the massive number of parallel threads on the GPU is discussed. This includes a proof of concept that a cell-based AMR code can be effectively implemented, even on a small scale, in the memory and threading model provided by OpenCL. Additionally, the program requires dynamic memory in order to properly implement the mesh; as this is not supported in the OpenCL 1.1 standard, a combination of CPU memory management and GPU computation effectively implements a dynamic memory allocation scheme. Load balancing is achieved through a new stencil-based implementation of a space-filling curve, eliminating the need for a complete recalculation of the indexing on the mesh. A cartesian grid hash table scheme to allow fast parallel neighbor accesses is also discussed. Finally, the relative speedup of the GPU-enabled AMR code is compared to the original serial version. We conclude that parallelization using the GPU provides significant speedup for typical numerical applications and is feasible for scientific applications in the next generation of supercomputing.
A Predictive Model of Fragmentation using Adaptive Mesh Refinement and a Hierarchical Material Model
Koniges, A E; Masters, N D; Fisher, A C; Anderson, R W; Eder, D C; Benson, D; Kaiser, T B; Gunney, B T; Wang, P; Maddox, B R; Hansen, J F; Kalantar, D H; Dixit, P; Jarmakani, H; Meyers, M A
2009-03-03
Fragmentation is a fundamental material process that naturally spans spatial scales from microscopic to macroscopic. We developed a mathematical framework using an innovative combination of hierarchical material modeling (HMM) and adaptive mesh refinement (AMR) to connect the continuum to microstructural regimes. This framework has been implemented in a new multi-physics, multi-scale, 3D simulation code, NIF ALE-AMR. New multi-material volume fraction and interface reconstruction algorithms were developed for this new code, which is leading the world effort in hydrodynamic simulations that combine AMR with ALE (Arbitrary Lagrangian-Eulerian) techniques. The interface reconstruction algorithm is also used to produce fragments following material failure. In general, the material strength and failure models have history vector components that must be advected along with other properties of the mesh during remap stage of the ALE hydrodynamics. The fragmentation models are validated against an electromagnetically driven expanding ring experiment and dedicated laser-based fragmentation experiments conducted at the Jupiter Laser Facility. As part of the exit plan, the NIF ALE-AMR code was applied to a number of fragmentation problems of interest to the National Ignition Facility (NIF). One example shows the added benefit of multi-material ALE-AMR that relaxes the requirement that material boundaries must be along mesh boundaries.
ADER-WENO Finite Volume Schemes with Space-Time Adaptive Mesh Refinement
Dumbser, Michael; Hidalgo, Arturo; Balsara, Dinshaw S
2012-01-01
We present the first high order one-step ADER-WENO finite volume scheme with Adaptive Mesh Refinement (AMR) in multiple space dimensions. High order spatial accuracy is obtained through a WENO reconstruction, while a high order one-step time discretization is achieved using a local space-time discontinuous Galerkin predictor method. Due to the one-step nature of the underlying scheme, the resulting algorithm is particularly well suited for an AMR strategy on space-time adaptive meshes, i.e.with time-accurate local time stepping. The AMR property has been implemented 'cell-by-cell', with a standard tree-type algorithm, while the scheme has been parallelized via the Message Passing Interface (MPI) paradigm. The new scheme has been tested over a wide range of examples for nonlinear systems of hyperbolic conservation laws, including the classical Euler equations of compressible gas dynamics and the equations of magnetohydrodynamics (MHD). High order in space and time have been confirmed via a numerical convergenc...
Simulation of tsunamis generated by landslides using adaptive mesh refinement on GPU
de la Asunción, M.; Castro, M. J.
2017-09-01
Adaptive mesh refinement (AMR) is a widely used technique to accelerate computationally intensive simulations, which consists of dynamically increasing the spatial resolution of the areas of interest of the domain as the simulation advances. During the last years there have appeared many publications that tackle the implementation of AMR-based applications in GPUs in order to take advantage of their massively parallel architecture. In this paper we present the first AMR-based application implemented on GPU for the simulation of tsunamis generated by landslides by using a two-layer shallow water system. We also propose a new strategy for the interpolation and projection of the values of the fine cells in the AMR algorithm based on the fluctuations of the state values instead of the usual approach of considering the current state values. Numerical experiments on artificial and realistic problems show the validity and efficiency of the solver.
Relativistic Flows Using Spatial and Temporal Adaptive Structured Mesh Refinement. I. Hydrodynamics
Wang, Peng; Zhang, Weiqun
2007-01-01
Astrophysical relativistic flow problems require high resolution three-dimensional numerical simulations. In this paper, we describe a new parallel three-dimensional code for simulations of special relativistic hydrodynamics (SRHD) using both spatially and temporally structured adaptive mesh refinement (AMR). We used method of lines to discrete SRHD equations spatially and used a total variation diminishing (TVD) Runge-Kutta scheme for time integration. For spatial reconstruction, we have implemented piecewise linear method (PLM), piecewise parabolic method (PPM), third order convex essentially non-oscillatory (CENO) and third and fifth order weighted essentially non-oscillatory (WENO) schemes. Flux is computed using either direct flux reconstruction or approximate Riemann solvers including HLL, modified Marquina flux, local Lax-Friedrichs flux formulas and HLLC. The AMR part of the code is built on top of the cosmological Eulerian AMR code {\\sl enzo}, which uses the Berger-Colella AMR algorithm and is parall...
A Survey of High Level Frameworks in Block-Structured Adaptive Mesh Refinement Packages
Dubey, Anshu; Bell, John; Berzins, Martin; Brandt, Steve; Bryan, Greg; Colella, Phillip; Graves, Daniel; Lijewski, Michael; Löffler, Frank; O'Shea, Brian; Schnetter, Erik; Van Straalen, Brian; Weide, Klaus
2016-01-01
Over the last decade block-structured adaptive mesh refinement (SAMR) has found increasing use in large, publicly available codes and frameworks. SAMR frameworks have evolved along different paths. Some have stayed focused on specific domain areas, others have pursued a more general functionality, providing the building blocks for a larger variety of applications. In this survey paper we examine a representative set of SAMR packages and SAMR-based codes that have been in existence for half a decade or more, have a reasonably sized and active user base outside of their home institutions, and are publicly available. The set consists of a mix of SAMR packages and application codes that cover a broad range of scientific domains. We look at their high-level frameworks, and their approach to dealing with the advent of radical changes in hardware architecture. The codes included in this survey are BoxLib, Cactus, Chombo, Enzo, FLASH, and Uintah.
On the Computation of Integral Curves in Adaptive Mesh Refinement Vector Fields
Deines, Eduard; Weber, Gunther H.; Garth, Christoph; Van Straalen, Brian; Borovikov, Sergey; Martin, Daniel F.; Joy, Kenneth I.
2011-06-27
Integral curves, such as streamlines, streaklines, pathlines, and timelines, are an essential tool in the analysis of vector field structures, offering straightforward and intuitive interpretation of visualization results. While such curves have a long-standing tradition in vector field visualization, their application to Adaptive Mesh Refinement (AMR) simulation results poses unique problems. AMR is a highly effective discretization method for a variety of physical simulation problems and has recently been applied to the study of vector fields in flow and magnetohydrodynamic applications. The cell-centered nature of AMR data and discontinuities in the vector field representation arising from AMR level boundaries complicate the application of numerical integration methods to compute integral curves. In this paper, we propose a novel approach to alleviate these problems and show its application to streamline visualization in an AMR model of the magnetic field of the solar system as well as to a simulation of two incompressible viscous vortex rings merging.
Lichtenberg, Tim
2015-01-01
The astonishing diversity in the observed planetary population requires theoretical efforts and advances in planet formation theories. Numerical approaches provide a method to tackle the weaknesses of current planet formation models and are an important tool to close gaps in poorly constrained areas. We present a global disk setup to model the first stages of giant planet formation via gravitational instabilities (GI) in 3D with the block-structured adaptive mesh refinement (AMR) hydrodynamics code ENZO. With this setup, we explore the impact of AMR techniques on the fragmentation and clumping due to large-scale instabilities using different AMR configurations. Additionally, we seek to derive general resolution criteria for global simulations of self-gravitating disks of variable extent. We run a grid of simulations with varying AMR settings, including runs with a static grid for comparison, and study the effects of varying the disk radius. Adopting a marginally stable disk profile (Q_init=1), we validate the...
GLAMER Part I: A Code for Gravitational Lensing Simulations with Adaptive Mesh Refinement
Metcalf, R Benton
2013-01-01
A computer code is described for the simulation of gravitational lensing data. The code incorporates adaptive mesh refinement in choosing which rays to shoot based on the requirements of the source size, location and surface brightness distribution or to find critical curves/caustics. A variety of source surface brightness models are implemented to represent galaxies and quasar emission regions. The lensing mass can be represented by point masses (stars), smoothed simulation particles, analytic halo models, pixelized mass maps or any combination of these. The deflection and beam distortions (convergence and shear) are calculated by modified tree algorithm when halos, point masses or particles are used and by FFT when mass maps are used. The combination of these methods allow for a very large dynamical range to be represented in a single simulation. Individual images of galaxies can be represented in a simulation that covers many square degrees. For an individual strongly lensed quasar, source sizes from the s...
The Singularity Threshold of the Nonlinear Sigma Model Using 3D Adaptive Mesh Refinement
Liebling, S L
2002-01-01
Numerical solutions to the nonlinear sigma model (NLSM), a wave map from 3+1 Minkowski space to S^3, are computed in three spatial dimensions (3D) using adaptive mesh refinement (AMR). For initial data with compact support the model is known to have two regimes, one in which regular initial data forms a singularity and another in which the energy is dispersed to infinity. The transition between these regimes has been shown in spherical symmetry to demonstrate threshold behavior similar to that between black hole formation and dispersal in gravitating theories. Here, I generalize the result by removing the assumption of spherical symmetry. The evolutions suggest that the spherically symmetric critical solution remains an intermediate attractor separating the two end states.
Mesh refinement strategy for optimal control problems
Paiva, Luis Tiago; Fontes, Fernando,
2013-01-01
International audience; Direct methods are becoming the most used technique to solve nonlinear optimal control problems. Regular time meshes having equidistant spacing are frequently used. However, in some cases these meshes cannot cope accurately with nonlinear behavior. One way to improve the solution is to select a new mesh with a greater number of nodes. Another way, involves adaptive mesh refinement. In this case, the mesh nodes have non equidistant spacing which allow a non uniform node...
A new MHD code with adaptive mesh refinement and parallelization for astrophysics
Jiang, R.-L.; Fang, C.; Chen, P.-F.
2012-08-01
A new code, named MAP, is written in FORTRAN language for magnetohydrodynamics (MHD) simulations with the adaptive mesh refinement (AMR) and Message Passing Interface (MPI) parallelization. There are several optional numerical schemes for computing the MHD part, namely, modified Mac Cormack Scheme (MMC), Lax-Friedrichs scheme (LF), and weighted essentially non-oscillatory (WENO) scheme. All of them are second-order, two-step, component-wise schemes for hyperbolic conservative equations. The total variation diminishing (TVD) limiters and approximate Riemann solvers are also equipped. A high resolution can be achieved by the hierarchical block-structured AMR mesh. We use the extended generalized Lagrange multiplier (EGLM) MHD equations to reduce the non-divergence free error produced by the scheme in the magnetic induction equation. The numerical algorithms for the non-ideal terms, e.g., the resistivity and the thermal conduction, are also equipped in the code. The details of the AMR and MPI algorithms are described in the paper.
Fakhari, Abbas; Lee, Taehun
2014-03-01
An adaptive-mesh-refinement (AMR) algorithm for the finite-difference lattice Boltzmann method (FDLBM) is presented in this study. The idea behind the proposed AMR is to remove the need for a tree-type data structure. Instead, pointer attributes are used to determine the neighbors of a certain block via appropriate adjustment of its children identifications. As a result, the memory and time required for tree traversal are completely eliminated, leaving us with an efficient algorithm that is easier to implement and use on parallel machines. To allow different mesh sizes at separate parts of the computational domain, the Eulerian formulation of the streaming process is invoked. As a result, there is no need for rescaling the distribution functions or using a temporal interpolation at the fine-coarse grid boundaries. The accuracy and efficiency of the proposed FDLBM AMR are extensively assessed by investigating a variety of vorticity-dominated flow fields, including Taylor-Green vortex flow, lid-driven cavity flow, thin shear layer flow, and the flow past a square cylinder.
The Numerical Simulation of Ship Waves Using Cartesian Grid Methods with Adaptive Mesh Refinement
Dommermuth, Douglas G; Beck, Robert F; O'Shea, Thomas T; Wyatt, Donald C; Olson, Kevin; MacNeice, Peter
2014-01-01
Cartesian-grid methods with Adaptive Mesh Refinement (AMR) are ideally suited for simulating the breaking of waves, the formation of spray, and the entrainment of air around ships. As a result of the cartesian-grid formulation, minimal input is required to describe the ships geometry. A surface panelization of the ship hull is used as input to automatically generate a three-dimensional model. No three-dimensional gridding is required. The AMR portion of the numerical algorithm automatically clusters grid points near the ship in regions where wave breaking, spray formation, and air entrainment occur. Away from the ship, where the flow is less turbulent, the mesh is coarser. The numerical computations are implemented using parallel algorithms. Together, the ease of input and usage, the ability to resolve complex free-surface phenomena, and the speed of the numerical algorithms provide a robust capability for simulating the free-surface disturbances near a ship. Here, numerical predictions, with and without AMR,...
Henshaw, W; Schwendeman, D
2007-11-15
This paper describes an approach for the numerical solution of time-dependent partial differential equations in complex three-dimensional domains. The domains are represented by overlapping structured grids, and block-structured adaptive mesh refinement (AMR) is employed to locally increase the grid resolution. In addition, the numerical method is implemented on parallel distributed-memory computers using a domain-decomposition approach. The implementation is flexible so that each base grid within the overlapping grid structure and its associated refinement grids can be independently partitioned over a chosen set of processors. A modified bin-packing algorithm is used to specify the partition for each grid so that the computational work is evenly distributed amongst the processors. All components of the AMR algorithm such as error estimation, regridding, and interpolation are performed in parallel. The parallel time-stepping algorithm is illustrated for initial-boundary-value problems involving a linear advection-diffusion equation and the (nonlinear) reactive Euler equations. Numerical results are presented for both equations to demonstrate the accuracy and correctness of the parallel approach. Exact solutions of the advection-diffusion equation are constructed, and these are used to check the corresponding numerical solutions for a variety of tests involving different overlapping grids, different numbers of refinement levels and refinement ratios, and different numbers of processors. The problem of planar shock diffraction by a sphere is considered as an illustration of the numerical approach for the Euler equations, and a problem involving the initiation of a detonation from a hot spot in a T-shaped pipe is considered to demonstrate the numerical approach for the reactive case. For both problems, the solutions are shown to be well resolved on the finest grid. The parallel performance of the approach is examined in detail for the shock diffraction problem.
Development of a scalable gas-dynamics solver with adaptive mesh refinement
Korkut, Burak
There are various computational physics areas in which Direct Simulation Monte Carlo (DSMC) and Particle in Cell (PIC) methods are being employed. The accuracy of results from such simulations depend on the fidelity of the physical models being used. The computationally demanding nature of these problems make them ideal candidates to make use of modern supercomputers. The software developed to run such simulations also needs special attention so that the maintainability and extendability is considered with the recent numerical methods and programming paradigms. Suited for gas-dynamics problems, a software called SUGAR (Scalable Unstructured Gas dynamics with Adaptive mesh Refinement) has recently been developed and written in C++ and MPI. Physical and numerical models were added to this framework to simulate ion thruster plumes. SUGAR is used to model the charge-exchange (CEX) reactions occurring between the neutral and ion species as well as the induced electric field effect due to ions. Multiple adaptive mesh refinement (AMR) meshes were used in order to capture different physical length scales present in the flow. A multiple-thruster configuration was run to extend the studies to cases for which there is no axial or radial symmetry present that could only be modeled with a three-dimensional simulation capability. The combined plume structure showed interactions between individual thrusters where AMR capability captured this in an automated way. The back flow for ions was found to occur when CEX and momentum-exchange (MEX) collisions are present and strongly enhanced when the induced electric field is considered. The ion energy distributions in the back flow region were obtained and it was found that the inclusion of the electric field modeling is the most important factor in determining its shape. The plume back flow structure was also examined for a triple-thruster, 3-D geometry case and it was found that the ion velocity in the back flow region appears to be
Effenberger, Frederic; Arnold, Lukas; Grauer, Rainer; Dreher, Jürgen
2011-01-01
The formation of a thin current sheet in a magnetic quasi-separatrix layer (QSL) is investigated by means of numerical simulation using a simplified ideal, low-$\\beta$, MHD model. The initial configuration and driving boundary conditions are relevant to phenomena observed in the solar corona and were studied earlier by Aulanier et al., A&A 444, 961 (2005). In extension to that work, we use the technique of adaptive mesh refinement (AMR) to significantly enhance the local spatial resolution of the current sheet during its formation, which enables us to follow the evolution into a later stage. Our simulations are in good agreement with the results of Aulanier et al. up to the calculated time in that work. In a later phase, we observe a basically unarrested collapse of the sheet to length scales that are more than one order of magnitude smaller than those reported earlier. The current density attains correspondingly larger maximum values within the sheet. During this thinning process, which is finally limite...
Multigroup radiation hydrodynamics with flux-limited diffusion and adaptive mesh refinement
González, Matthias; Commerçon, Benoît; Masson, Jacques
2015-01-01
Radiative transfer plays a key role in the star formation process. Due to a high computational cost, radiation-hydrodynamics simulations performed up to now have mainly been carried out in the grey approximation. In recent years, multi-frequency radiation-hydrodynamics models have started to emerge, in an attempt to better account for the large variations of opacities as a function of frequency. We wish to develop an efficient multigroup algorithm for the adaptive mesh refinement code RAMSES which is suited to heavy proto-stellar collapse calculations. Due to prohibitive timestep constraints of an explicit radiative transfer method, we constructed a time-implicit solver based on a stabilised bi-conjugate gradient algorithm, and implemented it in RAMSES under the flux-limited diffusion approximation. We present a series of tests which demonstrate the high performance of our scheme in dealing with frequency-dependent radiation-hydrodynamic flows. We also present a preliminary simulation of a three-dimensional p...
Hummels, Cameron
2011-01-01
We carry out adaptive mesh refinement (AMR) cosmological simulations of Milky-Way mass halos in order to investigate the formation of disk-like galaxies in a {\\Lambda}-dominated Cold Dark Matter model. We evolve a suite of five halos to z = 0 and find gaseous-disk formation in all; however, in agreement with previous SPH simulations (that did not include a subgrid feedback model), the rotation curves of all halos are centrally peaked due to a massive spheroidal component. Our standard model includes radiative cooling and star formation, but no feedback. We further investigate this angular momentum problem by systematically modifying various simulation parameters including: (i) spatial resolution, ranging from 1700 to 212 pc; (ii) an additional pressure component to ensure that the Jeans length is always resolved; (iii) low star formation efficiency, going down to 0.1%; (iv) fixed physical resolution as opposed to comoving resolution; (v) a supernova feedback model which injects thermal energy to the local cel...
GPU accelerated cell-based adaptive mesh refinement on unstructured quadrilateral grid
Luo, Xisheng; Wang, Luying; Ran, Wei; Qin, Fenghua
2016-10-01
A GPU accelerated inviscid flow solver is developed on an unstructured quadrilateral grid in the present work. For the first time, the cell-based adaptive mesh refinement (AMR) is fully implemented on GPU for the unstructured quadrilateral grid, which greatly reduces the frequency of data exchange between GPU and CPU. Specifically, the AMR is processed with atomic operations to parallelize list operations, and null memory recycling is realized to improve the efficiency of memory utilization. It is found that results obtained by GPUs agree very well with the exact or experimental results in literature. An acceleration ratio of 4 is obtained between the parallel code running on the old GPU GT9800 and the serial code running on E3-1230 V2. With the optimization of configuring a larger L1 cache and adopting Shared Memory based atomic operations on the newer GPU C2050, an acceleration ratio of 20 is achieved. The parallelized cell-based AMR processes have achieved 2x speedup on GT9800 and 18x on Tesla C2050, which demonstrates that parallel running of the cell-based AMR method on GPU is feasible and efficient. Our results also indicate that the new development of GPU architecture benefits the fluid dynamics computing significantly.
De Colle, Fabio; Lopez-Camara, Diego; Ramirez-Ruiz, Enrico
2011-01-01
We report on the development of Mezcal-SRHD, a new adaptive mesh refinement, special relativistic hydrodynamics (SRHD) code, developed with the aim of studying the highly relativistic flows in Gamma-Ray Burst sources. The SRHD equations are solved using finite volume conservative solvers. The correct implementation of the algorithms is verified by one-dimensional (1D) shock tube and multidimensional tests. The code is then applied to study the propagation of 1D spherical impulsive blast waves expanding in a stratified medium with $\\rho \\propto r^{-k}$, bridging between the relativistic and Newtonian phases, as well as to a two-dimensional (2D) cylindrically symmetric impulsive jet propagating in a constant density medium. It is shown that the deceleration to non-relativistic speeds in one-dimension occurs on scales significantly larger than the Sedov length. This transition is further delayed with respect to the Sedov length as the degree of stratification of the ambient medium is increased. This result, toge...
Relativistic Flows Using Spatial And Temporal Adaptive Structured Mesh Refinement. I. Hydrodynamics
Wang, Peng; Abel, Tom; Zhang, Weiqun; /KIPAC, Menlo Park
2007-04-02
Astrophysical relativistic flow problems require high resolution three-dimensional numerical simulations. In this paper, we describe a new parallel three-dimensional code for simulations of special relativistic hydrodynamics (SRHD) using both spatially and temporally structured adaptive mesh refinement (AMR). We used method of lines to discrete SRHD equations spatially and used a total variation diminishing (TVD) Runge-Kutta scheme for time integration. For spatial reconstruction, we have implemented piecewise linear method (PLM), piecewise parabolic method (PPM), third order convex essentially non-oscillatory (CENO) and third and fifth order weighted essentially non-oscillatory (WENO) schemes. Flux is computed using either direct flux reconstruction or approximate Riemann solvers including HLL, modified Marquina flux, local Lax-Friedrichs flux formulas and HLLC. The AMR part of the code is built on top of the cosmological Eulerian AMR code enzo, which uses the Berger-Colella AMR algorithm and is parallel with dynamical load balancing using the widely available Message Passing Interface library. We discuss the coupling of the AMR framework with the relativistic solvers and show its performance on eleven test problems.
Juan J. Garcia-Cantero
2017-06-01
Full Text Available Gaining a better understanding of the human brain continues to be one of the greatest challenges for science, largely because of the overwhelming complexity of the brain and the difficulty of analyzing the features and behavior of dense neural networks. Regarding analysis, 3D visualization has proven to be a useful tool for the evaluation of complex systems. However, the large number of neurons in non-trivial circuits, together with their intricate geometry, makes the visualization of a neuronal scenario an extremely challenging computational problem. Previous work in this area dealt with the generation of 3D polygonal meshes that approximated the cells’ overall anatomy but did not attempt to deal with the extremely high storage and computational cost required to manage a complex scene. This paper presents NeuroTessMesh, a tool specifically designed to cope with many of the problems associated with the visualization of neural circuits that are comprised of large numbers of cells. In addition, this method facilitates the recovery and visualization of the 3D geometry of cells included in databases, such as NeuroMorpho, and provides the tools needed to approximate missing information such as the soma’s morphology. This method takes as its only input the available compact, yet incomplete, morphological tracings of the cells as acquired by neuroscientists. It uses a multiresolution approach that combines an initial, coarse mesh generation with subsequent on-the-fly adaptive mesh refinement stages using tessellation shaders. For the coarse mesh generation, a novel approach, based on the Finite Element Method, allows approximation of the 3D shape of the soma from its incomplete description. Subsequently, the adaptive refinement process performed in the graphic card generates meshes that provide good visual quality geometries at a reasonable computational cost, both in terms of memory and rendering time. All the described techniques have been
Maric, Tomislav; Bothe, Dieter
2013-01-01
A new parallelized unsplit geometrical Volume of Fluid (VoF) algorithm with support for arbitrary unstructured meshes and dynamic local Adaptive Mesh Refinement (AMR), as well as for two and three dimensional computation is developed. The geometrical VoF algorithm supports arbitrary unstructured meshes in order to enable computations involving flow domains of arbitrary geometrical complexity. The implementation of the method is done within the framework of the OpenFOAM library for Computational Continuum Mechanics (CCM) using the C++ programming language with modern policy based design for high program code modularity. The development of the geometrical VoF algorithm significantly extends the method base of the OpenFOAM library by geometrical volumetric flux computation for two-phase flow simulations. For the volume fraction advection, a novel unsplit geometrical algorithm is developed, which inherently sustains volume conservation utilizing unique Lagrangian discrete trajectories located in the mesh points. ...
Application of adaptive mesh refinement to particle-in-cell simulations of plasmas and beams
Vay, J.-L.; Colella, P.; Kwan, J.W.; McCorquodale, P.; Serafini, D.B.; Friedman, A.; Grote, D.P.; Westenskow, G.; Adam, J.-C.; Heron, A.; Haber, I.
2003-11-04
Plasma simulations are often rendered challenging by the disparity of scales in time and in space which must be resolved. When these disparities are in distinctive zones of the simulation domain, a method which has proven to be effective in other areas (e.g. fluid dynamics simulations) is the mesh refinement technique. We briefly discuss the challenges posed by coupling this technique with plasma Particle-In-Cell simulations, and present examples of application in Heavy Ion Fusion and related fields which illustrate the effectiveness of the approach. We also report on the status of a collaboration under way at Lawrence Berkeley National Laboratory between the Applied Numerical Algorithms Group (ANAG) and the Heavy Ion Fusion group to upgrade ANAG's mesh refinement library Chombo to include the tools needed by Particle-In-Cell simulation codes.
Ravindran, Prashaanth
The unstable nature of detonation waves is a result of the critical relationship between the hydrodynamic shock and the chemical reactions sustaining the shock. A perturbative analysis of the critical point is quite challenging due to the multiple spatio-temporal scales involved along with the non-linear nature of the shock-reaction mechanism. The author's research attempts to provide detailed resolution of the instabilities at the shock front. Another key aspect of the present research is to develop an understanding of the causality between the non-linear dynamics of the front and the eventual breakdown of the sub-structures. An accurate numerical simulation of detonation waves requires a very efficient solution of the Euler equations in conservation form with detailed, non-equilibrium chemistry. The difference in the flow and reaction length scales results in very stiff source terms, requiring the problem to be solved with adaptive mesh refinement. For this purpose, Berger-Colella's block-structured adaptive mesh refinement (AMR) strategy has been developed and applied to time-explicit finite volume methods. The block-structured technique uses a hierarchy of parent-child sub-grids, integrated recursively over time. One novel approach to partition the problem within a large supercomputer was the use of modified Peano-Hilbert space filling curves. The AMR framework was merged with CLAWPACK, a package providing finite volume numerical methods tailored for wave-propagation problems. The stiffness problem is bypassed by using a 1st order Godunov or a 2nd order Strang splitting technique, where the flow variables and source terms are integrated independently. A linearly explicit fourth-order Runge-Kutta integrator is used for the flow, and an ODE solver was used to overcome the numerical stiffness. Second-order spatial resolution is obtained by using a second-order Roe-HLL scheme with the inclusion of numerical viscosity to stabilize the solution near the discontinuity
De Colle, Fabio; Ramirez-Ruiz, Enrico [Astronomy and Astrophysics Department, University of California, Santa Cruz, CA 95064 (United States); Granot, Jonathan [Racah Institute of Physics, Hebrew University, Jerusalem 91904 (Israel); Lopez-Camara, Diego, E-mail: fabio@ucolick.org [Instituto de Ciencias Nucleares, Universidad Nacional Autonoma de Mexico, Ap. 70-543, 04510 D.F. (Mexico)
2012-02-20
We report on the development of Mezcal-SRHD, a new adaptive mesh refinement, special relativistic hydrodynamics (SRHD) code, developed with the aim of studying the highly relativistic flows in gamma-ray burst sources. The SRHD equations are solved using finite-volume conservative solvers, with second-order interpolation in space and time. The correct implementation of the algorithms is verified by one-dimensional (1D) and multi-dimensional tests. The code is then applied to study the propagation of 1D spherical impulsive blast waves expanding in a stratified medium with {rho}{proportional_to}r{sup -k}, bridging between the relativistic and Newtonian phases (which are described by the Blandford-McKee and Sedov-Taylor self-similar solutions, respectively), as well as to a two-dimensional (2D) cylindrically symmetric impulsive jet propagating in a constant density medium. It is shown that the deceleration to nonrelativistic speeds in one dimension occurs on scales significantly larger than the Sedov length. This transition is further delayed with respect to the Sedov length as the degree of stratification of the ambient medium is increased. This result, together with the scaling of position, Lorentz factor, and the shock velocity as a function of time and shock radius, is explained here using a simple analytical model based on energy conservation. The method used for calculating the afterglow radiation by post-processing the results of the simulations is described in detail. The light curves computed using the results of 1D numerical simulations during the relativistic stage correctly reproduce those calculated assuming the self-similar Blandford-McKee solution for the evolution of the flow. The jet dynamics from our 2D simulations and the resulting afterglow light curves, including the jet break, are in good agreement with those presented in previous works. Finally, we show how the details of the dynamics critically depend on properly resolving the structure of the
De Colle, Fabio; Granot, Jonathan; López-Cámara, Diego; Ramirez-Ruiz, Enrico
2012-02-01
We report on the development of Mezcal-SRHD, a new adaptive mesh refinement, special relativistic hydrodynamics (SRHD) code, developed with the aim of studying the highly relativistic flows in gamma-ray burst sources. The SRHD equations are solved using finite-volume conservative solvers, with second-order interpolation in space and time. The correct implementation of the algorithms is verified by one-dimensional (1D) and multi-dimensional tests. The code is then applied to study the propagation of 1D spherical impulsive blast waves expanding in a stratified medium with ρvpropr -k , bridging between the relativistic and Newtonian phases (which are described by the Blandford-McKee and Sedov-Taylor self-similar solutions, respectively), as well as to a two-dimensional (2D) cylindrically symmetric impulsive jet propagating in a constant density medium. It is shown that the deceleration to nonrelativistic speeds in one dimension occurs on scales significantly larger than the Sedov length. This transition is further delayed with respect to the Sedov length as the degree of stratification of the ambient medium is increased. This result, together with the scaling of position, Lorentz factor, and the shock velocity as a function of time and shock radius, is explained here using a simple analytical model based on energy conservation. The method used for calculating the afterglow radiation by post-processing the results of the simulations is described in detail. The light curves computed using the results of 1D numerical simulations during the relativistic stage correctly reproduce those calculated assuming the self-similar Blandford-McKee solution for the evolution of the flow. The jet dynamics from our 2D simulations and the resulting afterglow light curves, including the jet break, are in good agreement with those presented in previous works. Finally, we show how the details of the dynamics critically depend on properly resolving the structure of the relativistic flow.
Donmez, Orhan
We present a general procedure to solve the General Relativistic Hydrodynamical (GRH) equations with Adaptive-Mesh Refinement (AMR) and model of an accretion disk around a black hole. To do this, the GRH equations are written in a conservative form to exploit their hyperbolic character. The numerical solutions of the general relativistic hydrodynamic equations is done by High Resolution Shock Capturing schemes (HRSC), specifically designed to solve non-linear hyperbolic systems of conservation laws. These schemes depend on the characteristic information of the system. We use Marquina fluxes with MUSCL left and right states to solve GRH equations. First, we carry out different test problems with uniform and AMR grids on the special relativistic hydrodynamics equations to verify the second order convergence of the code in 1D, 2 D and 3D. Second, we solve the GRH equations and use the general relativistic test problems to compare the numerical solutions with analytic ones. In order to this, we couple the flux part of general relativistic hydrodynamic equation with a source part using Strang splitting. The coupling of the GRH equations is carried out in a treatment which gives second order accurate solutions in space and time. The test problems examined include shock tubes, geodesic flows, and circular motion of particle around the black hole. Finally, we apply this code to the accretion disk problems around the black hole using the Schwarzschild metric at the background of the computational domain. We find spiral shocks on the accretion disk. They are observationally expected results. We also examine the star-disk interaction near a massive black hole. We find that when stars are grounded down or a hole is punched on the accretion disk, they create shock waves which destroy the accretion disk.
Composite-Grid Techniques and Adaptive Mesh Refinement in Computational Fluid Dynamics
1990-01-01
grid that would result from global refinement is .,. If the ratio Nfto, 0t/Nf9 ob~ is too large, then global refinement should bc ADAPTIVE SCENARIO 58...which has the disadvantage of being nonconservative, i.e., it does not preserve conservation properties integrally. More will bc said about this in...r2 converges to the solution of the equation if the (sub-) matriz splitting A = B - C is used and the following conditions hold: 1. A is a real
Wenjun Ying
2015-01-01
Full Text Available A both space and time adaptive algorithm is presented for simulating electrical wave propagation in the Purkinje system of the heart. The equations governing the distribution of electric potential over the system are solved in time with the method of lines. At each timestep, by an operator splitting technique, the space-dependent but linear diffusion part and the nonlinear but space-independent reactions part in the partial differential equations are integrated separately with implicit schemes, which have better stability and allow larger timesteps than explicit ones. The linear diffusion equation on each edge of the system is spatially discretized with the continuous piecewise linear finite element method. The adaptive algorithm can automatically recognize when and where the electrical wave starts to leave or enter the computational domain due to external current/voltage stimulation, self-excitation, or local change of membrane properties. Numerical examples demonstrating efficiency and accuracy of the adaptive algorithm are presented.
Simulations of recoiling black holes: adaptive mesh refinement and radiative transfer
Meliani, Zakaria; Mizuno, Yosuke; Olivares, Hector; Porth, Oliver; Rezzolla, Luciano; Younsi, Ziri
2017-02-01
Context. In many astrophysical phenomena, and especially in those that involve the high-energy regimes that always accompany the astronomical phenomenology of black holes and neutron stars, physical conditions that are achieved are extreme in terms of speeds, temperatures, and gravitational fields. In such relativistic regimes, numerical calculations are the only tool to accurately model the dynamics of the flows and the transport of radiation in the accreting matter. Aims: We here continue our effort of modelling the behaviour of matter when it orbits or is accreted onto a generic black hole by developing a new numerical code that employs advanced techniques geared towards solving the equations of general-relativistic hydrodynamics. Methods: More specifically, the new code employs a number of high-resolution shock-capturing Riemann solvers and reconstruction algorithms, exploiting the enhanced accuracy and the reduced computational cost of adaptive mesh-refinement (AMR) techniques. In addition, the code makes use of sophisticated ray-tracing libraries that, coupled with general-relativistic radiation-transfer calculations, allow us to accurately compute the electromagnetic emissions from such accretion flows. Results: We validate the new code by presenting an extensive series of stationary accretion flows either in spherical or axial symmetry that are performed either in two or three spatial dimensions. In addition, we consider the highly nonlinear scenario of a recoiling black hole produced in the merger of a supermassive black-hole binary interacting with the surrounding circumbinary disc. In this way, we can present for the first time ray-traced images of the shocked fluid and the light curve resulting from consistent general-relativistic radiation-transport calculations from this process. Conclusions: The work presented here lays the ground for the development of a generic computational infrastructure employing AMR techniques to accurately and self
Essadki Mohamed
2016-09-01
Full Text Available Predictive simulation of liquid fuel injection in automotive engines has become a major challenge for science and applications. The key issue in order to properly predict various combustion regimes and pollutant formation is to accurately describe the interaction between the carrier gaseous phase and the polydisperse evaporating spray produced through atomization. For this purpose, we rely on the EMSM (Eulerian Multi-Size Moment Eulerian polydisperse model. It is based on a high order moment method in size, with a maximization of entropy technique in order to provide a smooth reconstruction of the distribution, derived from a Williams-Boltzmann mesoscopic model under the monokinetic assumption [O. Emre (2014 PhD Thesis, École Centrale Paris; O. Emre, R.O. Fox, M. Massot, S. Chaisemartin, S. Jay, F. Laurent (2014 Flow, Turbulence and Combustion 93, 689-722; O. Emre, D. Kah, S. Jay, Q.-H. Tran, A. Velghe, S. de Chaisemartin, F. Laurent, M. Massot (2015 Atomization Sprays 25, 189-254; D. Kah, F. Laurent, M. Massot, S. Jay (2012 J. Comput. Phys. 231, 394-422; D. Kah, O. Emre, Q.-H. Tran, S. de Chaisemartin, S. Jay, F. Laurent, M. Massot (2015 Int. J. Multiphase Flows 71, 38-65; A. Vié, F. Laurent, M. Massot (2013 J. Comp. Phys. 237, 277-310]. The present contribution relies on a major extension of this model [M. Essadki, S. de Chaisemartin, F. Laurent, A. Larat, M. Massot (2016 Submitted to SIAM J. Appl. Math.], with the aim of building a unified approach and coupling with a separated phases model describing the dynamics and atomization of the interface near the injector. The novelty is to be found in terms of modeling, numerical schemes and implementation. A new high order moment approach is introduced using fractional moments in surface, which can be related to geometrical quantities of the gas-liquid interface. We also provide a novel algorithm for an accurate resolution of the evaporation. Adaptive mesh refinement properly scaling on massively
Patched based methods for adaptive mesh refinement solutions of partial differential equations
Saltzman, J.
1997-09-02
This manuscript contains the lecture notes for a course taught from July 7th through July 11th at the 1997 Numerical Analysis Summer School sponsored by C.E.A., I.N.R.I.A., and E.D.F. The subject area was chosen to support the general theme of that year`s school which is ``Multiscale Methods and Wavelets in Numerical Simulation.`` The first topic covered in these notes is a description of the problem domain. This coverage is limited to classical PDEs with a heavier emphasis on hyperbolic systems and constrained hyperbolic systems. The next topic is difference schemes. These schemes are the foundation for the adaptive methods. After the background material is covered, attention is focused on a simple patched based adaptive algorithm and its associated data structures for square grids and hyperbolic conservation laws. Embellishments include curvilinear meshes, embedded boundary and overset meshes. Next, several strategies for parallel implementations are examined. The remainder of the notes contains descriptions of elliptic solutions on the mesh hierarchy, elliptically constrained flow solution methods and elliptically constrained flow solution methods with diffusion.
Wang, Cheng; Dong, XinZhuang; Shu, Chi-Wang
2015-10-01
For numerical simulation of detonation, computational cost using uniform meshes is large due to the vast separation in both time and space scales. Adaptive mesh refinement (AMR) is advantageous for problems with vastly different scales. This paper aims to propose an AMR method with high order accuracy for numerical investigation of multi-dimensional detonation. A well-designed AMR method based on finite difference weighted essentially non-oscillatory (WENO) scheme, named as AMR&WENO is proposed. A new cell-based data structure is used to organize the adaptive meshes. The new data structure makes it possible for cells to communicate with each other quickly and easily. In order to develop an AMR method with high order accuracy, high order prolongations in both space and time are utilized in the data prolongation procedure. Based on the message passing interface (MPI) platform, we have developed a workload balancing parallel AMR&WENO code using the Hilbert space-filling curve algorithm. Our numerical experiments with detonation simulations indicate that the AMR&WENO is accurate and has a high resolution. Moreover, we evaluate and compare the performance of the uniform mesh WENO scheme and the parallel AMR&WENO method. The comparison results provide us further insight into the high performance of the parallel AMR&WENO method.
Mesh refinement strategy for optimal control problems
Paiva, L. T.; Fontes, F. A. C. C.
2013-10-01
Direct methods are becoming the most used technique to solve nonlinear optimal control problems. Regular time meshes having equidistant spacing are frequently used. However, in some cases these meshes cannot cope accurately with nonlinear behavior. One way to improve the solution is to select a new mesh with a greater number of nodes. Another way, involves adaptive mesh refinement. In this case, the mesh nodes have non equidistant spacing which allow a non uniform nodes collocation. In the method presented in this paper, a time mesh refinement strategy based on the local error is developed. After computing a solution in a coarse mesh, the local error is evaluated, which gives information about the subintervals of time domain where refinement is needed. This procedure is repeated until the local error reaches a user-specified threshold. The technique is applied to solve the car-like vehicle problem aiming minimum consumption. The approach developed in this paper leads to results with greater accuracy and yet with lower overall computational time as compared to using a time meshes having equidistant spacing.
Moura, R. C.; Silva, A. F. C.; Bigarella, E. D. V.; Fazenda, A. L.; Ortega, M. A.
2016-08-01
This paper proposes two important improvements to shock-capturing strategies using a discontinuous Galerkin scheme, namely, accurate shock identification via finite-time Lyapunov exponent (FTLE) operators and efficient shock treatment through a point-implicit discretization of a PDE-based artificial viscosity technique. The advocated approach is based on the FTLE operator, originally developed in the context of dynamical systems theory to identify certain types of coherent structures in a flow. We propose the application of FTLEs in the detection of shock waves and demonstrate the operator's ability to identify strong and weak shocks equally well. The detection algorithm is coupled with a mesh refinement procedure and applied to transonic and supersonic flows. While the proposed strategy can be used potentially with any numerical method, a high-order discontinuous Galerkin solver is used in this study. In this context, two artificial viscosity approaches are employed to regularize the solution near shocks: an element-wise constant viscosity technique and a PDE-based smooth viscosity model. As the latter approach is more sophisticated and preferable for complex problems, a point-implicit discretization in time is proposed to reduce the extra stiffness introduced by the PDE-based technique, making it more competitive in terms of computational cost.
Vertical Scan (V-SCAN) for 3-D Grid Adaptive Mesh Refinement for an atmospheric Model Dynamical Core
Andronova, N. G.; Vandenberg, D.; Oehmke, R.; Stout, Q. F.; Penner, J. E.
2009-12-01
One of the major building blocks of a rigorous representation of cloud evolution in global atmospheric models is a parallel adaptive grid MPI-based communication library (an Adaptive Blocks for Locally Cartesian Topologies library -- ABLCarT), which manages the block-structured data layout, handles ghost cell updates among neighboring blocks and splits a block as refinements occur. The library has several modules that provide a layer of abstraction for adaptive refinement: blocks, which contain individual cells of user data; shells - the global geometry for the problem, including a sphere, reduced sphere, and now a 3D sphere; a load balancer for placement of blocks onto processors; and a communication support layer which encapsulates all data movement. A major performance concern with adaptive mesh refinement is how to represent calculations that have need to be sequenced in a particular order in a direction, such as calculating integrals along a specific path (e.g. atmospheric pressure or geopotential in the vertical dimension). This concern is compounded if the blocks have varying levels of refinement, or are scattered across different processors, as can be the case in parallel computing. In this paper we describe an implementation in ABLCarT of a vertical scan operation, which allows computing along vertical paths in the correct order across blocks transparent to their resolution and processor location. We test this functionality on a 2D and a 3D advection problem, which tests the performance of the model’s dynamics (transport) and physics (sources and sinks) for different model resolutions needed for inclusion of cloud formation.
3-D grid refinement using the University of Michigan adaptive mesh library for a pure advective test
Oehmke, R.; Vandenberg, D.; Andronova, N.; Penner, J.; Stout, Q.; Zubov, V.; Jablonowski, C.
2008-05-01
The numerical representation of the partial differential equations (PDE) for high resolution atmospheric dynamical and physical features requires division of the atmospheric volume into a set of 3D grids, each of which has a not quite rectangular form. Each location on the grid contains multiple data that together represent the state of Earth's atmosphere. For successful numerical integration of the PDEs the size of each grid box is used to define the Courant-Friedrichs-Levi criterion in setting the time step. 3D adaptive representations of a sphere are needed to represent the evolution of clouds. In this paper we present the University of Michigan adaptive mesh library - a library that supports the production of parallel codes with use of adaptation on a sphere. The library manages the block-structured data layout, handles ghost cell updates among neighboring blocks and splits blocks as refinements occur. The library has several modules that provide a layer of abstraction for adaptive refinement: blocks, which contain individual cells of user data; shells — the global geometry for the problem, including a sphere, reduced sphere, and now a 3D sphere; a load balancer for placement of blocks onto processors; and a communication support layer which encapsulates all data movement. Users provide data manipulation functions for performing interpolation of user data when refining blocks. We rigorously test the library using refinement of the modeled vertical transport of a tracer with prescribed atmospheric sources and sinks. It is both a 2 and a 3D test, and bridges the performance of the model's dynamics and physics needed for inclusion of cloud formation.
Cunningham, Andrew J.; Frank, Adam; Varnière, Peggy; Mitran, Sorin; Jones, Thomas W.
2009-06-01
A description is given of the algorithms implemented in the AstroBEAR adaptive mesh-refinement code for ideal magnetohydrodynamics. The code provides several high-resolution shock-capturing schemes which are constructed to maintain conserved quantities of the flow in a finite-volume sense. Divergence-free magnetic field topologies are maintained to machine precision by collating the components of the magnetic field on a cell-interface staggered grid and utilizing the constrained transport approach for integrating the induction equations. The maintenance of magnetic field topologies on adaptive grids is achieved using prolongation and restriction operators which preserve the divergence and curl of the magnetic field across collocated grids of different resolutions. The robustness and correctness of the code is demonstrated by comparing the numerical solution of various tests with analytical solutions or previously published numerical solutions obtained by other codes.
Amaziane Brahim
2014-07-01
Full Text Available In this paper, we consider adaptive numerical simulation of miscible displacement problems in porous media, which are modeled by single phase flow equations. A vertex-centred finite volume method is employed to discretize the coupled system: the Darcy flow equation and the diffusion-convection concentration equation. The convection term is approximated with a Godunov scheme over the dual finite volume mesh, whereas the diffusion-dispersion term is discretized by piecewise linear conforming finite elements. We introduce two kinds of indicators, both of them of residual type. The first one is related to time discretization and is local with respect to the time discretization: thus, at each time, it provides an appropriate information for the choice of the next time step. The second is related to space discretization and is local with respect to both the time and space variable and the idea is that at each time it is an efficient tool for mesh adaptivity. An error estimation procedure evaluates where additional refinement is needed and grid generation procedures dynamically create or remove fine-grid patches as resolution requirements change. The method was implemented in the software MELODIE, developed by the French Institute for Radiological Protection and Nuclear Safety (IRSN, Institut de Radioprotection et de Sûreté Nucléaire. The algorithm is then used to simulate the evolution of radionuclide migration from the waste packages through a heterogeneous disposal, demonstrating its capability to capture complex behavior of the resulting flow.
Lopez-Camara, D.; Lazzati, Davide [Department of Physics, NC State University, 2401 Stinson Drive, Raleigh, NC 27695-8202 (United States); Morsony, Brian J. [Department of Astronomy, University of Wisconsin-Madison, 2535 Sterling Hall, 475 N. Charter Street, Madison, WI 53706-1582 (United States); Begelman, Mitchell C., E-mail: dlopezc@ncsu.edu [JILA, University of Colorado, 440 UCB, Boulder, CO 80309-0440 (United States)
2013-04-10
We present the results of special relativistic, adaptive mesh refinement, 3D simulations of gamma-ray burst jets expanding inside a realistic stellar progenitor. Our simulations confirm that relativistic jets can propagate and break out of the progenitor star while remaining relativistic. This result is independent of the resolution, even though the amount of turbulence and variability observed in the simulations is greater at higher resolutions. We find that the propagation of the jet head inside the progenitor star is slightly faster in 3D simulations compared to 2D ones at the same resolution. This behavior seems to be due to the fact that the jet head in 3D simulations can wobble around the jet axis, finding the spot of least resistance to proceed. Most of the average jet properties, such as density, pressure, and Lorentz factor, are only marginally affected by the dimensionality of the simulations and therefore results from 2D simulations can be considered reliable.
Angelidis, Dionysios; Sotiropoulos, Fotis
2015-11-01
The geometrical details of wind turbines determine the structure of the turbulence in the near and far wake and should be taken in account when performing high fidelity calculations. Multi-resolution simulations coupled with an immersed boundary method constitutes a powerful framework for high-fidelity calculations past wind farms located over complex terrains. We develop a 3D Immersed-Boundary Adaptive Mesh Refinement flow solver (IB-AMR) which enables turbine-resolving LES of wind turbines. The idea of using a hybrid staggered/non-staggered grid layout adopted in the Curvilinear Immersed Boundary Method (CURVIB) has been successfully incorporated on unstructured meshes and the fractional step method has been employed. The overall performance and robustness of the second order accurate, parallel, unstructured solver is evaluated by comparing the numerical simulations against conforming grid calculations and experimental measurements of laminar and turbulent flows over complex geometries. We also present turbine-resolving multi-scale LES considering all the details affecting the induced flow field; including the geometry of the tower, the nacelle and especially the rotor blades of a wind tunnel scale turbine. This material is based upon work supported by the Department of Energy under Award Number DE-EE0005482 and the Sandia National Laboratories.
D'Avillez, M A; Breitschwerdt, Dieter
2005-01-01
State of the art models of the ISM use adaptive mesh refinement to capture small scale structures, by refining on the fly those regions of the grid where density and pressure gradients occur, keeping at the same time the existing resolution in the other regions. With this technique it became possible to study the ISM in star-forming galaxies in a global way by following matter circulation between stars and the interstellar gas, and, in particular the energy input by random and clustered supernova explosions, which determine the dynamical and chemical evolution of the ISM, and hence of the galaxy as a whole. In this paper we review the conditions for a self-consistent modelling of the ISM and present the results from the latest developments in the 3D HD/MHD global models of the ISM. Special emphasis is put on the effects of the magnetic field with respect to the volume and mass fractions of the different ISM ``phases'', the relative importance of ram, thermal and magnetic pressures, and whether the field can p...
Moving Overlapping Grids with Adaptive Mesh Refinement for High-Speed Reactive and Non-reactive Flow
Henshaw, W D; Schwendeman, D W
2005-08-30
We consider the solution of the reactive and non-reactive Euler equations on two-dimensional domains that evolve in time. The domains are discretized using moving overlapping grids. In a typical grid construction, boundary-fitted grids are used to represent moving boundaries, and these grids overlap with stationary background Cartesian grids. Block-structured adaptive mesh refinement (AMR) is used to resolve fine-scale features in the flow such as shocks and detonations. Refinement grids are added to base-level grids according to an estimate of the error, and these refinement grids move with their corresponding base-level grids. The numerical approximation of the governing equations takes place in the parameter space of each component grid which is defined by a mapping from (fixed) parameter space to (moving) physical space. The mapped equations are solved numerically using a second-order extension of Godunov's method. The stiff source term in the reactive case is handled using a Runge-Kutta error-control scheme. We consider cases when the boundaries move according to a prescribed function of time and when the boundaries of embedded bodies move according to the surface stress exerted by the fluid. In the latter case, the Newton-Euler equations describe the motion of the center of mass of the each body and the rotation about it, and these equations are integrated numerically using a second-order predictor-corrector scheme. Numerical boundary conditions at slip walls are described, and numerical results are presented for both reactive and non-reactive flows in order to demonstrate the use and accuracy of the numerical approach.
Li, Pak Shing; Klein, Richard I. [Astronomy Department, University of California, Berkeley, CA 94720 (United States); Martin, Daniel F. [Lawrence Berkeley National Laboratory, 1 Cyclotron Road, Berkeley, CA 94720 (United States); McKee, Christopher F., E-mail: psli@astron.berkeley.edu, E-mail: klein@astron.berkeley.edu, E-mail: DFMartin@lbl.gov, E-mail: cmckee@astro.berkeley.edu [Physics Department and Astronomy Department, University of California, Berkeley, CA 94720 (United States)
2012-02-01
Performing a stable, long-duration simulation of driven MHD turbulence with a high thermal Mach number and a strong initial magnetic field is a challenge to high-order Godunov ideal MHD schemes because of the difficulty in guaranteeing positivity of the density and pressure. We have implemented a robust combination of reconstruction schemes, Riemann solvers, limiters, and constrained transport electromotive force averaging schemes that can meet this challenge, and using this strategy, we have developed a new adaptive mesh refinement (AMR) MHD module of the ORION2 code. We investigate the effects of AMR on several statistical properties of a turbulent ideal MHD system with a thermal Mach number of 10 and a plasma {beta}{sub 0} of 0.1 as initial conditions; our code is shown to be stable for simulations with higher Mach numbers (M{sub rms}= 17.3) and smaller plasma beta ({beta}{sub 0} = 0.0067) as well. Our results show that the quality of the turbulence simulation is generally related to the volume-averaged refinement. Our AMR simulations show that the turbulent dissipation coefficient for supersonic MHD turbulence is about 0.5, in agreement with unigrid simulations.
Li, Pak Shing; Klein, Richard I; McKee, Christopher F
2011-01-01
Performing a stable, long duration simulation of driven MHD turbulence with a high thermal Mach number and a strong initial magnetic field is a challenge to high-order Godunov ideal MHD schemes because of the difficulty in guaranteeing positivity of the density and pressure. We have implemented a robust combination of reconstruction schemes, Riemann solvers, limiters, and Constrained Transport EMF averaging schemes that can meet this challenge, and using this strategy, we have developed a new Adaptive Mesh Refinement (AMR) MHD module of the ORION2 code. We investigate the effects of AMR on several statistical properties of a turbulent ideal MHD system with a thermal Mach number of 10 and a plasma $\\beta_0$ of 0.1 as initial conditions; our code is shown to be stable for simulations with higher Mach numbers ($M_rms = 17.3$) and smaller plasma beta ($\\beta_0 = 0.0067$) as well. Our results show that the quality of the turbulence simulation is generally related to the volume-averaged refinement. Our AMR simulati...
Huang, Rongzong; Wu, Huiying
2016-06-01
A total enthalpy-based lattice Boltzmann (LB) method with adaptive mesh refinement (AMR) is developed in this paper to efficiently simulate solid-liquid phase change problem where variables vary significantly near the phase interface and thus finer grid is required. For the total enthalpy-based LB method, the velocity field is solved by an incompressible LB model with multiple-relaxation-time (MRT) collision scheme, and the temperature field is solved by a total enthalpy-based MRT LB model with the phase interface effects considered and the deviation term eliminated. With a kinetic assumption that the density distribution function for solid phase is at equilibrium state, a volumetric LB scheme is proposed to accurately realize the nonslip velocity condition on the diffusive phase interface and in the solid phase. As compared with the previous schemes, this scheme can avoid nonphysical flow in the solid phase. As for the AMR approach, it is developed based on multiblock grids. An indicator function is introduced to control the adaptive generation of multiblock grids, which can guarantee the existence of overlap area between adjacent blocks for information exchange. Since MRT collision schemes are used, the information exchange is directly carried out in the moment space. Numerical tests are firstly performed to validate the strict satisfaction of the nonslip velocity condition, and then melting problems in a square cavity with different Prandtl numbers and Rayleigh numbers are simulated, which demonstrate that the present method can handle solid-liquid phase change problem with high efficiency and accuracy.
Constrained-Transport Magnetohydrodynamics with Adaptive-Mesh-Refinement in CHARM
Miniatii, Francesco; Martin, Daniel
2011-05-24
We present the implementation of a three-dimensional, second order accurate Godunov-type algorithm for magneto-hydrodynamic (MHD), in the adaptivemesh-refinement (AMR) cosmological code CHARM. The algorithm is based on the full 12-solve spatially unsplit Corner-Transport-Upwind (CTU) scheme. Thefluid quantities are cell-centered and are updated using the Piecewise-Parabolic- Method (PPM), while the magnetic field variables are face-centered and areevolved through application of the Stokes theorem on cell edges via a Constrained- Transport (CT) method. The so-called ?multidimensional MHD source terms?required in the predictor step for high-order accuracy are applied in a simplified form which reduces their complexity in three dimensions without loss of accuracyor robustness. The algorithm is implemented on an AMR framework which requires specific synchronization steps across refinement levels. These includeface-centered restriction and prolongation operations and a reflux-curl operation, which maintains a solenoidal magnetic field across refinement boundaries. Thecode is tested against a large suite of test problems, including convergence tests in smooth flows, shock-tube tests, classical two- and three-dimensional MHD tests,a three-dimensional shock-cloud interaction problem and the formation of a cluster of galaxies in a fully cosmological context. The magnetic field divergence isshown to remain negligible throughout. Subject headings: cosmology: theory - methods: numerical
Zanotti, Olindo; Dumbser, Michael
2015-01-01
We present a new numerical tool for solving the special relativistic ideal MHD equations that is based on the combination of the following three key features: (i) a one-step ADER discontinuous Galerkin (DG) scheme that allows for an arbitrary order of accuracy in both space and time, (ii) an a posteriori subcell finite volume limiter that is activated to avoid spurious oscillations at discontinuities without destroying the natural subcell resolution capabilities of the DG finite element framework and finally (iii) a space-time adaptive mesh refinement (AMR) framework with time-accurate local time-stepping. The divergence-free character of the magnetic field is instead taken into account through the so-called 'divergence-cleaning' approach. The convergence of the new scheme is verified up to 5th order in space and time and the results for a sample of significant numerical tests including shock tube problems, the RMHD rotor problem and the Orszag-Tang vortex system are shown. We also consider a simple case of t...
Zhang, S.; Yuen, D.A.; Zhu, A.; Song, S.; George, D.L.
2011-01-01
We parallelized the GeoClaw code on one-level grid using OpenMP in March, 2011 to meet the urgent need of simulating tsunami waves at near-shore from Tohoku 2011 and achieved over 75% of the potential speed-up on an eight core Dell Precision T7500 workstation [1]. After submitting that work to SC11 - the International Conference for High Performance Computing, we obtained an unreleased OpenMP version of GeoClaw from David George, who developed the GeoClaw code as part of his PH.D thesis. In this paper, we will show the complementary characteristics of the two approaches used in parallelizing GeoClaw and the speed-up obtained by combining the advantage of each of the two individual approaches with adaptive mesh refinement (AMR), demonstrating the capabilities of running GeoClaw efficiently on many-core systems. We will also show a novel simulation of the Tohoku 2011 Tsunami waves inundating the Sendai airport and Fukushima Nuclear Power Plants, over which the finest grid distance of 20 meters is achieved through a 4-level AMR. This simulation yields quite good predictions about the wave-heights and travel time of the tsunami waves. ?? 2011 IEEE.
woptic: optical conductivity with Wannier functions and adaptive k-mesh refinement
Assmann, E; Kuneš, J; Toschi, A; Blaha, P; Held, K
2015-01-01
We present an algorithm for the adaptive tetrahedral integration over the Brillouin zone of crystalline materials, and apply it to compute the optical conductivity, dc conductivity, and thermopower. For these quantities, whose contributions are often localized in small portions of the Brillouin zone, adaptive integration is especially relevant. Our implementation, the woptic package, is tied into the wien2wannier framework and allows including a many-body self energy, e.g. from dynamical mean-field theory (DMFT). Wannier functions and dipole matrix elements are computed with the DFT package Wien2k and Wannier90. For illustration, we show DFT results for fcc-Al and DMFT results for the correlated metal SrVO$_3$.
Mesh Adaptation and Shape Optimization on Unstructured Meshes Project
National Aeronautics and Space Administration — In this SBIR CRM proposes to implement the entropy adjoint method for solution adaptive mesh refinement into the Loci/CHEM unstructured flow solver. The scheme will...
Autotuning of Adaptive Mesh Refinement PDE Solvers on Shared Memory Architectures
Nogina, Svetlana
2012-01-01
Many multithreaded, grid-based, dynamically adaptive solvers for partial differential equations permanently have to traverse subgrids (patches) of different and changing sizes. The parallel efficiency of this traversal depends on the interplay of the patch size, the architecture used, the operations triggered throughout the traversal, and the grain size, i.e. the size of the subtasks the patch is broken into. We propose an oracle mechanism delivering grain sizes on-the-fly. It takes historical runtime measurements for different patch and grain sizes as well as the traverse\\'s operations into account, and it yields reasonable speedups. Neither magic configuration settings nor an expensive pre-tuning phase are necessary. It is an autotuning approach. © 2012 Springer-Verlag.
Electrical impedance tomography using adaptive mesh refinement%利用自适应网格细分法的电阻抗成像
严佩敏; 王朔中
2006-01-01
In electrical impedance tomography ( EIT), distribution of the internal resistivity or conductivity of an unknown object is estimated using measured boundary voltage data induced by different current patterns with various reconstruction algorithms.The reconstruction algorithms usually employ the Newton-Raphson iteration scheme to visualize the resistivity distribution inside the object.Accuracy of the imaging process depends not only on the algorithm used, but also on the scheme of finite element discretization.In this paper an adaptive mesh refinement is used in a modified reconstruction algorithm for the regularized EIT.The method has a major impact on efficient solution of the forward problem as well as on achieving improved image resolution.Computer simulations indicate that the Newton-Raphson reconstruction algorithm for EIT using adaptive mesh refinement performs better than the classical Newton-Raphson algorithm in terms of reconstructed image resolution.
An Adaptive Mesh Algorithm: Mesh Structure and Generation
Scannapieco, Anthony J. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)
2016-06-21
The purpose of Adaptive Mesh Refinement is to minimize spatial errors over the computational space not to minimize the number of computational elements. The additional result of the technique is that it may reduce the number of computational elements needed to retain a given level of spatial accuracy. Adaptive mesh refinement is a computational technique used to dynamically select, over a region of space, a set of computational elements designed to minimize spatial error in the computational model of a physical process. The fundamental idea is to increase the mesh resolution in regions where the physical variables are represented by a broad spectrum of modes in k-space, hence increasing the effective global spectral coverage of those physical variables. In addition, the selection of the spatially distributed elements is done dynamically by cyclically adjusting the mesh to follow the spectral evolution of the system. Over the years three types of AMR schemes have evolved; block, patch and locally refined AMR. In block and patch AMR logical blocks of various grid sizes are overlaid to span the physical space of interest, whereas in locally refined AMR no logical blocks are employed but locally nested mesh levels are used to span the physical space. The distinction between block and patch AMR is that in block AMR the original blocks refine and coarsen entirely in time, whereas in patch AMR the patches change location and zone size with time. The type of AMR described herein is a locally refi ned AMR. In the algorithm described, at any point in physical space only one zone exists at whatever level of mesh that is appropriate for that physical location. The dynamic creation of a locally refi ned computational mesh is made practical by a judicious selection of mesh rules. With these rules the mesh is evolved via a mesh potential designed to concentrate the nest mesh in regions where the physics is modally dense, and coarsen zones in regions where the physics is modally
Guzik, S; McCorquodale, P; Colella, P
2011-12-16
A fourth-order accurate finite-volume method is presented for solving time-dependent hyperbolic systems of conservation laws on mapped grids that are adaptively refined in space and time. Novel considerations for formulating the semi-discrete system of equations in computational space combined with detailed mechanisms for accommodating the adapting grids ensure that conservation is maintained and that the divergence of a constant vector field is always zero (freestream-preservation property). Advancement in time is achieved with a fourth-order Runge-Kutta method.
An effective quadrilateral mesh adaptation
KHATTRI Sanjay Kumar
2006-01-01
Accuracy of a simulation strongly depends on the grid quality. Here, quality means orthogonality at the boundaries and quasi-orthogonality within the critical regions, smoothness, bounded aspect ratios and solution adaptive behaviour. It is not recommended to refine the parts of the domain where the solution shows little variation. It is desired to concentrate grid points and cells in the part of the domain where the solution shows strong gradients or variations. We present a simple, effective and computationally efficient approach for quadrilateral mesh adaptation. Several numerical examples are presented for supporting our claim.
Donmez, O
2004-01-01
In this paper, the general procedure to solve the General Relativistic Hydrodynamical(GRH) equations with Adaptive-Mesh Refinement (AMR) is presented. In order to achieve, the GRH equations are written in the conservation form to exploit their hyperbolic character. The numerical solutions of general relativistic hydrodynamic equations are done by High Resolution Shock Capturing schemes (HRSC), specifically designed to solve non-linear hyperbolic systems of conservation laws. These schemes depend on the characteristic information of the system. The Marquina fluxes with MUSCL left and right states are used to solve GRH equations. First, different test problems with uniform and AMR grids on the special relativistic hydrodynamics equations are carried out to verify the second order convergence of the code in 1D, 2D and 3D. Results from uniform and AMR grid are compared. It is found that adaptive grid does a better job when the number of resolution is increased. Second, the general relativistic hydrodynamical equa...
MPDATA error estimator for mesh adaptivity
Szmelter, Joanna; Smolarkiewicz, Piotr K.
2006-04-01
In multidimensional positive definite advection transport algorithm (MPDATA) the leading error as well as the first- and second-order solutions are known explicitly by design. This property is employed to construct refinement indicators for mesh adaptivity. Recent progress with the edge-based formulation of MPDATA facilitates the use of the method in an unstructured-mesh environment. In particular, the edge-based data structure allows for flow solvers to operate on arbitrary hybrid meshes, thereby lending itself to implementations of various mesh adaptivity techniques. A novel unstructured-mesh nonoscillatory forward-in-time (NFT) solver for compressible Euler equations is used to illustrate the benefits of adaptive remeshing as well as mesh movement and enrichment for the efficacy of MPDATA-based flow solvers. Validation against benchmark test cases demonstrates robustness and accuracy of the approach.
Valdivia, Valeska
2014-01-01
Context. Ultraviolet radiation plays a crucial role in molecular clouds. Radiation and matter are tightly coupled and their interplay influences the physical and chemical properties of gas. In particular, modeling the radiation propagation requires calculating column densities, which can be numerically expensive in high-resolution multidimensional simulations. Aims. Developing fast methods for estimating column densities is mandatory if we are interested in the dynamical influence of the radiative transfer. In particular, we focus on the effect of the UV screening on the dynamics and on the statistical properties of molecular clouds. Methods. We have developed a tree-based method for a fast estimate of column densities, implemented in the adaptive mesh refinement code RAMSES. We performed numerical simulations using this method in order to analyze the influence of the screening on the clump formation. Results. We find that the accuracy for the extinction of the tree-based method is better than 10%, while the ...
Toward a Consistent Framework for High Order Mesh Refinement Schemes in Numerical Relativity
Mongwane, Bishop
2015-01-01
It has now become customary in the field of numerical relativity to couple high order finite difference schemes to mesh refinement algorithms. To this end, different modifications to the standard Berger-Oliger adaptive mesh refinement algorithm have been proposed. In this work we present a fourth order stable mesh refinement scheme with sub-cycling in time for numerical relativity. We do not use buffer zones to deal with refinement boundaries but explicitly specify boundary data for refined grids. We argue that the incompatibility of the standard mesh refinement algorithm with higher order Runge Kutta methods is a manifestation of order reduction phenomena, caused by inconsistent application of boundary data in the refined grids. Our scheme also addresses the problem of spurious reflections that are generated when propagating waves cross mesh refinement boundaries. We introduce a transition zone on refined levels within which the phase velocity of propagating modes is allowed to decelerate in order to smoothl...
Performance of a streaming mesh refinement algorithm.
Thompson, David C.; Pebay, Philippe Pierre
2004-08-01
In SAND report 2004-1617, we outline a method for edge-based tetrahedral subdivision that does not rely on saving state or communication to produce compatible tetrahedralizations. This report analyzes the performance of the technique by characterizing (a) mesh quality, (b) execution time, and (c) traits of the algorithm that could affect quality or execution time differently for different meshes. It also details the method used to debug the several hundred subdivision templates that the algorithm relies upon. Mesh quality is on par with other similar refinement schemes and throughput on modern hardware can exceed 600,000 output tetrahedra per second. But if you want to understand the traits of the algorithm, you have to read the report!
Dönmez, Orhan
2004-09-01
In this paper, the general procedure to solve the general relativistic hydrodynamical (GRH) equations with adaptive-mesh refinement (AMR) is presented. In order to achieve, the GRH equations are written in the conservation form to exploit their hyperbolic character. The numerical solutions of GRH equations are obtained by high resolution shock Capturing schemes (HRSC), specifically designed to solve nonlinear hyperbolic systems of conservation laws. These schemes depend on the characteristic information of the system. The Marquina fluxes with MUSCL left and right states are used to solve GRH equations. First, different test problems with uniform and AMR grids on the special relativistic hydrodynamics equations are carried out to verify the second-order convergence of the code in one, two and three dimensions. Results from uniform and AMR grid are compared. It is found that adaptive grid does a better job when the number of resolution is increased. Second, the GRH equations are tested using two different test problems which are Geodesic flow and Circular motion of particle In order to do this, the flux part of GRH equations is coupled with source part using Strang splitting. The coupling of the GRH equations is carried out in a treatment which gives second order accurate solutions in space and time.
Hybrid Surface Mesh Adaptation for Climate Modeling
Ahmed Khamayseh; Valmor de Almeida; Glen Hansen
2008-01-01
Solution-driven mesh adaptation is becoming quite popular for spatial error control in the numerical simulation of complex computational physics applications, such as climate modeling. Typically, spatial adaptation is achieved by element subdivision (h adaptation) with a primary goal of resolving the local length scales of interest. A second, lesspopular method of spatial adaptivity is called "mesh motion" (r adaptation); the smooth repositioning of mesh node points aimed at resizing existing elements to capture the local length scales. This paper proposes an adaptation method based on a combination of both element subdivision and node point repositioning (rh adaptation). By combining these two methods using the notion of a mobility function, the proposed approach seeks to increase the flexibility and extensibility of mesh motion algorithms while providing a somewhat smoother transition between refined regions than is pro-duced by element subdivision alone. Further, in an attempt to support the requirements of a very general class of climate simulation applications, the proposed method is de-signed to accommodate unstructured, polygonal mesh topologies in addition to the most popular mesh types.
Schartmann, M; Burkert, A; Gillessen, S; Genzel, R; Pfuhl, O; Eisenhauer, F; Plewa, P M; Ott, T; George, E M; Habibi, M
2015-01-01
The dusty, ionized gas cloud G2 is currently passing the massive black hole in the Galactic Center at a distance of roughly 2400 Schwarzschild radii. We explore the possibility of a starting point of the cloud within the disks of young stars. We make use of the large amount of new observations in order to put constraints on G2's origin. Interpreting the observations as a diffuse cloud of gas, we employ three-dimensional hydrodynamical adaptive mesh refinement (AMR) simulations with the PLUTO code and do a detailed comparison with observational data. The simulations presented in this work update our previously obtained results in multiple ways: (1) high resolution three-dimensional hydrodynamical AMR simulations are used, (2) the cloud follows the updated orbit based on the Brackett-$\\gamma$ data, (3) a detailed comparison to the observed high-quality position-velocity diagrams and the evolution of the total Brackett-$\\gamma$ luminosity is done. We concentrate on two unsolved problems of the diffuse cloud scen...
Adaptive sampling for mesh spectrum editing
ZHAO Xiang-jun; ZHANG Hong-xin; BAO Hu-jun
2006-01-01
A mesh editing framework is presented in this paper, which integrates Free-Form Deformation (FFD) and geometry signal processing. By using simplified model from original mesh, the editing task can be accomplished with a few operations. We take the deformation of the proxy and the position coordinates of the mesh models as geometry signal. Wavelet analysis is employed to separate local detail information gracefully. The crucial innovation of this paper is a new adaptive regular sampling approach for our signal analysis based editing framework. In our approach, an original mesh is resampled and then refined iteratively which reflects optimization of our proposed spectrum preserving energy. As an extension of our spectrum editing scheme,the editing principle is applied to geometry details transferring, which brings satisfying results.
Electrostatic PIC with adaptive Cartesian mesh
Kolobov, Vladimir I
2016-01-01
We describe an initial implementation of an electrostatic Particle-in-Cell (ES-PIC) module with adaptive Cartesian mesh in our Unified Flow Solver framework. Challenges of PIC method with cell-based adaptive mesh refinement (AMR) are related to a decrease of the particle-per-cell number in the refined cells with a corresponding increase of the numerical noise. The developed ES-PIC solver is validated for capacitively coupled plasma, its AMR capabilities are demonstrated for simulations of streamer development during high-pressure gas breakdown. It is shown that cell-based AMR provides a convenient particle management algorithm for exponential multiplications of electrons and ions in the ionization events.
Adaptive and Unstructured Mesh Cleaving
Bronson, Jonathan R.; Sastry, Shankar P.; Levine, Joshua A.; Whitaker, Ross T.
2015-01-01
We propose a new strategy for boundary conforming meshing that decouples the problem of building tetrahedra of proper size and shape from the problem of conforming to complex, non-manifold boundaries. This approach is motivated by the observation that while several methods exist for adaptive tetrahedral meshing, they typically have difficulty at geometric boundaries. The proposed strategy avoids this conflict by extracting the boundary conforming constraint into a secondary step. We first build a background mesh having a desired set of tetrahedral properties, and then use a generalized stenciling method to divide, or “cleave”, these elements to get a set of conforming tetrahedra, while limiting the impacts cleaving has on element quality. In developing this new framework, we make several technical contributions including a new method for building graded tetrahedral meshes as well as a generalization of the isosurface stuffing and lattice cleaving algorithms to unstructured background meshes. PMID:26137171
Details of tetrahedral anisotropic mesh adaptation
Jensen, Kristian Ejlebjerg; Gorman, Gerard
2016-04-01
We have implemented tetrahedral anisotropic mesh adaptation using the local operations of coarsening, swapping, refinement and smoothing in MATLAB without the use of any for- N loops, i.e. the script is fully vectorised. In the process of doing so, we have made three observations related to details of the implementation: 1. restricting refinement to a single edge split per element not only simplifies the code, it also improves mesh quality, 2. face to edge swapping is unnecessary, and 3. optimising for the Vassilevski functional tends to give a little higher value for the mean condition number functional than optimising for the condition number functional directly. These observations have been made for a uniform and a radial shock metric field, both starting from a structured mesh in a cube. Finally, we compare two coarsening techniques and demonstrate the importance of applying smoothing in the mesh adaptation loop. The results pertain to a unit cube geometry, but we also show the effect of corners and edges by applying the implementation in a spherical geometry.
Valdivia, Valeska; Hennebelle, Patrick
2014-11-01
Context. Ultraviolet radiation plays a crucial role in molecular clouds. Radiation and matter are tightly coupled and their interplay influences the physical and chemical properties of gas. In particular, modeling the radiation propagation requires calculating column densities, which can be numerically expensive in high-resolution multidimensional simulations. Aims: Developing fast methods for estimating column densities is mandatory if we are interested in the dynamical influence of the radiative transfer. In particular, we focus on the effect of the UV screening on the dynamics and on the statistical properties of molecular clouds. Methods: We have developed a tree-based method for a fast estimate of column densities, implemented in the adaptive mesh refinement code RAMSES. We performed numerical simulations using this method in order to analyze the influence of the screening on the clump formation. Results: We find that the accuracy for the extinction of the tree-based method is better than 10%, while the relative error for the column density can be much more. We describe the implementation of a method based on precalculating the geometrical terms that noticeably reduces the calculation time. To study the influence of the screening on the statistical properties of molecular clouds we present the probability distribution function of gas and the associated temperature per density bin and the mass spectra for different density thresholds. Conclusions: The tree-based method is fast and accurate enough to be used during numerical simulations since no communication is needed between CPUs when using a fully threaded tree. It is then suitable to parallel computing. We show that the screening for far UV radiation mainly affects the dense gas, thereby favoring low temperatures and affecting the fragmentation. We show that when we include the screening, more structures are formed with higher densities in comparison to the case that does not include this effect. We
Parallel Mesh Adaptive Techniques for Complex Flow Simulation: Geometry Conservation
Angelo Casagrande
2012-01-01
Full Text Available Dynamic mesh adaptation on unstructured grids, by localised refinement and derefinement, is a very efficient tool for enhancing solution accuracy and optimising computational time. One of the major drawbacks, however, resides in the projection of the new nodes created, during the refinement process, onto the boundary surfaces. This can be addressed by the introduction of a library capable of handling geometric properties given by a CAD (computer-aided design description. This is of particular interest also to enhance the adaptation module when the mesh is being smoothed, and hence moved, to then reproject it onto the surface of the exact geometry.
Evolutions in 3D numerical relativity using fixed mesh refinement
Schnetter, E; Hawke, I; Schnetter, Erik; Hawley, Scott H.; Hawke, Ian
2004-01-01
We present results of 3D numerical simulations using a finite difference code featuring fixed mesh refinement (FMR), in which a subset of the computational domain is refined in space and time. We apply this code to a series of test cases including a robust stability test, a nonlinear gauge wave and an excised Schwarzschild black hole in an evolving gauge. We find that the mesh refinement results are comparable in accuracy, stability and convergence to unigrid simulations with the same effective resolution. At the same time, the use of FMR reduces the computational resources needed to obtain a given accuracy. Particular care must be taken at the interfaces between coarse and fine grids to avoid a loss of convergence at high resolutions. This FMR system, "Carpet", is a driver module in the freely available Cactus computational infrastructure, and is able to endow existing Cactus simulation modules ("thorns") with FMR with little or no extra effort.
h-Refinement for simple corner balance scheme of SN transport equation on distorted meshes
Yang, Rong; Yuan, Guangwei
2016-11-01
The transport sweep algorithm is a common method for solving discrete ordinate transport equation, but it breaks down once a concave cell appears in spatial meshes. To deal with this issue a local h-refinement for simple corner balance (SCB) scheme of SN transport equation on arbitrary quadrilateral meshes is presented in this paper by using a new subcell partition. It follows that a hybrid mesh with both triangle and quadrilateral cells is generated, and the geometric quality of these cells improves, especially it is ensured that all cells become convex. Combining with the original SCB scheme, an adaptive transfer algorithm based on the hybrid mesh is constructed. Numerical experiments are presented to verify the utility and accuracy of the new algorithm, especially for some application problems such as radiation transport coupled with Lagrangian hydrodynamic flow. The results show that it performs well on extremely distorted meshes with concave cells, on which the original SCB scheme does not work.
Core, X.
2002-02-01
The isobar approximation for the system of the balance equations of mass, momentum, energy and chemical species is a suitable approximation to represent low Mach number reactive flows. In this approximation, which neglects acoustics phenomena, the mixture is hydrodynamically incompressible and the thermodynamic effects lead to an uniform compression of the system. We present a novel numerical scheme for this approximation. An incremental projection method, which uses the original form of mass balance equation, discretizes in time the Navier-Stokes equations. Spatial discretization is achieved through a finite volume approach on MAC-type staggered mesh. A higher order de-centered scheme is used to compute the convective fluxes. We associate to this discretization a local mesh refinement method, based on Flux Interface Correction technique. A first application concerns a forced flow with variable density which mimics a combustion problem. The second application is natural convection with first small temperature variations and then beyond the limit of validity of the Boussinesq approximation. Finally, we treat a third application which is a laminar diffusion flame. For each of these test problems, we demonstrate the robustness of the proposed numerical scheme, notably for the density spatial variations. We analyze the gain in accuracy obtained with the local mesh refinement method. (author)
Sascha M. Schnepp
2012-01-01
Full Text Available An extension of the framework of the Finite Integration Technique (FIT including dynamic and adaptive mesh refinement is presented. After recalling the standard formulation of the FIT, the proposed mesh adaptation procedure is described. Besides the linear interpolation approach, a novel interpolation technique based on specialized spline functions for approximating the discrete electromagnetic field solution during mesh adaptation is introduced. The standard FIT on a fixed mesh and the new adaptive approach are applied to a simulation test case with a known analytical solution. The numerical accuracy of the two methods is shown to be comparable. The dynamic mesh approach is, however, much more efficient. This is demonstrated with the full scale modeling of the complete rf gun at the Photo Injector Test Facility DESY Zeuthen (PITZ on a single computer. Results of a detailed design study addressing the effects of individual components of the gun onto the beam emittance using a fully self-consistent approach are presented.
Schnepp, Sascha M; Weiland, Thomas
2011-01-01
An extension of the framework of the Finite Integration Technique (FIT) including dynamic and adaptive mesh refinement is presented. After recalling the standard formulation of the FIT, the proposed mesh adaptation procedure is described. Besides the linear interpolation approach, a novel interpolation technique based on specialized spline functions for approximating the discrete electromagnetic field solution during mesh adaptation is introduced. The standard FIT on a fixed mesh and the new adaptive approach are applied to a simulation test case with known analytical solution. The numerical accuracy of the two methods are shown to be comparable. The dynamic mesh approach is, however, much more efficient. This is also demonstrated for the full scale modeling of the complete RF gun at the Photo Injector Test Facility DESY Zeuthen (PITZ) on a single computer. Results of a detailed design study addressing the effects of individual components of the gun onto the beam emittance using a fully self-consistent approa...
Adaptive Mesh Fluid Simulations on GPU
Wang, Peng; Kaehler, Ralf
2009-01-01
We describe an implementation of compressible inviscid fluid solvers with block-structured adaptive mesh refinement on Graphics Processing Units using NVIDIA's CUDA. We show that a class of high resolution shock capturing schemes can be mapped naturally on this architecture. Using the method of lines approach with the second order total variation diminishing Runge-Kutta time integration scheme, piecewise linear reconstruction, and a Harten-Lax-van Leer Riemann solver, we achieve an overall speedup of approximately 10 times faster execution on one graphics card as compared to a single core on the host computer. We attain this speedup in uniform grid runs as well as in problems with deep AMR hierarchies. Our framework can readily be applied to more general systems of conservation laws and extended to higher order shock capturing schemes. This is shown directly by an implementation of a magneto-hydrodynamic solver and comparing its performance to the pure hydrodynamic case. Finally, we also combined our CUDA par...
A multilevel adaptive mesh generation scheme using Kd-trees
Alfonso Limon
2009-04-01
Full Text Available We introduce a mesh refinement strategy for PDE based simulations that benefits from a multilevel decomposition. Using Harten's MRA in terms of Schroder-Pander linear multiresolution analysis [20], we are able to bound discontinuities in $mathbb{R}$. This MRA is extended to $mathbb{R}^n$ in terms of n-orthogonal linear transforms and utilized to identify cells that contain a codimension-one discontinuity. These refinement cells become leaf nodes in a balanced Kd-tree such that a local dyadic MRA is produced in $mathbb{R}^n$, while maintaining a minimal computational footprint. The nodes in the tree form an adaptive mesh whose density increases in the vicinity of a discontinuity.
Yan, Su; Arslanbekov, Robert R; Kolobov, Vladimir I; Jin, Jian-Ming
2016-01-01
A discontinuous Galerkin time-domain (DGTD) method based on dynamically adaptive Cartesian meshes (ACM) is developed for a full-wave analysis of electromagnetic fields in dispersive media. Hierarchical Cartesian grids offer simplicity close to that of structured grids and the flexibility of unstructured grids while being highly suited for adaptive mesh refinement (AMR). The developed DGTD-ACM achieves a desired accuracy by refining non-conformal meshes near material interfaces to reduce stair-casing errors without sacrificing the high efficiency afforded with uniform Cartesian meshes. Moreover, DGTD-ACM can dynamically refine the mesh to resolve the local variation of the fields during propagation of electromagnetic pulses. A local time-stepping scheme is adopted to alleviate the constraint on the time-step size due to the stability condition of the explicit time integration. Simulations of electromagnetic wave diffraction over conducting and dielectric cylinders and spheres demonstrate that the proposed meth...
Local mesh refinement for incompressible fluid flow with free surfaces
Terasaka, H.; Kajiwara, H.; Ogura, K. [Tokyo Electric Power Company (Japan)] [and others
1995-09-01
A new local mesh refinement (LMR) technique has been developed and applied to incompressible fluid flows with free surface boundaries. The LMR method embeds patches of fine grid in arbitrary regions of interest. Hence, more accurate solutions can be obtained with a lower number of computational cells. This method is very suitable for the simulation of free surface movements because free surface flow problems generally require a finer computational grid to obtain adequate results. By using this technique, one can place finer grids only near the surfaces, and therefore greatly reduce the total number of cells and computational costs. This paper introduces LMR3D, a three-dimensional incompressible flow analysis code. Numerical examples calculated with the code demonstrate well the advantages of the LMR method.
An Adaptive Mesh Algorithm: Mapping the Mesh Variables
Scannapieco, Anthony J. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)
2016-07-25
Both thermodynamic and kinematic variables must be mapped. The kinematic variables are defined on a separate kinematic mesh; it is the duel mesh to the thermodynamic mesh. The map of the kinematic variables is done by calculating the contributions of kinematic variables on the old thermodynamic mesh, mapping the kinematic variable contributions onto the new thermodynamic mesh and then synthesizing the mapped kinematic variables on the new kinematic mesh. In this document the map of the thermodynamic variables will be described.
Mesh saliency with adaptive local patches
Nouri, Anass; Charrier, Christophe; Lézoray, Olivier
2015-03-01
3D object shapes (represented by meshes) include both areas that attract the visual attention of human observers and others less or not attractive at all. This visual attention depends on the degree of saliency exposed by these areas. In this paper, we propose a technique for detecting salient regions in meshes. To do so, we define a local surface descriptor based on local patches of adaptive size and filled with a local height field. The saliency of mesh vertices is then defined as its degree measure with edges weights computed from adaptive patch similarities. Our approach is compared to the state-of-the-art and presents competitive results. A study evaluating the influence of the parameters establishing this approach is also carried out. The strength and the stability of our approach with respect to noise and simplification are also studied.
Development and Verification of Unstructured Adaptive Mesh Technique with Edge Compatibility
Ito, Kei; Kunugi, Tomoaki; Ohshima, Hiroyuki
In the design study of the large-sized sodium-cooled fast reactor (JSFR), one key issue is suppression of gas entrainment (GE) phenomena at a gas-liquid interface. Therefore, the authors have been developed a high-precision CFD algorithm to evaluate the GE phenomena accurately. The CFD algorithm has been developed on unstructured meshes to establish an accurate modeling of JSFR system. For two-phase interfacial flow simulations, a high-precision volume-of-fluid algorithm is employed. It was confirmed that the developed CFD algorithm could reproduce the GE phenomena in a simple GE experiment. Recently, the authors have been developed an important technique for the simulation of the GE phenomena in JSFR. That is an unstructured adaptive mesh technique which can apply fine cells dynamically to the region where the GE occurs in JSFR. In this paper, as a part of the development, a two-dimensional unstructured adaptive mesh technique is discussed. In the two-dimensional adaptive mesh technique, each cell is refined isotropically to reduce distortions of the mesh. In addition, connection cells are formed to eliminate the edge incompatibility between refined and non-refined cells. The two-dimensional unstructured adaptive mesh technique is verified by solving well-known lid-driven cavity flow problem. As a result, the two-dimensional unstructured adaptive mesh technique succeeds in providing a high-precision solution, even though poor-quality distorted initial mesh is employed. In addition, the simulation error on the two-dimensional unstructured adaptive mesh is much less than the error on the structured mesh with a larger number of cells.
Automatic off-body overset adaptive Cartesian mesh method based on an octree approach
Peron, Stephanie, E-mail: stephanie.peron@onera.fr [ONERA - The French Aerospace Lab, F-92322 Chatillon (France); Benoit, Christophe, E-mail: christophe.benoit@onera.fr [ONERA - The French Aerospace Lab, F-92322 Chatillon (France)
2013-01-01
This paper describes a method for generating adaptive structured Cartesian grids within a near-body/off-body mesh partitioning framework for the flow simulation around complex geometries. The off-body Cartesian mesh generation derives from an octree structure, assuming each octree leaf node defines a structured Cartesian block. This enables one to take into account the large scale discrepancies in terms of resolution between the different bodies involved in the simulation, with minimum memory requirements. Two different conversions from the octree to Cartesian grids are proposed: the first one generates Adaptive Mesh Refinement (AMR) type grid systems, and the second one generates abutting or minimally overlapping Cartesian grid set. We also introduce an algorithm to control the number of points at each adaptation, that automatically determines relevant values of the refinement indicator driving the grid refinement and coarsening. An application to a wing tip vortex computation assesses the capability of the method to capture accurately the flow features.
Mesh Generation via Local Bisection Refinement of Triangulated Grids
2015-06-01
tb [Maubach 1995, Theorem 5.1]. This is exploited in the recursive algorithm RefineTriangle (Algorithm 2) to com- patibly refine a given triangle; the... recursion depth of RefineTriangle is bounded by the maximum level of refinement in T [Maubach 1995]. RefineTriangle calls itself repeatedly on a... sequence of triangles until a compatibly divisible triangle is found. This sequence of triangles is then bisected in reverse order to preserve
Nonhydrostatic adaptive mesh dynamics for multiscale climate models (Invited)
Collins, W.; Johansen, H.; McCorquodale, P.; Colella, P.; Ullrich, P. A.
2013-12-01
Many of the atmospheric phenomena with the greatest potential impact in future warmer climates are inherently multiscale. Such meteorological systems include hurricanes and tropical cyclones, atmospheric rivers, and other types of hydrometeorological extremes. These phenomena are challenging to simulate in conventional climate models due to the relatively coarse uniform model resolutions relative to the native nonhydrostatic scales of the phenomonological dynamics. To enable studies of these systems with sufficient local resolution for the multiscale dynamics yet with sufficient speed for climate-change studies, we have adapted existing adaptive mesh dynamics for the DOE-NSF Community Atmosphere Model (CAM). In this talk, we present an adaptive, conservative finite volume approach for moist non-hydrostatic atmospheric dynamics. The approach is based on the compressible Euler equations on 3D thin spherical shells, where the radial direction is treated implicitly (using a fourth-order Runga-Kutta IMEX scheme) to eliminate time step constraints from vertical acoustic waves. Refinement is performed only in the horizontal directions. The spatial discretization is the equiangular cubed-sphere mapping, with a fourth-order accurate discretization to compute flux averages on faces. By using both space-and time-adaptive mesh refinement, the solver allocates computational effort only where greater accuracy is needed. The resulting method is demonstrated to be fourth-order accurate for model problems, and robust at solution discontinuities and stable for large aspect ratios. We present comparisons using a simplified physics package for dycore comparisons of moist physics. Hadley cell lifting an advected tracer into upper atmosphere, with horizontal adaptivity
Adaptive meshing technique applied to an orthopaedic finite element contact problem.
Roarty, Colleen M; Grosland, Nicole M
2004-01-01
Finite element methods have been applied extensively and with much success in the analysis of orthopaedic implants. Recently a growing interest has developed, in the orthopaedic biomechanics community, in how numerical models can be constructed for the optimal solution of problems in contact mechanics. New developments in this area are of paramount importance in the design of improved implants for orthopaedic surgery. Finite element and other computational techniques are widely applied in the analysis and design of hip and knee implants, with additional joints (ankle, shoulder, wrist) attracting increased attention. The objective of this investigation was to develop a simplified adaptive meshing scheme to facilitate the finite element analysis of a dual-curvature total wrist implant. Using currently available software, the analyst has great flexibility in mesh generation, but must prescribe element sizes and refinement schemes throughout the domain of interest. Unfortunately, it is often difficult to predict in advance a mesh spacing that will give acceptable results. Adaptive finite-element mesh capabilities operate to continuously refine the mesh to improve accuracy where it is required, with minimal intervention by the analyst. Such mesh adaptation generally means that in certain areas of the analysis domain, the size of the elements is decreased (or increased) and/or the order of the elements may be increased (or decreased). In concept, mesh adaptation is very appealing. Although there have been several previous applications of adaptive meshing for in-house FE codes, we have coupled an adaptive mesh formulation with the pre-existing commercial programs PATRAN (MacNeal-Schwendler Corp., USA) and ABAQUS (Hibbit Karlson and Sorensen, Pawtucket, RI). In doing so, we have retained several attributes of the commercial software, which are very attractive for orthopaedic implant applications.
A REGIONAL REFINEMENT FOR FINITE ELEMENT MESH DESIGN USING COLLAPSIBLE ELEMENT
Priyo Suprobo
2000-01-01
Full Text Available A practical algorithm for automated mesh design in finite element analysis is developed. A regional mixed mesh improvement procedure is introduced. The error control%2C algorithm implementation%2C code development%2C and the solution accuracy are discussed. Numerical example includes automated mesh designs for plane elastic media with singularities. The efficiency of the procedure is demonstrated. Abstract in Bahasa Indonesia : regional+refinement%2C+mesh+generation%2C+isoparametric+element%2C+collapsible+element
A unified framework for mesh refinement in random and physical space
Li, Jing; Stinis, Panos
2016-10-01
In recent work we have shown how an accurate reduced model can be utilized to perform mesh refinement in random space. That work relied on the explicit knowledge of an accurate reduced model which is used to monitor the transfer of activity from the large to the small scales of the solution. Since this is not always available, we present in the current work a framework which shares the merits and basic idea of the previous approach but does not require an explicit knowledge of a reduced model. Moreover, the current framework can be applied for refinement in both random and physical space. In this manuscript we focus on the application to random space mesh refinement. We study examples of increasing difficulty (from ordinary to partial differential equations) which demonstrate the efficiency and versatility of our approach. We also provide some results from the application of the new framework to physical space mesh refinement.
Areias, P.; Rabczuk, T.; de Sá, J. César
2016-12-01
We propose an alternative crack propagation algorithm which effectively circumvents the variable transfer procedure adopted with classical mesh adaptation algorithms. The present alternative consists of two stages: a mesh-creation stage where a local damage model is employed with the objective of defining a crack-conforming mesh and a subsequent analysis stage with a localization limiter in the form of a modified screened Poisson equation which is exempt of crack path calculations. In the second stage, the crack naturally occurs within the refined region. A staggered scheme for standard equilibrium and screened Poisson equations is used in this second stage. Element subdivision is based on edge split operations using a constitutive quantity (damage). To assess the robustness and accuracy of this algorithm, we use five quasi-brittle benchmarks, all successfully solved.
Lober, R.R.; Tautges, T.J.; Vaughan, C.T.
1997-03-01
Paving is an automated mesh generation algorithm which produces all-quadrilateral elements. It can additionally generate these elements in varying sizes such that the resulting mesh adapts to a function distribution, such as an error function. While powerful, conventional paving is a very serial algorithm in its operation. Parallel paving is the extension of serial paving into parallel environments to perform the same meshing functions as conventional paving only on distributed, discretized models. This extension allows large, adaptive, parallel finite element simulations to take advantage of paving`s meshing capabilities for h-remap remeshing. A significantly modified version of the CUBIT mesh generation code has been developed to host the parallel paving algorithm and demonstrate its capabilities on both two dimensional and three dimensional surface geometries and compare the resulting parallel produced meshes to conventionally paved meshes for mesh quality and algorithm performance. Sandia`s {open_quotes}tiling{close_quotes} dynamic load balancing code has also been extended to work with the paving algorithm to retain parallel efficiency as subdomains undergo iterative mesh refinement.
Parallel, Gradient-Based Anisotropic Mesh Adaptation for Re-entry Vehicle Configurations
Bibb, Karen L.; Gnoffo, Peter A.; Park, Michael A.; Jones, William T.
2006-01-01
Two gradient-based adaptation methodologies have been implemented into the Fun3d refine GridEx infrastructure. A spring-analogy adaptation which provides for nodal movement to cluster mesh nodes in the vicinity of strong shocks has been extended for general use within Fun3d, and is demonstrated for a 70 sphere cone at Mach 2. A more general feature-based adaptation metric has been developed for use with the adaptation mechanics available in Fun3d, and is applicable to any unstructured, tetrahedral, flow solver. The basic functionality of general adaptation is explored through a case of flow over the forebody of a 70 sphere cone at Mach 6. A practical application of Mach 10 flow over an Apollo capsule, computed with the Felisa flow solver, is given to compare the adaptive mesh refinement with uniform mesh refinement. The examples of the paper demonstrate that the gradient-based adaptation capability as implemented can give an improvement in solution quality.
Ralf Deiterding
2011-01-01
Full Text Available Numerical simulation can be key to the understanding of the multidimensional nature of transient detonation waves. However, the accurate approximation of realistic detonations is demanding as a wide range of scales needs to be resolved. This paper describes a successful solution strategy that utilizes logically rectangular dynamically adaptive meshes. The hydrodynamic transport scheme and the treatment of the nonequilibrium reaction terms are sketched. A ghost fluid approach is integrated into the method to allow for embedded geometrically complex boundaries. Large-scale parallel simulations of unstable detonation structures of Chapman-Jouguet detonations in low-pressure hydrogen-oxygen-argon mixtures demonstrate the efficiency of the described techniques in practice. In particular, computations of regular cellular structures in two and three space dimensions and their development under transient conditions, that is, under diffraction and for propagation through bends are presented. Some of the observed patterns are classified by shock polar analysis, and a diagram of the transition boundaries between possible Mach reflection structures is constructed.
Dynamic mesh refinement for discrete models of jet electro-hydrodynamics
Lauricella, Marco; Pisignano, Dario; Succi, Sauro
2015-01-01
Nowadays, several models of unidimensional fluid jets exploit discrete element methods. In some cases, as for models aiming at describing the electrospinning nanofabrication process of polymer fibers, discrete element methods suffer a non constant resolution of the jet representation. We develop a dynamic mesh-refinement method for the numerical study of the electro-hydrodynamic behavior of charged jets using discrete element methods. To this purpose, we import ideas and techniques from the string method originally developed in the framework of free-energy landscape simulations. The mesh-refined discrete element method is demonstrated for the case of electrospinning applications.
Brislawn, Kristi D.; Brown, David L.; Chesshire, Geoffrey S.; Saltzman, Jeffrey S.
1995-01-01
Adaptive mesh refinement (AMR) in conjunction with higher-order upwind finite-difference methods have been used effectively on a variety of problems in two and three dimensions. In this paper we introduce an approach for resolving problems that involve complex geometries in which resolution of boundary geometry is important. The complex geometry is represented by using the method of overlapping grids, while local resolution is obtained by refining each component grid with the AMR algorithm, appropriately generalized for this situation. The CMPGRD algorithm introduced by Chesshire and Henshaw is used to automatically generate the overlapping grid structure for the underlying mesh.
AbouEisha, Hassan M.
2017-07-13
We consider a class of two-and three-dimensional h-refined meshes generated by an adaptive finite element method. We introduce an element partition tree, which controls the execution of the multi-frontal solver algorithm over these refined grids. We propose and study algorithms with polynomial computational cost for the optimization of these element partition trees. The trees provide an ordering for the elimination of unknowns. The algorithms automatically optimize the element partition trees using extensions of dynamic programming. The construction of the trees by the dynamic programming approach is expensive. These generated trees cannot be used in practice, but rather utilized as a learning tool to propose fast heuristic algorithms. In this first part of our paper we focus on the dynamic programming approach, and draw a sketch of the heuristic algorithm. The second part will be devoted to a more detailed analysis of the heuristic algorithm extended for the case of hp-adaptive
Numerical modeling of seismic waves using frequency-adaptive meshes
Hu, Jinyin; Jia, Xiaofeng
2016-08-01
An improved modeling algorithm using frequency-adaptive meshes is applied to meet the computational requirements of all seismic frequency components. It automatically adopts coarse meshes for low-frequency computations and fine meshes for high-frequency computations. The grid intervals are adaptively calculated based on a smooth inversely proportional function of grid size with respect to the frequency. In regular grid-based methods, the uniform mesh or non-uniform mesh is used for frequency-domain wave propagators and it is fixed for all frequencies. A too coarse mesh results in inaccurate high-frequency wavefields and unacceptable numerical dispersion; on the other hand, an overly fine mesh may cause storage and computational overburdens as well as invalid propagation angles of low-frequency wavefields. Experiments on the Padé generalized screen propagator indicate that the Adaptive mesh effectively solves these drawbacks of regular fixed-mesh methods, thus accurately computing the wavefield and its propagation angle in a wide frequency band. Several synthetic examples also demonstrate its feasibility for seismic modeling and migration.
Layer-adapted meshes for reaction-convection-diffusion problems
Linß, Torsten
2010-01-01
This book on numerical methods for singular perturbation problems - in particular, stationary reaction-convection-diffusion problems exhibiting layer behaviour is devoted to the construction and analysis of layer-adapted meshes underlying these numerical methods. A classification and a survey of layer-adapted meshes for reaction-convection-diffusion problems are included. This structured and comprehensive account of current ideas in the numerical analysis for various methods on layer-adapted meshes is addressed to researchers in finite element theory and perturbation problems. Finite differences, finite elements and finite volumes are all covered.
Adaptive-mesh algorithms for computational fluid dynamics
Powell, Kenneth G.; Roe, Philip L.; Quirk, James
1993-01-01
The basic goal of adaptive-mesh algorithms is to distribute computational resources wisely by increasing the resolution of 'important' regions of the flow and decreasing the resolution of regions that are less important. While this goal is one that is worthwhile, implementing schemes that have this degree of sophistication remains more of an art than a science. In this paper, the basic pieces of adaptive-mesh algorithms are described and some of the possible ways to implement them are discussed and compared. These basic pieces are the data structure to be used, the generation of an initial mesh, the criterion to be used to adapt the mesh to the solution, and the flow-solver algorithm on the resulting mesh. Each of these is discussed, with particular emphasis on methods suitable for the computation of compressible flows.
Kinetic Solvers with Adaptive Mesh in Phase Space
Arslanbekov, Robert R; Frolova, Anna A
2013-01-01
An Adaptive Mesh in Phase Space (AMPS) methodology has been developed for solving multi-dimensional kinetic equations by the discrete velocity method. A Cartesian mesh for both configuration (r) and velocity (v) spaces is produced using a tree of trees data structure. The mesh in r-space is automatically generated around embedded boundaries and dynamically adapted to local solution properties. The mesh in v-space is created on-the-fly for each cell in r-space. Mappings between neighboring v-space trees implemented for the advection operator in configuration space. We have developed new algorithms for solving the full Boltzmann and linear Boltzmann equations with AMPS. Several recent innovations were used to calculate the full Boltzmann collision integral with dynamically adaptive mesh in velocity space: importance sampling, multi-point projection method, and the variance reduction method. We have developed an efficient algorithm for calculating the linear Boltzmann collision integral for elastic and inelastic...
Adaptive refinement tools for tetrahedral unstructured grids
Pao, S. Paul (Inventor); Abdol-Hamid, Khaled S. (Inventor)
2011-01-01
An exemplary embodiment providing one or more improvements includes software which is robust, efficient, and has a very fast run time for user directed grid enrichment and flow solution adaptive grid refinement. All user selectable options (e.g., the choice of functions, the choice of thresholds, etc.), other than a pre-marked cell list, can be entered on the command line. The ease of application is an asset for flow physics research and preliminary design CFD analysis where fast grid modification is often needed to deal with unanticipated development of flow details.
A Hybrid Advection Scheme for Conserving Angular Momentum on a Refined Cartesian Mesh
Byerly, Zachary D; Tohline, Joel E; Marcello, Dominic C
2014-01-01
We test a new "hybrid" scheme for simulating dynamical fluid flows in which cylindrical components of the momentum are advected across a rotating Cartesian coordinate mesh. This hybrid scheme allows us to conserve angular momentum to machine precision while capitalizing on the advantages offered by a Cartesian mesh, such as a straightforward implementation of mesh refinement. Our test focuses on measuring the real and imaginary parts of the eigenfrequency of unstable axisymmetric modes that naturally arise in massless polytropic tori having a range of different aspect ratios, and quantifying the uncertainty in these measurements. Our measured eigenfrequencies show good agreement with the results obtained from the linear stability analysis of Kojima (1986) and from nonlinear hydrodynamic simulations performed on a cylindrical coordinate mesh by Woodward et al. (1994). When compared against results conducted with a traditional Cartesian advection scheme, the hybrid scheme achieves qualitative convergence at the...
Phase-field simulation of dendritic solidification using a full threaded tree with adaptive meshing
Yin Yajun; Zhou Jianxin; Liao Dunming; Pang Shengyong; Shen Xu
2014-01-01
Simulation of the microstructure evolution during solidification is greatly beneficial to the control of solidification microstructures. A phase-field method based on the ful threaded tree (FTT) for the simulation of casting solidification microstructure was proposed in this paper, and the structure of the ful threaded tree and the mesh refinement method was discussed. During dendritic growth in solidification, the mesh for simulation is adaptively refined at the liquid-solid interface, and coarsened in other areas. The numerical results of a three-dimension dendrite growth indicate that the phase-field method based on FTT is suitable for microstructure simulation. Most importantly, the FTT method can increase the spatial and temporal resolutions beyond the limits imposed by the available hardware compared with the conventional uniform mesh. At the simulation time of 0.03 s in this study, the computer memory used for computation is no more than 10 MB with the FTT method, while it is about 50 MB with the uniform mesh method. In addition, the proposed FTT method is more efficient in computation time when compared with the uniform mesh method. It would take about 20 h for the uniform mesh method, while only 2 h for the FTT method for computation when the solidification time is 0.17 s in this study.
Dimensional reduction as a tool for mesh refinement and trackingsingularities of PDEs
Stinis, Panagiotis
2007-06-10
We present a collection of algorithms which utilizedimensional reduction to perform mesh refinement and study possiblysingular solutions of time-dependent partial differential equations. Thealgorithms are inspired by constructions used in statistical mechanics toevaluate the properties of a system near a critical point. The firstalgorithm allows the accurate determination of the time of occurrence ofa possible singularity. The second algorithm is an adaptive meshrefinement scheme which can be used to approach efficiently the possiblesingularity. Finally, the third algorithm uses the second algorithm untilthe available resolution is exhausted (as we approach the possiblesingularity) and then switches to a dimensionally reduced model which,when accurate, can follow faithfully the solution beyond the time ofoccurrence of the purported singularity. An accurate dimensionallyreduced model should dissipate energy at the right rate. We construct twovariants of each algorithm. The first variant assumes that we have actualknowledge of the reduced model. The second variant assumes that we knowthe form of the reduced model, i.e., the terms appearing in the reducedmodel, but not necessarily their coefficients. In this case, we alsoprovide a way of determining the coefficients. We present numericalresults for the Burgers equation with zero and nonzero viscosity toillustrate the use of the algorithms.
Mosler, J.; Ortiz, M.
2009-01-01
A variational h-adaptive finite element formulation is proposed. The distinguishing feature of this method is that mesh refinement and coarsening are governed by the same minimization principle characterizing the underlying physical problem. Hence, no error estimates are invoked at any stage of the adaption procedure. As a consequence, linearity of the problem and a corresponding Hilbert-space functional framework are not required and the proposed formulation can be applied to hig...
Block-structured adaptive meshes and reduced grids for atmospheric general circulation models.
Jablonowski, Christiane; Oehmke, Robert C; Stout, Quentin F
2009-11-28
Adaptive mesh refinement techniques offer a flexible framework for future variable-resolution climate and weather models since they can focus their computational mesh on certain geographical areas or atmospheric events. Adaptive meshes can also be used to coarsen a latitude-longitude grid in polar regions. This allows for the so-called reduced grid setups. A spherical, block-structured adaptive grid technique is applied to the Lin-Rood finite-volume dynamical core for weather and climate research. This hydrostatic dynamics package is based on a conservative and monotonic finite-volume discretization in flux form with vertically floating Lagrangian layers. The adaptive dynamical core is built upon a flexible latitude-longitude computational grid and tested in two- and three-dimensional model configurations. The discussion is focused on static mesh adaptations and reduced grids. The two-dimensional shallow water setup serves as an ideal testbed and allows the use of shallow water test cases like the advection of a cosine bell, moving vortices, a steady-state flow, the Rossby-Haurwitz wave or cross-polar flows. It is shown that reduced grid configurations are viable candidates for pure advection applications but should be used moderately in nonlinear simulations. In addition, static grid adaptations can be successfully used to resolve three-dimensional baroclinic waves in the storm-track region.
Implementations of mesh refinement schemes for particle-in-cell plasma simulations
Vay, J.-L.; Colella, P.; Friedman, A.; Grote, D.P.; McCorquodale, P.; Serafini, D.B.
2003-10-20
Plasma simulations are often rendered challenging by the disparity of scales in time and in space which must be resolved. When these disparities are in distinctive zones of the simulation region, a method which has proven to be effective in other areas (e.g. fluid dynamics simulations) is the mesh refinement technique. We briefly discuss the challenges posed by coupling this technique with plasma Particle-In-Cell simulations and present two implementations in more detail, with examples.
Adaptive upscaling with the dual mesh method
Guerillot, D.; Verdiere, S.
1997-08-01
The objective of this paper is to demonstrate that upscaling should be calculated during the flow simulation instead of trying to enhance the a priori upscaling methods. Hence, counter-examples are given to motivate our approach, the so-called Dual Mesh Method. The main steps of this numerical algorithm are recalled. Applications illustrate the necessity to consider different average relative permeability values depending on the direction in space. Moreover, these values could be different for the same average saturation. This proves that an a priori upscaling cannot be the answer even in homogeneous cases because of the {open_quotes}dynamical heterogeneity{close_quotes} created by the saturation profile. Other examples show the efficiency of the Dual Mesh Method applied to heterogeneous medium and to an actual field case in South America.
GENERATION AND APPLICATION OF UNSTRUCTURED ADAPTIVE MESHES WITH MOVING BOUNDARIES
无
2001-01-01
This paper presents a method to generate unstructured adaptive meshes with moving boundaries and its application to CFD. Delaunay triangulation criterion in conjunction with the automatic point creation is used to generate 2-D and 3-D unstructured grids. A local grid regeneration method is proposed to cope with moving boundaries. Numerical examples include the interactions of shock waves with movable bodies and the movement of a projectile within a ram accelerator, illustrating an efficient and robust mesh generation method developed.``
Adaptive mesh generation for viscous flows using Delaunay triangulation
Mavriplis, Dimitri J.
1990-01-01
A method for generating an unstructured triangular mesh in two dimensions, suitable for computing high Reynolds number flows over arbitrary configurations is presented. The method is based on a Delaunay triangulation, which is performed in a locally stretched space, in order to obtain very high aspect ratio triangles in the boundary layer and the wake regions. It is shown how the method can be coupled with an unstructured Navier-Stokes solver to produce a solution adaptive mesh generation procedure for viscous flows.
Automatic mesh adaptivity for CADIS and FW-CADIS neutronics modeling of difficult shielding problems
Ibrahim, A. M.; Peplow, D. E.; Mosher, S. W.; Wagner, J. C.; Evans, T. M. [Oak Ridge National Laboratory, P.O. Box 2008, Oak Ridge, TN 37831 (United States); Wilson, P. P.; Sawan, M. E. [University of Wisconsin-Madison, 1500 Engineering Dr., Madison, WI 53706 (United States)
2013-07-01
The CADIS and FW-CADIS hybrid Monte Carlo/deterministic techniques dramatically increase the efficiency of neutronics modeling, but their use in the accurate design analysis of very large and geometrically complex nuclear systems has been limited by the large number of processors and memory requirements for their preliminary deterministic calculations and final Monte Carlo calculation. Three mesh adaptivity algorithms were developed to reduce the memory requirements of CADIS and FW-CADIS without sacrificing their efficiency improvement. First, a macro-material approach enhances the fidelity of the deterministic models without changing the mesh. Second, a deterministic mesh refinement algorithm generates meshes that capture as much geometric detail as possible without exceeding a specified maximum number of mesh elements. Finally, a weight window coarsening algorithm de-couples the weight window mesh and energy bins from the mesh and energy group structure of the deterministic calculations in order to remove the memory constraint of the weight window map from the deterministic mesh resolution. The three algorithms were used to enhance an FW-CADIS calculation of the prompt dose rate throughout the ITER experimental facility. Using these algorithms resulted in a 23.3% increase in the number of mesh tally elements in which the dose rates were calculated in a 10-day Monte Carlo calculation and, additionally, increased the efficiency of the Monte Carlo simulation by a factor of at least 3.4. The three algorithms enabled this difficult calculation to be accurately solved using an FW-CADIS simulation on a regular computer cluster, obviating the need for a world-class super computer. (authors)
Adaptive Mesh Computations with the PLUTO Code for Astrophysical Fluid Dynamics
Mignone, A.; Zanni, C.
2012-07-01
We present an overview of the current version of the PLUTO code for numerical simulations of astrophysical fluid flows over block-structured adaptive grids. The code preserves its modular framework for the solution of the classical or relativistic magnetohydrodynamics (MHD) equations while exploiting the distributed infrastructure of the Chombo library for multidimensional adaptive mesh refinement (AMR) parallel computations. Equations are evolved in time using an explicit second-order, dimensionally unsplit time stepping scheme based on a cell-centered discretization of the flow variables. Efficiency and robustness are shown through multidimensional benchmarks and applications to problems of astrophysical relevance.
Adaptively Refined Euler and Navier-Stokes Solutions with a Cartesian-Cell Based Scheme
Coirier, William J.; Powell, Kenneth G.
1995-01-01
A Cartesian-cell based scheme with adaptive mesh refinement for solving the Euler and Navier-Stokes equations in two dimensions has been developed and tested. Grids about geometrically complicated bodies were generated automatically, by recursive subdivision of a single Cartesian cell encompassing the entire flow domain. Where the resulting cells intersect bodies, N-sided 'cut' cells were created using polygon-clipping algorithms. The grid was stored in a binary-tree data structure which provided a natural means of obtaining cell-to-cell connectivity and of carrying out solution-adaptive mesh refinement. The Euler and Navier-Stokes equations were solved on the resulting grids using an upwind, finite-volume formulation. The inviscid fluxes were found in an upwinded manner using a linear reconstruction of the cell primitives, providing the input states to an approximate Riemann solver. The viscous fluxes were formed using a Green-Gauss type of reconstruction upon a co-volume surrounding the cell interface. Data at the vertices of this co-volume were found in a linearly K-exact manner, which ensured linear K-exactness of the gradients. Adaptively-refined solutions for the inviscid flow about a four-element airfoil (test case 3) were compared to theory. Laminar, adaptively-refined solutions were compared to accepted computational, experimental and theoretical results.
Numerical simulation of immiscible viscous fingering using adaptive unstructured meshes
Adam, A.; Salinas, P.; Percival, J. R.; Pavlidis, D.; Pain, C.; Muggeridge, A. H.; Jackson, M.
2015-12-01
Displacement of one fluid by another in porous media occurs in various settings including hydrocarbon recovery, CO2 storage and water purification. When the invading fluid is of lower viscosity than the resident fluid, the displacement front is subject to a Saffman-Taylor instability and is unstable to transverse perturbations. These instabilities can grow, leading to fingering of the invading fluid. Numerical simulation of viscous fingering is challenging. The physics is controlled by a complex interplay of viscous and diffusive forces and it is necessary to ensure physical diffusion dominates numerical diffusion to obtain converged solutions. This typically requires the use of high mesh resolution and high order numerical methods. This is computationally expensive. We demonstrate here the use of a novel control volume - finite element (CVFE) method along with dynamic unstructured mesh adaptivity to simulate viscous fingering with higher accuracy and lower computational cost than conventional methods. Our CVFE method employs a discontinuous representation for both pressure and velocity, allowing the use of smaller control volumes (CVs). This yields higher resolution of the saturation field which is represented CV-wise. Moreover, dynamic mesh adaptivity allows high mesh resolution to be employed where it is required to resolve the fingers and lower resolution elsewhere. We use our results to re-examine the existing criteria that have been proposed to govern the onset of instability.Mesh adaptivity requires the mapping of data from one mesh to another. Conventional methods such as consistent interpolation do not readily generalise to discontinuous fields and are non-conservative. We further contribute a general framework for interpolation of CV fields by Galerkin projection. The method is conservative, higher order and yields improved results, particularly with higher order or discontinuous elements where existing approaches are often excessively diffusive.
Anisotropic norm-oriented mesh adaptation for a Poisson problem
Brèthes, Gautier; Dervieux, Alain
2016-10-01
We present a novel formulation for the mesh adaptation of the approximation of a Partial Differential Equation (PDE). The discussion is restricted to a Poisson problem. The proposed norm-oriented formulation extends the goal-oriented formulation since it is equation-based and uses an adjoint. At the same time, the norm-oriented formulation somewhat supersedes the goal-oriented one since it is basically a solution-convergent method. Indeed, goal-oriented methods rely on the reduction of the error in evaluating a chosen scalar output with the consequence that, as mesh size is increased (more degrees of freedom), only this output is proven to tend to its continuous analog while the solution field itself may not converge. A remarkable quality of goal-oriented metric-based adaptation is the mathematical formulation of the mesh adaptation problem under the form of the optimization, in the well-identified set of metrics, of a well-defined functional. In the new proposed formulation, we amplify this advantage. We search, in the same well-identified set of metrics, the minimum of a norm of the approximation error. The norm is prescribed by the user and the method allows addressing the case of multi-objective adaptation like, for example in aerodynamics, adaptating the mesh for drag, lift and moment in one shot. In this work, we consider the basic linear finite-element approximation and restrict our study to L2 norm in order to enjoy second-order convergence. Numerical examples for the Poisson problem are computed.
ADAPTIVE MODEL REFINEMENT FOR THE IONOSPHERE AND THERMOSPHERE
National Aeronautics and Space Administration — ADAPTIVE MODEL REFINEMENT FOR THE IONOSPHERE AND THERMOSPHERE ANTHONY M. D’AMATO∗, AARON J. RIDLEY∗∗, AND DENNIS S. BERNSTEIN∗∗∗ Abstract. Mathematical models of...
White Dwarf Mergers on Adaptive Meshes I. Methodology and Code Verification
Katz, Max P; Calder, Alan C; Swesty, F Douglas; Almgren, Ann S; Zhang, Weiqun
2015-01-01
The Type Ia supernova progenitor problem is one of the most perplexing and exciting problems in astrophysics, requiring detailed numerical modeling to complement observations of these explosions. One possible progenitor that has merited recent theoretical attention is the white dwarf merger scenario, which has the potential to naturally explain many of the observed characteristics of Type Ia supernovae. To date there have been relatively few self-consistent simulations of merging white dwarf systems using mesh-based hydrodynamics. This is the first paper in a series describing simulations of these systems using a hydrodynamics code with adaptive mesh refinement. In this paper we describe our numerical methodology and discuss our implementation in the compressible hydrodynamics code CASTRO, which solves the Euler equations, and the Poisson equation for self-gravity, and couples the gravitational and rotation forces to the hydrodynamics. Standard techniques for coupling gravitation and rotation forces to the hy...
Adaptive radial basis function mesh deformation using data reduction
Gillebaart, T.; Blom, D. S.; van Zuijlen, A. H.; Bijl, H.
2016-09-01
Radial Basis Function (RBF) mesh deformation is one of the most robust mesh deformation methods available. Using the greedy (data reduction) method in combination with an explicit boundary correction, results in an efficient method as shown in literature. However, to ensure the method remains robust, two issues are addressed: 1) how to ensure that the set of control points remains an accurate representation of the geometry in time and 2) how to use/automate the explicit boundary correction, while ensuring a high mesh quality. In this paper, we propose an adaptive RBF mesh deformation method, which ensures the set of control points always represents the geometry/displacement up to a certain (user-specified) criteria, by keeping track of the boundary error throughout the simulation and re-selecting when needed. Opposed to the unit displacement and prescribed displacement selection methods, the adaptive method is more robust, user-independent and efficient, for the cases considered. Secondly, the analysis of a single high aspect ratio cell is used to formulate an equation for the correction radius needed, depending on the characteristics of the correction function used, maximum aspect ratio, minimum first cell height and boundary error. Based on the analysis two new radial basis correction functions are derived and proposed. This proposed automated procedure is verified while varying the correction function, Reynolds number (and thus first cell height and aspect ratio) and boundary error. Finally, the parallel efficiency is studied for the two adaptive methods, unit displacement and prescribed displacement for both the CPU as well as the memory formulation with a 2D oscillating and translating airfoil with oscillating flap, a 3D flexible locally deforming tube and deforming wind turbine blade. Generally, the memory formulation requires less work (due to the large amount of work required for evaluating RBF's), but the parallel efficiency reduces due to the limited
On adaptive refinements in discrete probabilistic fracture models
J. Eliáš
2017-01-01
Full Text Available The possibility to adaptively change discretization density is a well acknowledged and used feature of many continuum models. It is employed to save computational time and increase solution accuracy. Recently, adaptivity has been introduced also for discrete particle models. This contribution applies adaptive technique in probabilistic discrete modelling where material properties are varying in space according to a random field. The random field discretization is adaptively refined hand in hand with the model geometry.
ROAMing terrain (Real-time Optimally Adapting Meshes)
Duchaineau, M.; Wolinsky, M.; Sigeti, D.E.; Miller, M.C.; Aldrich, C.; Mineev, M.
1997-07-01
Terrain visualization is a difficult problem for applications requiring accurate images of large datasets at high frame rates, such as flight simulation and ground-based aircraft testing using synthetic sensor stimulation. On current graphics hardware, the problem is to maintain dynamic, view-dependent triangle meshes and texture maps that produce good images at the required frame rate. We present an algorithm for constructing triangle meshes that optimizes flexible view-dependent error metrics, produces guaranteed error bounds, achieves specified triangle counts directly, and uses frame-to-frame coherence to operate at high frame rates for thousands of triangles per frame. Our method, dubbed Real-time Optimally Adapting Meshes (ROAM), uses two priority queues to drive split and merge operations that maintain continuous triangulations built from pre-processed bintree triangles. We introduce two additional performance optimizations: incremental triangle stripping and priority-computation deferral lists. ROAM execution time is proportionate to the number of triangle changes per frame, which is typically a few percent of the output mesh size, hence ROAM performance is insensitive to the resolution and extent of the input terrain. Dynamic terrain and simple vertex morphing are supported.
Adaptive mesh generation for image registration and segmentation
Fogtmann, Mads; Larsen, Rasmus
2013-01-01
This paper deals with the problem of generating quality tetrahedral meshes for image registration. From an initial coarse mesh the approach matches the mesh to the image volume by combining red-green subdivision and mesh evolution through mesh-to-image matching regularized with a mesh quality...
An Application of the Mesh Generation and Refinement Tool to Mobile Bay, Alabama, USA
Aziz, Wali; Alarcon, Vladimir J.; McAnally, William; Martin, James; Cartwright, John
2009-08-01
A grid generation tool, called the Mesh Generation and Refinement Tool (MGRT), has been developed using Qt4. Qt4 is a comprehensive C++ application framework which includes GUI and container class-libraries and tools for cross-platform development. MGRT is capable of using several types of algorithms for grid generation. This paper presents an application of the MGRT grid generation tool for creating an unstructured grid of Mobile Bay (Alabama, USA) that will be used for hydrodynamics modeling. The algorithm used in this particular application is the Advancing-Front/Local-Reconnection (AFLR) [1] [2]. This research shows results of grids created with MGRT and compares them to grids (for the same geographical container) created using other grid generation tools. The superior quality of the grids generated by MGRT is shown.
Ibrahim, Ahmad M., E-mail: ibrahimam@ornl.gov [Oak Ridge National Laboratory, P.O. Box 2008, Oak Ridge, TN 37831 (United States); Wilson, Paul P. [University of Wisconsin-Madison, 1500 Engineering Dr., Madison, WI 53706 (United States); Sawan, Mohamed E., E-mail: sawan@engr.wisc.edu [University of Wisconsin-Madison, 1500 Engineering Dr., Madison, WI 53706 (United States); Mosher, Scott W.; Peplow, Douglas E.; Grove, Robert E. [Oak Ridge National Laboratory, P.O. Box 2008, Oak Ridge, TN 37831 (United States)
2014-10-15
Highlights: •Calculate the prompt dose rate everywhere throughout the entire fusion energy facility. •Utilize FW-CADIS to accurately perform difficult neutronics calculations for fusion energy systems. •Develop three mesh adaptivity algorithms to enhance FW-CADIS efficiency in fusion-neutronics calculations. -- Abstract: Three mesh adaptivity algorithms were developed to facilitate and expedite the use of the CADIS and FW-CADIS hybrid Monte Carlo/deterministic techniques in accurate full-scale neutronics simulations of fusion energy systems with immense sizes and complicated geometries. First, a macromaterial approach enhances the fidelity of the deterministic models without changing the mesh. Second, a deterministic mesh refinement algorithm generates meshes that capture as much geometric detail as possible without exceeding a specified maximum number of mesh elements. Finally, a weight window coarsening algorithm decouples the weight window mesh and energy bins from the mesh and energy group structure of the deterministic calculations in order to remove the memory constraint of the weight window map from the deterministic mesh resolution. The three algorithms were used to enhance an FW-CADIS calculation of the prompt dose rate throughout the ITER experimental facility and resulted in a 23.3% increase in the number of mesh tally elements in which the dose rates were calculated in a 10-day Monte Carlo calculation. Additionally, because of the significant increase in the efficiency of FW-CADIS simulations, the three algorithms enabled this difficult calculation to be accurately solved on a regular computer cluster, eliminating the need for a world-class super computer.
Stapleton, Scott; Gries, Thomas; Waas, Anthony M.; Pineda, Evan J.
2014-01-01
Enhanced finite elements are elements with an embedded analytical solution that can capture detailed local fields, enabling more efficient, mesh independent finite element analysis. The shape functions are determined based on the analytical model rather than prescribed. This method was applied to adhesively bonded joints to model joint behavior with one element through the thickness. This study demonstrates two methods of maintaining the fidelity of such elements during adhesive non-linearity and cracking without increasing the mesh needed for an accurate solution. The first method uses adaptive shape functions, where the shape functions are recalculated at each load step based on the softening of the adhesive. The second method is internal mesh adaption, where cracking of the adhesive within an element is captured by further discretizing the element internally to represent the partially cracked geometry. By keeping mesh adaptations within an element, a finer mesh can be used during the analysis without affecting the global finite element model mesh. Examples are shown which highlight when each method is most effective in reducing the number of elements needed to capture adhesive nonlinearity and cracking. These methods are validated against analogous finite element models utilizing cohesive zone elements.
A novel partitioning method for block-structured adaptive meshes
Fu, Lin; Litvinov, Sergej; Hu, Xiangyu Y.; Adams, Nikolaus A.
2017-07-01
We propose a novel partitioning method for block-structured adaptive meshes utilizing the meshless Lagrangian particle concept. With the observation that an optimum partitioning has high analogy to the relaxation of a multi-phase fluid to steady state, physically motivated model equations are developed to characterize the background mesh topology and are solved by multi-phase smoothed-particle hydrodynamics. In contrast to well established partitioning approaches, all optimization objectives are implicitly incorporated and achieved during the particle relaxation to stationary state. Distinct partitioning sub-domains are represented by colored particles and separated by a sharp interface with a surface tension model. In order to obtain the particle relaxation, special viscous and skin friction models, coupled with a tailored time integration algorithm are proposed. Numerical experiments show that the present method has several important properties: generation of approximately equal-sized partitions without dependence on the mesh-element type, optimized interface communication between distinct partitioning sub-domains, continuous domain decomposition which is physically localized and implicitly incremental. Therefore it is particularly suitable for load-balancing of high-performance CFD simulations.
A novel partitioning method for block-structured adaptive meshes
Fu, Lin, E-mail: lin.fu@tum.de; Litvinov, Sergej, E-mail: sergej.litvinov@aer.mw.tum.de; Hu, Xiangyu Y., E-mail: xiangyu.hu@tum.de; Adams, Nikolaus A., E-mail: nikolaus.adams@tum.de
2017-07-15
We propose a novel partitioning method for block-structured adaptive meshes utilizing the meshless Lagrangian particle concept. With the observation that an optimum partitioning has high analogy to the relaxation of a multi-phase fluid to steady state, physically motivated model equations are developed to characterize the background mesh topology and are solved by multi-phase smoothed-particle hydrodynamics. In contrast to well established partitioning approaches, all optimization objectives are implicitly incorporated and achieved during the particle relaxation to stationary state. Distinct partitioning sub-domains are represented by colored particles and separated by a sharp interface with a surface tension model. In order to obtain the particle relaxation, special viscous and skin friction models, coupled with a tailored time integration algorithm are proposed. Numerical experiments show that the present method has several important properties: generation of approximately equal-sized partitions without dependence on the mesh-element type, optimized interface communication between distinct partitioning sub-domains, continuous domain decomposition which is physically localized and implicitly incremental. Therefore it is particularly suitable for load-balancing of high-performance CFD simulations.
Adaptive remeshing method in 2D based on refinement and coarsening techniques
Giraud-Moreau, L.; Borouchaki, H.; Cherouat, A.
2007-04-01
The analysis of mechanical structures using the Finite Element Method, in the framework of large elastoplastic strains, needs frequent remeshing of the deformed domain during computation. Remeshing is necessary for two main reasons, the large geometric distortion of finite elements and the adaptation of the mesh size to the physical behavior of the solution. This paper presents an adaptive remeshing method to remesh a mechanical structure in two dimensions subjected to large elastoplastic deformations with damage. The proposed remeshing technique includes adaptive refinement and coarsening procedures, based on geometrical and physical criteria. The proposed method has been integrated in a computational environment using the ABAQUS solver. Numerical examples show the efficiency of the proposed approach.
Numerical relativity simulations of neutron star merger remnants using conservative mesh refinement
Dietrich, Tim; Ujevic, Maximiliano; Bruegmann, Bernd
2015-01-01
We study equal and unequal-mass neutron star mergers by means of new numerical relativity simulations in which the general relativistic hydrodynamics solver employs an algorithm that guarantees mass conservation across the refinement levels of the computational mesh. We consider eight binary configurations with total mass $M=2.7\\,M_\\odot$, mass-ratios $q=1$ and $q=1.16$, and four different equation of states (EOSs), and one configuration with a stiff EOS, $M=2.5M_\\odot$ and $q=1.5$. We focus on the post-merger dynamics and study the merger remnant, dynamical ejecta and the postmerger gravitational wave spectrum. Although most of the merger remnants form a hypermassive neutron star collapsing to a black hole+disk system on dynamical timescales, stiff EOSs can eventually produce a stable massive neutron star. Ejecta are mostly emitted around the orbital plane; favored by large mass ratios and softer EOS. The postmerger wave spectrum is mainly characterized by non-axisymmetric oscillations of the remnant. The st...
The PLUTO Code for Adaptive Mesh Computations in Astrophysical Fluid Dynamics
Mignone, A.; Zanni, C.; Tzeferacos, P.; van Straalen, B.; Colella, P.; Bodo, G.
2012-01-01
We present a description of the adaptive mesh refinement (AMR) implementation of the PLUTO code for solving the equations of classical and special relativistic magnetohydrodynamics (MHD and RMHD). The current release exploits, in addition to the static grid version of the code, the distributed infrastructure of the CHOMBO library for multidimensional parallel computations over block-structured, adaptively refined grids. We employ a conservative finite-volume approach where primary flow quantities are discretized at the cell center in a dimensionally unsplit fashion using the Corner Transport Upwind method. Time stepping relies on a characteristic tracing step where piecewise parabolic method, weighted essentially non-oscillatory, or slope-limited linear interpolation schemes can be handily adopted. A characteristic decomposition-free version of the scheme is also illustrated. The solenoidal condition of the magnetic field is enforced by augmenting the equations with a generalized Lagrange multiplier providing propagation and damping of divergence errors through a mixed hyperbolic/parabolic explicit cleaning step. Among the novel features, we describe an extension of the scheme to include non-ideal dissipative processes, such as viscosity, resistivity, and anisotropic thermal conduction without operator splitting. Finally, we illustrate an efficient treatment of point-local, potentially stiff source terms over hierarchical nested grids by taking advantage of the adaptivity in time. Several multidimensional benchmarks and applications to problems of astrophysical relevance assess the potentiality of the AMR version of PLUTO in resolving flow features separated by large spatial and temporal disparities.
THE PLUTO CODE FOR ADAPTIVE MESH COMPUTATIONS IN ASTROPHYSICAL FLUID DYNAMICS
Mignone, A.; Tzeferacos, P. [Dipartimento di Fisica Generale, Universita di Torino, via Pietro Giuria 1, 10125 Torino (Italy); Zanni, C.; Bodo, G. [INAF, Osservatorio Astronomico di Torino, Strada Osservatorio 20, Pino Torinese 10025 (Italy); Van Straalen, B.; Colella, P. [Lawrence Berkeley National Laboratory, 1 Cyclotron Road, MS 50A-1148, Berkeley, CA 94720 (United States)
2012-01-01
We present a description of the adaptive mesh refinement (AMR) implementation of the PLUTO code for solving the equations of classical and special relativistic magnetohydrodynamics (MHD and RMHD). The current release exploits, in addition to the static grid version of the code, the distributed infrastructure of the CHOMBO library for multidimensional parallel computations over block-structured, adaptively refined grids. We employ a conservative finite-volume approach where primary flow quantities are discretized at the cell center in a dimensionally unsplit fashion using the Corner Transport Upwind method. Time stepping relies on a characteristic tracing step where piecewise parabolic method, weighted essentially non-oscillatory, or slope-limited linear interpolation schemes can be handily adopted. A characteristic decomposition-free version of the scheme is also illustrated. The solenoidal condition of the magnetic field is enforced by augmenting the equations with a generalized Lagrange multiplier providing propagation and damping of divergence errors through a mixed hyperbolic/parabolic explicit cleaning step. Among the novel features, we describe an extension of the scheme to include non-ideal dissipative processes, such as viscosity, resistivity, and anisotropic thermal conduction without operator splitting. Finally, we illustrate an efficient treatment of point-local, potentially stiff source terms over hierarchical nested grids by taking advantage of the adaptivity in time. Several multidimensional benchmarks and applications to problems of astrophysical relevance assess the potentiality of the AMR version of PLUTO in resolving flow features separated by large spatial and temporal disparities.
Grigory I. Shishkin
2008-01-01
A boundary value problem is considered for a singularly perturbed parabolic convection-diffusion equation; we construct a finite difference scheme on a priori (se-quentially) adapted meshes and study its convergence. The scheme on a priori adapted meshes is constructed using a majorant function for the singular component of the discrete solution, which allows us to find a priori a subdomain where the computed solution requires a further improvement. This subdomain is defined by the perturbation parameter ε, the step-size of a uniform mesh in x, and also by the required accuracy of the discrete solution and the prescribed number of refinement iterations K for im-proving the solution. To solve the discrete problems aimed at the improvement of the solution, we use uniform meshes on the subdomains. The error of the numerical so-lution depends weakly on the parameter ε. The scheme converges almost ε-uniformly, precisely, under the condition N-1 = o(ev), where N denotes the number of nodes in the spatial mesh, and the value v=v(K) can be chosen arbitrarily small for suitable K.
LOAD AWARE ADAPTIVE BACKBONE SYNTHESIS IN WIRELESS MESH NETWORKS
Yuan Yuan; Zheng Baoyu
2009-01-01
Wireless Mesh Networks (WMNs) are envisioned to support the wired backbone with a wireless Backbone Networks (BNet) for providing internet connectivity to large-scale areas.With a wide range of internet-oriented applications with different Quality of Service (QoS) requirement,the large-scale WMNs should have good scalability and large bandwidth.In this paper,a Load Aware Adaptive Backbone Synthesis (LAABS) algorithm is proposed to automatically balance the traffic flow in the WMNs.The BNet will dynamically split into smaller size or merge into bigger one according to statistic load information of Backbone Nodes (BNs).Simulation results show LAABS generates moderate BNet size and converges quickly,thus providing scalable and stable BNet to facilitate traffic flow.
Adaptivity techniques for the computation of two-dimensional viscous flows using structured meshes
Szmelter, J.; Evans, A.; Weatherill, N. P.
In this paper three different adaptivity techniques have been investigated on the base of structured meshes. All the techniques indicate the significance of using adaptivity for improving computational results. In particular, the technique of combining point enrichment and node movement strategies offers the best compromise. Although, the work presented here used two-dimensional structured meshes, the techniques can be readily applied to hybrid and unstructured meshes. Also, preliminary three-dimensional numerical results have been already obtained by coauthors.
WHITE DWARF MERGERS ON ADAPTIVE MESHES. I. METHODOLOGY AND CODE VERIFICATION
Katz, Max P.; Zingale, Michael; Calder, Alan C.; Swesty, F. Douglas [Department of Physics and Astronomy, Stony Brook University, Stony Brook, NY, 11794-3800 (United States); Almgren, Ann S.; Zhang, Weiqun [Center for Computational Sciences and Engineering, Lawrence Berkeley National Laboratory, Berkeley, CA 94720 (United States)
2016-03-10
The Type Ia supernova (SN Ia) progenitor problem is one of the most perplexing and exciting problems in astrophysics, requiring detailed numerical modeling to complement observations of these explosions. One possible progenitor that has merited recent theoretical attention is the white dwarf (WD) merger scenario, which has the potential to naturally explain many of the observed characteristics of SNe Ia. To date there have been relatively few self-consistent simulations of merging WD systems using mesh-based hydrodynamics. This is the first paper in a series describing simulations of these systems using a hydrodynamics code with adaptive mesh refinement. In this paper we describe our numerical methodology and discuss our implementation in the compressible hydrodynamics code CASTRO, which solves the Euler equations, and the Poisson equation for self-gravity, and couples the gravitational and rotation forces to the hydrodynamics. Standard techniques for coupling gravitation and rotation forces to the hydrodynamics do not adequately conserve the total energy of the system for our problem, but recent advances in the literature allow progress and we discuss our implementation here. We present a set of test problems demonstrating the extent to which our software sufficiently models a system where large amounts of mass are advected on the computational domain over long timescales. Future papers in this series will describe our treatment of the initial conditions of these systems and will examine the early phases of the merger to determine its viability for triggering a thermonuclear detonation.
A versatile embedded boundary adaptive mesh method for compressible flow in complex geometry
Al-Marouf, M.
2017-02-25
We present an embedded ghost-fluid method for numerical solutions of the compressible Navier Stokes (CNS) equations in arbitrary complex domains. A PDE multidimensional extrapolation approach is used to reconstruct the solution in the ghost-fluid regions and imposing boundary conditions on the fluid-solid interface, coupled with a multi-dimensional algebraic interpolation for freshly cleared cells. The CNS equations are numerically solved by the second order multidimensional upwind method. Block-structured adaptive mesh refinement, implemented with the Chombo framework, is utilized to reduce the computational cost while keeping high resolution mesh around the embedded boundary and regions of high gradient solutions. The versatility of the method is demonstrated via several numerical examples, in both static and moving geometry, ranging from low Mach number nearly incompressible flows to supersonic flows. Our simulation results are extensively verified against other numerical results and validated against available experimental results where applicable. The significance and advantages of our implementation, which revolve around balancing between the solution accuracy and implementation difficulties, are briefly discussed as well.
Mirza, Anwar M. [Department of Computer Science, National University of Computer and Emerging Sciences, NUCES-FAST, A.K. Brohi Road, H-11, Islamabad (Pakistan)], E-mail: anwar.m.mirza@gmail.com; Iqbal, Shaukat [Faculty of Computer Science and Engineering, Ghulam Ishaq Khan (GIK) Institute of Engineering Science and Technology, Topi-23460, Swabi (Pakistan)], E-mail: shaukat@giki.edu.pk; Rahman, Faizur [Department of Physics, Allama Iqbal Open University, H-8 Islamabad (Pakistan)
2007-07-15
A spatially adaptive grid-refinement approach has been investigated to solve the even-parity Boltzmann transport equation. A residual based a posteriori error estimation scheme has been utilized for checking the approximate solutions for various finite element grids. The local particle balance has been considered as an error assessment criterion. To implement the adaptive approach, a computer program ADAFENT (adaptive finite elements for neutron transport) has been developed to solve the second order even-parity Boltzmann transport equation using K{sup +} variational principle for slab geometry. The program has a core K{sup +} module which employs Lagrange polynomials as spatial basis functions for the finite element formulation and Legendre polynomials for the directional dependence of the solution. The core module is called in by the adaptive grid generator to determine local gradients and residuals to explore the possibility of grid refinements in appropriate regions of the problem. The a posteriori error estimation scheme has been implemented in the outer grid refining iteration module. Numerical experiments indicate that local errors are large in regions where the flux gradients are large. A comparison of the spatially adaptive grid-refinement approach with that of uniform meshing approach for various benchmark cases confirms its superiority in greatly enhancing the accuracy of the solution without increasing the number of unknown coefficients. A reduction in the local errors of the order of 10{sup 2} has been achieved using the new approach in some cases.
The GeoClaw software for depth-averaged flows with adaptive refinement
Berger, Marsha J; LeVeque, Randall J; Mandli, Kyle
2010-01-01
Many geophysical flow or wave propagation problems can be modeled with two-dimensional depth-averaged equations, of which the shallow water equations are the simplest example. We describe the GeoClaw software that has been designed to solve problems of this nature, consisting of open source Fortran programs together with Python tools for the user interface and flow visualization. This software uses high-resolution shock-capturing finite volume methods on logically rectangular grids, including latitude--longitude grids on the sphere. Dry states are handled automatically to model inundation. The code incorporates adaptive mesh refinement to allow the efficient solution of large-scale geophysical problems. Examples are given illustrating its use for modeling tsunamis, dam break problems, and storm surge. Documentation and download information is available at www.clawpack.org/geoclaw
MECHANICAL DYNAMICS ANALYSIS OF PM GENERATOR USING H-ADAPTIVE REFINEMENT
AJAY KUMAR
2010-03-01
Full Text Available This paper describes the dynamic analysis of permanent magnet (PM rotor generator using COMSOL Multiphysics, a Finite Element Analysis (FEA based package and Simulink, a system simulation program. Model of PM rotor generator is developed for its mechanical dynamics and computational of torque resulting from magnetic force. For the model the mesh is constructed using first order Lagrange quadratic elements and h-adaptive refinement technique based upon bank bisection is used for improving accuracy of the model. Effect of rotor moment of inertia (MI on the winding resistance and winding inductance has been studied by using Simulink. It is shown that the system MI has a significant effect on optimal winding resistance and inductance to achieve steady state operation in shortest period of time.
The GeoClaw software for depth-averaged flows with adaptive refinement
Berger, M.J.; George, D.L.; LeVeque, R.J.; Mandli, K.T.
2011-01-01
Many geophysical flow or wave propagation problems can be modeled with two-dimensional depth-averaged equations, of which the shallow water equations are the simplest example. We describe the GeoClaw software that has been designed to solve problems of this nature, consisting of open source Fortran programs together with Python tools for the user interface and flow visualization. This software uses high-resolution shock-capturing finite volume methods on logically rectangular grids, including latitude-longitude grids on the sphere. Dry states are handled automatically to model inundation. The code incorporates adaptive mesh refinement to allow the efficient solution of large-scale geophysical problems. Examples are given illustrating its use for modeling tsunamis and dam-break flooding problems. Documentation and download information is available at www.clawpack.org/geoclaw. ?? 2011.
Pascucci, V
2004-02-18
This paper presents a simple approach for rendering isosurfaces of a scalar field. Using the vertex programming capability of commodity graphics cards, we transfer the cost of computing an isosurface from the Central Processing Unit (CPU), running the main application, to the Graphics Processing Unit (GPU), rendering the images. We consider a tetrahedral decomposition of the domain and draw one quadrangle (quad) primitive per tetrahedron. A vertex program transforms the quad into the piece of isosurface within the tetrahedron (see Figure 2). In this way, the main application is only devoted to streaming the vertices of the tetrahedra from main memory to the graphics card. For adaptively refined rectilinear grids, the optimization of this streaming process leads to the definition of a new 3D space-filling curve, which generalizes the 2D Sierpinski curve used for efficient rendering of triangulated terrains. We maintain the simplicity of the scheme when constructing view-dependent adaptive refinements of the domain mesh. In particular, we guarantee the absence of T-junctions by satisfying local bounds in our nested error basis. The expensive stage of fixing cracks in the mesh is completely avoided. We discuss practical tradeoffs in the distribution of the workload between the application and the graphics hardware. With current GPU's it is convenient to perform certain computations on the main CPU. Beyond the performance considerations that will change with the new generations of GPU's this approach has the major advantage of avoiding completely the storage in memory of the isosurface vertices and triangles.
Schnepp, Sascha M
2011-01-01
A framework for performing dynamic mesh adaptation with the discontinuous Galerkin method (DGM) is presented. Adaptations include modifications of the local mesh step size (h-adaptation) and the local degree of the approximating polynomials (p-adaptation) as well as their combination. The computation of the approximation within locally adapted elements is based on projections between finite element spaces (FES), which are shown to preserve the upper limit of the electromagnetic energy. The formulation supports high level hanging nodes and applies precomputation of surface integrals for increasing computational efficiency. A full wave simulation of electromagnetic scattering form a radar reflector demonstrates the applicability to large scale problems in three-dimensional space.
Welch, J. A.; Kópházi, J.; Owens, A. R.; Eaton, M. D.
2017-10-01
In this paper a method is presented for the application of energy-dependent spatial meshes applied to the multigroup, second-order, even-parity form of the neutron transport equation using Isogeometric Analysis (IGA). The computation of the inter-group regenerative source terms is based on conservative interpolation by Galerkin projection. The use of Non-Uniform Rational B-splines (NURBS) from the original computer-aided design (CAD) model allows for efficient implementation and calculation of the spatial projection operations while avoiding the complications of matching different geometric approximations faced by traditional finite element methods (FEM). The rate-of-convergence was verified using the method of manufactured solutions (MMS) and found to preserve the theoretical rates when interpolating between spatial meshes of different refinements. The scheme's numerical efficiency was then studied using a series of two-energy group pincell test cases where a significant saving in the number of degrees-of-freedom can be found if the energy group with a complex variation in the solution is refined more than an energy group with a simpler solution function. Finally, the method was applied to a heterogeneous, seven-group reactor pincell where the spatial meshes for each energy group were adaptively selected for refinement. It was observed that by refining selected energy groups a reduction in the total number of degrees-of-freedom for the same total L2 error can be obtained.
Geometrically Consistent Mesh Modification
Bonito, A.
2010-01-01
A new paradigm of adaptivity is to execute refinement, coarsening, and smoothing of meshes on manifolds with incomplete information about their geometry and yet preserve position and curvature accuracy. We refer to this collectively as geometrically consistent (GC) mesh modification. We discuss the concept of discrete GC, show the failure of naive approaches, and propose and analyze a simple algorithm that is GC and accuracy preserving. © 2010 Society for Industrial and Applied Mathematics.
Greene, Patrick T. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Schofield, Samuel P. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Nourgaliev, Robert [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)
2016-06-21
A new mesh smoothing method designed to cluster mesh cells near a dynamically evolving interface is presented. The method is based on weighted condition number mesh relaxation with the weight function being computed from a level set representation of the interface. The weight function is expressed as a Taylor series based discontinuous Galerkin projection, which makes the computation of the derivatives of the weight function needed during the condition number optimization process a trivial matter. For cases when a level set is not available, a fast method for generating a low-order level set from discrete cell-centered elds, such as a volume fraction or index function, is provided. Results show that the low-order level set works equally well for the weight function as the actual level set. Meshes generated for a number of interface geometries are presented, including cases with multiple level sets. Dynamic cases for moving interfaces are presented to demonstrate the method's potential usefulness to arbitrary Lagrangian Eulerian (ALE) methods.
Performance Evaluation of Various STL File Mesh Refining Algorithms Applied for FDM-RP Process
Ledalla, Siva Rama Krishna; Tirupathi, Balaji; Sriram, Venkatesh
2016-06-01
Layered manufacturing machines use the stereolithography (STL) file to build parts. When a curved surface is converted from a computer aided design (CAD) file to STL, it results in a geometrical distortion and chordal error. Parts manufactured with this file, might not satisfy geometric dimensioning and tolerance requirements due to approximated geometry. Current algorithms built in CAD packages have export options to globally reduce this distortion, which leads to an increase in the file size and pre-processing time. In this work, different mesh subdivision algorithms are applied on STL file of a complex geometric features using MeshLab software. The mesh subdivision algorithms considered in this work are modified butterfly subdivision technique, loops sub division technique and general triangular midpoint sub division technique. A comparative study is made with respect to volume and the build time using the above techniques. It is found that triangular midpoint sub division algorithm is more suitable for the geometry under consideration. Only the wheel cap part is then manufactured on Stratasys MOJO FDM machine. The surface roughness of the part is measured on Talysurf surface roughness tester.
Adaptive, Tactical Mesh Networking: Control Base MANET Model
2010-09-01
pp. 316–320 Available: IEEE Xplore , http://ieeexplore.ieee.org [Accessed: June 9, 2010]. [5] N. Sidiropoulos, “Multiuser Transmit Beamforming...Mobile Mesh Segments of TNT Testbed .......... 11 Figure 5. Infrastructure and Ad Hoc Mode of IEEE 802.11................................ 13 Figure...6. The Power Spectral Density of OFDM................................................ 14 Figure 7. A Typical IEEE 802.16 Network
Toth, J. [Inst. for Transportation Technologies, FAMU-FSU College of Engineering, Tallahassee, FL (United States); Ruge, P. [Inst. of Dynamics of Structures, Dresden Univ. of Technology (Germany)
2001-07-01
Extensive studies, concerning the longitudinal behavior of long railway bridges due to braking forces have been done by measurements in situ, and by statical, as well as dynamical simulations. Thereby, the only consistent numerical realization with respect to the measured data was the dynamical one. However, the consecutive discretizations in space and time with time-dependent system matrices are extremely time consuming due to the moving loads and varying stiffness of the ballast under, and in front of, the moving train. Therefore, every effort should be made to optimize the discretization in the space domain. This paper presents a strategy for assessing the quality of finite elements in space and for applying an adaptive mesh-refinement for this special engineering problem. The method is characterized by a spectral assessment, comparing a certain set of eigenvalues of the actual discretization with those of a very fine and rather exact numerical model. The error estimator introduced in this paper controls a whole set of global eigenvalues with corresponding natural vibration modes in order to assess certain types of shape functions. Thus, the procedure estimates local modifications on the one hand and p-properties on the other by means of global indication. (orig.)
Geostrophic balance preserving interpolation in mesh adaptive shallow-water ocean modelling
Maddison, James R; Farrell, Patrick E
2010-01-01
The accurate representation of geostrophic balance is an essential requirement for numerical modelling of geophysical flows. Significant effort is often put into the selection of accurate or optimal balance representation by the discretisation of the fundamental equations. The issue of accurate balance representation is particularly challenging when applying dynamic mesh adaptivity, where there is potential for additional imbalance injection when interpolating to new, optimised meshes. In the context of shallow-water modelling, we present a new method for preservation of geostrophic balance when applying dynamic mesh adaptivity. This approach is based upon interpolation of the Helmholtz decomposition of the Coriolis acceleration. We apply this in combination with a discretisation for which states in geostrophic balance are exactly steady solutions of the linearised equations on an f-plane; this method guarantees that a balanced and steady flow on a donor mesh remains balanced and steady after interpolation on...
Adaptive Meshing of Ship Air-Wake Flowfields
2014-10-21
Atkins ] Table 1. Coefficients for BDF schemes Order Φn+1 Φn Φn-1 1st 1 -1 0 2nd 3/2 -2 1/2 The control volume surrounding each node is the median...Enhancements To The FUN3D Flow Solver For Moving-Mesh Applications”, AIAA-2009-1360. Biedron, Robert T., Vatsa, Veer N., and Atkins , Harold L
Design of computer-generated beam-shaping holograms by iterative finite-element mesh adaption.
Dresel, T; Beyerlein, M; Schwider, J
1996-12-10
Computer-generated phase-only holograms can be used for laser beam shaping, i.e., for focusing a given aperture with intensity and phase distributions into a pregiven intensity pattern in their focal planes. A numerical approach based on iterative finite-element mesh adaption permits the design of appropriate phase functions for the task of focusing into two-dimensional reconstruction patterns. Both the hologram aperture and the reconstruction pattern are covered by mesh mappings. An iterative procedure delivers meshes with intensities equally distributed over the constituting elements. This design algorithm adds new elementary focuser functions to what we call object-oriented hologram design. Some design examples are discussed.
METHOD FOR ADAPTIVE MESH GENERATION BASED ON GEOMETRICAL FEATURES OF 3D SOLID
HUANG Xiaodong; DU Qungui; YE Bangyan
2006-01-01
In order to provide a guidance to specify the element size dynamically during adaptive finite element mesh generation, adaptive criteria are firstly defined according to the relationships between the geometrical features and the elements of 3D solid. Various modes based on different datum geometrical elements, such as vertex, curve, surface, and so on, are then designed for generating local refmed mesh. With the guidance of the defined criteria, different modes are automatically selected to apply on the appropriate datum objects to program the element size in the local special areas. As a result, the control information of element size is successfully programmed coveting the entire domain based on the geometrical features of 3D solid. A new algorithm based on Delaunay triangulation is then developed for generating 3D adaptive fmite element mesh, in which the element size is dynamically specified to catch the geometrical features and suitable tetrahedron facets are selected to locate interior nodes continuously. As a result, adaptive mesh with good-quality elements is generated. Examples show that the proposed method can be successfully applied to adaptive finite element mesh automatic generation based on the geometrical features of 3D solid.
A dynamic mesh refinement technique for Lattice Boltzmann simulations on octree-like grids
Neumann, Philipp
2012-04-27
In this contribution, we present our new adaptive Lattice Boltzmann implementation within the Peano framework, with special focus on nanoscale particle transport problems. With the continuum hypothesis not holding anymore on these small scales, new physical effects - such as Brownian fluctuations - need to be incorporated. We explain the overall layout of the application, including memory layout and access, and shortly review the adaptive algorithm. The scheme is validated by different benchmark computations in two and three dimensions. An extension to dynamically changing grids and a spatially adaptive approach to fluctuating hydrodynamics, allowing for the thermalisation of the fluid in particular regions of interest, is proposed. Both dynamic adaptivity and adaptive fluctuating hydrodynamics are validated separately in simulations of particle transport problems. The application of this scheme to an oscillating particle in a nanopore illustrates the importance of Brownian fluctuations in such setups. © 2012 Springer-Verlag.
A new anisotropic mesh adaptation method based upon hierarchical a posteriori error estimates
Huang, Weizhang; Kamenski, Lennard; Lang, Jens
2010-03-01
A new anisotropic mesh adaptation strategy for finite element solution of elliptic differential equations is presented. It generates anisotropic adaptive meshes as quasi-uniform ones in some metric space, with the metric tensor being computed based on hierarchical a posteriori error estimates. A global hierarchical error estimate is employed in this study to obtain reliable directional information of the solution. Instead of solving the global error problem exactly, which is costly in general, we solve it iteratively using the symmetric Gauß-Seidel method. Numerical results show that a few GS iterations are sufficient for obtaining a reasonably good approximation to the error for use in anisotropic mesh adaptation. The new method is compared with several strategies using local error estimators or recovered Hessians. Numerical results are presented for a selection of test examples and a mathematical model for heat conduction in a thermal battery with large orthotropic jumps in the material coefficients.
Finite element model for linear-elastic mixed mode loading using adaptive mesh strategy
无
2008-01-01
An adaptive mesh finite element model has been developed to predict the crack propagation direction as well as to calculate the stress intensity factors (SIFs), under linear-elastic assumption for mixed mode loading application. The finite element mesh is generated using the advancing front method. In order to suit the requirements of the fracture analysis, the generation of the background mesh and the construction of singular elements have been added to the developed program. The adaptive remeshing process is carried out based on the posteriori stress error norm scheme to obtain an optimal mesh. Previous works of the authors have proposed techniques for adaptive mesh generation of 2D cracked models. Facilitated by the singular elements, the displacement extrapolation technique is employed to calculate the SIF. The fracture is modeled by the splitting node approach and the trajectory follows the successive linear extensions of each crack increment. The SIFs values for two different case studies were estimated and validated by direct comparisons with other researchers work.
Essentials of finite element modeling and adaptive refinement
Dow, John O
2012-01-01
Finite Element Analysis is a very popular, computer-based tool that uses a complex system of points called nodes to make a grid called a ""mesh. "" The mesh contains the material and structural properties that define how the structure will react to certain loading conditions, allowing virtual testing and analysis of stresses or changes applied to the material or component design. This groundbreaking text extends the usefulness of finite element analysis by helping both beginners and advanced users alike. It simplifies, improves, and extends both the finite element method while at the same t
Shih, Chihhsiong; Yang, Yuanfan
2012-02-01
A novel three-dimensional (3-D) photorealistic texturing process is presented that applies a view-planning and view-sequencing algorithm to the 3-D coarse model to determine a set of best viewing angles for capturing the individual real-world objects/building's images. The best sequence of views will generate sets of visible edges in each view to serve as a guide for camera field shots by either manual adjustment or equipment alignment. The best view tries to cover as many objects/building surfaces as possible in one shot. This will lead to a smaller total number of shots taken for a complete model reconstruction requiring texturing with photo-realistic effects. The direct linear transformation method (DLT) is used for reprojection of 3-D model vertices onto a two-dimensional (2-D) images plane for actual texture mapping. Given this method, the actual camera orientations do not have to be unique and can be set arbitrarily without heavy and expensive positioning equipment. We also present results of a study on the texture-mapping precision as a function of the level of visible mesh subdivision. In addition, the control points selection for the DLT method used for reprojection of 3-D model vertices onto 2-D textured images is also investigated for its effects on mapping precision. By using DLT and perspective projection theories on a coarse model feature points, this technique will allow accurate 3-D texture mapping of refined model meshes of real-world buildings. The novel integration flow of this research not only greatly reduces the human labor and intensive equipment requirements of traditional methods, but also generates a more appealing photo-realistic appearance of reconstructed models, which is useful in many multimedia applications. The roles of view planning (VP) are multifold. VP can (1) reduce the repetitive texture-mapping computation load, (2) can present a set of visible model wireframe edges that can serve as a guide for images with sharp edges and
3D Simulation of Flow with Free Surface Based on Adaptive Octree Mesh System
Li Shaowu; Zhuang Qian; Huang Xiaoyun; Wang Dong
2015-01-01
The technique of adaptive tree mesh is an effective way to reduce computational cost through automatic adjustment of cell size according to necessity. In the present study, the 2D numerical N-S solver based on the adaptive quadtree mesh system was extended to a 3D one, in which a spatially adaptive octree mesh system and multiple parti-cle level set method were adopted for the convenience to deal with the air-water-structure multiple-medium coexisting domain. The stretching process of a dumbbell was simulated and the results indicate that the meshes are well adaptable to the free surface. The collapsing process of water column impinging a circle cylinder was simulated and from the results, it can be seen that the processes of fluid splitting and merging are properly simulated. The interaction of sec-ond-order Stokes waves with a square cylinder was simulated and the obtained drag force is consistent with the result by the Morison’s wave force formula with the coefficient values of the stable drag component and the inertial force component being set as 2.54.
STABILIZED FEM FOR CONVECTION-DIFFUSION PROBLEMS ON LAYER-ADAPTED MESHES
Hans-G(o)rg Roos
2009-01-01
The application of a standard Galerkin finite element method for convection-diflusion problems leads to oscillations in the discrete solution,therefore stabilization seems to be necessary.We discuss several recent stabilization methods,especially its combination with a Galerkin method on layer-adapted meshes.Supercloseness results obtained allow an improvement of the discrete solution using recovery techniques.
Barral, N.; Olivier, G.; Alauzet, F.
2017-02-01
Anisotropic metric-based mesh adaptation has proved its efficiency to reduce the CPU time of steady and unsteady simulations while improving their accuracy. However, its extension to time-dependent problems with body-fitted moving geometries is far from straightforward. This paper establishes a well-founded framework for multiscale mesh adaptation of unsteady problems with moving boundaries. This framework is based on a novel space-time analysis of the interpolation error, within the continuous mesh theory. An optimal metric field, called ALE metric field, is derived, which takes into account the movement of the mesh during the adaptation. Based on this analysis, the global fixed-point adaptation algorithm for time-dependent simulations is extended to moving boundary problems, within the range of body-fitted moving meshes and ALE simulations. Finally, three dimensional adaptive simulations with moving boundaries are presented to validate the proposed approach.
Jacobs, C. T.; Collins, G. S.; Piggott, M. D.; Kramer, S. C.; Wilson, C. R. G.
2013-02-01
Small-scale experiments of volcanic ash particle settling in water have demonstrated that ash particles can either settle slowly and individually, or rapidly and collectively as a gravitationally unstable ash-laden plume. This has important implications for the emplacement of tephra deposits on the seabed. Numerical modelling has the potential to extend the results of laboratory experiments to larger scales and explore the conditions under which plumes may form and persist, but many existing models are computationally restricted by the fixed mesh approaches that they employ. In contrast, this paper presents a new multiphase flow model that uses an adaptive unstructured mesh approach. As a simulation progresses, the mesh is optimized to focus numerical resolution in areas important to the dynamics and decrease it where it is not needed, thereby potentially reducing computational requirements. Model verification is performed using the method of manufactured solutions, which shows the correct solution convergence rates. Model validation and application considers 2-D simulations of plume formation in a water tank which replicate published laboratory experiments. The numerically predicted settling velocities for both individual particles and plumes, as well as instability behaviour, agree well with experimental data and observations. Plume settling is clearly hindered by the presence of a salinity gradient, and its influence must therefore be taken into account when considering particles in bodies of saline water. Furthermore, individual particles settle in the laminar flow regime while plume settling is shown (by plume Reynolds numbers greater than unity) to be in the turbulent flow regime, which has a significant impact on entrainment and settling rates. Mesh adaptivity maintains solution accuracy while providing a substantial reduction in computational requirements when compared to the same simulation performed using a fixed mesh, highlighting the benefits of an
Weitao LIU; Changchuan XIE; Michel Bellet; Hervé Combeau
2009-01-01
In order to improve the prediction accuracy of macrosegregation channel, an algorithm for dynamic remeshing is proposed. The basic idea is to generate fine elements near the liquidus isotherm. The norm of the gradient of solid fraction is used for piloting the remeshing in the mushy zone; whereas, the objective mesh size in the liquid is considered as a function of the distance to the liquidus isotherm. The efficiency of mesh adaptation is demonstrated by prediction of macrosegregation channel in a case of unidirectional solidification.
Burago, N. G.; Nikitin, I. S.; Yakushev, V. L.
2016-06-01
Techniques that improve the accuracy of numerical solutions and reduce their computational costs are discussed as applied to continuum mechanics problems with complex time-varying geometry. The approach combines shock-capturing computations with the following methods: (1) overlapping meshes for specifying complex geometry; (2) elastic arbitrarily moving adaptive meshes for minimizing the approximation errors near shock waves, boundary layers, contact discontinuities, and moving boundaries; (3) matrix-free implementation of efficient iterative and explicit-implicit finite element schemes; (4) balancing viscosity (version of the stabilized Petrov-Galerkin method); (5) exponential adjustment of physical viscosity coefficients; and (6) stepwise correction of solutions for providing their monotonicity and conservativeness.
Rosenberg, D.; Pouquet, A.; Germaschewski, K.; Ng, C. S.; Bhattacharjee, A.
2006-10-01
A recently developed spectral-element adaptive refinement incompressible magnetohydrodynamic (MHD) code is applied to simulate the problem of island coalescence instability (ICI) in 2D. The MHD solver is explicit, and uses the Elsasser formulation on high-order elements. It automatically takes advantage of the adaptive grid mechanics that have been described in [Rosenberg, Fournier, Fischer, Pouquet, J. Comp. Phys., 215, 59-80 (2006)], allowing both statically refined and dynamically refined grids. ICI is a MHD process that can produce strong current sheets and subsequent reconnection and heating in a high-Lundquist number plasma such as the solar corona [cf., Ng and Bhattacharjee, Phys. Plasmas, 5, 4028 (1998)]. Thus, it is desirable to use adaptive refinement grids to increase resolution, and to maintain accuracy at the same time. Results are compared with simulations using finite difference method with the same refinement grid, as well as pesudo-spectral simulations using uniform grid.
Kolobov, Vladimir [CFD Research Corporation, Huntsville, AL 35805, USA and The University of Alabama in Huntsville, Huntsville, AL 35805 (United States); Arslanbekov, Robert [CFD Research Corporation, Huntsville, AL 35805 (United States); Frolova, Anna [Computing Center of the Russian Academy of Sciences, Moscow, 119333 (Russian Federation)
2014-12-09
The paper describes an Adaptive Mesh in Phase Space (AMPS) technique for solving kinetic equations with deterministic mesh-based methods. The AMPS technique allows automatic generation of adaptive Cartesian mesh in both physical and velocity spaces using a Tree-of-Trees data structure. We illustrate advantages of AMPS for simulations of rarefied gas dynamics and electron kinetics on low temperature plasmas. In particular, we consider formation of the velocity distribution functions in hypersonic flows, particle kinetics near oscillating boundaries, and electron kinetics in a radio-frequency sheath. AMPS provide substantial savings in computational cost and increased efficiency of the mesh-based kinetic solvers.
A Simple Fault-Tolerant Adaptive and Minimal Routing Approach in 3-D Meshes
WU Jie(吴杰)
2003-01-01
In this paper we propose a sufficient condition for minimal routing in 3-dimensional (3-D) meshes with faulty nodes. It is based on an early work of the author on minimal routing in 2-dimensional (2-D) meshes. Unlike many traditional models that assume all the nodes know global fault distribution or just adjacent fault information, our approach is based on the concept of limited global fault information. First, we propose a fault model called faulty cube in which all faulty nodes in the system are contained in a set of faulty cubes. Fault information is then distributed to limited number of nodes while it is still sufficient to support minimal routing. The limited fault information collected at each node is represented by a vector called extended safety level. The extended safety level associated with a node can be used to determine the existence of a minimal path from this node to a given destination. Specifically, we study the existence of minimal paths at a given source node, limited distribution of fault information, minimal routing, and deadlock-free and livelock-free routing. Our results show that any minimal routing that is partially adaptive can be applied in our model as long as the destination node meets a certain condition. We also propose a dynamic planar-adaptive routing scheme that offers better fault tolerance and adaptivity than the planar-adaptive routing scheme in 3-D meshes. Our approach is the first attempt to address adaptive and minimal routing in 3-D meshes with faulty nodes using limited fault information.
Delaunay Refinement Mesh Generation
1997-05-18
determinant evaluation that considers floating- point operands, except for one limited example: Ottmann , Thiemt, and Ullrich [74] advocate the use of an...Sandia National Labo- ratories, October 1996. [74] Thomas Ottmann , Gerald Thiemt, and Christian Ullrich. Numerical Stability of Geometric Algorithms
DISCONTINUITY-CAPTURING FINITE ELEMENT COMPUTATION OF UNSTEADY FLOW WITH ADAPTIVE UNSTRUCTURED MESH
DONG Genjin; LU Xiyun; ZHUANG Lixian
2004-01-01
A discontinuity-capturing scheme of finite element method (FEM) is proposed. The unstructured-grid technique combined with a new type of adaptive mesh approach is developed for both compressible and incompressible unsteady flows, which exhibits the capability of capturing the shock waves and/or thin shear layers accurately in an unsteady viscous flow at high Reynolds number.In particular, a new testing variable, i.e., the disturbed kinetic energy E, is suggested and used in the adaptive mesh computation, which is universally applicable to the capturing of both shock waves and shear layers in the inviscid flow and viscous flow at high Reynolds number. Based on several calculated examples, this approach has been proved to be effective and efficient for the calculations of compressible and incompressible flows.
Optimal Channel Width Adaptation, Logical Topology Design, and Routing in Wireless Mesh Networks
Li Li
2009-01-01
Full Text Available Radio frequency spectrum is a finite and scarce resource. How to efficiently use the spectrum resource is one of the fundamental issues for multi-radio multi-channel wireless mesh networks. However, past research efforts that attempt to exploit multiple channels always assume channels of fixed predetermined width, which prohibits the further effective use of the spectrum resource. In this paper, we address how to optimally adapt channel width to more efficiently utilize the spectrum in IEEE802.11-based multi-radio multi-channel mesh networks. We mathematically formulate the channel width adaptation, logical topology design, and routing as a joint mixed 0-1 integer linear optimization problem, and we also propose our heuristic assignment algorithm. Simulation results show that our method can significantly improve spectrum use efficiency and network performance.
TRIM: A finite-volume MHD algorithm for an unstructured adaptive mesh
Schnack, D.D.; Lottati, I.; Mikic, Z. [Science Applications International Corp., San Diego, CA (United States)] [and others
1995-07-01
The authors describe TRIM, a MHD code which uses finite volume discretization of the MHD equations on an unstructured adaptive grid of triangles in the poloidal plane. They apply it to problems related to modeling tokamak toroidal plasmas. The toroidal direction is treated by a pseudospectral method. Care was taken to center variables appropriately on the mesh and to construct a self adjoint diffusion operator for cell centered variables.
FINITE VOLUME METHODS AND ADAPTIVE REFINEMENT FOR GLOBAL TSUNAMI PROPAGATION AND LOCAL INUNDATION
David L. George
2006-01-01
Full Text Available The shallow water equations are a commonly accepted approximation governing tsunami propagation. Numerically capturing certain features of local tsunami inundation requires solving these equations in their physically relevant conservative form, as integral con- servation laws for depth and momentum. This form of the equations presents challenges when trying to numerically model global tsunami propagation, so often the best numerical methods for the local inundation regime are not suitable for the global propagation regime. The different regimes of tsunami flow belong to different spatial scales as well, and re- quire correspondingly different grid resolutions. The long wavelength of deep ocean tsunamis requires a large global scale computing domain, yet near the shore the propa- gating energy is compressed and focused by bathymetry in unpredictable ways. This can lead to large variations in energy and run-up even over small localized regions.We have developed a finite volume method to deal with the diverse flow regimes of tsunamis. These methods are well suited for the inundation regime—they are robust in the presence of bores and steep gradients, or drying regions, and can capture the inundating shoreline and run-up features. Additionally, these methods are well-balanced, meaning that they can appropriately model global propagation.To deal with the disparate spatial scales, we have used adaptive refinement algorithms originally developed for gas dynamics, where often steep variation is highly localized at a given time, but moves throughout the domain. These algorithms allow evolving Cartesian sub-grids that can move with the propagating waves and highly resolve local inundation of impacted areas in a single global scale computation. Because the dry regions are part of the computing domain, simple rectangular cartesian grids eliminate the need for complex shoreline-fitted mesh generation.
Danaila, Ionut; Hecht, Frédéric
2009-01-01
to appear in J. Computational Physics; Numerical computations of stationary states of fast-rotating Bose-Einstein condensates require high spatial resolution due to the presence of a large number of quantized vortices. In this paper we propose a low-order finite element method with mesh adaptivity by metric control, as an alternative approach to the commonly used high order (finite difference or spectral) approximation methods. The mesh adaptivity is used with two different numerical algorith...
Danaila, Ionut; Hecht, Frédéric
2010-01-01
Numerical computations of stationary states of fast-rotating Bose-Einstein condensates re- quire high spatial resolution due to the presence of a large number of quantized vortices. In this paper we propose a low-order finite element method with mesh adaptivity by metric con- trol, as an alternative approach to the commonly used high order (finite difference or spectral) approximation methods. The mesh adaptivity is used with two different numerical algorithms to compute stationary vortex sta...
Danaila, Ionut; Hecht, Frédéric
2010-01-01
to appear in J. Computational Physics; Numerical computations of stationary states of fast-rotating Bose-Einstein condensates require high spatial resolution due to the presence of a large number of quantized vortices. In this paper we propose a low-order finite element method with mesh adaptivity by metric control, as an alternative approach to the commonly used high order (finite difference or spectral) approximation methods. The mesh adaptivity is used with two different numerical algorith...
Error Estimate and Adaptive Refinement in Mixed Discrete Least Squares Meshless Method
J. Amani
2014-01-01
Full Text Available The node moving and multistage node enrichment adaptive refinement procedures are extended in mixed discrete least squares meshless (MDLSM method for efficient analysis of elasticity problems. In the formulation of MDLSM method, mixed formulation is accepted to avoid second-order differentiation of shape functions and to obtain displacements and stresses simultaneously. In the refinement procedures, a robust error estimator based on the value of the least square residuals functional of the governing differential equations and its boundaries at nodal points is used which is inherently available from the MDLSM formulation and can efficiently identify the zones with higher numerical errors. The results are compared with the refinement procedures in the irreducible formulation of discrete least squares meshless (DLSM method and show the accuracy and efficiency of the proposed procedures. Also, the comparison of the error norms and convergence rate show the fidelity of the proposed adaptive refinement procedures in the MDLSM method.
Fluid flow and heat transfer investigation of pebble bed reactors using mesh-adaptive LES
Pavlidis, Dimitrios; Lathouwers, Danny, E-mail: d.lathouwers@tudelft.nl
2013-11-15
The very high temperature reactor is one of the designs currently being considered for nuclear power generation. One its variants is the pebble bed reactor in which the coolant passes through complex geometries (pores) at high Reynolds numbers. A computational fluid dynamics model with anisotropic mesh adaptivity is used to investigate coolant flow and heat transfer in such reactors. A novel method for implicitly incorporating solid boundaries based on multi-fluid flow modelling is adopted. The resulting model is able to resolve and simulate flow and heat transfer in randomly packed beds, regardless of the actual geometry, starting off with arbitrarily coarse meshes. The model is initially evaluated using an orderly stacked square channel of channel-height-to-particle diameter ratio of unity for a range of Reynolds numbers. The model is then applied to the face-centred cubical geometry. coolant flow and heat transfer patterns are investigated.
Ishikawa, Takeo; Matsunami, Michio
This paper proposes a method to generate adaptively 2D and 3D finite element meshes taking into account the continuity requirements of the magnetic field at the interface between two neighboring elements. First, this paper proposes a new error estimator that includes the Zienkiewicz and Zhu error norm estimator and the boundary rules in the electromagnetic field. Using a 2D simple model, this paper decides two parameters of the proposed estimator. Next, this paper presents a 3D mesh generation method based on the Voronoi-Delaunay theory, which ensures that the bounding surface of the domain is contained in the triangulation. The method has the capability to decrease the amount of information on the connectivity of boundary nodes by generating nodes not only in the interior of the domain but also on its surface. Two simple magnetostatic field problems are provided to illustrate the usefulness of the proposed method.
Abdolrasoul Khosravi
2015-05-01
Findings: the results show that there is no significant difference between documents relevance when Pubmed suggestions or MESH terms are used during each search session. According to the findings of this study both Pubmed suggestions and MESH terms help the user with better understanding the concept and they both broaden the terminology of users. At the same time both methods are efficient in relevancy of the retrieved documents.
Moving mesh generation with a sequential approach for solving PDEs
of physical and mesh equations suffers typically from long computation time due to highly nonlinear coupling between the two equations. Moreover, the extended system (physical and mesh equations) may be sensitive to the tuning parameters such as a temporal relaxation factor. It is therefore useful to design...... adaptive grid method (local refinement by adding/deleting the meshes at a discrete time level) as well as of efficiency for the dynamic adaptive grid method (or moving mesh method) where the number of meshes is not changed. For illustration, a phase change problem is solved with the decomposition algorithm.......In moving mesh methods, physical PDEs and a mesh equation derived from equidistribution of an error metrics (so-called the monitor function) are simultaneously solved and meshes are dynamically concentrated on steep regions (Lim et al., 2001). However, the simultaneous solution procedure...
Bajc, Iztok; Hecht, Frédéric; Žumer, Slobodan
2016-09-01
This paper presents a 3D mesh adaptivity strategy on unstructured tetrahedral meshes by a posteriori error estimates based on metrics derived from the Hessian of a solution. The study is made on the case of a nonlinear finite element minimization scheme for the Landau-de Gennes free energy functional of nematic liquid crystals. Newton's iteration for tensor fields is employed with steepest descent method possibly stepping in. Aspects relating the driving of mesh adaptivity within the nonlinear scheme are considered. The algorithmic performance is found to depend on at least two factors: when to trigger each single mesh adaptation, and the precision of the correlated remeshing. Each factor is represented by a parameter, with its values possibly varying for every new mesh adaptation. We empirically show that the time of the overall algorithm convergence can vary considerably when different sequences of parameters are used, thus posing a question about optimality. The extensive testings and debugging done within this work on the simulation of systems of nematic colloids substantially contributed to the upgrade of an open source finite element-oriented programming language to its 3D meshing possibilities, as also to an outer 3D remeshing module.
Guo, Zhikui; Chen, Chao; Tao, Chunhui
2016-04-01
Since 2007, there are four China Da yang cruises (CDCs), which have been carried out to investigate polymetallic sulfides in the southwest Indian ridge (SWIR) and have acquired both gravity data and bathymetry data on the corresponding survey lines(Tao et al., 2014). Sandwell et al. (2014) published a new global marine gravity model including the free air gravity data and its first order vertical gradient (Vzz). Gravity data and its gradient can be used to extract unknown density structure information(e.g. crust thickness) under surface of the earth, but they contain all the mass effect under the observation point. Therefore, how to get accurate gravity and its gradient effect of the existing density structure (e.g. terrain) has been a key issue. Using the bathymetry data or ETOPO1 (http://www.ngdc.noaa.gov/mgg/global/global.html) model at a full resolution to calculate the terrain effect could spend too much computation time. We expect to develop an effective method that takes less time but can still yield the desired accuracy. In this study, a constant-density polyhedral model is used to calculate the gravity field and its vertical gradient, which is based on the work of Tsoulis (2012). According to gravity field attenuation with distance and variance of bathymetry, we present an adaptive mesh refinement and coarsening strategies to merge both global topography data and multi-beam bathymetry data. The local coarsening or size of mesh depends on user-defined accuracy and terrain variation (Davis et al., 2011). To depict terrain better, triangular surface element and rectangular surface element are used in fine and coarse mesh respectively. This strategy can also be applied to spherical coordinate in large region and global scale. Finally, we applied this method to calculate Bouguer gravity anomaly (BGA), mantle Bouguer anomaly(MBA) and their vertical gradient in SWIR. Further, we compared the result with previous results in the literature. Both synthetic model
Abdulnaser M. Alshoaibi
2009-01-01
Full Text Available The purpose of this study is on the determination of 2D crack paths and surfaces as well as on the evaluation of the stress intensity factors as a part of the damage tolerant assessment. Problem statement: The evaluation of SIFs and crack tip singular stresses for arbitrary fracture structure are a challenging problem, involving the calculation of the crack path and the crack propagation rates at each step especially under mixed mode loading. Approach: This study was provided a finite element code which produces results comparable to the current available commercial software. Throughout the simulation of crack propagation an automatic adaptive mesh was carried out in the vicinity of the crack front nodes and in the elements which represent the higher stresses distribution. The finite element mesh was generated using the advancing front method. The adaptive remising process carried out based on the posteriori stress error norm scheme to obtain an optimal mesh. The onset criterion of crack propagation was based on the stress intensity factors which provide as the most important parameter that must be accurately estimated. Facilitated by the singular elements, the displacement extrapolation technique is employed to calculate the stress intensity factor. Crack direction is predicted using the maximum circumferential stress theory. The fracture was modeled by the splitting node approach and the trajectory follows the successive linear extensions of each crack increment. The propagation process is driven by Linear Elastic Fracture Mechanics (LEFM approach with minimum user interaction. Results: In evaluating the accuracy of the estimated stress intensity factors and the crack path predictions, the results were compared with sets of experimental data, benchmark analytical solutions as well as numerical results of other researchers. Conclusion/Recommendations: The assessment indicated that the program was highly reliable to evaluate the stress intensity
Numerical study of three-dimensional liquid jet breakup with adaptive unstructured meshes
Xie, Zhihua; Pavlidis, Dimitrios; Salinas, Pablo; Pain, Christopher; Matar, Omar
2016-11-01
Liquid jet breakup is an important fundamental multiphase flow, often found in many industrial engineering applications. The breakup process is very complex, involving jets, liquid films, ligaments, and small droplets, featuring tremendous complexity in interfacial topology and a large range of spatial scales. The objective of this study is to investigate the fluid dynamics of three-dimensional liquid jet breakup problems, such as liquid jet primary breakup and gas-sheared liquid jet breakup. An adaptive unstructured mesh modelling framework is employed here, which can modify and adapt unstructured meshes to optimally represent the underlying physics of multiphase problems and reduce computational effort without sacrificing accuracy. The numerical framework consists of a mixed control volume and finite element formulation, a 'volume of fluid' type method for the interface capturing based on a compressive control volume advection method and second-order finite element methods, and a force-balanced algorithm for the surface tension implementation. Numerical examples of some benchmark tests and the dynamics of liquid jet breakup with and without ambient gas are presented to demonstrate the capability of this method.
Adaptive Fault-Tolerant Routing in 2D Mesh with Cracky Rectangular Model
Yi Yang
2014-01-01
Full Text Available This paper mainly focuses on routing in two-dimensional mesh networks. We propose a novel faulty block model, which is cracky rectangular block, for fault-tolerant adaptive routing. All the faulty nodes and faulty links are surrounded in this type of block, which is a convex structure, in order to avoid routing livelock. Additionally, the model constructs the interior spanning forest for each block in order to keep in touch with the nodes inside of each block. The procedure for block construction is dynamically and totally distributed. The construction algorithm is simple and ease of implementation. And this is a fully adaptive block which will dynamically adjust its scale in accordance with the situation of networks, either the fault emergence or the fault recovery, without shutdown of the system. Based on this model, we also develop a distributed fault-tolerant routing algorithm. Then we give the formal proof for this algorithm to guarantee that messages will always reach their destinations if and only if the destination nodes keep connecting with these mesh networks. So the new model and routing algorithm maximize the availability of the nodes in networks. This is a noticeable overall improvement of fault tolerability of the system.
Commercon, Benoit; Audit, Edouard; Hennebelle, Patrick; Chabrier, Gilles
2011-01-01
Radiative transfer has a strong impact on the collapse and the fragmentation of prestellar dense cores. We present the radiation-hydrodynamics solver we designed for the RAMSES code. The method is designed for astrophysical purposes, and in particular for protostellar collapse. We present the solver, using the co-moving frame to evaluate the radiative quantities. We use the popular flux limited diffusion approximation, under the grey approximation (one group of photon). The solver is based on the second-order Godunov scheme of RAMSES for its hyperbolic part, and on an implicit scheme for the radiation diffusion and the coupling between radiation and matter. We report in details our methodology to integrate the RHD solver into RAMSES. We test successfully the method against several conventional tests. For validation in 3D, we perform calculations of the collapse of an isolated 1 M_sun prestellar dense core, without rotation. We compare successfully the results with previous studies using different models for r...
Simulations of recoiling black holes: adaptive mesh refinement and radiative transfer
Meliani, Zakaria; Olivares, Hector; Porth, Oliver; Rezzolla, Luciano; Younsi, Ziri
2016-01-01
(Abridged) We here continue our effort to model the behaviour of matter when orbiting or accreting onto a generic black hole by developing a new numerical code employing advanced techniques geared solve the equations of in general-relativistic hydrodynamics. The new code employs a number of high-resolution shock-capturing Riemann-solvers and reconstruction algorithms, exploiting the enhanced accuracy and the reduced computational cost of AMR techniques. In addition, the code makes use of sophisticated ray-tracing libraries that, coupled with general-relativistic radiation-transfer calculations, allow us to compute accurately the electromagnetic emissions from such accretion flows. We validate the new code by presenting an extensive series of stationary accretion flows either in spherical or axial symmetry and performed either in 2D or 3D. In addition, we consider the highly nonlinear scenario of a recoiling black hole produced in the merger of a supermassive black hole binary interacting with the surrounding ...
Gutowski, William J.; Prusa, Joseph M.; Smolarkiewicz, Piotr K.
2012-05-08
This project had goals of advancing the performance capabilities of the numerical general circulation model EULAG and using it to produce a fully operational atmospheric global climate model (AGCM) that can employ either static or dynamic grid stretching for targeted phenomena. The resulting AGCM combined EULAG's advanced dynamics core with the "physics" of the NCAR Community Atmospheric Model (CAM). Effort discussed below shows how we improved model performance and tested both EULAG and the coupled CAM-EULAG in several ways to demonstrate the grid stretching and ability to simulate very well a wide range of scales, that is, multi-scale capability. We leveraged our effort through interaction with an international EULAG community that has collectively developed new features and applications of EULAG, which we exploited for our own work summarized here. Overall, the work contributed to over 40 peer-reviewed publications and over 70 conference/workshop/seminar presentations, many of them invited. 3a. EULAG Advances EULAG is a non-hydrostatic, parallel computational model for all-scale geophysical flows. EULAG's name derives from its two computational options: EULerian (flux form) or semi-LAGrangian (advective form). The model combines nonoscillatory forward-in-time (NFT) numerical algorithms with a robust elliptic Krylov solver. A signature feature of EULAG is that it is formulated in generalized time-dependent curvilinear coordinates. In particular, this enables grid adaptivity. In total, these features give EULAG novel advantages over many existing dynamical cores. For EULAG itself, numerical advances included refining boundary conditions and filters for optimizing model performance in polar regions. We also added flexibility to the model's underlying formulation, allowing it to work with the pseudo-compressible equation set of Durran in addition to EULAG's standard anelastic formulation. Work in collaboration with others also extended the
Adaptive local surface refinement based on LR NURBS and its application to contact
Zimmermann, Christopher; Sauer, Roger A.
2017-08-01
A novel adaptive local surface refinement technique based on Locally Refined Non-Uniform Rational B-Splines (LR NURBS) is presented. LR NURBS can model complex geometries exactly and are the rational extension of LR B-splines. The local representation of the parameter space overcomes the drawback of non-existent local refinement in standard NURBS-based isogeometric analysis. For a convenient embedding into general finite element codes, the Bézier extraction operator for LR NURBS is formulated. An automatic remeshing technique is presented that allows adaptive local refinement and coarsening of LR NURBS. In this work, LR NURBS are applied to contact computations of 3D solids and membranes. For solids, LR NURBS-enriched finite elements are used to discretize the contact surfaces with LR NURBS finite elements, while the rest of the body is discretized by linear Lagrange finite elements. For membranes, the entire surface is discretized by LR NURBS. Various numerical examples are shown, and they demonstrate the benefit of using LR NURBS: Compared to uniform refinement, LR NURBS can achieve high accuracy at lower computational cost.
Multi-dimensional upwind fluctuation splitting scheme with mesh adaption for hypersonic viscous flow
Wood, William Alfred, III
production is shown relative to DMFDSFV. Remarkably the fluctuation splitting scheme shows grid converged skin friction coefficients with only five points in the boundary layer for this case. A viscous Mach 17.6 (perfect gas) cylinder case demonstrates solution monotonicity and heat transfer capability with the fluctuation splitting scheme. While fluctuation splitting is recommended over DMFDSFV, the difference in performance between the schemes is not so great as to obsolete DMFDSFV. The second half of the dissertation develops a local, compact, anisotropic unstructured mesh adaption scheme in conjunction with the multi-dimensional upwind solver, exhibiting a characteristic alignment behavior for scalar problems. This alignment behavior stands in contrast to the curvature clustering nature of the local, anisotropic unstructured adaption strategy based upon a posteriori error estimation that is used for comparison. The characteristic alignment is most pronounced for linear advection, with reduced improvement seen for the more complex non-linear advection and advection-diffusion cases. The adaption strategy is extended to the two-dimensional and axisymmetric Navier-Stokes equations of motion through the concept of fluctuation minimization. The system test case for the adaption strategy is a sting mounted capsule at Mach-10 wind tunnel conditions, considered in both two-dimensional and axisymmetric configurations. For this complex flowfield the adaption results are disappointing since feature alignment does not emerge from the local operations. Aggressive adaption is shown to result in a loss of robustness for the solver, particularly in the bow shock/stagnation point interaction region. Reducing the adaption strength maintains solution robustness but fails to produce significant improvement in the surface heat transfer predictions.
6th International Meshing Roundtable '97
White, D.
1997-09-01
The goal of the 6th International Meshing Roundtable is to bring together researchers and developers from industry, academia, and government labs in a stimulating, open environment for the exchange of technical information related to the meshing process. In the pas~ the Roundtable has enjoyed significant participation born each of these groups from a wide variety of countries. The Roundtable will consist of technical presentations from contributed papers and abstracts, two invited speakers, and two invited panels of experts discussing topics related to the development and use of automatic mesh generation tools. In addition, this year we will feature a "Bring Your Best Mesh" competition and poster session to encourage discussion and participation from a wide variety of mesh generation tool users. The schedule and evening social events are designed to provide numerous opportunities for informal dialog. A proceedings will be published by Sandia National Laboratories and distributed at the Roundtable. In addition, papers of exceptionally high quaIity will be submitted to a special issue of the International Journal of Computational Geometry and Applications. Papers and one page abstracts were sought that present original results on the meshing process. Potential topics include but are got limited to: Unstructured triangular and tetrahedral mesh generation Unstructured quadrilateral and hexahedral mesh generation Automated blocking and structured mesh generation Mixed element meshing Surface mesh generation Geometry decomposition and clean-up techniques Geometry modification techniques related to meshing Adaptive mesh refinement and mesh quality control Mesh visualization Special purpose meshing algorithms for particular applications Theoretical or novel ideas with practical potential Technical presentations from industrial researchers.
Scalable and Adaptive Streaming of 3D Mesh to Heterogeneous Devices
Abderrahim, Zeineb; Bouhlel, Mohamed Salim
2016-12-01
This article comprises a presentation of a web platform for the diffusion and visualization of 3D compressed data on the web. Indeed, the major goal of this work resides in the proposal of the transfer adaptation of the three-dimensional data to resources (network bandwidth, the type of visualization terminals, display resolution, user's preferences...). Also, it is an attempt to provide an effective consultation adapted to the user's request (preferences, levels of the requested detail, etc.). Such a platform can adapt the levels of detail to the change in the bandwidth and the rendering time when loading the mesh at the client level. In addition, the levels of detail are adapted to the distance between the object and the camera. These features are able to minimize the latency time and to make the real time interaction possible. The experiences as well as the comparison with the existing solutions show auspicious results in terms of latency, scalability and the quality of the experience offered to the users.
Gutowski, William J.; Prusa, Joseph M.; Smolarkiewicz, Piotr K.
2012-05-08
This project had goals of advancing the performance capabilities of the numerical general circulation model EULAG and using it to produce a fully operational atmospheric global climate model (AGCM) that can employ either static or dynamic grid stretching for targeted phenomena. The resulting AGCM combined EULAG's advanced dynamics core with the "physics" of the NCAR Community Atmospheric Model (CAM). Effort discussed below shows how we improved model performance and tested both EULAG and the coupled CAM-EULAG in several ways to demonstrate the grid stretching and ability to simulate very well a wide range of scales, that is, multi-scale capability. We leveraged our effort through interaction with an international EULAG community that has collectively developed new features and applications of EULAG, which we exploited for our own work summarized here. Overall, the work contributed to over 40 peer-reviewed publications and over 70 conference/workshop/seminar presentations, many of them invited. 3a. EULAG Advances EULAG is a non-hydrostatic, parallel computational model for all-scale geophysical flows. EULAG's name derives from its two computational options: EULerian (flux form) or semi-LAGrangian (advective form). The model combines nonoscillatory forward-in-time (NFT) numerical algorithms with a robust elliptic Krylov solver. A signature feature of EULAG is that it is formulated in generalized time-dependent curvilinear coordinates. In particular, this enables grid adaptivity. In total, these features give EULAG novel advantages over many existing dynamical cores. For EULAG itself, numerical advances included refining boundary conditions and filters for optimizing model performance in polar regions. We also added flexibility to the model's underlying formulation, allowing it to work with the pseudo-compressible equation set of Durran in addition to EULAG's standard anelastic formulation. Work in collaboration with others also extended the
Goal functional evaluations for phase-field fracture using PU-based DWR mesh adaptivity
Wick, Thomas
2016-06-01
In this study, a posteriori error estimation and goal-oriented mesh adaptivity are developed for phase-field fracture propagation. Goal functionals are computed with the dual-weighted residual (DWR) method, which is realized by a recently introduced novel localization technique based on a partition-of-unity (PU). This technique is straightforward to apply since the weak residual is used. The influence of neighboring cells is gathered by the PU. Consequently, neither strong residuals nor jumps over element edges are required. Therefore, this approach facilitates the application of the DWR method to coupled (nonlinear) multiphysics problems such as fracture propagation. These developments then allow for a systematic investigation of the discretization error for certain quantities of interest. Specifically, our focus on the relationship between the phase-field regularization and the spatial discretization parameter in terms of goal functional evaluations is novel.
Multiphase flow modelling of explosive volcanic eruptions using adaptive unstructured meshes
Jacobs, Christian T.; Collins, Gareth S.; Piggott, Matthew D.; Kramer, Stephan C.
2014-05-01
Explosive volcanic eruptions generate highly energetic plumes of hot gas and ash particles that produce diagnostic deposits and pose an extreme environmental hazard. The formation, dispersion and collapse of these volcanic plumes are complex multiscale processes that are extremely challenging to simulate numerically. Accurate description of particle and droplet aggregation, movement and settling requires a model capable of capturing the dynamics on a range of scales (from cm to km) and a model that can correctly describe the important multiphase interactions that take place. However, even the most advanced models of eruption dynamics to date are restricted by the fixed mesh-based approaches that they employ. The research presented herein describes the development of a compressible multiphase flow model within Fluidity, a combined finite element / control volume computational fluid dynamics (CFD) code, for the study of explosive volcanic eruptions. Fluidity adopts a state-of-the-art adaptive unstructured mesh-based approach to discretise the domain and focus numerical resolution only in areas important to the dynamics, while decreasing resolution where it is not needed as a simulation progresses. This allows the accurate but economical representation of the flow dynamics throughout time, and potentially allows large multi-scale problems to become tractable in complex 3D domains. The multiphase flow model is verified with the method of manufactured solutions, and validated by simulating published gas-solid shock tube experiments and comparing the numerical results against pressure gauge data. The application of the model considers an idealised 7 km by 7 km domain in which the violent eruption of hot gas and volcanic ash high into the atmosphere is simulated. Although the simulations do not correspond to a particular eruption case study, the key flow features observed in a typical explosive eruption event are successfully captured. These include a shock wave resulting
Kennedy, A. T.; Cornford, S. L.; Lunt, D. J.; Payne, A. J.
2016-12-01
Paleo ice sheet models provide valuable tools to test our understanding of the cryosphere-ocean-climate system and how it might respond under warm conditions. However, the long time scales and uncertainty in boundary conditions required for paleo simulations usually necessitates the use of highly parameterised and simplified model techniques, and not the use of state-of-the-art models. One such state-of-the-art model which has been used for the present day is BISICLES which, due to its adaptive mesh refinement capabilities, can explicitly model highly localised and dynamic features such as grounding line migration and ice streams. We will show results testing the suitability of using such a model for a paleo application, including the model's sensitivity to uncertainty in the ice volume, bedrock properties and climatic/oceanic forcing. We will also show preliminary results of modelling the Antarctic Ice Sheet state at the Mid-Pliocene, a period when the ice sheet is expected to have contributed many metres worth of global mean sea level increase. We will highlight the range of ice mass loss under different parameterisation and forcing schemes and the level of agreement with previous data and modelling studies, e.g. Miller et al (2012) and DeConto & Pollard (2016). It remains too computationally expensive to run BISICLES for a full glacial-interglacial cycle for example, but this model could still prove valuable for assessing the role of highly dynamic features in past ice sheets. This work will help bridge a gap in understanding of the strengths and weaknesses of the simpler models used in paleo ice sheet modelling compared to the state-of-the-art models used for present day and future prediction.
Development of Adaptive Model Refinement (AMoR) for Multiphysics and Multifidelity Problems
Turinsky, Paul [North Carolina State Univ., Raleigh, NC (United States)
2015-02-09
This project investigated the development and utilization of Adaptive Model Refinement (AMoR) for nuclear systems simulation applications. AMoR refers to utilization of several models of physical phenomena which differ in prediction fidelity. If the highest fidelity model is judged to always provide or exceeded the desired fidelity, than if one can determine the difference in a Quantity of Interest (QoI) between the highest fidelity model and lower fidelity models, one could utilize the fidelity model that would just provide the magnitude of the QoI desired. Assuming lower fidelity models require less computational resources, in this manner computational efficiency can be realized provided the QoI value can be accurately and efficiently evaluated. This work utilized Generalized Perturbation Theory (GPT) to evaluate the QoI, by convoluting the GPT solution with the residual of the highest fidelity model determined using the solution from lower fidelity models. Specifically, a reactor core neutronics problem and thermal-hydraulics problem were studied to develop and utilize AMoR. The highest fidelity neutronics model was based upon the 3D space-time, two-group, nodal diffusion equations as solved in the NESTLE computer code. Added to the NESTLE code was the ability to determine the time-dependent GPT neutron flux. The lower fidelity neutronics model was based upon the point kinetics equations along with utilization of a prolongation operator to determine the 3D space-time, two-group flux. The highest fidelity thermal-hydraulics model was based upon the space-time equations governing fluid flow in a closed channel around a heat generating fuel rod. The Homogenous Equilibrium Mixture (HEM) model was used for the fluid and Finite Difference Method was applied to both the coolant and fuel pin energy conservation equations. The lower fidelity thermal-hydraulic model was based upon the same equations as used for the highest fidelity model but now with coarse spatial
A comparison of tetrahedral mesh improvement techniques
Freitag, L.A.; Ollivier-Gooch, C. [Argonne National Lab., IL (United States). Mathematics and Computer Science Div.
1996-12-01
Automatic mesh generation and adaptive refinement methods for complex three-dimensional domains have proven to be very successful tools for the efficient solution of complex applications problems. These methods can, however, produce poorly shaped elements that cause the numerical solution to be less accurate and more difficult to compute. Fortunately, the shape of the elements can be improved through several mechanisms, including face-swapping techniques that change local connectivity and optimization-based mesh smoothing methods that adjust grid point location. The authors consider several criteria for each of these two methods and compare the quality of several meshes obtained by using different combinations of swapping and smoothing. Computational experiments show that swapping is critical to the improvement of general mesh quality and that optimization-based smoothing is highly effective in eliminating very small and very large angles. The highest quality meshes are obtained by using a combination of swapping and smoothing techniques.
Output-Based Adaptive Meshing Applied to Space Launch System Booster Separation Analysis
Dalle, Derek J.; Rogers, Stuart E.
2015-01-01
This paper presents details of Computational Fluid Dynamic (CFD) simulations of the Space Launch System during solid-rocket booster separation using the Cart3D inviscid code with comparisons to Overflow viscous CFD results and a wind tunnel test performed at NASA Langley Research Center's Unitary PlanWind Tunnel. The Space Launch System (SLS) launch vehicle includes two solid-rocket boosters that burn out before the primary core stage and thus must be discarded during the ascent trajectory. The main challenges for creating an aerodynamic database for this separation event are the large number of basis variables (including orientation of the core, relative position and orientation of the boosters, and rocket thrust levels) and the complex flow caused by the booster separation motors. The solid-rocket boosters are modified from their form when used with the Space Shuttle Launch Vehicle, which has a rich flight history. However, the differences between the SLS core and the Space Shuttle External Tank result in the boosters separating with much narrower clearances, and so reducing aerodynamic uncertainty is necessary to clear the integrated system for flight. This paper discusses an approach that has been developed to analyze about 6000 wind tunnel simulations and 5000 flight vehicle simulations using Cart3D in adaptive-meshing mode. In addition, a discussion is presented of Overflow viscous CFD runs used for uncertainty quantification. Finally, the article presents lessons learned and improvements that will be implemented in future separation databases.
Zhou, Lu; Yazici, Birsen
2011-04-01
Quantitatively accurate fluorescence diffuse optical tomographic (FDOT) image reconstruction is a computationally demanding problem that requires repeated numerical solutions of two coupled partial differential equations and an associated inverse problem. Recently, adaptive finite element methods have been explored to reduce the computation requirements of the FDOT image reconstruction. However, existing approaches ignore the ubiquitous presence of noise in boundary measurements. In this paper, we analyze the effect of finite element discretization on the FDOT forward and inverse problems in the presence of measurement noise and develop novel adaptive meshing algorithms for FDOT that take into account noise statistics. We formulate the FDOT inverse problem as an optimization problem in the maximum a posteriori framework to estimate the fluorophore concentration in a bounded domain. We use the mean-square-error (MSE) between the exact solution and the discretized solution as a figure of merit to evaluate the image reconstruction accuracy, and derive an upper bound on the MSE which depends upon the forward and inverse problem discretization parameters, noise statistics, a priori information of fluorophore concentration, source and detector geometry, as well as background optical properties. Next, we use this error bound to develop adaptive meshing algorithms for the FDOT forward and inverse problems to reduce the MSE due to discretization in the reconstructed images. Finally, we present a set of numerical simulations to illustrate the practical advantages of our adaptive meshing algorithms for FDOT image reconstruction.
Hayek, Mohamed; Ackerer, Philippe; Sonnendrücker, Éric
2009-02-01
We propose a new refinement indicator (NRI) for adaptive parameterization to determine the diffusion coefficient in an elliptic equation in two-dimensional space. The diffusion coefficient is assumed to be a piecewise constant space function. The unknowns are both the parameter values and the zonation. Refinement indicators are used to localize parameter discontinuities in order to construct iteratively the zonation (parameterization). The refinement indicator is obtained usually by using the first-order effect on the objective function of removing degrees of freedom for a current set of parameters. In this work, in order to reduce the computation costs, we propose a new refinement indicator based on the second-order effect on the objective function. This new refinement indicator depends on the objective function, and its first and second derivatives with respect to the parameter constraints. Numerical experiments show the high efficiency of the new refinement indicator compared to the standard one.
Rosenberg, Duane; Fournier, Aimé; Fischer, Paul; Pouquet, Annick
2006-06-01
An object-oriented geophysical and astrophysical spectral-element adaptive refinement (GASpAR) code is introduced. Like most spectral-element codes, GASpAR combines finite-element efficiency with spectral-method accuracy. It is also designed to be flexible enough for a range of geophysics and astrophysics applications where turbulence or other complex multiscale problems arise. The formalism accommodates both conforming and non-conforming elements. Several aspects of this code derive from existing methods, but here are synthesized into a new formulation of dynamic adaptive refinement (DARe) of non-conforming h-type. As a demonstration of the code, several new 2D test cases are introduced that have time-dependent analytic solutions and exhibit localized flow features, including the 2D Burgers equation with straight, curved-radial and oblique-colliding fronts. These are proposed as standard test problems for comparable DARe codes. Quantitative errors are reported for 2D spatial and temporal convergence of DARe.
MGGHAT: Elliptic PDE software with adaptive refinement, multigrid and high order finite elements
Mitchell, William F.
1993-01-01
MGGHAT (MultiGrid Galerkin Hierarchical Adaptive Triangles) is a program for the solution of linear second order elliptic partial differential equations in two dimensional polygonal domains. This program is now available for public use. It is a finite element method with linear, quadratic or cubic elements over triangles. The adaptive refinement via newest vertex bisection and the multigrid iteration are both based on a hierarchical basis formulation. Visualization is available at run time through an X Window display, and a posteriori through output files that can be used as GNUPLOT input. In this paper, we describe the methods used by MGGHAT, define the problem domain for which it is appropriate, illustrate use of the program, show numerical and graphical examples, and explain how to obtain the software.
Adam, A.; Pavlidis, D.; Percival, J. R.; Salinas, P.; Xie, Z.; Fang, F.; Pain, C. C.; Muggeridge, A. H.; Jackson, M. D.
2016-09-01
A general, higher-order, conservative and bounded interpolation for the dynamic and adaptive meshing of control-volume fields dual to continuous and discontinuous finite element representations is presented. Existing techniques such as node-wise interpolation are not conservative and do not readily generalise to discontinuous fields, whilst conservative methods such as Grandy interpolation are often too diffusive. The new method uses control-volume Galerkin projection to interpolate between control-volume fields. Bounded solutions are ensured by using a post-interpolation diffusive correction. Example applications of the method to interface capturing during advection and also to the modelling of multiphase porous media flow are presented to demonstrate the generality and robustness of the approach.
Axisymmetric modeling of cometary mass loading on an adaptively refined grid: MHD results
Gombosi, Tamas I.; Powell, Kenneth G.; De Zeeuw, Darren L.
1994-01-01
The first results of an axisymmetric magnetohydrodynamic (MHD) model of the interaction of an expanding cometary atmosphere with the solar wind are presented. The model assumes that far upstream the plasma flow lines are parallel to the magnetic field vector. The effects of mass loading and ion-neutral friction are taken into account by the governing equations, whcih are solved on an adaptively refined unstructured grid using a Monotone Upstream Centered Schemes for Conservative Laws (MUSCL)-type numerical technique. The combination of the adaptive refinement with the MUSCL-scheme allows the entire cometary atmosphere to be modeled, while still resolving both the shock and the near nucleus of the comet. The main findingsare the following: (1) A shock is formed approximately = 0.45 Mkm upstream of the comet (its location is controlled by the sonic and Alfvenic Mach numbers of the ambient solar wind flow and by the cometary mass addition rate). (2) A contact surface is formed approximately = 5,600 km upstream of the nucleus separating an outward expanding cometary ionosphere from the nearly stagnating solar wind flow. The location of the contact surface is controlled by the upstream flow conditions, the mass loading rate and the ion-neutral drag. The contact surface is also the boundary of the diamagnetic cavity. (3) A closed inner shock terminates the supersonic expansion of the cometary ionosphere. This inner shock is closer to the nucleus on dayside than on the nightside.
SU-D-207-04: GPU-Based 4D Cone-Beam CT Reconstruction Using Adaptive Meshing Method
Zhong, Z; Gu, X; Iyengar, P; Mao, W; Wang, J [UT Southwestern Medical Center, Dallas, TX (United States); Guo, X [University of Texas at Dallas, Richardson, TX (United States)
2015-06-15
Purpose: Due to the limited number of projections at each phase, the image quality of a four-dimensional cone-beam CT (4D-CBCT) is often degraded, which decreases the accuracy of subsequent motion modeling. One of the promising methods is the simultaneous motion estimation and image reconstruction (SMEIR) approach. The objective of this work is to enhance the computational speed of the SMEIR algorithm using adaptive feature-based tetrahedral meshing and GPU-based parallelization. Methods: The first step is to generate the tetrahedral mesh based on the features of a reference phase 4D-CBCT, so that the deformation can be well captured and accurately diffused from the mesh vertices to voxels of the image volume. After the mesh generation, the updated motion model and other phases of 4D-CBCT can be obtained by matching the 4D-CBCT projection images at each phase with the corresponding forward projections of the deformed reference phase of 4D-CBCT. The entire process of this 4D-CBCT reconstruction method is implemented on GPU, resulting in significantly increasing the computational efficiency due to its tremendous parallel computing ability. Results: A 4D XCAT digital phantom was used to test the proposed mesh-based image reconstruction algorithm. The image Result shows both bone structures and inside of the lung are well-preserved and the tumor position can be well captured. Compared to the previous voxel-based CPU implementation of SMEIR, the proposed method is about 157 times faster for reconstructing a 10 -phase 4D-CBCT with dimension 256×256×150. Conclusion: The GPU-based parallel 4D CBCT reconstruction method uses the feature-based mesh for estimating motion model and demonstrates equivalent image Result with previous voxel-based SMEIR approach, with significantly improved computational speed.
Numerical Modelling of Volcanic Ash Settling in Water Using Adaptive Unstructured Meshes
Jacobs, C. T.; Collins, G. S.; Piggott, M. D.; Kramer, S. C.; Wilson, C. R.
2011-12-01
At the bottom of the world's oceans lies layer after layer of ash deposited from past volcanic eruptions. Correct interpretation of these layers can provide important constraints on the duration and frequency of volcanism, but requires a full understanding of the complex multi-phase settling and deposition process. Analogue experiments of tephra settling through a tank of water demonstrate that small ash particles can either settle individually, or collectively as a gravitationally unstable ash-laden plume. These plumes are generated when the concentration of particles exceeds a certain threshold such that the density of the tephra-water mixture is sufficiently large relative to the underlying particle-free water for a gravitational Rayleigh-Taylor instability to develop. These ash-laden plumes are observed to descend as a vertical density current at a velocity much greater than that of single particles, which has important implications for the emplacement of tephra deposits on the seabed. To extend the results of laboratory experiments to large scales and explore the conditions under which vertical density currents may form and persist, we have developed a multi-phase extension to Fluidity, a combined finite element / control volume CFD code that uses adaptive unstructured meshes. As a model validation, we present two- and three-dimensional simulations of tephra plume formation in a water tank that replicate laboratory experiments (Carey, 1997, doi:10.1130/0091-7613(1997)0252.3.CO;2). An inflow boundary condition at the top of the domain allows particles to flux in at a constant rate of 0.472 gm-2s-1, forming a near-surface layer of tephra particles, which initially settle individually at the predicted Stokes velocity of 1.7 mms-1. As more tephra enters the water and the particle concentration increases, the layer eventually becomes unstable and plumes begin to form, descending with velocities more than ten times greater than those of individual particles. The
Mesh Generation and Adaption for High Reynolds Number RANS Computations Project
National Aeronautics and Space Administration — The innovation of our Phase II STTR program is to develop and provide to NASA automatic mesh generation software for the simulation of fluid flows using...
Mesh Generation and Adaption for High Reynolds Number RANS Computations Project
National Aeronautics and Space Administration — This proposal offers to provide NASA with an automatic mesh generator for the simulation of aerodynamic flows using Reynolds-Averages Navier-Stokes (RANS) models....
Pavlidis, D.; Lathouwers, D. [Delft Univ. of Tech., Section Physics of Nuclear Reactors, Delft (Netherlands)
2011-07-01
A computational fluid dynamics model with anisotropic mesh adaptivity is used to investigate coolant flow and heat transfer in pebble bed reactors. A novel method for implicitly incorporating solid boundaries based on multi-fluid flow modelling is adopted. The resulting model is able to resolve and simulate flow and heat transfer in randomly packed beds, regardless of the actual geometry, starting off with arbitrarily coarse meshes. The model is initially evaluated using an orderly stacked square channel of channel-height-to-particle diameter ratio of unity for a range of Reynolds numbers. The model is then applied to the face-centred cubical geometry. Coolant flow and heat transfer patterns are investigated. (author)
Greene, Patrick T.; Schofield, Samuel P.; Nourgaliev, Robert
2017-04-01
A new mesh smoothing method designed to cluster cells near a dynamically evolving interface is presented. The method is based on weighted condition number mesh relaxation with the weight function computed from a level set representation of the interface. The weight function is expressed as a Taylor series based discontinuous Galerkin projection, which makes the computation of the derivatives of the weight function needed during the condition number optimization process a trivial matter. For cases when a level set is not available, a fast method for generating a low-order level set from discrete cell-centered fields, such as a volume fraction or index function, is provided. Results show that the low-order level set works equally well as the actual level set for mesh smoothing. Meshes generated for a number of interface geometries are presented, including cases with multiple level sets. Dynamic cases with moving interfaces show the new method is capable of maintaining a desired resolution near the interface with an acceptable number of relaxation iterations per time step, which demonstrates the method's potential to be used as a mesh relaxer for arbitrary Lagrangian Eulerian (ALE) methods.
Li, Xianping
2010-01-01
Heterogeneous anisotropic diffusion problems arise in the various areas of science and engineering including plasma physics, petroleum engineering, and image processing. Standard numerical methods can produce spurious oscillations when they are used to solve those problems. A common approach to avoid this difficulty is to design a proper numerical scheme and/or a proper mesh so that the numerical solution validates the discrete counterpart (DMP) of the maximum principle satisfied by the continuous solution. A well known mesh condition for the DMP satisfaction by the linear finite element solution of isotropic diffusion problems is the non-obtuse angle condition that requires the dihedral angles of mesh elements to be non-obtuse. In this paper, a generalization of the condition, the so-called anisotropic non-obtuse angle condition, is developed for the finite element solution of heterogeneous anisotropic diffusion problems. The new condition is essentially the same as the existing one except that the dihedral ...
Lipnikov, Konstantin [Los Alamos National Laboratory; Agouzal, Abdellatif [UNIV DE LYON; Vassilevski, Yuri [Los Alamos National Laboratory
2009-01-01
We present a new technology for generating meshes minimizing the interpolation and discretization errors or their gradients. The key element of this methodology is construction of a space metric from edge-based error estimates. For a mesh with N{sub h} triangles, the error is proportional to N{sub h}{sup -1} and the gradient of error is proportional to N{sub h}{sup -1/2} which are optimal asymptotics. The methodology is verified with numerical experiments.
2012-01-22
ICES REPORT 12-05 January 2012 An Isogeometric Design-through-analysis Methodology based on Adaptive Hierarchical Refinement of NURBS , Immersed...M.J. Borden, E. Rank, T.J.R. Hughes, An Isogeometric Design-through-analysis Methodology based on Adaptive Hierarchical Refinement of NURBS , Immersed...analysis Methodology based on Adaptive Hierarchical Refinement of NURBS , Immersed Boundary Methods, and T-spline CAD Surfaces 5a. CONTRACT NUMBER 5b
THE ADAPTIVE FINITE ELEMENT REMESHING IN LARGE DEFORMATION OF METAL FORMING
X.X. Guo; J.S. Liu
2004-01-01
The adaptive remeshing technique for quadrilateral elements consists of modules the trigger of remeshing, the new mesh generation, adaptive refinement and interpolation of field variables. The new adaptive mesh generation is the key problem. First, a coarse mesh is created by using "loop algorithm". Subsequent local mesh adaptive refinement is performed based on effective strain. Finally, a typical example of upsetting is given to test efficient of techniques, from which it is verified that the remeshing algorithm developed here exhibits good performance and has high accuracy.
Greene, Patrick; Schofield, Sam; Nourgaliev, Robert
2016-11-01
A new mesh smoothing method designed to cluster cells near a dynamically evolving interface is presented. The method is based on weighted condition number mesh relaxation with the weight function being computed from a level set representation of the interface. The weight function is expressed as a Taylor series based discontinuous Galerkin (DG) projection, which makes the computation of the derivatives of the weight function needed during the condition number optimization process a trivial matter. For cases when a level set is not available, a fast method for generating a low-order level set from discrete cell-centered fields, such as a volume fraction or index function, is provided. Results show that the low-order level set works equally well for the weight function as the actual level set. The method retains the excellent smoothing capabilities of condition number relaxation, while providing a method for clustering mesh cells near regions of interest. Dynamic cases for moving interfaces are presented to demonstrate the method's potential usefulness as a mesh relaxer for arbitrary Lagrangian Eulerian (ALE) methods. This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
Foks, Nathan Leon
The interpretation of geophysical data plays an important role in the analysis of potential field data in resource exploration industries. Two categories of interpretation techniques are discussed in this thesis; boundary detection and geophysical inversion. Fault or boundary detection is a method to interpret the locations of subsurface boundaries from measured data, while inversion is a computationally intensive method that provides 3D information about subsurface structure. My research focuses on these two aspects of interpretation techniques. First, I develop a method to aid in the interpretation of faults and boundaries from magnetic data. These processes are traditionally carried out using raster grid and image processing techniques. Instead, I use unstructured meshes of triangular facets that can extract inferred boundaries using mesh edges. Next, to address the computational issues of geophysical inversion, I develop an approach to reduce the number of data in a data set. The approach selects the data points according to a user specified proxy for its signal content. The approach is performed in the data domain and requires no modification to existing inversion codes. This technique adds to the existing suite of compressive inversion algorithms. Finally, I develop an algorithm to invert gravity data for an interfacing surface using an unstructured mesh of triangular facets. A pertinent property of unstructured meshes is their flexibility at representing oblique, or arbitrarily oriented structures. This flexibility makes unstructured meshes an ideal candidate for geometry based interface inversions. The approaches I have developed provide a suite of algorithms geared towards large-scale interpretation of potential field data, by using an unstructured representation of both the data and model parameters.
Ramos, Zorangelí; Alegría, Margarita
2014-04-01
Few studies addressing the mental health needs of Latinos describe how interventions are tailored or culturally adapted to address the needs of their target population. Without reference to this process, efforts to replicate results and provide working models of the adaptation process for other researchers are thwarted. The purpose of this article is to describe the process of a cultural adaptation that included accommodations for health literacy of a brief telephone cognitive-behavioral depression intervention for Latinos in low-resource settings. We followed a five-stage approach (i.e., information gathering, preliminary adaptation, preliminary testing, adaptation, and refinement) as described by Barrera, Castro, Strycker, and Toobert (2013) to structure our process. Cultural adaptations included condensation of the sessions, review, and modifications of materials presented to participants including the addition of visual aids, culturally relevant metaphors, values, and proverbs. Feedback from key stakeholders, including clinician and study participants, was fundamental to the adaptation process. Areas for further inquiry and adaptation identified in our process include revisions to the presentation of "cognitive restructuring" to participants and the inclusion of participant beliefs about the cause of their depression. Cultural adaptation is a dynamic process, requiring numerous refinements to ensure that an intervention is tailored and relevant to the target population.
Design of Finite Element Tools for Coupled Surface and Volume Meshes
Daniel K(o)ster; Oliver Kriessl; Kunibert G. Siebert
2008-01-01
Many problems with underlying variational structure involve a coupling of volume with surface effects. A straight-forward approach in a finite element discretization is to make use of the surface triangulation that is naturally induced by the volume triangulation. In an adaptive method one wants to facilitate "matching" local mesh modifications, i.e., local refinement and/or coarsening, of volume and surface mesh with standard tools such that the surface grid is always induced by the volume grid. We describe the concepts behind this approach for bisectional refinement and describe new tools incorporated in the finite element toolbox ALBERTA. We also present several important applications of the mesh coupling.
Faosan Mapa
2014-01-01
Full Text Available Normal 0 false false false IN X-NONE X-NONE MicrosoftInternetExplorer4 /* Style Definitions */ table.MsoNormalTable {mso-style-name:"Table Normal"; mso-tstyle-rowband-size:0; mso-tstyle-colband-size:0; mso-style-noshow:yes; mso-style-priority:99; mso-style-qformat:yes; mso-style-parent:""; mso-padding-alt:0cm 5.4pt 0cm 5.4pt; mso-para-margin:0cm; mso-para-margin-bottom:.0001pt; mso-pagination:widow-orphan; font-size:11.0pt; font-family:"Calibri","sans-serif"; mso-ascii-font-family:Calibri; mso-ascii-theme-font:minor-latin; mso-fareast-font-family:"Times New Roman"; mso-fareast-theme-font:minor-fareast; mso-hansi-font-family:Calibri; mso-hansi-theme-font:minor-latin; mso-bidi-font-family:"Times New Roman"; mso-bidi-theme-font:minor-bidi;} Wireless Mesh Network (WMN adalah suatu konektivitas jaringan yang self-organized, self-configured dan multi-hop. Tujuan dari WMN adalah menawarkan pengguna suatu bentuk jaringan nirkabel yang dapat dengan mudah berkomunikasi dengan jaringan konvensional dengan kecepatan tinggi dan dengan cakupan yang lebih luas serta biaya awal yang minimal. Diperlukan suatu desain protokol routing yang efisien untuk WMN yang secara adaptif dapat mendukung mesh routers dan mesh clients. Dalam tulisan ini, diusulkan untuk mengoptimalkan protokol OLSR, yang merupakan protokol routing proaktif. Digunakan heuristik yang meningkatkan protokol OLSR melalui adaptive refreshing time interval dan memperbaiki metode MPR selecting algorithm. Suatu analisa dalam meningkatkan protokol OLSR melalui adaptive refreshing time interval dan memperbaiki algoritma pemilihan MPR menunjukkan kinerja yang signifikan dalam hal throughput jika dibandingkan dengan protokol OLSR yang asli. Akan tetapi, terdapat kenaikan dalam hal delay. Pada simulasi yang dilakukan dapat disimpulkan bahwa OLSR dapat dioptimalkan dengan memodifikasi pemilihan node MPR berdasarkan cost effective dan penyesuaian waktu interval refreshing hello message sesuai dengan keadaan
Sutthisak Phongthanapanich; Suthee Traivivatana; Parinya Boonmaruth; Pramote Dechaumphai
2006-01-01
Based on flux-based formulation,a nodeless variable element method is developed to analyze two-dimensional steady-state and transient heat transfer problems.The nodeless variable element employs quadratic interpolation functions to provide higher solution accuracy without necessity to actually generate additional nodes.The flux-based formulation is applied to reduce the complexity in deriving the finite element equations as compared to the conventional finite element method.The solution accuracy is further improved by implementing an adaptive meshing technique to generate finite element mesh that can adapt and move along corresponding to the solution behavior.The technique generates small elements in the regions of steep solution gradients to provide accurate solution,and mean while it generates larger elements in the other regions where the solution gradients are slight to reduce the computational time and the computer memory.The effectiveness of the combined procedure is demonstrated by heat transfer problems that have exact solutions.These problems are:(a) a steady-state heat conduction analysis in a square plate subjected to a highly localized surface heating,and (b) a transient heat conduction analysis in a long pate subjected to a moving heat source.
Adaptive Ridge Point Refinement for Seeds Detection in X-Ray Coronary Angiogram
Ruoxiu Xiao
2015-01-01
Full Text Available Seed point is prerequired condition for tracking based method for extracting centerline or vascular structures from the angiogram. In this paper, a novel seed point detection method for coronary artery segmentation is proposed. Vessels on the image are first enhanced according to the distribution of Hessian eigenvalue in multiscale space; consequently, centerlines of tubular vessels are also enhanced. Ridge point is extracted as candidate seed point, which is then refined according to its mathematical definition. The theoretical feasibility of this method is also proven. Finally, all the detected ridge points are checked using a self-adaptive threshold to improve the robustness of results. Clinical angiograms are used to evaluate the performance of the proposed algorithm, and the results show that the proposed algorithm can detect a large set of true seed points located on most branches of vessels. Compared with traditional seed point detection algorithms, the proposed method can detect a larger number of seed points with higher precision. Considering that the proposed method can achieve accurate seed detection without any human interaction, it can be utilized for several clinical applications, such as vessel segmentation, centerline extraction, and topological identification.
殷亚军; 李阳东; 涂志新; 李文; 沈旭; 周建新
2016-01-01
Level Set方法因能有效地处理界面处复杂的拓扑结构变化以及大变形问题，广泛应用于界面追踪领域。在Level Set方法追踪运动界面时引入八叉树网格技术，通过八叉树网格的细化和粗化技术减少计算网格数量和计算内存并提高计算效率和计算精度。因为八叉树网格为非均匀网格，其相邻网格的层数值可能不相同，所以不能直接采用WENO格式离散Level Set函数得到网格处的函数值，进而提出八叉树网格离散模型解决这一问题，并提出基于八叉树网格距离场重新初始化方法减少Level Set方法的质量损失，最后将基于八叉树网格技术的Level Set方法应用于两个给定速度场的运动界面模拟算例以及基准件方腔的铸造充型过程的模拟。模拟结果表明该方法可以提高界面的精度，同时改善质量守恒性。%Level Set method is one of the dominant methods to calculate and track evolution of interfaces because it can efficiently handle complex topology change of interfaces and large deformation. This method often suffers from large amount of numerical dissipation and high order accurate numerical discretization in time and space is used to improve mass conservation properties, which leads to an increase of computational overhead. When octree adaptive mesh technology is incorporated into Level Set to capture moving interfaces, the amount of computational meshes and memory usage is reduced and the calculation precision and efficiency are enhanced by adaptive octree mesh refining or coarsening according to interface distance. Because octree mesh is non-uniform and layer numbers of neighbor grids may be different, the WENO formula could not be directly applied to set values of octree grids. A discretization model and a technique of re-initializing distance field were proposed to calculate value of neighbor octree grids as well as to reduce mass loss in Level Set function. Numerical
Kimura, Satoshi; Candy, Adam S.; Holland, Paul R.; Piggott, Matthew D.; Jenkins, Adrian
2013-07-01
Several different classes of ocean model are capable of representing floating glacial ice shelves. We describe the incorporation of ice shelves into Fluidity-ICOM, a nonhydrostatic finite-element ocean model with the capacity to utilize meshes that are unstructured and adaptive in three dimensions. This geometric flexibility offers several advantages over previous approaches. The model represents melting and freezing on all ice-shelf surfaces including vertical faces, treats the ice shelf topography as continuous rather than stepped, and does not require any smoothing of the ice topography or any of the additional parameterisations of the ocean mixed layer used in isopycnal or z-coordinate models. The model can also represent a water column that decreases to zero thickness at the 'grounding line', where the floating ice shelf is joined to its tributary ice streams. The model is applied to idealised ice-shelf geometries in order to demonstrate these capabilities. In these simple experiments, arbitrarily coarsening the mesh outside the ice-shelf cavity has little effect on the ice-shelf melt rate, while the mesh resolution within the cavity is found to be highly influential. Smoothing the vertical ice front results in faster flow along the smoothed ice front, allowing greater exchange with the ocean than in simulations with a realistic ice front. A vanishing water-column thickness at the grounding line has little effect in the simulations studied. We also investigate the response of ice shelf basal melting to variations in deep water temperature in the presence of salt stratification.
Finite element mesh generation
Lo, Daniel SH
2014-01-01
Highlights the Progression of Meshing Technologies and Their ApplicationsFinite Element Mesh Generation provides a concise and comprehensive guide to the application of finite element mesh generation over 2D domains, curved surfaces, and 3D space. Organised according to the geometry and dimension of the problem domains, it develops from the basic meshing algorithms to the most advanced schemes to deal with problems with specific requirements such as boundary conformity, adaptive and anisotropic elements, shape qualities, and mesh optimization. It sets out the fundamentals of popular techniques
Bode, P; Bode, Paul; Ostriker, Jeremiah P.
2003-01-01
An improved implementation of an N-body code for simulating collisionless cosmological dynamics is presented. TPM (Tree-Particle-Mesh) combines the PM method on large scales with a tree code to handle particle-particle interactions at small separations. After the global PM forces are calculated, spatially distinct regions above a given density contrast are located; the tree code calculates the gravitational interactions inside these denser objects at higher spatial and temporal resolution. The new implementation includes individual particle time steps within trees, an improved treatment of tidal forces on trees, new criteria for higher force resolution and choice of time step, and parallel treatment of large trees. TPM is compared to P^3M and a tree code (GADGET) and is found to give equivalent results in significantly less time. The implementation is highly portable (requiring a Fortran compiler and MPI) and efficient on parallel machines. The source code can be found at http://astro.princeton.edu/~bode/TPM/
An accuracy assessment of Cartesian-mesh approaches for the Euler equations
Coirier, William J.; Powell, Kenneth G.
1995-01-01
A critical assessment of the accuracy of Cartesian-mesh approaches for steady, transonic solutions of the Euler equations of gas dynamics is made. An exact solution of the Euler equations (Ringleb's flow) is used not only to infer the order of the truncation error of the Cartesian-mesh approaches, but also to compare the magnitude of the discrete error directly to that obtained with a structured mesh approach. Uniformly and adaptively refined solutions using a Cartesian-mesh approach are obtained and compared to each other and to uniformly refined structured mesh results. The effect of cell merging is investigated as well as the use of two different K-exact reconstruction procedures. The solution methodology of the schemes is explained and tabulated results are presented to compare the solution accuracies.
Surface meshing with curvature convergence
Li, Huibin
2014-06-01
Surface meshing plays a fundamental role in graphics and visualization. Many geometric processing tasks involve solving geometric PDEs on meshes. The numerical stability, convergence rates and approximation errors are largely determined by the mesh qualities. In practice, Delaunay refinement algorithms offer satisfactory solutions to high quality mesh generations. The theoretical proofs for volume based and surface based Delaunay refinement algorithms have been established, but those for conformal parameterization based ones remain wide open. This work focuses on the curvature measure convergence for the conformal parameterization based Delaunay refinement algorithms. Given a metric surface, the proposed approach triangulates its conformal uniformization domain by the planar Delaunay refinement algorithms, and produces a high quality mesh. We give explicit estimates for the Hausdorff distance, the normal deviation, and the differences in curvature measures between the surface and the mesh. In contrast to the conventional results based on volumetric Delaunay refinement, our stronger estimates are independent of the mesh structure and directly guarantee the convergence of curvature measures. Meanwhile, our result on Gaussian curvature measure is intrinsic to the Riemannian metric and independent of the embedding. In practice, our meshing algorithm is much easier to implement and much more efficient. The experimental results verified our theoretical results and demonstrated the efficiency of the meshing algorithm. © 2014 IEEE.
Hydrodynamic simulations on a moving Voronoi mesh
Springel, Volker
2011-01-01
At the heart of any method for computational fluid dynamics lies the question of how the simulated fluid should be discretized. Traditionally, a fixed Eulerian mesh is often employed for this purpose, which in modern schemes may also be adaptively refined during a calculation. Particle-based methods on the other hand discretize the mass instead of the volume, yielding an approximately Lagrangian approach. It is also possible to achieve Lagrangian behavior in mesh-based methods if the mesh is allowed to move with the flow. However, such approaches have often been fraught with substantial problems related to the development of irregularity in the mesh topology. Here we describe a novel scheme that eliminates these weaknesses. It is based on a moving unstructured mesh defined by the Voronoi tessellation of a set of discrete points. The mesh is used to solve the hyperbolic conservation laws of ideal hydrodynamics with a finite volume approach, based on a second-order Godunov scheme with an exact Riemann solver. A...
Zhang, Hong
2016-01-01
Motivated by observations of saturation overshoot, this paper investigates numerical modeling of two-phase flow incorporating dynamic capillary pressure. The effects of the dynamic capillary coefficient, the infiltrating flux rate and the initial and boundary values are systematically studied using a travelling wave ansatz and efficient numerical methods. The travelling wave solutions may exhibit monotonic, non-monotonic or plateau-shaped behaviour. Special attention is paid to the non-monotonic profiles. The travelling wave results are confirmed by numerically solving the partial differential equation using an accurate adaptive moving mesh solver. Comparisons between the computed solutions using the Brooks-Corey model and the laboratory measurements of saturation overshoot verify the effectiveness of our approach.
2013-03-01
applications [3, 21, 22], including antenna, microwave circuits , geophysics, optics, etc. The Ground Penetrating Radar (GPR) is a popular and...Räisänen. An efficient FDTD algorithm for the analysis of microstrip patch antennas printed on a general anisotropic dielectric substrate. IEEE...45, 1995. [17] S. Gedney, F. Lansing, and D. Rascoe. Full wave analysis of microwave monolithic circuit devices using a generalized Yee-algorithm
2013-01-01
ξi be the Legendre- Gauss -Lobatto (LGL) points defined as the roots of (1 − ξ2)P ′N (ξ) = 0, where PN (ξ) is the N th order Legendre polynomial. The...techniques in grid generation, data structures, and numerical operations with applications, Springer, 2006. [8] G. W. Ley , R. L. Elsberry, Forecasts of typhoon
轨迹优化的 LASSO 网格自适应加密方法%LASSO-based node adaptive refinement in trajectory optimization
张松; 侯明善
2016-01-01
A least absolute shrinkage and selection operator (LASSO)based node adaptive refinement ap-proach for the direct method in trajectory optimization is proposed.Firstly,a higher resolution grid and its asso-ciated radial basis function set are created.Sequentially,the control variables are approximated using the resul-ting radial basis function set,and its sampling sequence is generated by interpolation.Finally,the coefficients of the formulated approximation function are estimated based on the statistical variable selection method-LASSO. The higher multi-resolution nodes associated with radial basis functions with non-zero coefficient are selected as new nodes.The proposed method refines the mesh without estimation of states and/or errors controls,and few extra parameters are involved.Therefore,the formulated trajectory optimization algorithm behaves strong adapt-ability and generality.The validity of this method is demonstrated by several typical examples.%针对轨迹优化直接方法，提出了以控制变量曲率为基础的最小绝对收缩与选择算子（least absolute shrinkage and selection operator，LASSO）网格自适应加密策略，用于提高优化精度。以高分辨率二分网格节点为中心，构造径向基函数逼近控制曲线，利用 LASSO 方法估计径向基函数系数，并自动筛选出位于控制曲线曲率极大区间的高分辨率节点加密当前网格。本文方法不需要进行状态和控制误差估计，适应性和通用性强。两组典型算例验证了方法的有效性。
Robust, multidimensional mesh motion based on Monge-Kantorovich equidistribution
Delzanno, G L [Los Alamos National Laboratory; Finn, J M [Los Alamos National Laboratory
2009-01-01
Mesh-motion (r-refinement) grid adaptivity schemes are attractive due to their potential to minimize the numerical error for a prescribed number of degrees of freedom. However, a key roadblock to a widespread deployment of the technique has been the formulation of robust, reliable mesh motion governing principles, which (1) guarantee a solution in multiple dimensions (2D and 3D), (2) avoid grid tangling (or folding of the mesh, whereby edges of a grid cell cross somewhere in the domain), and (3) can be solved effectively and efficiently. In this study, we formulate such a mesh-motion governing principle, based on volume equidistribution via Monge-Kantorovich optimization (MK). In earlier publications [1, 2], the advantages of this approach in regards to these points have been demonstrated for the time-independent case. In this study, demonstrate that Monge-Kantorovich equidistribution can in fact be used effectively in a time stepping context, and delivers an elegant solution to the otherwise pervasive problem of grid tangling in mesh motion approaches, without resorting to ad-hoc time-dependent terms (as in moving-mesh PDEs, or MMPDEs [3, 4]). We explore two distinct r-refinement implementations of MK: direct, where the current mesh relates to an initial, unchanging mesh, and sequential, where the current mesh is related to the previous one in time. We demonstrate that the direct approach is superior in regards to mesh distortion and robustness. The properties of the approach are illustrated with a paradigmatic hyperbolic PDE, the advection of a passive scalar. Imposed velocity flow fields or varying vorticity levels and flow shears are considered.
Hundebøll, Martin; Pedersen, Morten Videbæk; Roetter, Daniel Enrique Lucani
2014-01-01
This work studies the potential and impact of the FRANC network coding protocol for delivering high quality Dynamic Adaptive Streaming over HTTP (DASH) in wireless networks. Although DASH aims to tailor the video quality rate based on the available throughput to the destination, it relies...
Hybrid direct and iterative solvers for h refined grids with singularities
Paszyński, Maciej R.
2015-04-27
This paper describes a hybrid direct and iterative solver for two and three dimensional h adaptive grids with point singularities. The point singularities are eliminated by using a sequential linear computational cost solver O(N) on CPU [1]. The remaining Schur complements are submitted to incomplete LU preconditioned conjugated gradient (ILUPCG) iterative solver. The approach is compared to the standard algorithm performing static condensation over the entire mesh and executing the ILUPCG algorithm on top of it. The hybrid solver is applied for two or three dimensional grids automatically h refined towards point or edge singularities. The automatic refinement is based on the relative error estimations between the coarse and fine mesh solutions [2], and the optimal refinements are selected using the projection based interpolation. The computational mesh is partitioned into sub-meshes with local point and edge singularities separated. This is done by using the following greedy algorithm.
Medical case-based retrieval: integrating query MeSH terms for query-adaptive multi-modal fusion
Seco de Herrera, Alba G.; Foncubierta-Rodríguez, Antonio; Müller, Henning
2015-03-01
Advances in medical knowledge give clinicians more objective information for a diagnosis. Therefore, there is an increasing need for bibliographic search engines that can provide services helping to facilitate faster information search. The ImageCLEFmed benchmark proposes a medical case-based retrieval task. This task aims at retrieving articles from the biomedical literature that are relevant for differential diagnosis of query cases including a textual description and several images. In the context of this campaign many approaches have been investigated showing that the fusion of visual and text information can improve the precision of the retrieval. However, fusion does not always lead to better results. In this paper, a new query-adaptive fusion criterion to decide when to use multi-modal (text and visual) or only text approaches is presented. The proposed method integrates text information contained in MeSH (Medical Subject Headings) terms extracted and visual features of the images to find synonym relations between them. Given a text query, the query-adaptive fusion criterion decides when it is suitable to also use visual information for the retrieval. Results show that this approach can decide if a text or multi{modal approach should be used with 77.15% of accuracy.
E pur si muove: Galiliean-invariant cosmological hydrodynamical simulations on a moving mesh
Springel, Volker
2009-01-01
Hydrodynamic cosmological simulations at present usually employ either the Lagrangian SPH technique, or Eulerian hydrodynamics on a Cartesian mesh with adaptive mesh refinement. Both of these methods have disadvantages that negatively impact their accuracy in certain situations. We here propose a novel scheme which largely eliminates these weaknesses. It is based on a moving unstructured mesh defined by the Voronoi tessellation of a set of discrete points. The mesh is used to solve the hyperbolic conservation laws of ideal hydrodynamics with a finite volume approach, based on a second-order unsplit Godunov scheme with an exact Riemann solver. The mesh-generating points can in principle be moved arbitrarily. If they are chosen to be stationary, the scheme is equivalent to an ordinary Eulerian method with second order accuracy. If they instead move with the velocity of the local flow, one obtains a Lagrangian formulation of hydrodynamics that does not suffer from the mesh distortion limitations inherent in othe...
Adaptive Finite Volume Method for the Shallow Water Equations on Triangular Grids
Sudi Mungkasi
2016-01-01
Full Text Available This paper presents a numerical entropy production (NEP scheme for two-dimensional shallow water equations on unstructured triangular grids. We implement NEP as the error indicator for adaptive mesh refinement or coarsening in solving the shallow water equations using a finite volume method. Numerical simulations show that NEP is successful to be a refinement/coarsening indicator in the adaptive mesh finite volume method, as the method refines the mesh or grids around nonsmooth regions and coarsens them around smooth regions.
Solution adaptive grids applied to low Reynolds number flow
de With, G.; Holdø, A. E.; Huld, T. A.
2003-08-01
A numerical study has been undertaken to investigate the use of a solution adaptive grid for flow around a cylinder in the laminar flow regime. The main purpose of this work is twofold. The first aim is to investigate the suitability of a grid adaptation algorithm and the reduction in mesh size that can be obtained. Secondly, the uniform asymmetric flow structures are ideal to validate the mesh structures due to mesh refinement and consequently the selected refinement criteria. The refinement variable used in this work is a product of the rate of strain and the mesh cell size, and contains two variables Cm and Cstr which determine the order of each term. By altering the order of either one of these terms the refinement behaviour can be modified.
2015-01-01
Mesh generation and visualization software based on the CGAL library. Folder content: drawmesh Visualize slices of the mesh (surface/volumetric) as wireframe on top of an image (3D). drawsurf Visualize surfaces of the mesh (surface/volumetric). img2mesh Convert isosurface in image to volumetric...
Sarkis, C.; Silva, L.; Gandin, Ch-A.; Plapp, M.
2016-03-01
Dendritic growth is computed with automatic adaptation of an anisotropic and unstructured finite element mesh. The energy conservation equation is formulated for solid and liquid phases considering an interface balance that includes the Gibbs-Thomson effect. An equation for a diffuse interface is also developed by considering a phase field function with constant negative value in the liquid and constant positive value in the solid. Unknowns are the phase field function and a dimensionless temperature, as proposed by [1]. Linear finite element interpolation is used for both variables, and discretization stabilization techniques ensure convergence towards a correct non-oscillating solution. In order to perform quantitative computations of dendritic growth on a large domain, two additional numerical ingredients are necessary: automatic anisotropic unstructured adaptive meshing [2,[3] and parallel implementations [4], both made available with the numerical platform used (CimLib) based on C++ developments. Mesh adaptation is found to greatly reduce the number of degrees of freedom. Results of phase field simulations for dendritic solidification of a pure material in two and three dimensions are shown and compared with reference work [1]. Discussion on algorithm details and the CPU time will be outlined.
Neale, Richard B. [University Corporation For Atmospheric Research, Boulder, CO (United States)
2015-12-01
In this project we analyze climate simulations using the Community Earth System Model (CESM) in order to determine the modeled response and sensitivity to horizontal resolution. Simple aqua-planet configurations were used to provide a clean comparison of the response to resolution in CESM. This enables us to easily examine all aspects of the model sensitivity to resolution including mean quantities, variability and physical parameterization tendencies: the chief reflection of resolution sensitivity. An extension to the global resolution sensitivity study is the examination of regional grid refinement where resolution changes are prescribed in a single global simulation. We examine the relevance of the global resolution sensitivity results as applied to these regional refinement simulations. In particular we examine how variations in the grid resolution, centered on different parts of the globe, lead to differences in the parameterized response and the potential to generate residual circulations as a result. Given the potential to generate this resolution sensitivity we examine simple modifications to the parameterized physics that are able to moderate any residual circulations. Finally, we transfer the framework to the standard AMIP configuration to examine the resolution sensitivity in the presence of compounding effects such as land-sea distributions, orography and seasonal variation.
Algorithm refinement for fluctuating hydrodynamics
Williams, Sarah A.; Bell, John B.; Garcia, Alejandro L.
2007-07-03
This paper introduces an adaptive mesh and algorithmrefinement method for fluctuating hydrodynamics. This particle-continuumhybrid simulates the dynamics of a compressible fluid with thermalfluctuations. The particle algorithm is direct simulation Monte Carlo(DSMC), a molecular-level scheme based on the Boltzmann equation. Thecontinuum algorithm is based on the Landau-Lifshitz Navier-Stokes (LLNS)equations, which incorporate thermal fluctuations into macroscopichydrodynamics by using stochastic fluxes. It uses a recently-developedsolver for LLNS, based on third-order Runge-Kutta. We present numericaltests of systems in and out of equilibrium, including time-dependentsystems, and demonstrate dynamic adaptive refinement by the computationof a moving shock wave. Mean system behavior and second moment statisticsof our simulations match theoretical values and benchmarks well. We findthat particular attention should be paid to the spectrum of the flux atthe interface between the particle and continuum methods, specificallyfor the non-hydrodynamic (kinetic) time scales.
Skutnik, Steven E., E-mail: sskutnik@utk.edu; Davis, David R.
2016-05-01
The use of passive gamma and neutron signatures from fission indicators is a common means of estimating used fuel burnup, enrichment, and cooling time. However, while characteristic fission product signatures such as {sup 134}Cs, {sup 137}Cs, {sup 154}Eu, and others are generally reliable estimators for used fuel burnup within the context where the assembly initial enrichment and the discharge time are known, in the absence of initial enrichment and/or cooling time information (such as when applying NDA measurements in a safeguards/verification context), these fission product indicators no longer yield a unique solution for assembly enrichment, burnup, and cooling time after discharge. Through the use of a new Mesh-Adaptive Direct Search (MADS) algorithm, it is possible to directly probe the shape of this “degeneracy space” characteristic of individual nuclides (and combinations thereof), both as a function of constrained parameters (such as the assembly irradiation history) and unconstrained parameters (e.g., the cooling time before measurement and the measurement precision for particular indicator nuclides). In doing so, this affords the identification of potential means of narrowing the uncertainty space of potential assembly enrichment, burnup, and cooling time combinations, thereby bounding estimates of assembly plutonium content. In particular, combinations of gamma-emitting nuclides with distinct half-lives (e.g., {sup 134}Cs with {sup 137}Cs and {sup 154}Eu) in conjunction with gross neutron counting (via {sup 244}Cm) are able to reasonably constrain the degeneracy space of possible solutions to a space small enough to perform useful discrimination and verification of fuel assemblies based on their irradiation history.
Maltby, John; Day, Liz; Hall, Sophie
2015-01-01
The current paper presents a new measure of trait resilience derived from three common mechanisms identified in ecological theory: Engineering, Ecological and Adaptive (EEA) resilience. Exploratory and confirmatory factor analyses of five existing resilience scales suggest that the three trait resilience facets emerge, and can be reduced to a 12-item scale. The conceptualization and value of EEA resilience within the wider trait and well-being psychology is illustrated in terms of differing relationships with adaptive expressions of the traits of the five-factor personality model and the contribution to well-being after controlling for personality and coping, or over time. The current findings suggest that EEA resilience is a useful and parsimonious model and measure of trait resilience that can readily be placed within wider trait psychology and that is found to contribute to individual well-being.
Maltby, John; Day, Liz; Hall, Sophie
2015-01-01
The current paper presents a new measure of trait resilience derived from three common mechanisms identified in ecological theory: Engineering, Ecological and Adaptive (EEA) resilience. Exploratory and confirmatory factor analyses of five existing resilience scales suggest that the three trait resilience facets emerge, and can be reduced to a 12-item scale. The conceptualization and value of EEA resilience within the wider trait and well-being psychology is illustrated in terms of differing relationships with adaptive expressions of the traits of the five-factor personality model and the contribution to well-being after controlling for personality and coping, or over time. The current findings suggest that EEA resilience is a useful and parsimonious model and measure of trait resilience that can readily be placed within wider trait psychology and that is found to contribute to individual well-being. PMID:26132197
Determination of an Initial Mesh Density for Finite Element Computations via Data Mining
Kanapady, R; Bathina, S K; Tamma, K K; Kamath, C; Kumar, V
2001-07-23
Numerical analysis software packages which employ a coarse first mesh or an inadequate initial mesh need to undergo a cumbersome and time consuming mesh refinement studies to obtain solutions with acceptable accuracy. Hence, it is critical for numerical methods such as finite element analysis to be able to determine a good initial mesh density for the subsequent finite element computations or as an input to a subsequent adaptive mesh generator. This paper explores the use of data mining techniques for obtaining an initial approximate finite element density that avoids significant trial and error to start finite element computations. As an illustration of proof of concept, a square plate which is simply supported at its edges and is subjected to a concentrated load is employed for the test case. Although simplistic, the present study provides insight into addressing the above considerations.
John Maltby; Liz Day; Sophie Hall
2015-01-01
The current paper presents a new measure of trait resilience derived from three common mechanisms identified in ecological theory: Engineering, Ecological and Adaptive (EEA) resilience. Exploratory and confirmatory factor analyses of five existing resilience scales suggest that the three trait resilience facets emerge, and can be reduced to a 12-item scale. The conceptualization and value of EEA resilience within the wider trait and well-being psychology is illustrated in terms of differing r...
Gill, S P D; Gibson, B K; Flynn, C; Ibata, R A; Lewis, G F; Gill, Stuart P.D.; Knebe, Alexander; Gibson, Brad K.; Flynn, Chris; Ibata, Rodrigo A.; Lewis, Geraint F.
2002-01-01
An adaptive multi grid approach to simulating the formation of structure from collisionless dark matter is described. MLAPM (Multi-Level Adaptive Particle Mesh) is one of the most efficient serial codes available on the cosmological 'market' today. As part of Swinburne University's role in the development of the Square Kilometer Array, we are implementing hydrodynamics, feedback, and radiative transfer within the MLAPM adaptive mesh, in order to simulate baryonic processes relevant to the interstellar and intergalactic media at high redshift. We will outline our progress to date in applying the existing MLAPM to a study of the decay of satellite galaxies within massive host potentials.
Gallardo-Pujol, David; Kramp, Uwe; García-Forero, Carlos; Pérez-Ramírez, Meritxell; Andrés-Pueyo, Antonio
2006-10-01
The assessment of aggressiveness and the prediction of aggression has become a relevant research and applied topic in Psychiatry and Psychology. There have been many attempts in order to get a fast and reliable tool to measure aggression. Buss and Durkee started the pathway, and recently Bryant and Smith developed a tool with an enormous potential, a fast-applicable, reliable and valid test. We herein report a Spanish adaptation of this test and we show that aggressiveness can be measured rapidly, and in a simple, valid and reliable way across different populations. We focus on the discriminant capacity of this test to detect aggressive individuals.
Type 2 immunity and wound healing: evolutionary refinement of adaptive immunity by helminths
Gause, William C.; Wynn, Thomas A.; Allen, Judith E.
2013-01-01
Helminth-induced type 2 immune responses, which are characterized by the T helper 2 cell-associated cytokines interleukin-4 (IL-4) and IL-13, mediate host protection through enhanced tissue repair, the control of inflammation and worm expulsion. In this Opinion article, we consider type 2 immunity in the context of helminth-mediated tissue damage. We examine the relationship between the control of helminth infection and the mechanisms of wound repair, and we provide a new understanding of the adaptive type 2 immune response and its contribution to both host tolerance and resistance. PMID:23827958
Model refinements of transformers via a subproblem finite element method
Dular, Patrick; Kuo-Peng, Patrick; Ferreira Da Luz, Mauricio,; Krähenbühl, Laurent
2015-01-01
International audience; A progressive modeling of transformers is performed via a subproblem finite element method. A complete problem is split into subproblems with different adapted overlapping meshes. Model refinements are performed from ideal to real flux tubes, 1-D to 2-D to 3-D models, linear to nonlinear materials, perfect to real materials, single wire to volume conductor windings, and homogenized to fine models of cores and coils, with any coupling of these changes. The proposed unif...
Error Estimation and h-Adaptivity for Optimal Finite Element Analysis
Cwik, Tom; Lou, John
1997-01-01
The objective of adaptive meshing and automatic error control in finite element analysis is to eliminate the need for the application engineer from re-meshing and re-running design simulations to verify numerical accuracy. The user should only need to enter the component geometry and a coarse finite element mesh. The software will then autonomously and adaptively refine this mesh where needed, reducing the error in the fields to a user prescribed value. The ideal end result of the simulation is a measurable quantity (e.g. scattered field, input impedance), calculated to a prescribed error, in less time and less machine memory than if the user applied typical uniform mesh refinement by hand. It would also allow for the simulation of larger objects since an optimal mesh is created.
The finite cell method for polygonal meshes: poly-FCM
Duczek, Sascha; Gabbert, Ulrich
2016-10-01
In the current article, we extend the two-dimensional version of the finite cell method (FCM), which has so far only been used for structured quadrilateral meshes, to unstructured polygonal discretizations. Therefore, the adaptive quadtree-based numerical integration technique is reformulated and the notion of generalized barycentric coordinates is introduced. We show that the resulting polygonal (poly-)FCM approach retains the optimal rates of convergence if and only if the geometry of the structure is adequately resolved. The main advantage of the proposed method is that it inherits the ability of polygonal finite elements for local mesh refinement and for the construction of transition elements (e.g. conforming quadtree meshes without hanging nodes). These properties along with the performance of the poly-FCM are illustrated by means of several benchmark problems for both static and dynamic cases.
Davis, M. W.
1984-01-01
A Real-Time Self-Adaptive (RTSA) active vibration controller was used as the framework in developing a computer program for a generic controller that can be used to alleviate helicopter vibration. Based upon on-line identification of system parameters, the generic controller minimizes vibration in the fuselage by closed-loop implementation of higher harmonic control in the main rotor system. The new generic controller incorporates a set of improved algorithms that gives the capability to readily define many different configurations by selecting one of three different controller types (deterministic, cautious, and dual), one of two linear system models (local and global), and one or more of several methods of applying limits on control inputs (external and/or internal limits on higher harmonic pitch amplitude and rate). A helicopter rotor simulation analysis was used to evaluate the algorithms associated with the alternative controller types as applied to the four-bladed H-34 rotor mounted on the NASA Ames Rotor Test Apparatus (RTA) which represents the fuselage. After proper tuning all three controllers provide more effective vibration reduction and converge more quickly and smoothly with smaller control inputs than the initial RTSA controller (deterministic with external pitch-rate limiting). It is demonstrated that internal limiting of the control inputs a significantly improves the overall performance of the deterministic controller.
无线Mesh网络中的自适应队列调度算法研究%Research on Adaptive Queue Scheduling Algorithm in Wireless Mesh Networks
夏汉铸; 王志刚
2014-01-01
针对无线mesh网络的网络特性，分析了无线网络中的队列调度算法，提出了一种自适应的队列调度算法AQSM，详细讨论了该算法的具体实现过程及参数变化规则，通过仿真验证了该算法在提高网络性能的同时还可以实现对不同业务流的业务区分。%Wireless Mesh networks (WMNs) have emerged as a key technology for next-generation wireless networking. Queue scheduling algorithm is an important research area in wireless mesh network. In order to improve the performance of WMN and realize DiffServ between different traffic flows, an adaptive queue scheduling algorithm(AQSM)is presented, and the specific implementation process and parameter variation rules are discussed in detail. The analysis and simulation results show that the AQSM algorithm can increase the performance of the wireless Mesh networks and realize the service division of different service flows.
Grayver, Alexander V.
2015-07-01
This paper presents a distributed magnetotelluric inversion scheme based on adaptive finite-element method (FEM). The key novel aspect of the introduced algorithm is the use of automatic mesh refinement techniques for both forward and inverse modelling. These techniques alleviate tedious and subjective procedure of choosing a suitable model parametrization. To avoid overparametrization, meshes for forward and inverse problems were decoupled. For calculation of accurate electromagnetic (EM) responses, automatic mesh refinement algorithm based on a goal-oriented error estimator has been adopted. For further efficiency gain, EM fields for each frequency were calculated using independent meshes in order to account for substantially different spatial behaviour of the fields over a wide range of frequencies. An automatic approach for efficient initial mesh design in inverse problems based on linearized model resolution matrix was developed. To make this algorithm suitable for large-scale problems, it was proposed to use a low-rank approximation of the linearized model resolution matrix. In order to fill a gap between initial and true model complexities and resolve emerging 3-D structures better, an algorithm for adaptive inverse mesh refinement was derived. Within this algorithm, spatial variations of the imaged parameter are calculated and mesh is refined in the neighborhoods of points with the largest variations. A series of numerical tests were performed to demonstrate the utility of the presented algorithms. Adaptive mesh refinement based on the model resolution estimates provides an efficient tool to derive initial meshes which account for arbitrary survey layouts, data types, frequency content and measurement uncertainties. Furthermore, the algorithm is capable to deliver meshes suitable to resolve features on multiple scales while keeping number of unknowns low. However, such meshes exhibit dependency on an initial model guess. Additionally, it is demonstrated
An arbitrary boundary triangle mesh generation method for multi-modality imaging
Zhang, Xuanxuan; Deng, Yong; Gong, Hui; Meng, Yuanzheng; Yang, Xiaoquan; Luo, Qingming
2012-03-01
Low-resolution and ill-posedness are the major challenges in diffuse optical tomography(DOT)/fluorescence molecular tomography(FMT). Recently, the multi-modality imaging technology that combines micro-computed tomography (micro-CT) with DOT/FMT is developed to improve resolution and ill-posedness. To take advantage of the fine priori anatomical maps obtained from micro-CT, we present an arbitrary boundary triangle mesh generation method for FMT/DOT/micro-CT multi-modality imaging. A planar straight line graph (PSLG) based on the image of micro-CT is obtained by an adaptive boundary sampling algorithm. The subregions of mesh are accurately matched with anatomical structures by a two-step solution, firstly, the triangles and nodes during mesh refinement are labeled respectively, and then a revising algorithm is used to modifying meshes of each subregion. The triangle meshes based on a regular model and a micro-CT image are generated respectively. The results show that the subregions of triangle meshes can match with anatomical structures accurately and triangle meshes have good quality. This provides an arbitrary boundaries triangle mesh generation method with the ability to incorporate the fine priori anatomical information into DOT/FMT reconstructions.
Bogdanov, P. B.; Gorobets, A. V.; Sukov, S. A.
2013-08-01
The design of efficient algorithms for large-scale gas dynamics computations with hybrid (heterogeneous) computing systems whose high performance relies on massively parallel accelerators is addressed. A high-order accurate finite volume algorithm with polynomial reconstruction on unstructured hybrid meshes is used to compute compressible gas flows in domains of complex geometry. The basic operations of the algorithm are implemented in detail for massively parallel accelerators, including AMD and NVIDIA graphics processing units (GPUs). Major optimization approaches and a computation transfer technique are covered. The underlying programming tool is the Open Computing Language (OpenCL) standard, which performs on accelerators of various architectures, both existing and emerging.
Georg eLayher
2014-12-01
Full Text Available The categorization of real world objects is often reflected in the similarity of their visual appearances. Such categories of objects do not necessarily form disjunct sets of objects, neither semantically nor visually. The relationship between categories can often be described in terms of a hierarchical structure. For instance, tigers and leopards build two separate mammalian categories, but both belong to the category of felines. In other words, tigers and leopards are subcategories of the category Felidae. In the last decades, the unsupervised learning of categories of visual input stimuli has been addressed by numerous approaches in machine learning as well as in the computational neurosciences. However, the question of what kind of mechanisms might be involved in the process of subcategory learning, or category refinement, remains a topic of active investigation. We propose a recurrent computational network architecture for the unsupervised learning of categorial and subcategorial visual input representations. During learning, the connection strengths of bottom-up weights from input to higher-level category representations are adapted according to the input activity distribution. In a similar manner, top-down weights learn to encode the characteristics of a specific stimulus category. Feedforward and feedback learning in combination realize an associative memory mechanism, enabling the selective top-down propagation of a category's feedback weight distribution. We suggest that the difference between the expected input encoded in the projective field of a category node and the current input pattern controls the amplification of feedforward-driven representations. Large enough differences trigger the recruitment of new representational resources and the establishment of (sub- category representations. We demonstrate the temporal evolution of such learning and show how the approach successully establishes category and subcategory
1994-01-01
element approximations of singular solutions with quasiuniform meshes the pollution error may be significant (depending on the strength of the singu...3 4. For singular solutions computed using quasi-uniform meshes, the pollution error may be significant and ii%-superconvergence regions may not...superconvergence for singular solutions : L-shaped domain. 3 Fig. 24. Pollution effect and 71%-superconvergence for singular solutions : L-shaped domain meshed
toolkit computational mesh conceptual model.
Baur, David G.; Edwards, Harold Carter; Cochran, William K.; Williams, Alan B.; Sjaardema, Gregory D.
2010-03-01
The Sierra Toolkit computational mesh is a software library intended to support massively parallel multi-physics computations on dynamically changing unstructured meshes. This domain of intended use is inherently complex due to distributed memory parallelism, parallel scalability, heterogeneity of physics, heterogeneous discretization of an unstructured mesh, and runtime adaptation of the mesh. Management of this inherent complexity begins with a conceptual analysis and modeling of this domain of intended use; i.e., development of a domain model. The Sierra Toolkit computational mesh software library is designed and implemented based upon this domain model. Software developers using, maintaining, or extending the Sierra Toolkit computational mesh library must be familiar with the concepts/domain model presented in this report.
Venkatachari, Balaji Shankar; Chang, Chau-Lyan
2016-11-01
The focus of this study is scale-resolving simulations of the canonical normal shock- isotropic turbulence interaction using unstructured tetrahedral meshes and the space-time conservation element solution element (CESE) method. Despite decades of development in unstructured mesh methods and its potential benefits of ease of mesh generation around complex geometries and mesh adaptation, direct numerical or large-eddy simulations of turbulent flows are predominantly carried out using structured hexahedral meshes. This is due to the lack of consistent multi-dimensional numerical formulations in conventional schemes for unstructured meshes that can resolve multiple physical scales and flow discontinuities simultaneously. The CESE method - due to its Riemann-solver-free shock capturing capabilities, non-dissipative baseline schemes, and flux conservation in time as well as space - has the potential to accurately simulate turbulent flows using tetrahedral meshes. As part of the study, various regimes of the shock-turbulence interaction (wrinkled and broken shock regimes) will be investigated along with a study on how adaptive refinement of tetrahedral meshes benefits this problem. The research funding for this paper has been provided by Revolutionary Computational Aerosciences (RCA) subproject under the NASA Transformative Aeronautics Concepts Program (TACP).
A tetrahedral mesh generation approach for 3D marine controlled-source electromagnetic modeling
Um, Evan Schankee; Kim, Seung-Sep; Fu, Haohuan
2017-03-01
3D finite-element (FE) mesh generation is a major hurdle for marine controlled-source electromagnetic (CSEM) modeling. In this paper, we present a FE discretization operator (FEDO) that automatically converts a 3D finite-difference (FD) model into reliable and efficient tetrahedral FE meshes for CSEM modeling. FEDO sets up wireframes of a background seabed model that precisely honors the seafloor topography. The wireframes are then partitioned into multiple regions. Outer regions of the wireframes are discretized with coarse tetrahedral elements whose maximum size is as large as a skin depth of the regions. We demonstrate that such coarse meshes can produce accurate FE solutions because numerical dispersion errors of tetrahedral meshes do not accumulate but oscillates. In contrast, central regions of the wireframes are discretized with fine tetrahedral elements to describe complex geology in detail. The conductivity distribution is mapped from FD to FE meshes in a volume-averaged sense. To avoid excessive mesh refinement around receivers, we introduce an effective receiver size. Major advantages of FEDO are summarized as follow. First, FEDO automatically generates reliable and economic tetrahedral FE meshes without adaptive meshing or interactive CAD workflows. Second, FEDO produces FE meshes that precisely honor the boundaries of the seafloor topography. Third, FEDO derives multiple sets of FE meshes from a given FD model. Each FE mesh is optimized for a different set of sources and receivers and is fed to a subgroup of processors on a parallel computer. This divide and conquer approach improves the parallel scalability of the FE solution. Both accuracy and effectiveness of FEDO are demonstrated with various CSEM examples.
Schroeder, Philipp W
2016-01-01
This paper presents heavily grad-div and pressure jump stabilised, equal- and mixed-order discontinuous Galerkin finite element methods for non-isothermal incompressible flows based on the Oberbeck-Boussinesq approximation. In this framework, the enthalpy-porosity model for multiphase flow in melting and solidification problems can be employed. By considering the differentially heated cavity and the melting of pure gallium in a rectangular enclosure, it is shown that both boundary layers and sharp moving interior layers can be handled naturally by the proposed class of non-conforming methods. Due to the stabilising effect of the grad-div term and the robustness of discontinuous Galerkin methods, it is possible to solve the underlying problems accurately on coarse, non-adapted meshes. The interaction of heavy grad-div stabilisation and discontinuous Galerkin methods significantly improves the mass conservation properties and the overall accuracy of the numerical scheme which is observed for the first time. Hen...
Masterlark, Timothy; Lu, Zhong; Rykhus, Russell P.
2006-01-01
Interferometric synthetic aperture radar (InSAR) imagery documents the consistent subsidence, during the interval 1992–1999, of a pyroclastic flow deposit (PFD) emplaced during the 1986 eruption of Augustine Volcano, Alaska. We construct finite element models (FEMs) that simulate thermoelastic contraction of the PFD to account for the observed subsidence. Three-dimensional problem domains of the FEMs include a thermoelastic PFD embedded in an elastic substrate. The thickness of the PFD is initially determined from the difference between post- and pre-eruption digital elevation models (DEMs). The initial excess temperature of the PFD at the time of deposition, 640 °C, is estimated from FEM predictions and an InSAR image via standard least-squares inverse methods. Although the FEM predicts the major features of the observed transient deformation, systematic prediction errors (RMSE = 2.2 cm) are most likely associated with errors in the a priori PFD thickness distribution estimated from the DEM differences. We combine an InSAR image, FEMs, and an adaptive mesh algorithm to iteratively optimize the geometry of the PFD with respect to a minimized misfit between the predicted thermoelastic deformation and observed deformation. Prediction errors from an FEM, which includes an optimized PFD geometry and the initial excess PFD temperature estimated from the least-squares analysis, are sub-millimeter (RMSE = 0.3 mm). The average thickness (9.3 m), maximum thickness (126 m), and volume (2.1 × 107m3) of the PFD, estimated using the adaptive mesh algorithm, are about twice as large as the respective estimations for the a priori PFD geometry. Sensitivity analyses suggest unrealistic PFD thickness distributions are required for initial excess PFD temperatures outside of the range 500–800 °C.
A Dynamically Adaptive Arbitrary Lagrangian-Eulerian Method for Hydrodynamics
Anderson, R W; Pember, R B; Elliott, N S
2002-10-19
A new method that combines staggered grid Arbitrary Lagrangian-Eulerian (ALE) techniques with structured local adaptive mesh refinement (AMR) has been developed for solution of the Euler equations. The novel components of the combined ALE-AMR method hinge upon the integration of traditional AMR techniques with both staggered grid Lagrangian operators as well as elliptic relaxation operators on moving, deforming mesh hierarchies. Numerical examples demonstrate the utility of the method in performing detailed three-dimensional shock-driven instability calculations.
A Dynamically Adaptive Arbitrary Lagrangian-Eulerian Method for Hydrodynamics
Anderson, R W; Pember, R B; Elliott, N S
2004-01-28
A new method that combines staggered grid Arbitrary Lagrangian-Eulerian (ALE) techniques with structured local adaptive mesh refinement (AMR) has been developed for solution of the Euler equations. The novel components of the combined ALE-AMR method hinge upon the integration of traditional AMR techniques with both staggered grid Lagrangian operators as well as elliptic relaxation operators on moving, deforming mesh hierarchies. Numerical examples demonstrate the utility of the method in performing detailed three-dimensional shock-driven instability calculations.
Anisotropic Boundary Layer Adaptivity of Multi-Element Wings
Chitale, Kedar C; Sahni, Onkar; Shephard, Mark S; Jansen, Kenneth E
2014-01-01
Multi-element wings are popular in the aerospace community due to their high lift performance. Turbulent flow simulations of these configurations require very fine mesh spacings especially near the walls, thereby making use of a boundary layer mesh necessary. However, it is difficult to accurately determine the required mesh resolution a priori to the simulations. In this paper we use an anisotropic adaptive meshing approach including adaptive control of elements in the boundary layers and study its effectiveness for two multi-element wing configurations. The results are compared with experimental data as well as nested refinements to show the efficiency of adaptivity driven by error indicators, where superior resolution in wakes and near the tip region through adaptivity are highlighted.
Hermann, Verena; Käser, Martin; Castro, Cristóbal E.
2011-02-01
We present a Discontinuous Galerkin finite element method using a high-order time integration technique for seismic wave propagation modelling on non-conforming hybrid meshes in two space dimensions. The scheme can be formulated to achieve the same approximation order in space and time and avoids numerical artefacts due to non-conforming mesh transitions or the change of the element type. A point-wise Gaussian integration along partially overlapping edges of adjacent elements is used to preserve the schemes accuracy while providing a higher flexibility in the problem-adapted mesh generation process. We describe the domain decomposition strategy of the parallel implementation and validate the performance of the new scheme by numerical convergence test and experiments with comparisons to independent reference solutions. The advantage of non-conforming hybrid meshes is the possibility to choose the mesh spacing proportional to the seismic velocity structure, which allows for simple refinement or coarsening methods even for regular quadrilateral meshes. For particular problems of strong material contrasts and geometrically thin structures, the scheme reduces the computational cost in the sense of memory and run-time requirements. The presented results promise to achieve a similar behaviour for an extension to three space dimensions where the coupling of tetrahedral and hexahedral elements necessitates non-conforming mesh transitions to avoid linking elements in form of pyramids.
Adaptive numerical methods for partial differential equations
Cololla, P. [Univ. of California, Berkeley, CA (United States)
1995-07-01
This review describes a structured approach to adaptivity. The Automated Mesh Refinement (ARM) algorithms developed by M Berger are described, touching on hyperbolic and parabolic applications. Adaptivity is achieved by overlaying finer grids only in areas flagged by a generalized error criterion. The author discusses some of the issues involved in abutting disparate-resolution grids, and demonstrates that suitable algorithms exist for dissipative as well as hyperbolic systems.
Coirier, William John
1994-01-01
A Cartesian, cell-based scheme for solving the Euler and Navier-Stokes equations in two dimensions is developed and tested. Grids about geometrically complicated bodies are generated automatically, by recursive subdivision of a single Cartesian cell encompassing the entire flow domain. Where the resulting cells intersect bodies, polygonal 'cut' cells are created. The geometry of the cut cells is computed using polygon-clipping algorithms. The grid is stored in a binary-tree data structure which provides a natural means of obtaining cell-to-cell connectivity and of carrying out solution-adaptive refinement. The Euler and Navier-Stokes equations are solved on the resulting grids using a finite-volume formulation. The convective terms are upwinded, with a limited linear reconstruction of the primitive variables used to provide input states to an approximate Riemann solver for computing the fluxes between neighboring cells. A multi-stage time-stepping scheme is used to reach a steady-state solution. Validation of the Euler solver with benchmark numerical and exact solutions is presented. An assessment of the accuracy of the approach is made by uniform and adaptive grid refinements for a steady, transonic, exact solution to the Euler equations. The error of the approach is directly compared to a structured solver formulation. A non smooth flow is also assessed for grid convergence, comparing uniform and adaptively refined results. Several formulations of the viscous terms are assessed analytically, both for accuracy and positivity. The two best formulations are used to compute adaptively refined solutions of the Navier-Stokes equations. These solutions are compared to each other, to experimental results and/or theory for a series of low and moderate Reynolds numbers flow fields. The most suitable viscous discretization is demonstrated for geometrically-complicated internal flows. For flows at high Reynolds numbers, both an altered grid-generation procedure and a
Adaptive Finite Element Method Assisted by Stochastic Simulation of Chemical Systems
Cotter, Simon L.
2013-01-01
Stochastic models of chemical systems are often analyzed by solving the corresponding Fokker-Planck equation, which is a drift-diffusion partial differential equation for the probability distribution function. Efficient numerical solution of the Fokker-Planck equation requires adaptive mesh refinements. In this paper, we present a mesh refinement approach which makes use of a stochastic simulation of the underlying chemical system. By observing the stochastic trajectory for a relatively short amount of time, the areas of the state space with nonnegligible probability density are identified. By refining the finite element mesh in these areas, and coarsening elsewhere, a suitable mesh is constructed and used for the computation of the stationary probability density. Numerical examples demonstrate that the presented method is competitive with existing a posteriori methods. © 2013 Society for Industrial and Applied Mathematics.
A multidimensional grid-adaptive relativistic magnetofluid code
van der Holst, B.; Keppens, R.; Meliani, Z.
2008-01-01
A robust second order, shock-capturing numerical scheme for multidimensional special relativistic magnetohydrodynamics on computational domains with adaptive mesh refinement is presented. The base solver is a total variation diminishing Lax-Friedrichs scheme in a finite volume setting and is combine
2015-04-09
of the vortex winds. However, all simulations do a reasonable job at269 capturing the vortex core wind velocity magnitude.270 c. Case 3271 The initial...the National Science Foundation (Division of410 Mathematical Sciences) through program element 121670, and the Air Force Office of Scien -411 tific
Multi-level adaptive simulation of transient two-phase flow in heterogeneous porous media
Chueh, C.C.
2010-10-01
An implicit pressure and explicit saturation (IMPES) finite element method (FEM) incorporating a multi-level shock-type adaptive refinement technique is presented and applied to investigate transient two-phase flow in porous media. Local adaptive mesh refinement is implemented seamlessly with state-of-the-art artificial diffusion stabilization allowing simulations that achieve both high resolution and high accuracy. Two benchmark problems, modelling a single crack and a random porous medium, are used to demonstrate the robustness of the method and illustrate the capabilities of the adaptive refinement technique in resolving the saturation field and the complex interaction (transport phenomena) between two fluids in heterogeneous media. © 2010 Elsevier Ltd.
Algorithm refinement for stochastic partial differential equations.
Alexander, F. J. (Francis J.); Garcia, Alejandro L.,; Tartakovsky, D. M. (Daniel M.)
2001-01-01
A hybrid particle/continuum algorithm is formulated for Fickian diffusion in the fluctuating hydrodynamic limit. The particles are taken as independent random walkers; the fluctuating diffusion equation is solved by finite differences with deterministic and white-noise fluxes. At the interface between the particle and continuum computations the coupling is by flux matching, giving exact mass conservation. This methodology is an extension of Adaptive Mesh and Algorithm Refinement to stochastic partial differential equations. A variety of numerical experiments were performed for both steady and time-dependent scenarios. In all cases the mean and variance of density are captured correctly by the stochastic hybrid algorithm. For a non-stochastic version (i.e., using only deterministic continuum fluxes) the mean density is correct, but the variance is reduced except within the particle region, far from the interface. Extensions of the methodology to fluid mechanics applications are discussed.
A multi-mesh finite element method for Lagrange elements of arbitrary degree
Witkowski, Thomas
2010-01-01
We consider within a finite element approach the usage of different adaptively refined meshes for different variables in systems of nonlinear, time-depended PDEs. To resolve different solution behaviours of these variables, the meshes can be independently adapted. The resulting linear systems are usually much smaller, when compared to the usage of a single mesh, and the overall computational runtime can be more than halved in such cases. Our multi-mesh method works for Lagrange finite elements of arbitrary degree and is independent of the spatial dimension. The approach is well defined, and can be implemented in existing adaptive finite element codes with minimal effort. We show computational examples in 2D and 3D ranging from dendritic growth to solid-solid phase-transitions. A further application comes from fluid dynamics where we demonstrate the applicability of the approach for solving the incompressible Navier-Stokes equations with Lagrange finite elements of the same order for velocity and pressure. The...
Niski, K; Purnomo, B; Cohen, J
2006-11-06
Previous algorithms for view-dependent level of detail provide local mesh refinements either at the finest granularity or at a fixed, coarse granularity. The former provides triangle-level adaptation, often at the expense of heavy CPU usage and low triangle rendering throughput; the latter improves CPU usage and rendering throughput by operating on groups of triangles. We present a new multiresolution hierarchy and associated algorithms that provide adaptive granularity. This multi-grained hierarchy allows independent control of the number of hierarchy nodes processed on the CPU and the number of triangles to be rendered on the GPU. We employ a seamless texture atlas style of geometry image as a GPU-friendly data organization, enabling efficient rendering and GPU-based stitching of patch borders. We demonstrate our approach on both large triangle meshes and terrains with up to billions of vertices.
... Prosthetics Hernia Surgical Mesh Implants Hernia Surgical Mesh Implants Share Tweet Linkedin Pin it More sharing options ... majority of tissue used to produce these mesh implants are from a pig (porcine) or cow (bovine) ...
Urogynecologic Surgical Mesh Implants
... Prosthetics Urogynecologic Surgical Mesh Implants Urogynecologic Surgical Mesh Implants Share Tweet Linkedin Pin it More sharing options ... majority of tissue used to produce these mesh implants are from a pig (porcine) or cow (bovine). ...
Spherical geodesic mesh generation
Fung, Jimmy [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Kenamond, Mark Andrew [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Burton, Donald E. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Shashkov, Mikhail Jurievich [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)
2015-02-27
In ALE simulations with moving meshes, mesh topology has a direct influence on feature representation and code robustness. In three-dimensional simulations, modeling spherical volumes and features is particularly challenging for a hydrodynamics code. Calculations on traditional spherical meshes (such as spin meshes) often lead to errors and symmetry breaking. Although the underlying differencing scheme may be modified to rectify this, the differencing scheme may not be accessible. This work documents the use of spherical geodesic meshes to mitigate solution-mesh coupling. These meshes are generated notionally by connecting geodesic surface meshes to produce triangular-prismatic volume meshes. This mesh topology is fundamentally different from traditional mesh topologies and displays superior qualities such as topological symmetry. This work describes the geodesic mesh topology as well as motivating demonstrations with the FLAG hydrocode.
A moving mesh unstaggered constrained transport scheme for magnetohydrodynamics
Mocz, Philip; Pakmor, Rüdiger; Springel, Volker; Vogelsberger, Mark; Marinacci, Federico; Hernquist, Lars
2016-11-01
We present a constrained transport (CT) algorithm for solving the 3D ideal magnetohydrodynamic (MHD) equations on a moving mesh, which maintains the divergence-free condition on the magnetic field to machine-precision. Our CT scheme uses an unstructured representation of the magnetic vector potential, making the numerical method simple and computationally efficient. The scheme is implemented in the moving mesh code AREPO. We demonstrate the performance of the approach with simulations of driven MHD turbulence, a magnetized disc galaxy, and a cosmological volume with primordial magnetic field. We compare the outcomes of these experiments to those obtained with a previously implemented Powell divergence-cleaning scheme. While CT and the Powell technique yield similar results in idealized test problems, some differences are seen in situations more representative of astrophysical flows. In the turbulence simulations, the Powell cleaning scheme artificially grows the mean magnetic field, while CT maintains this conserved quantity of ideal MHD. In the disc simulation, CT gives slower magnetic field growth rate and saturates to equipartition between the turbulent kinetic energy and magnetic energy, whereas Powell cleaning produces a dynamically dominant magnetic field. Such difference has been observed in adaptive-mesh refinement codes with CT and smoothed-particle hydrodynamics codes with divergence-cleaning. In the cosmological simulation, both approaches give similar magnetic amplification, but Powell exhibits more cell-level noise. CT methods in general are more accurate than divergence-cleaning techniques, and, when coupled to a moving mesh can exploit the advantages of automatic spatial/temporal adaptivity and reduced advection errors, allowing for improved astrophysical MHD simulations.
Dębski Roman
2016-06-01
Full Text Available A new dynamic programming based parallel algorithm adapted to on-board heterogeneous computers for simulation based trajectory optimization is studied in the context of “high-performance sailing”. The algorithm uses a new discrete space of continuously differentiable functions called the multi-splines as its search space representation. A basic version of the algorithm is presented in detail (pseudo-code, time and space complexity, search space auto-adaptation properties. Possible extensions of the basic algorithm are also described. The presented experimental results show that contemporary heterogeneous on-board computers can be effectively used for solving simulation based trajectory optimization problems. These computers can be considered micro high performance computing (HPC platforms-they offer high performance while remaining energy and cost efficient. The simulation based approach can potentially give highly accurate results since the mathematical model that the simulator is built upon may be as complex as required. The approach described is applicable to many trajectory optimization problems due to its black-box represented performance measure and use of OpenCL.
Eldred, Michael Scott; Subia, Samuel Ramirez; Neckels, David; Hopkins, Matthew Morgan; Notz, Patrick K.; Adams, Brian M.; Carnes, Brian; Wittwer, Jonathan W.; Bichon, Barron J.; Copps, Kevin D.
2006-10-01
This report documents the results for an FY06 ASC Algorithms Level 2 milestone combining error estimation and adaptivity, uncertainty quantification, and probabilistic design capabilities applied to the analysis and design of bistable MEMS. Through the use of error estimation and adaptive mesh refinement, solution verification can be performed in an automated and parameter-adaptive manner. The resulting uncertainty analysis and probabilistic design studies are shown to be more accurate, efficient, reliable, and convenient.
MHD simulations on an unstructured mesh
Strauss, H.R. [New York Univ., NY (United States); Park, W.; Belova, E.; Fu, G.Y. [Princeton Univ., NJ (United States). Plasma Physics Lab.; Longcope, D.W. [Univ. of Montana, Missoula, MT (United States); Sugiyama, L.E. [Massachusetts Inst. of Tech., Cambridge, MA (United States)
1998-12-31
Two reasons for using an unstructured computational mesh are adaptivity, and alignment with arbitrarily shaped boundaries. Two codes which use finite element discretization on an unstructured mesh are described. FEM3D solves 2D and 3D RMHD using an adaptive grid. MH3D++, which incorporates methods of FEM3D into the MH3D generalized MHD code, can be used with shaped boundaries, which might be 3D.
An adaptive learning approach for 3-D surface reconstruction from point clouds.
Junior, Agostinho de Medeiros Brito; Neto, Adrião Duarte Dória; de Melo, Jorge Dantas; Goncalves, Luiz Marcos Garcia
2008-06-01
In this paper, we propose a multiresolution approach for surface reconstruction from clouds of unorganized points representing an object surface in 3-D space. The proposed method uses a set of mesh operators and simple rules for selective mesh refinement, with a strategy based on Kohonen's self-organizing map (SOM). Basically, a self-adaptive scheme is used for iteratively moving vertices of an initial simple mesh in the direction of the set of points, ideally the object boundary. Successive refinement and motion of vertices are applied leading to a more detailed surface, in a multiresolution, iterative scheme. Reconstruction was experimented on with several point sets, including different shapes and sizes. Results show generated meshes very close to object final shapes. We include measures of performance and discuss robustness.
Adaptive Kinetic-Fluid Solvers for Heterogeneous Computing Architectures
Zabelok, Sergey; Kolobov, Vladimir
2015-01-01
This paper describes recent progress towards porting a Unified Flow Solver (UFS) to heterogeneous parallel computing. UFS is an adaptive kinetic-fluid simulation tool, which combines Adaptive Mesh Refinement (AMR) with automatic cell-by-cell selection of kinetic or fluid solvers based on continuum breakdown criteria. The main challenge of porting UFS to graphics processing units (GPUs) comes from the dynamically adapted mesh, which causes irregular data access. We describe the implementation of CUDA kernels for three modules in UFS: the direct Boltzmann solver using discrete velocity method (DVM), the Direct Simulation Monte Carlo (DSMC) module, and the Lattice Boltzmann Method (LBM) solver, all using octree Cartesian mesh with AMR. Double digit speedups on single GPU and good scaling for multi-GPU have been demonstrated.
Overlay Share Mesh for Interactive Group Communication with High Dynamic
WU Yan-hua; CAI Yun-ze; XU Xiao-ming
2007-01-01
An overlay share mesh infrastructure is presented for high dynamic group communication systems, such as distributed interactive simulation (DIS) and distributed virtual environments (DVE). Overlay share mesh infrastructure can own better adapting ability for high dynamic group than tradition multi-tree multicast infrastructure by sharing links among different groups. The mechanism of overlay share mesh based on area of interest (AOI) was discussed in detail in this paper. A large number of simulation experiments were done and the permance of mesh infrastructure was studied. Experiments results proved that overlay mesh infrastructure owns better adaptability than traditional multi-tree infrastructure for high dynamic group communication systems.
HE JiFeng
2008-01-01
This paper presents a refinement calculus for service components. We model the behaviour of individual service by a guarded design, which enables one to separate the responsibility of clients from the commitment made by the system, and to iden-tify a component by a set of failures and divergences. Protocols are introduced to coordinate the interactions between a component with the external environment. We adopt the notion of process refinement to formalize the substitutivity of components, and provide a complete proof method based on the notion of simulations.
Lewis R. W.
2006-11-01
Full Text Available This paper describes the application of adaptive mesh methods to the numerical simulation of one and two-dimensional petroleum reservoir waterfloods. The method uses current information on the solution to adapt the mesh to the solution as the computation proceeds. It is shown that this leads to significant improvements in accuracy at a marginal increase in computational cost. Cet article décrit l'application des méthodes de maillages évolutifs à la simulation numérique dinjection d'eau à une ou deux dimensions dans des réservoirs pétroliers. La méthode utilise des informations disponibles sur la solution pour adapter le maillage à la solution pendant que se déroule le calcul. On montre que cela conduit à des améliorations significatives en ce qui concerne la précision avec une augmentation marginale du coût des calculs.
Multiple Staggered Mesh Ewald: Boosting the Accuracy of the Smooth Particle Mesh Ewald Method
Wang, Han; Fang, Jun
2016-01-01
The smooth particle mesh Ewald (SPME) method is the standard method for computing the electrostatic interactions in the molecular simulations. In this work, the multiple staggered mesh Ewald (MSME) method is proposed to boost the accuracy of the SPME method. Unlike the SPME that achieves higher accuracy by refining the mesh, the MSME improves the accuracy by averaging the standard SPME forces computed on, e.g. $M$, staggered meshes. We prove, from theoretical perspective, that the MSME is as accurate as the SPME, but uses $M^2$ times less mesh points in a certain parameter range. In the complementary parameter range, the MSME is as accurate as the SPME with twice of the interpolation order. The theoretical conclusions are numerically validated both by a uniform and uncorrelated charge system, and by a three-point-charge water system that is widely used as solvent for the bio-macromolecules.
An h-adaptive local discontinuous Galerkin method for the Navier-Stokes-Korteweg equations
Tian, Lulu; Xu, Yan; Kuerten, J. G. M.; van der Vegt, J. J. W.
2016-08-01
In this article, we develop a mesh adaptation algorithm for a local discontinuous Galerkin (LDG) discretization of the (non)-isothermal Navier-Stokes-Korteweg (NSK) equations modeling liquid-vapor flows with phase change. This work is a continuation of our previous research, where we proposed LDG discretizations for the (non)-isothermal NSK equations with a time-implicit Runge-Kutta method. To save computing time and to capture the thin interfaces more accurately, we extend the LDG discretization with a mesh adaptation method. Given the current adapted mesh, a criterion for selecting candidate elements for refinement and coarsening is adopted based on the locally largest value of the density gradient. A strategy to refine and coarsen the candidate elements is then provided. We emphasize that the adaptive LDG discretization is relatively simple and does not require additional stabilization. The use of a locally refined mesh in combination with an implicit Runge-Kutta time method is, however, non-trivial, but results in an efficient time integration method for the NSK equations. Computations, including cases with solid wall boundaries, are provided to demonstrate the accuracy, efficiency and capabilities of the adaptive LDG discretizations.
Coupling of non-conforming meshes in a component mode synthesis method
Akcay-Perdahcioglu, D.; Doreille, M.; Boer, de A.; Ludwig, T.
2013-01-01
A common mesh refinement-based coupling technique is embedded into a component mode synthesis method, Craig–Bampton. More specifically, a common mesh is generated between the non-conforming interfaces of the coupled structures, and the compatibility constraints are enforced on that mesh via L2-minim
A new class of accurate, mesh-free hydrodynamic simulation methods
Hopkins, Philip F.
2015-06-01
We present two new Lagrangian methods for hydrodynamics, in a systematic comparison with moving-mesh, smoothed particle hydrodynamics (SPH), and stationary (non-moving) grid methods. The new methods are designed to simultaneously capture advantages of both SPH and grid-based/adaptive mesh refinement (AMR) schemes. They are based on a kernel discretization of the volume coupled to a high-order matrix gradient estimator and a Riemann solver acting over the volume `overlap'. We implement and test a parallel, second-order version of the method with self-gravity and cosmological integration, in the code GIZMO:1 this maintains exact mass, energy and momentum conservation; exhibits superior angular momentum conservation compared to all other methods we study; does not require `artificial diffusion' terms; and allows the fluid elements to move with the flow, so resolution is automatically adaptive. We consider a large suite of test problems, and find that on all problems the new methods appear competitive with moving-mesh schemes, with some advantages (particularly in angular momentum conservation), at the cost of enhanced noise. The new methods have many advantages versus SPH: proper convergence, good capturing of fluid-mixing instabilities, dramatically reduced `particle noise' and numerical viscosity, more accurate sub-sonic flow evolution, and sharp shock-capturing. Advantages versus non-moving meshes include: automatic adaptivity, dramatically reduced advection errors and numerical overmixing, velocity-independent errors, accurate coupling to gravity, good angular momentum conservation and elimination of `grid alignment' effects. We can, for example, follow hundreds of orbits of gaseous discs, while AMR and SPH methods break down in a few orbits. However, fixed meshes minimize `grid noise'. These differences are important for a range of astrophysical problems.
Robust moving mesh algorithms for hybrid stretched meshes: Application to moving boundaries problems
Landry, Jonathan; Soulaïmani, Azzeddine; Luke, Edward; Ben Haj Ali, Amine
2016-12-01
A robust Mesh-Mover Algorithm (MMA) approach is designed to adapt meshes of moving boundaries problems. A new methodology is developed from the best combination of well-known algorithms in order to preserve the quality of initial meshes. In most situations, MMAs distribute mesh deformation while preserving a good mesh quality. However, invalid meshes are generated when the motion is complex and/or involves multiple bodies. After studying a few MMA limitations, we propose the following approach: use the Inverse Distance Weighting (IDW) function to produce the displacement field, then apply the Geometric Element Transformation Method (GETMe) smoothing algorithms to improve the resulting mesh quality, and use an untangler to revert negative elements. The proposed approach has been proven efficient to adapt meshes for various realistic aerodynamic motions: a symmetric wing that has suffered large tip bending and twisting and the high-lift components of a swept wing that has moved to different flight stages. Finally, the fluid flow problem has been solved on meshes that have moved and they have produced results close to experimental ones. However, for situations where moving boundaries are too close to each other, more improvements need to be made or other approaches should be taken, such as an overset grid method.
Refinement for Administrative Policies
Dekker, MAC; Etalle, S Sandro
2007-01-01
Flexibility of management is an important requisite for access control systems as it allows users to adapt the access control system in accordance with practical requirements. This paper builds on earlier work where we defined administrative policies for a general class of RBAC models. We present a formal definition of administrative refinnement and we show that there is an ordering for administrative privileges which yields administrative refinements of policies. We argue (by giving an examp...
Effects of mesh style and grid convergence on numerical simulation accuracy of centrifugal pump
刘厚林; 刘明明; 白羽; 董亮
2015-01-01
In order to evaluate the effects of mesh generation techniques and grid convergence on pump performance in centrifugal pump model, three widely used mesh styles including structured hexahedral, unstructured tetrahedral and hybrid prismatic/tetrahedral meshes were generated for a centrifugal pump model. And quantitative grid convergence was assessed based on a grid convergence index (GCI), which accounts for the degree of grid refinement. The structured, unstructured or hybrid meshes are found to have certain difference for velocity distributions in impeller with the change of grid cell number. And the simulation results have errors to different degrees compared with experimental data. The GCI-value for structured meshes calculated is lower than that for the unstructured and hybrid meshes. Meanwhile, the structured meshes are observed to get more vortexes in impeller passage. Nevertheless, the hybrid meshes are found to have larger low-velocity area at outlet and more secondary vortexes at a specified location than structured meshes and unstructured meshes.
Pei Ping; YURY N. PETRENKO
2015-01-01
A Mesh network simulation framework which provides a powerful and concise modeling chain for a network structure will be introduce in this report. Mesh networks has a special topologic structure. The paper investigates a message transfer in wireless mesh network simulation and how does it works in cellular network simulation. Finally the experimental result gave us the information that mesh networks have different principle in transmission way with cellular networks in transmission, and multi...
Mesh generation in archipelagos
Terwisscha van Scheltinga, A.; Myers, P.G.; Pietrzak, J.D.
2012-01-01
A new mesh size field is presented that is specifically designed for efficient meshing of highly irregular oceanic domains: archipelagos. The new approach is based on the standard mesh size field that uses the proximity to the nearest coastline. Here, the proximities to the two nearest coastlines
MeshEZW: an image coder using mesh and finite elements
Landais, Thomas; Bonnaud, Laurent; Chassery, Jean-Marc
2003-08-01
In this paper, we present a new method to compress the information in an image, called MeshEZW. The proposed approach is based on the finite elements method, a mesh construction and a zerotree method. The zerotree method is an adaptive of the EZW algorithm with two new symbols for increasing the performance. These steps allow a progressive representation of the image by the automatic construction of a bitstream. The mesh structure is adapted to the image compression domain and is defined to allow video comrpession. The coder is described and some preliminary results are discussed.
Fournier, D.; Le Tellier, R.; Suteau, C., E-mail: damien.fournier@cea.fr, E-mail: romain.le-tellier@cea.fr, E-mail: christophe.suteau@cea.fr [CEA, DEN, DER/SPRC/LEPh, Cadarache, Saint Paul-lez-Durance (France); Herbin, R., E-mail: raphaele.herbin@cmi.univ-mrs.fr [Laboratoire d' Analyse et de Topologie de Marseille, Centre de Math´ematiques et Informatique (CMI), Universit´e de Provence, Marseille Cedex (France)
2011-07-01
The solution of the time-independent neutron transport equation in a deterministic way invariably consists in the successive discretization of the three variables: energy, angle and space. In the SNATCH solver used in this study, the energy and the angle are respectively discretized with a multigroup approach and the discrete ordinate method. A set of spatial coupled transport equations is obtained and solved using the Discontinuous Galerkin Finite Element Method (DGFEM). Within this method, the spatial domain is decomposed into elements and the solution is approximated by a hierarchical polynomial basis in each one. This approach is time and memory consuming when the mesh becomes fine or the basis order high. To improve the computational time and the memory footprint, adaptive algorithms are proposed. These algorithms are based on an error estimation in each cell. If the error is important in a given region, the mesh has to be refined (h−refinement) or the polynomial basis order increased (p−refinement). This paper is related to the choice between the two types of refinement. Two ways to estimate the error are compared on different benchmarks. Analyzing the differences, a hp−refinement method is proposed and tested. (author)
Earth As An Unstructured Mesh and Its Recovery from Seismic Waveform Data
De Hoop, M. V.
2015-12-01
We consider multi-scale representations of Earth's interior from thepoint of view of their possible recovery from multi- andhigh-frequency seismic waveform data. These representations areintrinsically connected to (geologic, tectonic) structures, that is,geometric parametrizations of Earth's interior. Indeed, we address theconstruction and recovery of such parametrizations using localiterative methods with appropriately designed data misfits andguaranteed convergence. The geometric parametrizations containinterior boundaries (defining, for example, faults, salt bodies,tectonic blocks, slabs) which can, in principle, be obtained fromsuccessive segmentation. We make use of unstructured meshes. For the adaptation and recovery of an unstructured mesh we introducean energy functional which is derived from the Hausdorff distance. Viaan augmented Lagrangian method, we incorporate the mentioned datamisfit. The recovery is constrained by shape optimization of theinterior boundaries, and is reminiscent of Hausdorff warping. We useelastic deformation via finite elements as a regularization whilefollowing a two-step procedure. The first step is an update determinedby the energy functional; in the second step, we modify the outcome ofthe first step where necessary to ensure that the new mesh isregular. This modification entails an array of techniques includingtopology correction involving interior boundary contacting andbreakup, edge warping and edge removal. We implement this as afeed-back mechanism from volume to interior boundary meshesoptimization. We invoke and apply a criterion of mesh quality controlfor coarsening, and for dynamical local multi-scale refinement. Wepresent a novel (fluid-solid) numerical framework based on theDiscontinuous Galerkin method.
ADAPTIVITY IN SPACE AND TIME FOR MAGNETOQUASISTATICS
Markus Clemens; Jens Lang; Delia Teleaga; Georg Wimmer
2009-01-01
This paper addresses fully space-time adaptive magnetic field computations. We de-scribe an adaptive Whitney finite element method for solving the magnetoquasistatic for-mulation of Maxwell's equations on unstructured 3D tetrahedral grids. Spatial mesh re-finement and coarsening are based on hierarchical error estimators especially designed for combining tetrahedral H (curl)-conforming edge elements in space with linearly implicit Rosenbrock methods in time. An embedding technique is applied to get efficiency in time through variable time steps. Finally, we present numerical results for the magnetic recording write head benchmark problem proposed by the Storage Research Consortium in Japan.
Workshop on adaptive grid methods for fusion plasmas
Wiley, J.C. [Univ. of Texas, Austin, TX (United States)
1995-07-01
The author describes a general `hp` finite element method with adaptive grids. The code was based on the work of Oden, et al. The term `hp` refers to the method of spatial refinement (h), in conjunction with the order of polynomials used as a part of the finite element discretization (p). This finite element code seems to handle well the different mesh grid sizes occuring between abuted grids with different resolutions.
Parameterization adaption for 3D shape optimization in aerodynamics
Badr Abou El Majd
2013-10-01
Full Text Available When solving a PDE problem numerically, a certain mesh-refinement process is always implicit, and very classically, mesh adaptivity is a very effective means to accelerate grid convergence. Similarly, when optimizing a shape by means of an explicit geometrical representation, it is natural to seek for an analogous concept of parameterization adaptivity. We propose here an adaptive parameterization for three-dimensional optimum design in aerodynamics by using the so-called “Free-Form Deformation” approach based on 3D tensorial Bézier parameterization. The proposed procedure leads to efficient numerical simulations with highly reduced computational costs.[How to cite this article: Majd, B.A.. 2014. Parameterization adaption for 3D shape optimization in aerodynamics. International Journal of Science and Engineering, 6(1:61-69. Doi: 10.12777/ijse.6.1.61-69
Algorithm Refinement for Stochastic Partial Differential Equations. I. Linear Diffusion
Alexander, Francis J.; Garcia, Alejandro L.; Tartakovsky, Daniel M.
2002-10-01
A hybrid particle/continuum algorithm is formulated for Fickian diffusion in the fluctuating hydrodynamic limit. The particles are taken as independent random walkers; the fluctuating diffusion equation is solved by finite differences with deterministic and white-noise fluxes. At the interface between the particle and continuum computations the coupling is by flux matching, giving exact mass conservation. This methodology is an extension of Adaptive Mesh and Algorithm Refinement to stochastic partial differential equations. Results from a variety of numerical experiments are presented for both steady and time-dependent scenarios. In all cases the mean and variance of density are captured correctly by the stochastic hybrid algorithm. For a nonstochastic version (i.e., using only deterministic continuum fluxes) the mean density is correct, but the variance is reduced except in particle regions away from the interface. Extensions of the methodology to fluid mechanics applications are discussed.
Algorithm refinement for stochastic partial differential equations I. linear diffusion
Alexander, F J; Tartakovsky, D M
2002-01-01
A hybrid particle/continuum algorithm is formulated for Fickian diffusion in the fluctuating hydrodynamic limit. The particles are taken as independent random walkers; the fluctuating diffusion equation is solved by finite differences with deterministic and white-noise fluxes. At the interface between the particle and continuum computations the coupling is by flux matching, giving exact mass conservation. This methodology is an extension of Adaptive Mesh and Algorithm Refinement to stochastic partial differential equations. Results from a variety of numerical experiments are presented for both steady and time-dependent scenarios. In all cases the mean and variance of density are captured correctly by the stochastic hybrid algorithm. For a nonstochastic version (i.e., using only deterministic continuum fluxes) the mean density is correct, but the variance is reduced except in particle regions away from the interface. Extensions of the methodology to fluid mechanics applications are discussed.
Multi criteria anisotropic adaptive remeshing for sheet metal forming in FORGEÂ®
Perchat, Etienne; François, Guillaume; Coupez, Thierry
2013-12-01
In this paper we present an innovative automatic adaptive anisotropic remeshing technique that has been introduced in the commercial FEM software FORGE®. It enables the full 3D simulation of industrial applications of parts with a high aspect ratio such as sheets. An anisotropic mesh is generated in order to adapt to the part and tools geometries, and in order to minimize interpolation error on the velocity field and/or on any other user defined function (eg. temperature, strain …). By minimizing the estimated error, the anisotropic adapted meshes provide a highly accurate solution, often better than those obtained on globally-refined isotropic meshes and at a much lower cost due to the small total number of nodes..
2015-01-01
With the advances in mobile computing technologies and the growth of the Net, mobile mesh networks are going through a set of important evolutionary steps. In this paper, we survey architectural aspects of mobile mesh networks and their use cases and deployment models. Also, we survey challenging areas of mobile mesh networks and describe our vision of promising mobile services. This paper presents a basic introductory material for Masters of Open Information Technologies Lab, interested in m...
Adaptive-Anisotropic Wavelet Collocation Method on general curvilinear coordinate systems
Brown-Dymkoski, Eric; Vasilyev, Oleg V.
2017-03-01
A new general framework for an Adaptive-Anisotropic Wavelet Collocation Method (A-AWCM) for the solution of partial differential equations is developed. This proposed framework addresses two major shortcomings of existing wavelet-based adaptive numerical methodologies, namely the reliance on a rectangular domain and the "curse of anisotropy", i.e. drastic over-resolution of sheet- and filament-like features arising from the inability of the wavelet refinement mechanism to distinguish highly correlated directional information in the solution. The A-AWCM addresses both of these challenges by incorporating coordinate transforms into the Adaptive Wavelet Collocation Method for the solution of PDEs. The resulting integrated framework leverages the advantages of both the curvilinear anisotropic meshes and wavelet-based adaptive refinement in a complimentary fashion, resulting in greatly reduced cost of resolution for anisotropic features. The proposed Adaptive-Anisotropic Wavelet Collocation Method retains the a priori error control of the solution and fully automated mesh refinement, while offering new abilities through the flexible mesh geometry, including body-fitting. The new A-AWCM is demonstrated for a variety of cases, including parabolic diffusion, acoustic scattering, and unsteady external flow.
Cheng, Siu-Wing; Shewchuk, Jonathan
2013-01-01
Written by authors at the forefront of modern algorithms research, Delaunay Mesh Generation demonstrates the power and versatility of Delaunay meshers in tackling complex geometric domains ranging from polyhedra with internal boundaries to piecewise smooth surfaces. Covering both volume and surface meshes, the authors fully explain how and why these meshing algorithms work.The book is one of the first to integrate a vast amount of cutting-edge material on Delaunay triangulations. It begins with introducing the problem of mesh generation and describing algorithms for constructing Delaunay trian
2011-11-01
triangles in two dimensions and tetrahedra ( tets ) in three dimensions. There are many other ways to discretize a region using unstructured meshes, but this...The boundary points associated with the airfoil surface were moved, but all of the interior points remained stationary , which resulted in a mesh
Reaction rates for reaction-diffusion kinetics on unstructured meshes
Hellander, Stefan; Petzold, Linda
2017-02-01
The reaction-diffusion master equation is a stochastic model often utilized in the study of biochemical reaction networks in living cells. It is applied when the spatial distribution of molecules is important to the dynamics of the system. A viable approach to resolve the complex geometry of cells accurately is to discretize space with an unstructured mesh. Diffusion is modeled as discrete jumps between nodes on the mesh, and the diffusion jump rates can be obtained through a discretization of the diffusion equation on the mesh. Reactions can occur when molecules occupy the same voxel. In this paper, we develop a method for computing accurate reaction rates between molecules occupying the same voxel in an unstructured mesh. For large voxels, these rates are known to be well approximated by the reaction rates derived by Collins and Kimball, but as the mesh is refined, no analytical expression for the rates exists. We reduce the problem of computing accurate reaction rates to a pure preprocessing step, depending only on the mesh and not on the model parameters, and we devise an efficient numerical scheme to estimate them to high accuracy. We show in several numerical examples that as we refine the mesh, the results obtained with the reaction-diffusion master equation approach those of a more fine-grained Smoluchowski particle-tracking model.
Domingues M. O.
2013-12-01
Full Text Available We present a new adaptive multiresoltion method for the numerical simulation of ideal magnetohydrodynamics. The governing equations, i.e., the compressible Euler equations coupled with the Maxwell equations are discretized using a finite volume scheme on a two-dimensional Cartesian mesh. Adaptivity in space is obtained via Harten’s cell average multiresolution analysis, which allows the reliable introduction of a locally refined mesh while controlling the error. The explicit time discretization uses a compact Runge–Kutta method for local time stepping and an embedded Runge-Kutta scheme for automatic time step control. An extended generalized Lagrangian multiplier approach with the mixed hyperbolic-parabolic correction type is used to control the incompressibility of the magnetic field. Applications to a two-dimensional problem illustrate the properties of the method. Memory savings and numerical divergences of magnetic field are reported and the accuracy of the adaptive computations is assessed by comparing with the available exact solution.
New adaptive routing pattern for 2D Mesh structure%2D Mesh结构中一种新的自适应路由模式
张麟麟; 李东生; 雍爱霞
2012-01-01
针对2D Mesh拓扑结构中的损坏节点会导致静态XY路由算法无法有效传输的问题,提出了一种新的路由算法——FTXY路由算法.首先对网络拓扑结构中的平均延时、理想平均吞吐量和能量消耗进行了理论评估,然后在NIRGAM仿真软平台上采用FTXY路由算法对上述三个参数进行验证,并与XY路由算法进行比较.实验结果显示,新的路由算法可以有效地绕过损坏节点,并且不会造成阻塞,提高了网络的传输性能.%For the problem that the static XY routing algorithm could not transmit effectively on account of the damaged nodes in 2D Mesh topology, this paper introduced a new routing algorithm; FTXY routing algorithm. Firstly,it evaluated the average latency, average throughput and the energy consumption theoretically, then used FTXY routing algorithm to validate the above three parameters in NIRGAM software, and compared with the XY routing algorithm. The experimental results show that the new routing algorithm can effectively round the damaged nodes and it cannot induce block, thus improving the transmission performance of the network.
Bucki, Marek; Payan, Yohan; 10.1016/j.media.2010.02.003
2010-01-01
Finite Element mesh generation remains an important issue for patient specific biomechanical modeling. While some techniques make automatic mesh generation possible, in most cases, manual mesh generation is preferred for better control over the sub-domain representation, element type, layout and refinement that it provides. Yet, this option is time consuming and not suited for intraoperative situations where model generation and computation time is critical. To overcome this problem we propose a fast and automatic mesh generation technique based on the elastic registration of a generic mesh to the specific target organ in conjunction with element regularity and quality correction. This Mesh-Match-and-Repair (MMRep) approach combines control over the mesh structure along with fast and robust meshing capabilities, even in situations where only partial organ geometry is available. The technique was successfully tested on a database of 5 pre-operatively acquired complete femora CT scans, 5 femoral heads partially...
Feischl, Michael; Gantner, Gregor; Praetorius, Dirk
2015-06-01
We consider the Galerkin boundary element method (BEM) for weakly-singular integral equations of the first-kind in 2D. We analyze some residual-type a posteriori error estimator which provides a lower as well as an upper bound for the unknown Galerkin BEM error. The required assumptions are weak and allow for piecewise smooth parametrizations of the boundary, local mesh-refinement, and related standard piecewise polynomials as well as NURBS. In particular, our analysis gives a first contribution to adaptive BEM in the frame of isogeometric analysis (IGABEM), for which we formulate an adaptive algorithm which steers the local mesh-refinement and the multiplicity of the knots. Numerical experiments underline the theoretical findings and show that the proposed adaptive strategy leads to optimal convergence.
Three-dimensional h-adaptivity for the multigroup neutron diffusion equations
Wang, Yaqi
2009-04-01
Adaptive mesh refinement (AMR) has been shown to allow solving partial differential equations to significantly higher accuracy at reduced numerical cost. This paper presents a state-of-the-art AMR algorithm applied to the multigroup neutron diffusion equation for reactor applications. In order to follow the physics closely, energy group-dependent meshes are employed. We present a novel algorithm for assembling the terms coupling shape functions from different meshes and show how it can be made efficient by deriving all meshes from a common coarse mesh by hierarchic refinement. Our methods are formulated using conforming finite elements of any order, for any number of energy groups. The spatial error distribution is assessed with a generalization of an error estimator originally derived for the Poisson equation. Our implementation of this algorithm is based on the widely used Open Source adaptive finite element library deal.II and is made available as part of this library\\'s extensively documented tutorial. We illustrate our methods with results for 2-D and 3-D reactor simulations using 2 and 7 energy groups, and using conforming finite elements of polynomial degree up to 6. © 2008 Elsevier Ltd. All rights reserved.
An Evaluation Method for Distortion Energy Parameterization of Triangular Meshes
SHI Jing; ZHAO Xiu-yang; ZHANG Cai-ming; YANG Bo
2013-01-01
Parameterization of triangle meshes is a fundamental problem for texture mapping, surface fitting, surface reconstruction, and mesh editing. The deformation of triangular meshes caused by the parameterized process is the measurement of parameterization. Traditional standard method has its limitation when evaluating mixture distortion energy parameterizations. Thus an evaluation method bases on distortion energy parameterization of triangular meshes is introduced for the limitation. The novel method employs an adaptive expression form to the mixture energy, and uses a weight factor to represent distortion energy distribution. By using this method, we can evaluate all kinds of parameterization in a uniform measurement and acquire a more intuitive and clear evaluation.
Held, Gilbert
2005-01-01
Wireless mesh networking is a new technology that has the potential to revolutionize how we access the Internet and communicate with co-workers and friends. Wireless Mesh Networks examines the concept and explores its advantages over existing technologies. This book explores existing and future applications, and examines how some of the networking protocols operate.The text offers a detailed analysis of the significant problems affecting wireless mesh networking, including network scale issues, security, and radio frequency interference, and suggests actual and potential solutions for each pro
A study on moving mesh finite element solution of the porous medium equation
Ngo, Cuong; Huang, Weizhang
2017-02-01
An adaptive moving mesh finite element method is studied for the numerical solution of the porous medium equation with and without variable exponents and absorption. The method is based on the moving mesh partial differential equation approach and employs its newly developed implementation. The implementation has several improvements over the traditional one, including its explicit, compact form of the mesh velocities, ease to program, and less likelihood of producing singular meshes. Three types of metric tensor that correspond to uniform and arclength-based and Hessian-based adaptive meshes are considered. The method shows first-order convergence for uniform and arclength-based adaptive meshes, and second-order convergence for Hessian-based adaptive meshes. It is also shown that the method can be used for situations with complex free boundaries, emerging and splitting of free boundaries, and the porous medium equation with variable exponents and absorption. Two-dimensional numerical results are presented.
Mesh implants: An overview of crucial mesh parameters
Lei-Ming; Zhu; Philipp; Schuster; Uwe; Klinge
2015-01-01
Hernia repair is one of the most frequently performed surgical interventions that use mesh implants. This article evaluates crucial mesh parameters to facilitate selection of the most appropriate mesh implant, considering raw materials, mesh composition, structure parameters and mechanical parameters. A literature review was performed using the Pub Med database. The most important mesh parameters in the selection of a mesh implant are the raw material, structural parameters and mechanical parameters, which should match the physiological conditions. The structural parameters, especially the porosity, are the most important predictors of the biocompatibility performance of synthetic meshes. Meshes with large pores exhibit less inflammatory infiltrate, connective tissue and scar bridging, which allows increased soft tissue ingrowth. The raw material and combination of raw materials of the used mesh, including potential coatings and textile design, strongly impact the inflammatory reaction to the mesh. Synthetic meshes made from innovative polymers combined with surface coating have been demonstrated to exhibit advantageous behavior in specialized fields. Monofilament, largepore synthetic meshes exhibit advantages. The value of mesh classification based on mesh weight seems to be overestimated. Mechanical properties of meshes, such as anisotropy/isotropy, elasticity and tensile strength, are crucial parameters for predicting mesh performance after implantation.
An a posteriori-driven adaptive Mixed High-Order method with application to electrostatics
Di Pietro, Daniele A.; Specogna, Ruben
2016-12-01
In this work we propose an adaptive version of the recently introduced Mixed High-Order method and showcase its performance on a comprehensive set of academic and industrial problems in computational electromagnetism. The latter include, in particular, the numerical modeling of comb-drive and MEMS devices. Mesh adaptation is driven by newly derived, residual-based error estimators. The resulting method has several advantageous features: It supports fairly general meshes, it enables arbitrary approximation orders, and has a moderate computational cost thanks to hybridization and static condensation. The a posteriori-driven mesh refinement is shown to significantly enhance the performance on problems featuring singular solutions, allowing to fully exploit the high-order of approximation.
A Dynamically Adaptive Arbitrary Lagrangian-Eulerian Method for Solution of the Euler Equations
Anderson, R W; Elliott, N S; Pember, R B
2003-02-14
A new method that combines staggered grid arbitrary Lagrangian-Eulerian (ALE) techniques with structured local adaptive mesh refinement (AMR) has been developed for solution of the Euler equations. The novel components of the methods are driven by the need to reconcile traditional AMR techniques with the staggered variables and moving, deforming meshes associated with Lagrange based ALE schemes. We develop interlevel solution transfer operators and interlevel boundary conditions first in the case of purely Lagrangian hydrodynamics, and then extend these ideas into an ALE method by developing adaptive extensions of elliptic mesh relaxation techniques. Conservation properties of the method are analyzed, and a series of test problem calculations are presented which demonstrate the utility and efficiency of the method.
Botsch, Mario; Pauly, Mark; Alliez, Pierre; Levy, Bruno
2010-01-01
Geometry processing, or mesh processing, is a fast-growing area of research that uses concepts from applied mathematics, computer science, and engineering to design efficient algorithms for the acquisition, reconstruction, analysis, manipulation, simulation, and transmission of complex 3D models. Applications of geometry processing algorithms already cover a wide range of areas from multimedia, entertainment, and classical computer-aided design, to biomedical computing, reverse engineering, and scientific computing. Over the last several years, triangle meshes have become increasingly popular,
石繁荣; 黄玉清; 任珍文; 伍春
2013-01-01
Wireless sensor networks have been widely utilized in agricultural production in such as crop information monitoring systems, agricultural facilities’wireless control systems, etc. The wireless sensor networks could promote the development of agricultural information and intelligence, and more research has been focused on using ZigBee wireless technology to build the networks in recent years. To collect the base crop status and environmental information of greenhouses in a wireless way, a wireless sensor monitoring network system was designed. The basic work of this paper was the software and hardware system design; further work is projected to be low-power adaptive mechanism design. In analyzing the distribution characteristics of the greenhouse base, it could be seen that the greenhouses were concentrated in their distribution, but independent from each other. So the network topology architecture was designed as clustering Tree-Mesh hybrid topology architecture, and the nodes of the cluster belonged to the same greenhouse. The network was built up by a coordinator, and a large number of routers and sensor nodes were joined in. The coordinator was a sink node, it was designed as a gateway, and there were some routers which played the role of cluster head in the network. The clustering Tree-Mesh hybrid network was built in two steps: First, the mesh network was established by the coordinator and cluster head. Then, the tree network was built by the cluster head, and the tree was a cluster with routers and sensor nodes. The system utilized ZigBee to build the wireless sensor network and multi-hop communication, and the hardware of a single chip multi-sensor wireless node based CC2530 was designed. The modular design of the hardware subsystem was composed of a radio module, sensor module and power module. The finite state machine node software and the low-power improvement were designed based on Z-Stack. The stack ran on a task allocation mechanism that was similar
Error computation for adaptive finite element analysis
Khan, A A; Memon, I R; Ming, X Y
2002-01-01
The paper gives a simple numerical procedure for computations of errors generated by the discretisation process of finite element method. The procedure given is based on the ZZ error estimator which is believed to be reasonable accurate and thus can be readily implemented in any existing finite element codes. The devised procedure not only estimates the global energy norm error but also evaluates the local errors in individual elements. In the example, the given procedure is combined with an adaptive refinement procedure, which provides guidance for optimal mesh designing and allows the user to obtain a desired accuracy with a limited number of interaction. (author)
Atkey, Robert; Ghani, Neil
2012-01-01
Dependently typed programming languages allow sophisticated properties of data to be expressed within the type system. Of particular use in dependently typed programming are indexed types that refine data by computationally useful information. For example, the N-indexed type of vectors refines lists by their lengths. Other data types may be refined in similar ways, but programmers must produce purpose-specific refinements on an ad hoc basis, developers must anticipate which refinements to include in libraries, and implementations must often store redundant information about data and their refinements. In this paper we show how to generically derive inductive characterisations of refinements of inductive types, and argue that these characterisations can alleviate some of the aforementioned difficulties associated with ad hoc refinements. Our characterisations also ensure that standard techniques for programming with and reasoning about inductive types are applicable to refinements, and that refinements can the...
Retrofitting Masonry Walls with Carbon Mesh
Patrick Bischof
2014-01-01
Full Text Available Static-cyclic shear load tests and tensile tests on retrofitted masonry walls were conducted at UAS Fribourg for an evaluation of the newly developed retrofitting system, the S&P ARMO-System. This retrofitting system consists of a composite of carbon mesh embedded in a specially adapted high quality spray mortar. It can be applied with established construction techniques using traditional construction materials. The experimental study has shown that masonry walls reinforced by this retrofitting system reach a similar strength and a higher ductility than retrofits by means of bonded carbon fiber reinforced polymer sheets. Hence, the retrofitting system using carbon fiber meshes embedded in a high quality mortar constitutes a good option for static or seismic retrofits or reinforcements for masonry walls. However, the experimental studies also revealed that the mechanical anchorage of carbon mesh may be delicate depending on its design.
Optimal Throughput and Self-adaptability of Robust Real-Time IEEE 802.15.4 MAC for AMI Mesh Network
Shabani, Hikma; Mohamud Ahmed, Musse; Khan, Sheroz; Hameed, Shahab Ahmed; Hadi Habaebi, Mohamed
2013-12-01
A smart grid refers to a modernization of the electricity system that brings intelligence, reliability, efficiency and optimality to the power grid. To provide an automated and widely distributed energy delivery, the smart grid will be branded by a two-way flow of electricity and information system between energy suppliers and their customers. Thus, the smart grid is a power grid that integrates data communication networks which provide the collected and analysed data at all levels in real time. Therefore, the performance of communication systems is so vital for the success of smart grid. Merit to the ZigBee/IEEE802.15.4std low cost, low power, low data rate, short range, simplicity and free licensed spectrum that makes wireless sensor networks (WSNs) the most suitable wireless technology for smart grid applications. Unfortunately, almost all ZigBee channels overlap with wireless local area network (WLAN) channels, resulting in severe performance degradation due to interference. In order to improve the performance of communication systems, this paper proposes an optimal throughput and self-adaptability of ZigBee/IEEE802.15.4std for smart grid.
A parallel adaptive finite difference algorithm for petroleum reservoir simulation
Hoang, Hai Minh
2005-07-01
Adaptive finite differential for problems arising in simulation of flow in porous medium applications are considered. Such methods have been proven useful for overcoming limitations of computational resources and improving the resolution of the numerical solutions to a wide range of problems. By local refinement of the computational mesh where it is needed to improve the accuracy of solutions, yields better solution resolution representing more efficient use of computational resources than is possible with traditional fixed-grid approaches. In this thesis, we propose a parallel adaptive cell-centered finite difference (PAFD) method for black-oil reservoir simulation models. This is an extension of the adaptive mesh refinement (AMR) methodology first developed by Berger and Oliger (1984) for the hyperbolic problem. Our algorithm is fully adaptive in time and space through the use of subcycling, in which finer grids are advanced at smaller time steps than the coarser ones. When coarse and fine grids reach the same advanced time level, they are synchronized to ensure that the global solution is conservative and satisfy the divergence constraint across all levels of refinement. The material in this thesis is subdivided in to three overall parts. First we explain the methodology and intricacies of AFD scheme. Then we extend a finite differential cell-centered approximation discretization to a multilevel hierarchy of refined grids, and finally we are employing the algorithm on parallel computer. The results in this work show that the approach presented is robust, and stable, thus demonstrating the increased solution accuracy due to local refinement and reduced computing resource consumption. (Author)
GIZMO: A New Class of Accurate, Mesh-Free Hydrodynamic Simulation Methods
Hopkins, Philip F
2014-01-01
We present and study two new Lagrangian numerical methods for solving the equations of hydrodynamics, in a systematic comparison with moving-mesh, SPH, and non-moving grid methods. The new methods are designed to capture many advantages of both smoothed-particle hydrodynamics (SPH) and grid-based or adaptive mesh refinement (AMR) schemes. They are based on a kernel discretization of the volume coupled to a high-order matrix gradient estimator and a Riemann solver acting over the volume 'overlap.' We implement and test a parallel, second-order version of the method with coupled self-gravity & cosmological integration, in the code GIZMO: this maintains exact mass, energy and momentum conservation; exhibits superior angular momentum conservation compared to all other methods we study; does not require 'artificial diffusion' terms; and allows fluid elements to move with the flow so resolution is automatically adaptive. We consider a large suite of test problems, and find that on all problems the new methods a...
A LAGUERRE VORONOI BASED SCHEME FOR MESHING PARTICLE SYSTEMS.
Bajaj, Chandrajit
2005-06-01
We present Laguerre Voronoi based subdivision algorithms for the quadrilateral and hexahedral meshing of particle systems within a bounded region in two and three dimensions, respectively. Particles are smooth functions over circular or spherical domains. The algorithm first breaks the bounded region containing the particles into Voronoi cells that are then subsequently decomposed into an initial quadrilateral or an initial hexahedral scaffold conforming to individual particles. The scaffolds are subsequently refined via applications of recursive subdivision (splitting and averaging rules). Our choice of averaging rules yield a particle conforming quadrilateral/hexahedral mesh, of good quality, along with being smooth and differentiable in the limit. Extensions of the basic scheme to dynamic re-meshing in the case of addition, deletion, and moving particles are also discussed. Motivating applications of the use of these static and dynamic meshes for particle systems include the mechanics of epoxy/glass composite materials, bio-molecular force field calculations, and gas hydrodynamics simulations in cosmology.
Documentation for MeshKit - Reactor Geometry (&mesh) Generator
Jain, Rajeev [Argonne National Lab. (ANL), Argonne, IL (United States); Mahadevan, Vijay [Argonne National Lab. (ANL), Argonne, IL (United States)
2015-09-30
This report gives documentation for using MeshKit’s Reactor Geometry (and mesh) Generator (RGG) GUI and also briefly documents other algorithms and tools available in MeshKit. RGG is a program designed to aid in modeling and meshing of complex/large hexagonal and rectilinear reactor cores. RGG uses Argonne’s SIGMA interfaces, Qt and VTK to produce an intuitive user interface. By integrating a 3D view of the reactor with the meshing tools and combining them into one user interface, RGG streamlines the task of preparing a simulation mesh and enables real-time feedback that reduces accidental scripting mistakes that could waste hours of meshing. RGG interfaces with MeshKit tools to consolidate the meshing process, meaning that going from model to mesh is as easy as a button click. This report is designed to explain RGG v 2.0 interface and provide users with the knowledge and skills to pilot RGG successfully. Brief documentation of MeshKit source code, tools and other algorithms available are also presented for developers to extend and add new algorithms to MeshKit. RGG tools work in serial and parallel and have been used to model complex reactor core models consisting of conical pins, load pads, several thousands of axially varying material properties of instrumentation pins and other interstices meshes.
Coloured Petri Net Refinement Specification and Correctness Proof with Coq
Choppy, Christine; Mayero, Micaela; Petrucci, Laure
2009-01-01
In this work, we address the formalisation of symmetric nets, a subclass of coloured Petri nets, refinement in COQ. We first provide a formalisation of the net models, and of their type refinement in COQ. Then the COQ proof assistant is used to prove the refinement correctness lemma. An example adapted from a protocol example illustrates our work.
Isotopic Implicit Surface Meshing
Boissonnat, Jean-Daniel; Cohen-Steiner, David; Vegter, Gert
2004-01-01
This paper addresses the problem of piecewise linear approximation of implicit surfaces. We first give a criterion ensuring that the zero-set of a smooth function and the one of a piecewise linear approximation of it are isotopic. Then, we deduce from this criterion an implicit surface meshing algor
Feature-preserving surface mesh smoothing via suboptimal Delaunay triangulation.
Gao, Zhanheng; Yu, Zeyun; Holst, Michael
2013-01-01
A method of triangular surface mesh smoothing is presented to improve angle quality by extending the original optimal Delaunay triangulation (ODT) to surface meshes. The mesh quality is improved by solving a quadratic optimization problem that minimizes the approximated interpolation error between a parabolic function and its piecewise linear interpolation defined on the mesh. A suboptimal problem is derived to guarantee a unique, analytic solution that is significantly faster with little loss in accuracy as compared to the optimal one. In addition to the quality-improving capability, the proposed method has been adapted to remove noise while faithfully preserving sharp features such as edges and corners of a mesh. Numerous experiments are included to demonstrate the performance of the method.
Vickers, Trevor
1992-01-01
On the Refinement Calculus gives one view of the development of the refinement calculus and its attempt to bring together - among other things - Z specifications and Dijkstra's programming language. It is an excellent source of reference material for all those seeking the background and mathematical underpinnings of the refinement calculus.
Mesh Resolution Effect on 3D RANS Turbomachinery Flow Simulations
Yershov, Sergiy
2016-01-01
The paper presents the study of the effect of a mesh refinement on numerical results of 3D RANS computations of turbomachinery flows. The CFD solver F, which based on the second-order accurate ENO scheme, is used in this study. The simplified multigrid algorithm and local time stepping permit decreasing computational time. The flow computations are performed for a number of turbine and compressor cascades and stages. In all flow cases, the successively refined meshes of H-type with an approximate orthogonalization near the solid walls were generated. The results obtained are compared in order to estimate their both mesh convergence and ability to resolve the transonic flow pattern. It is concluded that for thorough studying the fine phenomena of the 3D turbomachinery flows, it makes sense to use the computational meshes with the number of cells from several millions up to several hundred millions per a single turbomachinery blade channel, while for industrial computations, a mesh of about or less than one mil...
无
2011-01-01
The scaled boundary finite element method(SBFEM) is a semi-analytical numerical method,which models an analysis domain by a small number of large-sized subdomains and discretises subdomain boundaries only.In a subdomain,all fields of state variables including displacement,stress,velocity and acceleration are semi-analytical,and the kinetic energy,strain energy and energy error are all integrated semi-analytically.These advantages are taken in this study to develop a posteriori h-hierarchical adaptive SBFEM for transient elastodynamic problems using a mesh refinement procedure which subdivides subdomains.Because only a small number of subdomains are subdivided,mesh refinement is very simple and efficient,and mesh mapping to transfer state variables from an old mesh to a new one is also very simple but accurate.Two 2D examples with stress wave propagation were modelled.The results show that the developed method is capable of capturing propagation of steep stress regions and calculating accurate dynamic responses,using only a fraction of degrees of freedom required by adaptive finite element method.
Advanced Automatic Hexahedral Mesh Generation from Surface Quad Meshes
Kremer, Michael; Bommes, David; Lim, Isaak; Kobbelt, Leif
2013-01-01
International audience; A purely topological approach for the generation of hexahedral meshes from quadrilateral surface meshes of genus zero has been proposed by M. MÃ¼ller-Hannemann: in a first stage, the input surface mesh is reduced to a single hexahedron by successively eliminating loops from the dual graph of the quad mesh; in the second stage, the hexahedral mesh is constructed by extruding a layer of hexahedra for each dual loop from the first stage in reverse elimination order. In th...
A GPU-accelerated adaptive discontinuous Galerkin method for level set equation
Karakus, A.; Warburton, T.; Aksel, M. H.; Sert, C.
2016-01-01
This paper presents a GPU-accelerated nodal discontinuous Galerkin method for the solution of two- and three-dimensional level set (LS) equation on unstructured adaptive meshes. Using adaptive mesh refinement, computations are localised mostly near the interface location to reduce the computational cost. Small global time step size resulting from the local adaptivity is avoided by local time-stepping based on a multi-rate Adams-Bashforth scheme. Platform independence of the solver is achieved with an extensible multi-threading programming API that allows runtime selection of different computing devices (GPU and CPU) and different threading interfaces (CUDA, OpenCL and OpenMP). Overall, a highly scalable, accurate and mass conservative numerical scheme that preserves the simplicity of LS formulation is obtained. Efficiency, performance and local high-order accuracy of the method are demonstrated through distinct numerical test cases.
Feature detection of triangular meshes vianeighbor supporting
Xiao-chao WANG; Jun-jie CAO; Xiu-ping LIU; Bao-jun LI; Xi-quan SHI; Yi-zhen SUN
2012-01-01
We propose a robust method for detecting features on triangular meshes by combining normal tensor voting with neighbor supporting.Our method contains two stages:feature detection and feature refinement.First,the normal tensor voting method is modified to detect the initial features,which may include some pseudo features.Then,at the feature refinement stage,a novel salient measure deriving from the idea of neighbor supporting is developed. Benefiting from the integrated reliable salient measure feature,pseudo features can be effectively discriminated from the initially detected features and removed. Compared to previous methods based on the differential geometric property,the main advantage of our method is that it can detect both sharp and weak features.Numerical experiments show that our algorithm is robust,effective,and can produce more accurate results.We also discuss how detected features are incorporated into applications,such as feature-preserving mesh denoising and hole-filling,and present visually appealing results by integrating feature information.
OPTIMIZING EUCALYPTUS PULP REFINING
Vail Manfredi
2004-01-01
This paper discusses the refining of bleached eucalyptus kraft pulp (BEKP).Pilot plant tests were carried out in to optimize the refining process and to identify the effects of refining variables on final paper quality and process costs.The following parameters are discussed: pulp consistency, disk pattern design, refiner speed,energy input, refiner configuration (parallel or serial)and refining intensity.The effects of refining on pulp fibers were evaluated against the pulp quality properties, such as physical strengths, bulk, opacity and porosity, as well as the interactions with papermaking process, such as paper machine runnability, paper breaks and refining control.The results showed that process optimization,considering pulp quality and refining costs, were obtained when eucalyptus pulp is refined under the lowest intensity and the highest pulp consistency possible. Changes on the operational refining conditions will have the highest impact on total energy requirements (costs) without any significant effect on final paper properties.It was also observed that classical ways to control the industrial operation, such as those based on drainage measurements, do not represent the best alternative to maximize the final paper properties neither the paper machine runability.
An adaptive numerical method for free surface flows passing rigidly mounted obstacles
Nikitin, Kirill D; Terekhov, Kirill M; Vassilevski, Yuri V; Yanbarisov, Ruslan
2016-01-01
The paper develops a method for the numerical simulation of a free-surface flow of incompressible viscous fluid around a streamlined body. The body is a rigid stationary construction partially submerged in the fluid. The application we are interested in the paper is a flow around a surface mounted offshore oil platform. The numerical method builds on a hybrid finite volume / finite difference discretization using adaptive octree cubic meshes. The mesh is dynamically refined towards the free surface and the construction. Special care is taken to devise a discretization for the case of curvilinear boundaries and interfaces immersed in the octree Cartesian background computational mesh. To demonstrate the accuracy of the method, we show the results for two benchmark problems: the sloshing 3D container and the channel laminar flow passing the 3D cylinder of circular cross-section. Further, we simulate numerically a flow with surface waves around an offshore oil platform for the realistic set of geophysical data.
A general boundary capability embedded in an orthogonal mesh
Hewett, D.W.; Yu-Jiuan Chen [Lawrence Livermore National Lab., CA (United States)
1995-07-01
The authors describe how they hold onto orthogonal mesh discretization when dealing with curved boundaries. Special difference operators were constructed to approximate numerical zones split by the domain boundary; the operators are particularly simple for this rectangular mesh. The authors demonstrated that this simple numerical approach, termed Dynamic Alternating Direction Implicit, turned out to be considerably more efficient than more complex grid-adaptive algorithms that were tried previously.
Efficient Packet Forwarding in Mesh Network
Soumen Kanrar
2012-01-01
Wireless Mesh Network (WMN) is a multi hop low cost, with easy maintenance robust network providing reliable service coverage. WMNs consist of mesh routers and mesh clients. In this architecture, while static mesh routers form the wireless backbone, mesh clients access the network through mesh routers as well as directly meshing with each other. Different from traditional wireless networks, WMN is dynamically self-organized and self-configured. In other words, the nodes in the mesh network au...
Tang, Zhao; Wei, Qingshan; Wei, Alexander
2011-12-01
Metal-mesh lithography (MML) is a practical hybrid of microcontact printing and capillary force lithography that can be applied over millimeter-sized areas with a high level of uniformity. MML can be achieved by blotting various inks onto substrates through thin copper grids, relying on preferential wetting and capillary interactions between template and substrate for pattern replication. The resulting mesh patterns, which are inverted relative to those produced by stenciling or serigraphy, can be reproduced with low micrometer resolution. MML can be combined with other surface chemistry and lift-off methods to create functional microarrays for diverse applications, such as periodic islands of gold nanorods and patterned corrals for fibroblast cell cultures.
Semi-structured meshes for axial turbomachinery blades
Sbardella, L.; Sayma, A. I.; Imregun, M.
2000-03-01
This paper describes the development and application of a novel mesh generator for the flow analysis of turbomachinery blades. The proposed method uses a combination of structured and unstructured meshes, the former in the radial direction and the latter in the axial and tangential directions, in order to exploit the fact that blade-like structures are not strongly three-dimensional since the radial variation is usually small. The proposed semi-structured mesh formulation was found to have a number of advantages over its structured counterparts. There is a significant improvement in the smoothness of the grid spacing and also in capturing particular aspects of the blade passage geometry. It was also found that the leading- and trailing-edge regions could be discretized without generating superfluous points in the far field, and that further refinements of the mesh to capture wake and shock effects were relatively easy to implement. The capability of the method is demonstrated in the case of a transonic fan blade for which the steady state flow is predicted using both structured and semi-structured meshes. A totally unstructured mesh is also generated for the same geometry to illustrate the disadvantages of using such an approach for turbomachinery blades. Copyright
Arnold, Alexander; Bruhns, Otto T.; Mosler, Jörn
2011-07-01
A novel finite element formulation suitable for computing efficiently the stiffness distribution in soft biological tissue is presented in this paper. For that purpose, the inverse problem of finite strain hyperelasticity is considered and solved iteratively. In line with Arnold et al (2010 Phys. Med. Biol. 55 2035), the computing time is effectively reduced by using adaptive finite element methods. In sharp contrast to previous approaches, the novel mesh adaption relies on an r-adaption (re-allocation of the nodes within the finite element triangulation). This method allows the detection of material interfaces between healthy and diseased tissue in a very effective manner. The evolution of the nodal positions is canonically driven by the same minimization principle characterizing the inverse problem of hyperelasticity. Consequently, the proposed mesh adaption is variationally consistent. Furthermore, it guarantees that the quality of the numerical solution is improved. Since the proposed r-adaption requires only a relatively coarse triangulation for detecting material interfaces, the underlying finite element spaces are usually not rich enough for predicting the deformation field sufficiently accurately (the forward problem). For this reason, the novel variational r-refinement is combined with the variational h-adaption (Arnold et al 2010) to obtain a variational hr-refinement algorithm. The resulting approach captures material interfaces well (by using r-adaption) and predicts a deformation field in good agreement with that observed experimentally (by using h-adaption).
A parallel direct solver for the self-adaptive hp Finite Element Method
Paszyński, Maciej R.
2010-03-01
In this paper we present a new parallel multi-frontal direct solver, dedicated for the hp Finite Element Method (hp-FEM). The self-adaptive hp-FEM generates in a fully automatic mode, a sequence of hp-meshes delivering exponential convergence of the error with respect to the number of degrees of freedom (d.o.f.) as well as the CPU time, by performing a sequence of hp refinements starting from an arbitrary initial mesh. The solver constructs an initial elimination tree for an arbitrary initial mesh, and expands the elimination tree each time the mesh is refined. This allows us to keep track of the order of elimination for the solver. The solver also minimizes the memory usage, by de-allocating partial LU factorizations computed during the elimination stage of the solver, and recomputes them for the backward substitution stage, by utilizing only about 10% of the computational time necessary for the original computations. The solver has been tested on 3D Direct Current (DC) borehole resistivity measurement simulations problems. We measure the execution time and memory usage of the solver over a large regular mesh with 1.5 million degrees of freedom as well as on the highly non-regular mesh, generated by the self-adaptive h p-FEM, with finite elements of various sizes and polynomial orders of approximation varying from p = 1 to p = 9. From the presented experiments it follows that the parallel solver scales well up to the maximum number of utilized processors. The limit for the solver scalability is the maximum sequential part of the algorithm: the computations of the partial LU factorizations over the longest path, coming from the root of the elimination tree down to the deepest leaf. © 2009 Elsevier Inc. All rights reserved.
Mesh Algorithms for PDE with Sieve I: Mesh Distribution
Matthew G. Knepley
2009-01-01
Full Text Available We have developed a new programming framework, called Sieve, to support parallel numerical partial differential equation(s (PDE algorithms operating over distributed meshes. We have also developed a reference implementation of Sieve in C++ as a library of generic algorithms operating on distributed containers conforming to the Sieve interface. Sieve makes instances of the incidence relation, or arrows, the conceptual first-class objects represented in the containers. Further, generic algorithms acting on this arrow container are systematically used to provide natural geometric operations on the topology and also, through duality, on the data. Finally, coverings and duality are used to encode not only individual meshes, but all types of hierarchies underlying PDE data structures, including multigrid and mesh partitions. In order to demonstrate the usefulness of the framework, we show how the mesh partition data can be represented and manipulated using the same fundamental mechanisms used to represent meshes. We present the complete description of an algorithm to encode a mesh partition and then distribute a mesh, which is independent of the mesh dimension, element shape, or embedding. Moreover, data associated with the mesh can be similarly distributed with exactly the same algorithm. The use of a high level of abstraction within the Sieve leads to several benefits in terms of code reuse, simplicity, and extensibility. We discuss these benefits and compare our approach to other existing mesh libraries.
ALEGRA -- A massively parallel h-adaptive code for solid dynamics
Summers, R.M.; Wong, M.K.; Boucheron, E.A.; Weatherby, J.R. [Sandia National Labs., Albuquerque, NM (United States)
1997-12-31
ALEGRA is a multi-material, arbitrary-Lagrangian-Eulerian (ALE) code for solid dynamics designed to run on massively parallel (MP) computers. It combines the features of modern Eulerian shock codes, such as CTH, with modern Lagrangian structural analysis codes using an unstructured grid. ALEGRA is being developed for use on the teraflop supercomputers to conduct advanced three-dimensional (3D) simulations of shock phenomena important to a variety of systems. ALEGRA was designed with the Single Program Multiple Data (SPMD) paradigm, in which the mesh is decomposed into sub-meshes so that each processor gets a single sub-mesh with approximately the same number of elements. Using this approach the authors have been able to produce a single code that can scale from one processor to thousands of processors. A current major effort is to develop efficient, high precision simulation capabilities for ALEGRA, without the computational cost of using a global highly resolved mesh, through flexible, robust h-adaptivity of finite elements. H-adaptivity is the dynamic refinement of the mesh by subdividing elements, thus changing the characteristic element size and reducing numerical error. The authors are working on several major technical challenges that must be met to make effective use of HAMMER on MP computers.
Danilov, A. A.; Salamatova, V. Yu; Vassilevski, Yu V.
2012-12-01
Here, a workflow for high-resolution efficient numerical modeling of bioimpedance measurements is suggested that includes 3D image segmentation, adaptive mesh generation, finite-element discretization, and the analysis of simulation results. Using the adaptive unstructured tetrahedral meshes enables to decrease significantly a number of mesh elements while keeping model accuracy. The numerical results illustrate current, potential, and sensitivity field distributions for a conventional Kubicek-like scheme of bioimpedance measurements using segmented geometric model of human torso based on Visible Human Project data. The whole body VHP man computational mesh is constructed that contains 574 thousand vertices and 3.3 million tetrahedrons.
Adaptive-grid methods for time-dependent partial differential equations
Hedstrom, G.W.; Rodrique, G.H.
1981-01-01
This paper contains a survey of recent developments of adaptive-grid algorithms for time-dependent partial differential equations. Two lines of research are discussed. One involves the automatic selection of moving grids to follow propagating waves. The other is based on stationary grids but uses local mesh refinement in both space and time. Advantages and disadvantages of both approaches are discussed. The development of adaptive-grid schemes shows promise of greatly increasing our ability to solve problems in several spatial dimensions.
An application of MeSH enrichment analysis in livestock.
Morota, G; Peñagaricano, F; Petersen, J L; Ciobanu, D C; Tsuyuzaki, K; Nikaido, I
2015-08-01
An integral part of functional genomics studies is to assess the enrichment of specific biological terms in lists of genes found to be playing an important role in biological phenomena. Contrasting the observed frequency of annotated terms with those of the background is at the core of overrepresentation analysis (ORA). Gene Ontology (GO) is a means to consistently classify and annotate gene products and has become a mainstay in ORA. Alternatively, Medical Subject Headings (MeSH) offers a comprehensive life science vocabulary including additional categories that are not covered by GO. Although MeSH is applied predominantly in human and model organism research, its full potential in livestock genetics is yet to be explored. In this study, MeSH ORA was evaluated to discern biological properties of identified genes and contrast them with the results obtained from GO enrichment analysis. Three published datasets were employed for this purpose, representing a gene expression study in dairy cattle, the use of SNPs for genome-wide prediction in swine and the identification of genomic regions targeted by selection in horses. We found that several overrepresented MeSH annotations linked to these gene sets share similar concepts with those of GO terms. Moreover, MeSH yielded unique annotations, which are not directly provided by GO terms, suggesting that MeSH has the potential to refine and enrich the representation of biological knowledge. We demonstrated that MeSH can be regarded as another choice of annotation to draw biological inferences from genes identified via experimental analyses. When used in combination with GO terms, our results indicate that MeSH can enhance our functional interpretations for specific biological conditions or the genetic basis of complex traits in livestock species.
Wireless Mesh Network Routing Under Uncertain Demands
Wellons, Jonathan; Dai, Liang; Chang, Bin; Xue, Yuan
Traffic routing plays a critical role in determining the performance of a wireless mesh network. Recent research results usually fall into two ends of the spectrum. On one end are the heuristic routing algorithms, which are highly adaptive to the dynamic environments of wireless networks yet lack the analytical properties of how well the network performs globally. On the other end are the optimal routing algorithms that are derived from the optimization problem formulation of mesh network routing. They can usually claim analytical properties such as resource use optimality and throughput fairness. However, traffic demand is usually implicitly assumed as static and known a priori in these problem formulations. In contrast, recent studies of wireless network traces show that the traffic demand, even being aggregated at access points, is highly dynamic and hard to estimate. Thus, to apply the optimization-based routing solution in practice, one must take into account the dynamic and uncertain nature of wireless traffic demand. There are two basic approaches to address the traffic uncertainty in optimal mesh network routing (1) predictive routing that infers the traffic demand with maximum possibility based in its history and optimizes the routing strategy based on the predicted traffic demand and (2) oblivious routing that considers all the possible traffic demands and selects the routing strategy where the worst-case network performance could be optimized. This chapter provides an overview of the optimal routing strategies for wireless mesh networks with a focus on the above two strategies that explicitly consider the traffic uncertainty. It also identifies the key factors that affect the performance of each routing strategy and provides guidelines towards the strategy selection in mesh network routing under uncertain traffic demands.
An edge-based unstructured mesh discretisation in geospherical framework
Szmelter, Joanna; Smolarkiewicz, Piotr K.
2010-07-01
An arbitrary finite-volume approach is developed for discretising partial differential equations governing fluid flows on the sphere. Unconventionally for unstructured-mesh global models, the governing equations are cast in the anholonomic geospherical framework established in computational meteorology. The resulting discretisation retains proven properties of the geospherical formulation, while it offers the flexibility of unstructured meshes in enabling irregular spatial resolution. The latter allows for a global enhancement of the spatial resolution away from the polar regions as well as for a local mesh refinement. A class of non-oscillatory forward-in-time edge-based solvers is developed and applied to numerical examples of three-dimensional hydrostatic flows, including shallow-water benchmarks, on a rotating sphere.
Energy-efficient wireless mesh networks
Ntlatlapa, N
2007-06-01
Full Text Available deficient areas such as rural areas in Africa. Index Terms—Energy efficient design, Wireless Mesh networks, Network Protocols I. INTRODUCTION The objectives of this research group are the application and adaptation of existing wireless local area... networks, especially those based on 802.11 standard, for the energyefficient wireless mesh network (EEWMN) architectures, protocols and controls. In addition to the WMN regular features of self...
Gamra: Simple Meshes for Complex Earthquakes
Landry, Walter
2016-01-01
The static offsets caused by earthquakes are well described by elastostatic models with a discontinuity in the displacement along the fault. A traditional approach to model this discontinuity is to align the numerical mesh with the fault and solve the equations using finite elements. However, this distorted mesh can be difficult to generate and update. We present a new numerical method, inspired by the Immersed Interface Method, for solving the elastostatic equations with embedded discontinuities. This method has been carefully designed so that it can be used on parallel machines on an adapted finite difference grid. We have implemented this method in Gamra, a new code for earth modelling. We demonstrate the correctness of the method with analytic tests, and we demonstrate its practical performance by solving a realistic earthquake model to extremely high precision.
Multiresolution mesh segmentation based on surface roughness and wavelet analysis
Roudet, Céline; Dupont, Florent; Baskurt, Atilla
2007-01-01
During the last decades, the three-dimensional objects have begun to compete with traditional multimedia (images, sounds and videos) and have been used by more and more applications. The common model used to represent them is a surfacic mesh due to its intrinsic simplicity and efficacity. In this paper, we present a new algorithm for the segmentation of semi-regular triangle meshes, via multiresolution analysis. Our method uses several measures which reflect the roughness of the surface for all meshes resulting from the decomposition of the initial model into different fine-to-coarse multiresolution meshes. The geometric data decomposition is based on the lifting scheme. Using that formulation, we have compared various interpolant prediction operators, associated or not with an update step. For each resolution level, the resulting approximation mesh is then partitioned into classes having almost constant roughness thanks to a clustering algorithm. Resulting classes gather regions having the same visual appearance in term of roughness. The last step consists in decomposing the mesh into connex groups of triangles using region growing ang merging algorithms. These connex surface patches are of particular interest for adaptive mesh compression, visualisation, indexation or watermarking.
Sibeyn, J.; Rao, P; Juurlink, B.
1996-01-01
Algorithms for performing gossiping on one- and higher dimensional meshes are presented. As a routing model, we assume the practically important worm-hole routing. For one-dimensional arrays and rings, we give a novel lower bound and an asymptotically optimal gossiping algorithm for all choices of the parameters involved. For two-dimensional meshes and tori, several simple algorithms composed of one-dimensional phases are presented. For an important range of packet and mesh sizes it gives cle...
Adaptive computational methods for SSME internal flow analysis
Oden, J. T.
1986-01-01
Adaptive finite element methods for the analysis of classes of problems in compressible and incompressible flow of interest in SSME (space shuttle main engine) analysis and design are described. The general objective of the adaptive methods is to improve and to quantify the quality of numerical solutions to the governing partial differential equations of fluid dynamics in two-dimensional cases. There are several different families of adaptive schemes that can be used to improve the quality of solutions in complex flow simulations. Among these are: (1) r-methods (node-redistribution or moving mesh methods) in which a fixed number of nodal points is allowed to migrate to points in the mesh where high error is detected; (2) h-methods, in which the mesh size h is automatically refined to reduce local error; and (3) p-methods, in which the local degree p of the finite element approximation is increased to reduce local error. Two of the three basic techniques have been studied in this project: an r-method for steady Euler equations in two dimensions and a p-method for transient, laminar, viscous incompressible flow. Numerical results are presented. A brief introduction to residual methods of a-posterior error estimation is also given and some pertinent conclusions of the study are listed.
Paszyński, Maciej R.
2013-04-01
This paper describes a direct solver algorithm for a sequence of finite element meshes that are h-refined towards one or several point singularities. For such a sequence of grids, the solver delivers linear computational cost O(N) in terms of CPU time and memory with respect to the number of unknowns N. The linear computational cost is achieved by utilizing the recursive structure provided by the sequence of h-adaptive grids with a special construction of the elimination tree that allows for reutilization of previously computed partial LU (or Cholesky) factorizations over the entire unrefined part of the computational mesh. The reutilization technique reduces the computational cost of the entire sequence of h-refined grids from O(N2) down to O(N). Theoretical estimates are illustrated with numerical results on two- and three-dimensional model problems exhibiting one or several point singularities. © 2013 Elsevier Ltd. All rights reserved.
Synthesized Optimization of Triangular Mesh
HU Wenqiang; YANG Wenyu
2006-01-01
Triangular mesh is often used to describe geometric object as computed model in digital manufacture, thus the mesh model with both uniform triangular shape and excellent geometric shape is expected. But in fact, the optimization of triangular shape often is contrary with that of geometric shape. In this paper, one synthesized optimizing algorithm is presented through subdividing triangles to achieve the trade-off solution between the geometric and triangular shape optimization of mesh model. The result mesh with uniform triangular shape and excellent topology are obtained.
Quality Tetrahedral Mesh Smoothing via Boundary-Optimized Delaunay Triangulation.
Gao, Zhanheng; Yu, Zeyun; Holst, Michael
2012-12-01
Despite its great success in improving the quality of a tetrahedral mesh, the original optimal Delaunay triangulation (ODT) is designed to move only inner vertices and thus cannot handle input meshes containing "bad" triangles on boundaries. In the current work, we present an integrated approach called boundary-optimized Delaunay triangulation (B-ODT) to smooth (improve) a tetrahedral mesh. In our method, both inner and boundary vertices are repositioned by analytically minimizing the error between a paraboloid function and its piecewise linear interpolation over the neighborhood of each vertex. In addition to the guaranteed volume-preserving property, the proposed algorithm can be readily adapted to preserve sharp features in the original mesh. A number of experiments are included to demonstrate the performance of our method.
Planet-disc interaction on a freely moving mesh
Munoz, Diego J; Springel, Volker; Hernquist, Lars
2014-01-01
General-purpose, moving-mesh schemes for hydrodynamics have opened the possibility of combining the accuracy of grid-based numerical methods with the flexibility and automatic resolution adaptivity of particle-based methods. Due to their supersonic nature, Keplerian accretion discs are in principle a very attractive system for applying such freely moving mesh techniques. However, the high degree of symmetry of simple accretion disc models can be difficult to capture accurately by these methods, due to the generation of geometric grid noise and associated numerical diffusion, which is absent in polar grids. To explore these and other issues, in this work we study the idealized problem of two-dimensional planet-disc interaction with the moving-mesh code AREPO. We explore the hydrodynamic evolution of discs with planets through a series of numerical experiments that vary the planet mass, the disc viscosity and the mesh resolution, and compare the resulting surface density, vortensity field and tidal torque with ...
A multigrid method for steady Euler equations on unstructured adaptive grids
Riemslagh, Kris; Dick, Erik
1993-01-01
A flux-difference splitting type algorithm is formulated for the steady Euler equations on unstructured grids. The polynomial flux-difference splitting technique is used. A vertex-centered finite volume method is employed on a triangular mesh. The multigrid method is in defect-correction form. A relaxation procedure with a first order accurate inner iteration and a second-order correction performed only on the finest grid, is used. A multi-stage Jacobi relaxation method is employed as a smoother. Since the grid is unstructured a Jacobi type is chosen. The multi-staging is necessary to provide sufficient smoothing properties. The domain is discretized using a Delaunay triangular mesh generator. Three grids with more or less uniform distribution of nodes but with different resolution are generated by successive refinement of the coarsest grid. Nodes of coarser grids appear in the finer grids. The multigrid method is started on these grids. As soon as the residual drops below a threshold value, an adaptive refinement is started. The solution on the adaptively refined grid is accelerated by a multigrid procedure. The coarser multigrid grids are generated by successive coarsening through point removement. The adaption cycle is repeated a few times. Results are given for the transonic flow over a NACA-0012 airfoil.
Tangle-Free Mesh Motion for Ablation Simulations
Droba, Justin
2016-01-01
Problems involving mesh motion-which should not be mistakenly associated with moving mesh methods, a class of adaptive mesh redistribution techniques-are of critical importance in numerical simulations of the thermal response of melting and ablative materials. Ablation is the process by which material vaporizes or otherwise erodes due to strong heating. Accurate modeling of such materials is of the utmost importance in design of passive thermal protection systems ("heatshields") for spacecraft, the layer of the vehicle that ensures survival of crew and craft during re-entry. In an explicit mesh motion approach, a complete thermal solve is first performed. Afterwards, the thermal response is used to determine surface recession rates. These values are then used to generate boundary conditions for an a posteriori correction designed to update the location of the mesh nodes. Most often, linear elastic or biharmonic equations are used to model this material response, traditionally in a finite element framework so that complex geometries can be simulated. A simple scheme for moving the boundary nodes involves receding along the surface normals. However, for all but the simplest problem geometries, evolution in time following such a scheme will eventually bring the mesh to intersect and "tangle" with itself, inducing failure. This presentation demonstrates a comprehensive and sophisticated scheme that analyzes the local geometry of each node with help from user-provided clues to eliminate the tangle and enable simulations on a wide-class of difficult problem geometries. The method developed is demonstrated for linear elastic equations but is general enough that it may be adapted to other modeling equations. The presentation will explicate the inner workings of the tangle-free mesh motion algorithm for both two and three-dimensional meshes. It will show abstract examples of the method's success, including a verification problem that demonstrates its accuracy and
Adaptive and Iterative Methods for Simulations of Nanopores with the PNP-Stokes Equations
Mitscha-Baude, Gregor; Tulzer, Gerhard; Heitzinger, Clemens
2016-01-01
We present a 3D finite element solver for the nonlinear Poisson-Nernst-Planck (PNP) equations for electrodiffusion, coupled to the Stokes system of fluid dynamics. The model serves as a building block for the simulation of macromolecule dynamics inside nanopore sensors. We add to existing numerical approaches by deploying goal-oriented adaptive mesh refinement. To reduce the computation overhead of mesh adaptivity, our error estimator uses the much cheaper Poisson-Boltzmann equation as a simplified model, which is justified on heuristic grounds but shown to work well in practice. To address the nonlinearity in the full PNP-Stokes system, three different linearization schemes are proposed and investigated, with two segregated iterative approaches both outperforming a naive application of Newton's method. Numerical experiments are reported on a real-world nanopore sensor geometry. We also investigate two different models for the interaction of target molecules with the nanopore sensor through the PNP-Stokes equ...
Oskay Kaya
2012-01-01
Full Text Available We present a case of immediate abdominal wall reconstruction with biologic mesh following the resection of locally advanced colonic cancer. The tumor in the right colon did not respond to neoadjuvant chemotherapy. Surgical enbloc excision, including excision of the invasion in the abdominal wall, was achieved, and the defect was reconstructed with porcine dermal collagen mesh. The patient was discharged with no complication, and adaptation of the mesh was excellent at the six-month followup.
Mesh Algorithms for PDE with Sieve I: Mesh Distribution
Knepley, Matthew G
2009-01-01
We have developed a new programming framework, called Sieve, to support parallel numerical PDE algorithms operating over distributed meshes. We have also developed a reference implementation of Sieve in C++ as a library of generic algorithms operating on distributed containers conforming to the Sieve interface. Sieve makes instances of the incidence relation, or \\emph{arrows}, the conceptual first-class objects represented in the containers. Further, generic algorithms acting on this arrow container are systematically used to provide natural geometric operations on the topology and also, through duality, on the data. Finally, coverings and duality are used to encode not only individual meshes, but all types of hierarchies underlying PDE data structures, including multigrid and mesh partitions. In order to demonstrate the usefulness of the framework, we show how the mesh partition data can be represented and manipulated using the same fundamental mechanisms used to represent meshes. We present the complete des...
Prevention of Adhesion to Prosthetic Mesh
van ’t Riet, Martijne; de Vos van Steenwijk, Peggy J.; Bonthuis, Fred; Marquet, Richard L.; Steyerberg, Ewout W.; Jeekel, Johannes; Bonjer, H. Jaap
2003-01-01
Objective To assess whether use of antiadhesive liquids or coatings could prevent adhesion formation to prosthetic mesh. Summary Background Data Incisional hernia repair frequently involves the use of prosthetic mesh. However, concern exists about development of adhesions between viscera and the mesh, predisposing to intestinal obstruction or enterocutaneous fistulas. Methods In 91 rats, a defect in the muscular abdominal wall was created, and mesh was fixed intraperitoneally to cover the defect. Rats were divided in five groups: polypropylene mesh only (control group), addition of Sepracoat or Icodextrin solution to polypropylene mesh, Sepramesh (polypropylene mesh with Seprafilm coating), and Parietex composite mesh (polyester mesh with collagen coating). Seven and 30 days postoperatively, adhesions were assessed and wound healing was studied by microscopy. Results Intraperitoneal placement of polypropylene mesh was followed by bowel adhesions to the mesh in 50% of the cases. A mean of 74% of the mesh surface was covered by adhesions after 7 days, and 48% after 30 days. Administration of Sepracoat or Icodextrin solution had no influence on adhesion formation. Coated meshes (Sepramesh and Parietex composite mesh) had no bowel adhesions. Sepramesh was associated with a significant reduction of the mesh surface covered by adhesions after 7 and 30 days. Infection was more prevalent with Parietex composite mesh, with concurrent increased mesh surface covered by adhesions after 30 days (78%). Conclusions Sepramesh significantly reduced mesh surface covered by adhesions and prevented bowel adhesion to the mesh. Parietex composite mesh prevented bowel adhesions as well but increased infection rates in the current model. PMID:12496539
Risk Factors for Mesh Exposure after Transvaginal Mesh Surgery
Ke Niu; Yong-Xian Lu; Wen-Jie Shen; Ying-Hui Zhang; Wen-Ying Wang
2016-01-01
Background:Mesh exposure after surgery continues to be a clinical challenge for urogynecological surgeons.The purpose of this study was to explore the risk factors for polypropylene (PP) mesh exposure after transvaginal mesh (TVM) surgery.Methods:This study included 195 patients with advanced pelvic organ prolapse (POP),who underwent TVM from January 2004to December 2012 at the First Affiliated Hospital of Chinese PLA General Hospital.Clinical data were evaluated including patient's demography,TVM type,concomitant procedures,operation time,blood loss,postoperative morbidity,and mesh exposure.Mesh exposure was identified through postoperative vaginal examination.Statistical analysis was performed to identify risk factors for mesh exposure.Results:Two-hundred and nine transvaginal PP meshes were placed,including 194 in the anterior wall and 15 in the posterior wall.Concomitant tension-free vaginal tape was performed in 61 cases.The mean follow-up time was 35.1 ± 23.6 months.PP mesh exposure was identified in 32 cases (16.4％),with 31 in the anterior wall and 1 in the posterior wall.Significant difference was found in operating time and concomitant procedures between exposed and nonexposed groups (F =7.443,P =0.007;F =4.307,P =0.039,respectively).Binary logistic regression revealed that the number of concomitant procedures and operation time were risk factors for mesh exposure (P =0.001,P =0.043).Conclusion:Concomitant procedures and increased operating time increase the risk for postoperative mesh exposure in patients undergoing TVM surgery for POP.
OPTIMIZING EUCALYPTUS PULP REFINING
VailManfredi
2004-01-01
This paper discusses the refining of bleachedeucalyptus kraft pulp (BEKP).Pilot plant tests were carded out in to optimize therefining process and to identify the effects of refiningvariables on final paper quality and process costs.The following parameters are discussed: pulpconsistency, disk pattern design, refiner speed,energy input, refiner configuration (parallel or serial)and refining intensity.The effects of refining on pulp fibers were evaluatedagainst the pulp quality properties, such as physicalstrengths, bulk, opacity and porosity, as well as theinteractions with papermaking process, such as papermachine runnability, paper breaks and refiningcontrol.The results showed that process optimization,considering pulp quality and refining costs, wereobtained when eucalyptus pulp is refined under thelowest intensity and the highest pulp consistencypossible. Changes on the operational refiningconditions will have the highest impact on totalenergy requirements (costs) without any significanteffect on final paper properties.It was also observed that classical ways to control theindustrial operation, such as those based on drainagemeasurements, do not represent the best alternative tomaximize the final paper properties neither the papermachine runability.
Refinement of the Computerized Adaptive Screening Test
1990-11-01
Representative Clinton B. Walker I Selection and Classification Technical Area Michael Rumsey, Chief Manpower and Personnel Research Zita M. Simutis... Crichton (1979) added random values to item parameters to study the effects of item parameter estimation errors. Wetzel and McBride (1983) simulated reality
Parameterization for fitting triangular mesh
LIN Hongwei; WANG Guojin; LIU Ligang; BAO Hujun
2006-01-01
In recent years, with the development of 3D data acquisition equipments, the study on reverse engineering has become more and more important. However, the existing methods for parameterization can hardly ensure that the parametric domain is rectangular, and the parametric curve grid is regular. In order to overcome these limitations, we present a novel method for parameterization of triangular meshes in this paper. The basic idea is twofold: first, because the isotherms in the steady temperature do not intersect with each other, and are distributed uniformly, no singularity (fold-over) exists in the parameterization; second, a 3D harmonic equation is solved by the finite element method to obtain the steady temperature field on a 2D triangular mesh surface with four boundaries. Therefore, our proposed method avoids the embarrassment that it is impossible to solve the 2D quasi-harmonic equation on the 2D triangular mesh without the parametric values at mesh vertices. Furthermore, the isotherms on the temperature field are taken as a set of iso-parametric curves on the triangular mesh surface. The other set of iso-parametric curves can be obtained by connecting the points with the same chord-length on the isotherms sequentially. The obtained parametric curve grid is regular, and distributed uniformly, and can map the triangular mesh surface to the unit square domain with boundaries of mesh surface to boundaries of parametric domain, which ensures that the triangular mesh surface or point cloud can be fitted with the NURBS surface.
Guaranteed-Quality Triangular Meshes
1989-04-01
Defense Ad, : ed Research P: jects Pgency or the U.S- Gower ment° iI Guaranteed-Quality Triangular Meshes DTIC ELECTE L. Paul Chew* JUL 1419891 TR 89-983 S... Wittchen , M. S. Shephard, K. R. Grice, and M. A. Yerry, Robust, geometrically based, automatic two-dimensional mesh generation, International Journal for
An Improved Moving Mesh Algorithm
无
2001-01-01
we consider an iterative algorithm of mesh optimization for finite element solution, and give an improved moving mesh strategy that reduces rapidly the complexity and cost of solving variational problems.A numerical result is presented for a 2-dimensional problem by the improved algorithm.
Particle Mesh Hydrodynamics for Astrophysics Simulations
Chatelain, Philippe; Cottet, Georges-Henri; Koumoutsakos, Petros
We present a particle method for the simulation of three dimensional compressible hydrodynamics based on a hybrid Particle-Mesh discretization of the governing equations. The method is rooted on the regularization of particle locations as in remeshed Smoothed Particle Hydrodynamics (rSPH). The rSPH method was recently introduced to remedy problems associated with the distortion of computational elements in SPH, by periodically re-initializing the particle positions and by using high order interpolation kernels. In the PMH formulation, the particles solely handle the convective part of the compressible Euler equations. The particle quantities are then interpolated onto a mesh, where the pressure terms are computed. PMH, like SPH, is free of the convection CFL condition while at the same time it is more efficient as derivatives are computed on a mesh rather than particle-particle interactions. PMH does not detract from the adaptive character of SPH and allows for control of its accuracy. We present simulations of a benchmark astrophysics problem demonstrating the capabilities of this approach.
Refined Semilattices of Semigroups
Liang Zhang; K.P. Shum; Ronghua Zhang
2001-01-01
In this paper, we introduce the concept of refined semilattices of semigroups. This is a modified concept of the generally strong semilattice of semigroups initiated by Zhang and Huang. By using the concept of generally strong semilattice, Zhang and Huang showed that a regular band can be expressed by a generally strong semilattice of rectangular bands. However, the proof of the associativity for the multiplication is not complete and there exist some gaps in their construction of regular bands. We now revise the generally strong semilattices and call them refined semilattices. In this way, we are able to remove the gaps,and the associative law of the multiplication can be verified. As an application, we prove that a band is regular if and only if it is a refined semilattice of rectangular bands. In fact, refined semilattices provide a new device in the construction of new semigroups from the old ones.
Lyra, Paulo Roberto Maciel [Pernambuco Univ., Recife, PE (Brazil). Dept. de Engenharia Civil
1990-12-31
This work describes a procedure for the adaptive time dependent Finite Element Method using an automatic mesh refinement (H-Version) that efficiently reduces estimated errors ( a posteriori) below pre-assigned limits. Classical model problem for steady-state heat transfer are investigated, and the results are compared with the analytical solution. Then some typical time-dependent problem are qualitatively analysed. (author) 10 refs., 7 figs.
A two-dimensional adaptive spectral element method for the direct simulation of incompressible flow
Hsu, Li-Chieh
The spectral element method is a high order discretization scheme for the solution of nonlinear partial differential equations. The method draws its strengths from the finite element method for geometrical flexibility and spectral methods for high accuracy. Although the method is, in theory, very powerful for complex phenomena such as transitional flows, its practical implementation is limited by the arbitrary choice of domain discretization. For instance, it is hard to estimate the appropriate number of elements for a specific case. Selection of regions to be refined or coarsened is difficult especially as the flow becomes more complex and memory limits of the computer are stressed. We present an adaptive spectral element method in which the grid is automatically refined or coarsened in order to capture underresolved regions of the domain and to follow regions requiring high resolution as they develop in time. The objective is to provide the best and most efficient solution to a time-dependent nonlinear problem by continually optimizing resource allocation. The adaptivity is based on an error estimator which determines which regions need more resolution. The solution strategy is as follows: compute an initial solution with a suitable initial mesh, estimate errors in the solution locally in each element, modify the mesh according to the error estimators, interpolate old mesh solutions onto the new elements, and resume the numerical solution process. A two-dimensional adaptive spectral element method for the direct simulation of incompressible flows has been developed. The adaptive algorithm effectively diagnoses and refines regions of the flow where complexity of the solution requires increased resolution. The method has been demonstrated on two-dimensional examples in heat conduction, Stokes and Navier-Stokes flows.
An adaptive projection method for the incompressible Euler equations
Almgren, Ann S.; Bell, John B.; Colella, Phillip; Howell, Louis H.
A new adaptive projection method has been developed for time-dependent incompressible variable density flow. The levels in the adaptive mesh hierarchy are refined in both space and time. The advection step takes place on individual grids in an approach similar to that of the single grid method. The projection at each level is similar to the uniform grid projection but must now incorporate multiple grids per level. A sync projection is introduced which is needed to synchronize the solution at each level l with the data at the levels above it at the end of each level l time step. This adaptive projection method is second-order accurate and provides an accurate and efficient tool for modeling variable density flows.
Refinement by interface instantiation
Hallerstede, Stefan; Hoang, Thai Son
2012-01-01
be easily refined. Our first contribution hence is a proposal for a new construct called interface that encapsulates the external variables, along with a mechanism for interface instantiation. Using the new construct and mechanism, external variables can be refined consistently. Our second contribution...... is an approach for verifying the correctness of Event-B extensions using the supporting Rodin tool. We illustrate our approach by proving the correctness of interface instantiation....
NAFTA opportunities: Petroleum refining
1993-01-01
The North American Free Trade Agreement (NAFTA) creates a more transparent environment for the sale of refined petroleum products to Mexico, and locks in access to Canada's relatively open market for these products. Canada and Mexico are sizable United States export markets for refined petroleum products, with exports of $556 million and $864 million, respectively, in 1992. These markets represent approximately 24 percent of total U.S. exports of these goods.
Bozzelli, Laura; French, Tim; Hales, James; Pinchinat, Sophie
2012-01-01
In this paper we present refinement modal logic. A refinement is like a bisimulation, except that from the three relational requirements only 'atoms' and 'back' need to be satisfied. Our logic contains a new operator 'forall' in additional to the standard modalities 'Box' for each agent. The operator 'forall' acts as a quantifier over the set of all refinements of a given model. We call it the refinement operator. As a variation on a bisimulation quantifier, it can be seen as a refinement quantifier over a variable not occurring in the formula bound by the operator. The logic combines the simplicity of multi-agent modal logic with some powers of monadic second order quantification. We present a sound and complete axiomatization of multiagent refinement modal logic. We also present an extension of the logic to the modal mu-calculus, and an axiomatization for the single-agent version of this logic. Examples and applications are also discussed: to software verification and design (the set of agents can also be s...
Streaming Compression of Hexahedral Meshes
Isenburg, M; Courbet, C
2010-02-03
We describe a method for streaming compression of hexahedral meshes. Given an interleaved stream of vertices and hexahedral our coder incrementally compresses the mesh in the presented order. Our coder is extremely memory efficient when the input stream documents when vertices are referenced for the last time (i.e. when it contains topological finalization tags). Our coder then continuously releases and reuses data structures that no longer contribute to compressing the remainder of the stream. This means in practice that our coder has only a small fraction of the whole mesh in memory at any time. We can therefore compress very large meshes - even meshes that do not file in memory. Compared to traditional, non-streaming approaches that load the entire mesh and globally reorder it during compression, our algorithm trades a less compact compressed representation for significant gains in speed, memory, and I/O efficiency. For example, on the 456k hexahedra 'blade' mesh, our coder is twice as fast and uses 88 times less memory (only 3.1 MB) with the compressed file increasing about 3% in size. We also present the first scheme for predictive compression of properties associated with hexahedral cells.
Advances in the development of wire mesh reactor for coal gasification studies.
Zeng, Cai; Chen, Lei; Liu, Gang; Li, Wenhua; Huang, Baoming; Zhu, Hongdong; Zhang, Bing; Zamansky, Vladimir
2008-08-01
In an effort to further understand the coal gasification behavior in entrained-flow gasifiers, a high pressure and high temperature wire mesh reactor with new features was recently built. An advanced LABVIEW-based temperature measurement and control system were adapted. Molybdenum wire mesh with aperture smaller than 70 mum and type D thermocouple were used to enable high carbon conversion (>90%) at temperatures >1000 degrees C. Gaseous species from wire mesh reactor were quantified using a high sensitivity gas chromatography. The material balance of coal pyrolysis in wire mesh reactor was demonstrated for the first time by improving the volatile's quantification techniques.
Nanowire mesh solar fuels generator
Yang, Peidong; Chan, Candace; Sun, Jianwei; Liu, Bin
2016-05-24
This disclosure provides systems, methods, and apparatus related to a nanowire mesh solar fuels generator. In one aspect, a nanowire mesh solar fuels generator includes (1) a photoanode configured to perform water oxidation and (2) a photocathode configured to perform water reduction. The photocathode is in electrical contact with the photoanode. The photoanode may include a high surface area network of photoanode nanowires. The photocathode may include a high surface area network of photocathode nanowires. In some embodiments, the nanowire mesh solar fuels generator may include an ion conductive polymer infiltrating the photoanode and the photocathode in the region where the photocathode is in electrical contact with the photoanode.
Goal-Oriented Self-Adaptive hp Finite Element Simulation of 3D DC Borehole Resistivity Simulations
Calo, Victor M.
2011-05-14
In this paper we present a goal-oriented self-adaptive hp Finite Element Method (hp-FEM) with shared data structures and a parallel multi-frontal direct solver. The algorithm automatically generates (without any user interaction) a sequence of meshes delivering exponential convergence of a prescribed quantity of interest with respect to the number of degrees of freedom. The sequence of meshes is generated from a given initial mesh, by performing h (breaking elements into smaller elements), p (adjusting polynomial orders of approximation) or hp (both) refinements on the finite elements. The new parallel implementation utilizes a computational mesh shared between multiple processors. All computational algorithms, including automatic hp goal-oriented adaptivity and the solver work fully in parallel. We describe the parallel self-adaptive hp-FEM algorithm with shared computational domain, as well as its efficiency measurements. We apply the methodology described to the three-dimensional simulation of the borehole resistivity measurement of direct current through casing in the presence of invasion.
Khayat, Roger E.; Genouvrier, Delphine
2001-05-01
An adaptive (Lagrangian) boundary element approach is proposed for the general three-dimensional simulation of confined free-surface Stokes flow. The method is stable as it includes remeshing capabilities of the deforming free surface and thus can handle large deformations. A simple algorithm is developed for mesh refinement of the deforming free-surface mesh. Smooth transition between large and small elements is achieved without significant degradation of the aspect ratio of the elements in the mesh. Several flow problems are presented to illustrate the utility of the approach, particularly as encountered in polymer processing and rheology. These problems illustrate the transient nature of the flow during the processes of extrusion and thermoforming, the elongation of a fluid sample in an extensional rheometer, and the coating of a sphere. Surface tension effects are also explored. Copyright
An unstructured-mesh atmospheric model for nonhydrostatic dynamics: Towards optimal mesh resolution
Szmelter, Joanna; Zhang, Zhao; Smolarkiewicz, Piotr K.
2015-08-01
The paper advances the limited-area anelastic model (Smolarkiewicz et al. (2013) [45]) for investigation of nonhydrostatic dynamics in mesoscale atmospheric flows. New developments include the extension to a tetrahedral-based median-dual option for unstructured meshes and a static mesh adaptivity technique using an error indicator based on inherent properties of the Multidimensional Positive Definite Advection Transport Algorithm (MPDATA). The model employs semi-implicit nonoscillatory forward-in-time integrators for soundproof PDEs, built on MPDATA and a robust non-symmetric Krylov-subspace elliptic solver. Finite-volume spatial discretisation adopts an edge-based data structure. Simulations of stratified orographic flows and the associated gravity-wave phenomena in media with uniform and variable dispersive properties verify the advancement and demonstrate the potential of heterogeneous anisotropic discretisation with large variation in spatial resolution for study of complex stratified flows that can be computationally unattainable with regular grids.
Grid refinement for entropic lattice Boltzmann models.
Dorschner, B; Frapolli, N; Chikatamarla, S S; Karlin, I V
2016-11-01
We propose a multidomain grid refinement technique with extensions to entropic incompressible, thermal, and compressible lattice Boltzmann models. Its validity and accuracy are assessed by comparison to available direct numerical simulation and experiment for the simulation of isothermal, thermal, and viscous supersonic flow. In particular, we investigate the advantages of grid refinement for the setups of turbulent channel flow, flow past a sphere, Rayleigh-Bénard convection, as well as the supersonic flow around an airfoil. Special attention is paid to analyzing the adaptive features of entropic lattice Boltzmann models for multigrid simulations.
Grid refinement for entropic lattice Boltzmann models
Dorschner, B; Chikatamarla, S S; Karlin, I V
2016-01-01
We propose a novel multi-domain grid refinement technique with extensions to entropic incompressible, thermal and compressible lattice Boltzmann models. Its validity and accuracy are accessed by comparison to available direct numerical simulation and experiment for the simulation of isothermal, thermal and viscous supersonic flow. In particular, we investigate the advantages of grid refinement for the set-ups of turbulent channel flow, flow past a sphere, Rayleigh-Benard convection as well as the supersonic flow around an airfoil. Special attention is payed to analyzing the adaptive features of entropic lattice Boltzmann models for multi-grid simulations.
Managing chronic pelvic pain following reconstructive pelvic surgery with transvaginal mesh.
Gyang, Anthony N; Feranec, Jessica B; Patel, Rakesh C; Lamvu, Georgine M
2014-03-01
In 2001, the US Food and Drug Administration (FDA) approved the first transvaginal mesh kit to treat pelvic organ prolapse (POP). Since the introduction of vaginal mesh kits, some vaginal meshes have been associated with chronic pelvic pain after reconstructive pelvic floor surgery. Pelvic pain results in between 0 % and 30 % of patients following transvaginal mesh placement. Common causes of chronic pelvic pain include pelvic floor muscle spasm, pudendal neuralgia, and infection. Paucity of data exists on the effective management of chronic pelvic pain after pelvic reconstructive surgery with mesh. We outline the management of chronic pelvic pain after transvaginal mesh placement for reconstructive pelvic floor repair based on our clinical experience and adaptation of data used in other aspects of managing chronic pelvic pain conditions.
Toward Interoperable Mesh, Geometry and Field Components for PDE Simulation Development
Chand, K K; Diachin, L F; Li, X; Ollivier-Gooch, C; Seol, E S; Shephard, M; Tautges, T; Trease, H
2005-07-11
Mesh-based PDE simulation codes are becoming increasingly sophisticated and rely on advanced meshing and discretization tools. Unfortunately, it is still difficult to interchange or interoperate tools developed by different communities to experiment with various technologies or to develop new capabilities. To address these difficulties, we have developed component interfaces designed to support the information flow of mesh-based PDE simulations. We describe this information flow and discuss typical roles and services provided by the geometry, mesh, and field components of the simulation. Based on this delineation for the roles of each component, we give a high-level description of the abstract data model and set of interfaces developed by the Department of Energy's Interoperable Tools for Advanced Petascale Simulation (ITAPS) center. These common interfaces are critical to our interoperability goal, and we give examples of several services based upon these interfaces including mesh adaptation and mesh improvement.
Constancio, Silva
2006-07-01
In 2004, refining margins showed a clear improvement that persisted throughout the first three quarters of 2005. This enabled oil companies to post significantly higher earnings for their refining activity in 2004 compared to 2003, with the results of the first half of 2005 confirming this trend. As for petrochemicals, despite a steady rise in the naphtha price, higher cash margins enabled a turnaround in 2004 as well as a clear improvement in oil company financial performance that should continue in 2005, judging by the net income figures reported for the first half-year. Despite this favorable business environment, capital expenditure in refining and petrochemicals remained at a low level, especially investment in new capacity, but a number of projects are being planned for the next five years. (author)
Mersiline mesh in premaxillary augmentation.
Foda, Hossam M T
2005-01-01
Premaxillary retrusion may distort the aesthetic appearance of the columella, lip, and nasal tip. This defect is characteristically seen in, but not limited to, patients with cleft lip nasal deformity. This study investigated 60 patients presenting with premaxillary deficiencies in which Mersiline mesh was used to augment the premaxilla. All the cases had surgery using the external rhinoplasty technique. Two methods of augmentation with Mersiline mesh were used: the Mersiline roll technique, for the cases with central symmetric deficiencies, and the Mersiline packing technique, for the cases with asymmetric deficiencies. Premaxillary augmentation with Mersiline mesh proved to be simple technically, easy to perform, and not associated with any complications. Periodic follow-up evaluation for a mean period of 32 months (range, 12-98 months) showed that an adequate degree of premaxillary augmentation was maintained with no clinically detectable resorption of the mesh implant.
Unterweger, K.
2015-01-01
© Springer International Publishing Switzerland 2015. We propose to couple our adaptive mesh refinement software PeanoClaw with existing solvers for complex overland flows that are tailored to regular Cartesian meshes. This allows us to augment them with spatial adaptivity and local time-stepping without altering the computational kernels. FullSWOF2D—Full Shallow Water Overland Flows—here is our software of choice though all paradigms hold for other solvers as well.We validate our hybrid simulation software in an artificial test scenario before we provide results for a large-scale flooding scenario of the Mecca region. The latter demonstrates that our coupling approach enables the simulation of complex “real-world” scenarios.
YUAN Si; HE Xue-feng
2006-01-01
Based on the newly-developed element energy projection (EEP) method for computation of super-convergent results in one-dimensional finite element method (FEM),the task of self-adaptive FEM analysis was converted into the task of adaptive piecewise polynomial interpolation. As a result, a satisfactory FEM mesh can be obtained, and further FEM analysis on this mesh would immediately produce an FEM solution which usually satisfies the user specified error tolerance. Even though the error tolerance was not completely satisfied, one or two steps of further local refinements would be sufficient.This strategy was found to be very simple, rapid, cheap and efficient. Taking the elliptical ordinary differential equation of second order as the model problem, the fundamental idea,implementation strategy and detailed algorithm are described. Representative numerical examples are given to show the effectiveness and reliability of the proposed approach.
GENERATION OF IRREGULAR HEXAGONAL MESHES
Vlasov Aleksandr Nikolaevich
2012-07-01
Decomposition is performed in a constructive way and, as option, it involves meshless representation. Further, this mapping method is used to generate the calculation mesh. In this paper, the authors analyze different cases of mapping onto simply connected and bi-connected canonical domains. They represent forward and backward mapping techniques. Their potential application for generation of nonuniform meshes within the framework of the asymptotic homogenization theory is also performed to assess and project effective characteristics of heterogeneous materials (composites.
Verborgh, Ruben
2013-01-01
The book is styled on a Cookbook, containing recipes - combined with free datasets - which will turn readers into proficient OpenRefine users in the fastest possible way.This book is targeted at anyone who works on or handles a large amount of data. No prior knowledge of OpenRefine is required, as we start from the very beginning and gradually reveal more advanced features. You don't even need your own dataset, as we provide example data to try out the book's recipes.
Image-driven mesh optimization
Lindstrom, P; Turk, G
2001-01-05
We describe a method of improving the appearance of a low vertex count mesh in a manner that is guided by rendered images of the original, detailed mesh. This approach is motivated by the fact that greedy simplification methods often yield meshes that are poorer than what can be represented with a given number of vertices. Our approach relies on edge swaps and vertex teleports to alter the mesh connectivity, and uses the downhill simplex method to simultaneously improve vertex positions and surface attributes. Note that this is not a simplification method--the vertex count remains the same throughout the optimization. At all stages of the optimization the changes are guided by a metric that measures the differences between rendered versions of the original model and the low vertex count mesh. This method creates meshes that are geometrically faithful to the original model. Moreover, the method takes into account more subtle aspects of a model such as surface shading or whether cracks are visible between two interpenetrating parts of the model.
Method and system for mesh network embedded devices
Wang, Ray (Inventor)
2009-01-01
A method and system for managing mesh network devices. A mesh network device with integrated features creates an N-way mesh network with a full mesh network topology or a partial mesh network topology.
Improved AFEM algorithm for bioluminescence tomography based on dual-mesh alternation strategy
Wei Li; Heng Zhao; Xiaochao Qu; Yanbin Hou; Xueli Chen; Duofang Chen; Xiaowei He; Qitan Zhang; Jimin Liang
2012-01-01
Adaptive finite element method (AFEM) is broadly adopted to recover the internal source in biological tissues.In this letter,a novel dual-mesh alternation strategy (dual-mesh AFEM) is developed for bioluminescence tomography.By comprehensively considering the error estimation of the finite element method solution on each mesh,two different adaptive strategies based on the error indicator of the reconstructed source and the photon flux density are used alternately in the process.Combined with the constantly adjusted permissible region in the adaptive process,the new algorithm can achieve a more accurate source location compared with the AFEM in the previous experiments.%Adaptive finite element method (AFEM) is broadly adopted to recover the internal source in biological tissues. In this letter, a novel dual-mesh alternation strategy (dual-mesh AFEM) is developed for biolumi-nescence tomography. By comprehensively considering the error estimation of the finite element method solution on each mesh, two different adaptive strategies based on the error indicator of the reconstructed source and the photon flux density are used alternately in the process. Combined with the constantly adjusted permissible region in the adaptive process, the new algorithm can achieve a more accurate source location compared with the AFEM in the previous experiments.
Incremental Bisimulation Abstraction Refinement
Godskesen, Jens Christian; Song, Lei; Zhang, Lijun
2013-01-01
an abstraction refinement approach for the probabilistic computation tree logic (PCTL), which is based on incrementally computing a sequence of may- and must-quotient automata. These are induced by depth-bounded bisimulation equivalences of increasing depth. The approach is both sound and complete, since...
User Manual for the PROTEUS Mesh Tools
Smith, Micheal A. [Argonne National Lab. (ANL), Argonne, IL (United States); Shemon, Emily R. [Argonne National Lab. (ANL), Argonne, IL (United States)
2015-06-01
This report describes the various mesh tools that are provided with the PROTEUS code giving both descriptions of the input and output. In many cases the examples are provided with a regression test of the mesh tools. The most important mesh tools for any user to consider using are the MT_MeshToMesh.x and the MT_RadialLattice.x codes. The former allows the conversion between most mesh types handled by PROTEUS while the second allows the merging of multiple (assembly) meshes into a radial structured grid. Note that the mesh generation process is recursive in nature and that each input specific for a given mesh tool (such as .axial or .merge) can be used as “mesh” input for any of the mesh tools discussed in this manual.
Kim, D.; Ghanem, R. [State Univ. of New York, Buffalo, NY (United States)
1994-12-31
Multigrid solution technique to solve a material nonlinear problem in a visual programming environment using the finite element method is discussed. The nonlinear equation of equilibrium is linearized to incremental form using Newton-Rapson technique, then multigrid solution technique is used to solve linear equations at each Newton-Rapson step. In the process, adaptive mesh refinement, which is based on the bisection of a pair of triangles, is used to form grid hierarchy for multigrid iteration. The solution process is implemented in a visual programming environment with distributed computing capability, which enables more intuitive understanding of solution process, and more effective use of resources.
Parallel Programming Strategies for Irregular Adaptive Applications
Biswas, Rupak; Biegel, Bryan (Technical Monitor)
2001-01-01
Achieving scalable performance for dynamic irregular applications is eminently challenging. Traditional message-passing approaches have been making steady progress towards this goal; however, they suffer from complex implementation requirements. The use of a global address space greatly simplifies the programming task, but can degrade the performance for such computations. In this work, we examine two typical irregular adaptive applications, Dynamic Remeshing and N-Body, under competing programming methodologies and across various parallel architectures. The Dynamic Remeshing application simulates flow over an airfoil, and refines localized regions of the underlying unstructured mesh. The N-Body experiment models two neighboring Plummer galaxies that are about to undergo a merger. Both problems demonstrate dramatic changes in processor workloads and interprocessor communication with time; thus, dynamic load balancing is a required component.
Tetrahedral meshing via maximal Poisson-disk sampling
Guo, Jianwei
2016-02-15
In this paper, we propose a simple yet effective method to generate 3D-conforming tetrahedral meshes from closed 2-manifold surfaces. Our approach is inspired by recent work on maximal Poisson-disk sampling (MPS), which can generate well-distributed point sets in arbitrary domains. We first perform MPS on the boundary of the input domain, we then sample the interior of the domain, and we finally extract the tetrahedral mesh from the samples by using 3D Delaunay or regular triangulation for uniform or adaptive sampling, respectively. We also propose an efficient optimization strategy to protect the domain boundaries and to remove slivers to improve the meshing quality. We present various experimental results to illustrate the efficiency and the robustness of our proposed approach. We demonstrate that the performance and quality (e.g., minimal dihedral angle) of our approach are superior to current state-of-the-art optimization-based approaches.
The evolution of oil refining in Europe
Reid, A. [CONCAWE, Brussels (Belgium)
2013-04-01
Back in 1963 when CONCAWE was founded, the world looked very different from what it is today, and so did the global and European refining industry. Oil product markets were expanding fast and new refineries were being built at a steady rate. The oil crisis of the 1970s brought an abrupt end to this, heralding a long era of consolidation and stepwise adaptation. At the same time the nature of the global oil business shifted from fully integrated companies producing, transporting and refining their own oil to a much more diversified situation where oil production ('upstream') and refining/distribution ('downstream') gradually became two essentially separate businesses. From being purely a 'cost centre' in an integrated chain, refining has become a separate activity in its own right, operating as a 'profit centre' between two global markets - crude oil and products - which, although not entirely independent, have their own dynamics and influences. In addition demand gradually shifted towards lighter products while the quality requirements on all products were considerably tightened. This article explores the new challenges that these changes have imposed on EU refiners, and describes CONCAWE's contributions to understanding their impact on refinery production and investments.
Connectivity editing for quadrilateral meshes
Peng, Chihan
2011-12-12
We propose new connectivity editing operations for quadrilateral meshes with the unique ability to explicitly control the location, orientation, type, and number of the irregular vertices (valence not equal to four) in the mesh while preserving sharp edges. We provide theoretical analysis on what editing operations are possible and impossible and introduce three fundamental operations to move and re-orient a pair of irregular vertices. We argue that our editing operations are fundamental, because they only change the quad mesh in the smallest possible region and involve the fewest irregular vertices (i.e., two). The irregular vertex movement operations are supplemented by operations for the splitting, merging, canceling, and aligning of irregular vertices. We explain how the proposed highlevel operations are realized through graph-level editing operations such as quad collapses, edge flips, and edge splits. The utility of these mesh editing operations are demonstrated by improving the connectivity of quad meshes generated from state-of-art quadrangulation techniques. © 2011 ACM.
Connectivity editing for quadrilateral meshes
Peng, Chihan
2011-12-01
We propose new connectivity editing operations for quadrilateral meshes with the unique ability to explicitly control the location, orientation, type, and number of the irregular vertices (valence not equal to four) in the mesh while preserving sharp edges. We provide theoretical analysis on what editing operations are possible and impossible and introduce three fundamental operations to move and re-orient a pair of irregular vertices. We argue that our editing operations are fundamental, because they only change the quad mesh in the smallest possible region and involve the fewest irregular vertices (i.e., two). The irregular vertex movement operations are supplemented by operations for the splitting, merging, canceling, and aligning of irregular vertices. We explain how the proposed high-level operations are realized through graph-level editing operations such as quad collapses, edge flips, and edge splits. The utility of these mesh editing operations are demonstrated by improving the connectivity of quad meshes generated from state-of-art quadrangulation techniques.
Mesh Router Nodes placement in Rural Wireless Mesh Networks
Ebongue, Jean Louis Fendji Kedieng; Thron, Christopher; Nlong, Jean Michel
2015-01-01
The problem of placement of mesh router nodes in Wireless Mesh Networks is known to be a NP hard problem. In this paper, the problem is addressed under a constraint of network model tied to rural regions where we usually observe low density and sparse population. We consider the area to cover as decomposed into a set of elementary areas which can be required or optional in terms of coverage and where a node can be placed or not. We propose an effective algorithm to ensure the coverage. This a...
Application of new boundary method in Cartesian-mesh%一种新的边界处理方法在笛卡尔网格中的应用
韩玉琪; 江立军; 张常贤; 高歌
2012-01-01
采用自适应笛卡尔网格方法求解Euler方程,固壁边界条件通过一种离散强迫浸入边界方法引入,边界切割网格与流场网格保持一致,从而有效解决了小切割网格单元的时间步长限制问题.针对自适应笛卡尔网格的特点提出了新的虚拟点反射方法和插值方法,解决了数值吸力峰值问题.对RAE2822翼型、双NACA0012翼型和Suddhoo三元翼型的流动状况进行了无黏数值模拟,并与现有的理论解、单域以及分区结构网格解进行了对比.结果表明：该离散强迫浸入边界方法有效地引进了固壁边界条件,结合自适应笛卡尔网格技术能够准确模拟复杂几何外形下的流动状况.%Adaptively-refined Cartesian-mesh approaches were employed for the computation of Euler equations,solid wall boundary was introduced by a discrete forcing immersed boundary approach,and shape of cut cell kept in accordance with that in the flow field,which efficiently relieved the time step restriction of small cut cell.To remedy the numerical suction peak,a novel method for reflection of ghost point along with interpolation scheme were devised which were applied for adaptively-refined Cartesian-mesh.Inviscid flows around RAE 2822 airfoil,double NACA0012 airfoil and Suddhoo three-element airfoil were simulated and compared with published theoretical and single/multi domain structured mesh results.The results show that solid wall is efficiently resolved by the current discrete forcing immersed boundary approach and flows around complex geometries are adequately simulated by adaptively-refined Cartesian-mesh approaches.
Acquiring Plausible Predications from MEDLINE by Clustering MeSH Annotations.
Miñarro-Giménez, Jose Antonio; Kreuzthaler, Markus; Bernhardt-Melischnig, Johannes; Martínez-Costa, Catalina; Schulz, Stefan
2015-01-01
The massive accumulation of biomedical knowledge is reflected by the growth of the literature database MEDLINE with over 23 million bibliographic records. All records are manually indexed by MeSH descriptors, many of them refined by MeSH subheadings. We use subheading information to cluster types of MeSH descriptor co-occurrences in MEDLINE by processing co-occurrence information provided by the UMLS. The goal is to infer plausible predicates to each resulting cluster. In an initial experiment this was done by grouping disease-pharmacologic substance co-occurrences into six clusters. Then, a domain expert manually performed the assignment of meaningful predicates to the clusters. The mean accuracy of the best ten generated biomedical facts of each cluster was 85%. This result supports the evidence of the potential of MeSH subheadings for extracting plausible medical predications from MEDLINE.
Perspicuity and Granularity in Refinement
Boiten, Eerke
2011-01-01
This paper reconsiders refinements which introduce actions on the concrete level which were not present at the abstract level. It draws a distinction between concrete actions which are "perspicuous" at the abstract level, and changes of granularity of actions between different levels of abstraction. The main contribution of this paper is in exploring the relation between these different methods of "action refinement", and the basic refinement relation that is used. In particular, it shows how the "refining skip" method is incompatible with failures-based refinement relations, and consequently some decisions in designing Event-B refinement are entangled.
Adaptive particle-cell algorithm for Fokker-Planck based rarefied gas flow simulations
Pfeiffer, M.; Gorji, M. H.
2017-04-01
Recently, the Fokker-Planck (FP) kinetic model has been devised on the basis of the Boltzmann equation (Jenny et al., 2010; Gorji et al., 2011). Particle Monte-Carlo schemes are then introduced for simulations of rarefied gas flows based on the FP kinetics. Here the particles follow independent stochastic paths and thus a spatio-temporal resolution coarser than the collisional scales becomes possible. In contrast to the direct simulation Monte-Carlo (DSMC), the computational cost is independent of the Knudsen number resulting in efficient simulations at moderate/low Knudsen flows. In order to further exploit the efficiency of the FP method, the required particle-cell resolutions should be found, and a cell refinement strategy has to be developed accordingly. In this study, an adaptive particle-cell scheme applicable to a general unstructured mesh is derived for the FP model. Virtual sub cells are introduced for the adaptive mesh refinement. Moreover a sub cell-merging algorithm is provided to honor the minimum required number of particles per cell. For assessments, the 70 degree blunted cone reentry flow (Allgre et al., 1997) is studied. Excellent agreement between the introduced adaptive FP method and DSMC is achieved.
Roberts, Nathan V.; Demkowiz, Leszek; Moser, Robert
2015-11-15
The discontinuous Petrov-Galerkin methodology with optimal test functions (DPG) of Demkowicz and Gopalakrishnan [18, 20] guarantees the optimality of the solution in an energy norm, and provides several features facilitating adaptive schemes. Whereas Bubnov-Galerkin methods use identical trial and test spaces, Petrov-Galerkin methods allow these function spaces to differ. In DPG, test functions are computed on the fly and are chosen to realize the supremum in the inf-sup condition; the method is equivalent to a minimum residual method. For well-posed problems with sufficiently regular solutions, DPG can be shown to converge at optimal rates—the inf-sup constants governing the convergence are mesh-independent, and of the same order as those governing the continuous problem [48]. DPG also provides an accurate mechanism for measuring the error, and this can be used to drive adaptive mesh refinements. We employ DPG to solve the steady incompressible Navier-Stokes equations in two dimensions, building on previous work on the Stokes equations, and focusing particularly on the usefulness of the approach for automatic adaptivity starting from a coarse mesh. We apply our approach to a manufactured solution due to Kovasznay as well as the lid-driven cavity flow, backward-facing step, and flow past a cylinder problems.
Efficient Packet Forwarding in Mesh Network
Kanrar, Soumen
2012-01-01
Wireless Mesh Network (WMN) is a multi hop low cost, with easy maintenance robust network providing reliable service coverage. WMNs consist of mesh routers and mesh clients. In this architecture, while static mesh routers form the wireless backbone, mesh clients access the network through mesh routers as well as directly meshing with each other. Different from traditional wireless networks, WMN is dynamically self-organized and self-configured. In other words, the nodes in the mesh network automatically establish and maintain network connectivity. Over the years researchers have worked, to reduce the redundancy in broadcasting packet in the mesh network in the wireless domain for providing reliable service coverage, the source node deserves to broadcast or flood the control packets. The redundant control packet consumes the bandwidth of the wireless medium and significantly reduces the average throughput and consequently reduces the overall system performance. In this paper I study the optimization problem in...
On Linear Spaces of Polyhedral Meshes.
Poranne, Roi; Chen, Renjie; Gotsman, Craig
2015-05-01
Polyhedral meshes (PM)-meshes having planar faces-have enjoyed a rise in popularity in recent years due to their importance in architectural and industrial design. However, they are also notoriously difficult to generate and manipulate. Previous methods start with a smooth surface and then apply elaborate meshing schemes to create polyhedral meshes approximating the surface. In this paper, we describe a reverse approach: given the topology of a mesh, we explore the space of possible planar meshes having that topology. Our approach is based on a complete characterization of the maximal linear spaces of polyhedral meshes contained in the curved manifold of polyhedral meshes with a given topology. We show that these linear spaces can be described as nullspaces of differential operators, much like harmonic functions are nullspaces of the Laplacian operator. An analysis of this operator provides tools for global and local design of a polyhedral mesh, which fully expose the geometric possibilities and limitations of the given topology.
Implementation of tetrahedral-mesh geometry in Monte Carlo radiation transport code PHITS
Furuta, Takuya; Sato, Tatsuhiko; Han, Min Cheol; Yeom, Yeon Soo; Kim, Chan Hyeong; Brown, Justin L.; Bolch, Wesley E.
2017-06-01
A new function to treat tetrahedral-mesh geometry was implemented in the particle and heavy ion transport code systems. To accelerate the computational speed in the transport process, an original algorithm was introduced to initially prepare decomposition maps for the container box of the tetrahedral-mesh geometry. The computational performance was tested by conducting radiation transport simulations of 100 MeV protons and 1 MeV photons in a water phantom represented by tetrahedral mesh. The simulation was repeated with varying number of meshes and the required computational times were then compared with those of the conventional voxel representation. Our results show that the computational costs for each boundary crossing of the region mesh are essentially equivalent for both representations. This study suggests that the tetrahedral-mesh representation offers not only a flexible description of the transport geometry but also improvement of computational efficiency for the radiation transport. Due to the adaptability of tetrahedrons in both size and shape, dosimetrically equivalent objects can be represented by tetrahedrons with a much fewer number of meshes as compared its voxelized representation. Our study additionally included dosimetric calculations using a computational human phantom. A significant acceleration of the computational speed, about 4 times, was confirmed by the adoption of a tetrahedral mesh over the traditional voxel mesh geometry.
Final Report: Symposium on Adaptive Methods for Partial Differential Equations
Pernice, M.; Johnson, C.R.; Smith, P.J.; Fogelson, A.
1998-12-10
OAK-B135 Final Report: Symposium on Adaptive Methods for Partial Differential Equations. Complex physical phenomena often include features that span a wide range of spatial and temporal scales. Accurate simulation of such phenomena can be difficult to obtain, and computations that are under-resolved can even exhibit spurious features. While it is possible to resolve small scale features by increasing the number of grid points, global grid refinement can quickly lead to problems that are intractable, even on the largest available computing facilities. These constraints are particularly severe for three dimensional problems that involve complex physics. One way to achieve the needed resolution is to refine the computational mesh locally, in only those regions where enhanced resolution is required. Adaptive solution methods concentrate computational effort in regions where it is most needed. These methods have been successfully applied to a wide variety of problems in computational science and engineering. Adaptive methods can be difficult to implement, prompting the development of tools and environments to facilitate their use. To ensure that the results of their efforts are useful, algorithm and tool developers must maintain close communication with application specialists. Conversely it remains difficult for application specialists who are unfamiliar with the methods to evaluate the trade-offs between the benefits of enhanced local resolution and the effort needed to implement an adaptive solution method.
Improved Butterfly Subdivision Scheme for Meshes with Arbitrary Topology
ZHANG Hui; MA Yong-you; ZHANG Cheng; JIANG Shou-wei
2005-01-01
Based on the butterfly subdivision scheme and the modified butterfly subdivision scheme, an improved butterfly subdivision scheme is proposed. The scheme uses a small stencil of six points to calculate new inserting vertex, 2n new vertices are inserted in the 2n triangle faces in each recursion, and the n old vertices are kept, special treatment is given to the boundary, achieving higher smoothness while using small stencils is realized. With the proposed scheme, the number of triangle faces increases only by a factor of 3 in each refinement step. Compared with the butterfly subdivision scheme and the modified butterfly subdivision scheme, the size of triangle faces changes more gradually, which allows one to have greater control over the resolution of a refined mesh.
A new self-adaptive remeshing approach
Wu Yong; He Yuanjun; Zhang Lin
2006-01-01
This paper proposes a self-adaptive approach to converting irregular genus-0 meshes into those with subdivision connectivity. To assure a maximal utilization of the multiresolution techniques on the remesh,we map the original mesh onto the unit sphere and construct a base mesh with only four vertices. We also introduce a self-adaptive relocation operation, which is used to adapt the vertex distribution of the spherical subdivision mesh to that of the parameterized mesh, to improve the visual appearance of the remesh.The experimental results show that our method can not only make the number of irregular vertices in the remesh as small as possible, but also preserve the details of the original mesh well.
Capelli, Silvia C; Bürgi, Hans-Beat; Dittrich, Birger; Grabowsky, Simon; Jayatilaka, Dylan
2014-09-01
Hirshfeld atom refinement (HAR) is a method which determines structural parameters from single-crystal X-ray diffraction data by using an aspherical atom partitioning of tailor-made ab initio quantum mechanical molecular electron densities without any further approximation. Here the original HAR method is extended by implementing an iterative procedure of successive cycles of electron density calculations, Hirshfeld atom scattering factor calculations and structural least-squares refinements, repeated until convergence. The importance of this iterative procedure is illustrated via the example of crystalline ammonia. The new HAR method is then applied to X-ray diffraction data of the dipeptide Gly-l-Ala measured at 12, 50, 100, 150, 220 and 295 K, using Hartree-Fock and BLYP density functional theory electron densities and three different basis sets. All positions and anisotropic displacement parameters (ADPs) are freely refined without constraints or restraints - even those for hydrogen atoms. The results are systematically compared with those from neutron diffraction experiments at the temperatures 12, 50, 150 and 295 K. Although non-hydrogen-atom ADPs differ by up to three combined standard uncertainties (csu's), all other structural parameters agree within less than 2 csu's. Using our best calculations (BLYP/cc-pVTZ, recommended for organic molecules), the accuracy of determining bond lengths involving hydrogen atoms from HAR is better than 0.009 Å for temperatures of 150 K or below; for hydrogen-atom ADPs it is better than 0.006 Å(2) as judged from the mean absolute X-ray minus neutron differences. These results are among the best ever obtained. Remarkably, the precision of determining bond lengths and ADPs for the hydrogen atoms from the HAR procedure is comparable with that from the neutron measurements - an outcome which is obtained with a routinely achievable resolution of the X-ray data of 0.65 Å.
Fitting polynomial surfaces to triangular meshes with Voronoi squared distance minimization
Nivoliers, Vincent
2012-11-06
This paper introduces Voronoi squared distance minimization (VSDM), an algorithm that fits a surface to an input mesh. VSDM minimizes an objective function that corresponds to a Voronoi-based approximation of the overall squared distance function between the surface and the input mesh (SDM). This objective function is a generalization of the one minimized by centroidal Voronoi tessellation, and can be minimized by a quasi-Newton solver. VSDM naturally adapts the orientation of the mesh elements to best approximate the input, without estimating any differential quantities. Therefore, it can be applied to triangle soups or surfaces with degenerate triangles, topological noise and sharp features. Applications of fitting quad meshes and polynomial surfaces to input triangular meshes are demonstrated. © 2012 Springer-Verlag London.
Fitting polynomial surfaces to triangular meshes with Voronoi Squared Distance Minimization
Nivoliers, Vincent
2011-12-01
This paper introduces Voronoi Squared Distance Minimization (VSDM), an algorithm that fits a surface to an input mesh. VSDM minimizes an objective function that corresponds to a Voronoi-based approximation of the overall squared distance function between the surface and the input mesh (SDM). This objective function is a generalization of Centroidal Voronoi Tesselation (CVT), and can be minimized by a quasi-Newton solver. VSDM naturally adapts the orientation of the mesh to best approximate the input, without estimating any differential quantities. Therefore it can be applied to triangle soups or surfaces with degenerate triangles, topological noise and sharp features. Applications of fitting quad meshes and polynomial surfaces to input triangular meshes are demonstrated.
Feature-preserving surface mesh smoothing via suboptimal Delaunay triangulation ☆
Gao, Zhanheng; Yu, Zeyun; Holst, Michael
2012-01-01
A method of triangular surface mesh smoothing is presented to improve angle quality by extending the original optimal Delaunay triangulation (ODT) to surface meshes. The mesh quality is improved by solving a quadratic optimization problem that minimizes the approximated interpolation error between a parabolic function and its piecewise linear interpolation defined on the mesh. A suboptimal problem is derived to guarantee a unique, analytic solution that is significantly faster with little loss in accuracy as compared to the optimal one. In addition to the quality-improving capability, the proposed method has been adapted to remove noise while faithfully preserving sharp features such as edges and corners of a mesh. Numerous experiments are included to demonstrate the performance of the method. PMID:23580890
Benazzi, E.; Alario, F
2004-07-01
In 2003, refining margins showed a clear improvement that continued throughout the first three quarters of 2004. Oil companies posted significantly higher earnings in 2003 compared to 2002, with the results of first quarter 2004 confirming this trend. Due to higher feedstock prices, the implementation of new capacity and more intense competition, the petrochemicals industry was not able to boost margins in 2003. In such difficult business conditions, aggravated by soaring crude prices, the petrochemicals industry is not likely to see any improvement in profitability before the second half of 2004. (author)
Benazzi, E
2003-07-01
Down sharply in 2002, refining margins showed a clear improvement in the first half-year of 2003. As a result, the earnings reported by oil companies for financial year 2002 were significantly lower than in 2001, but the prospects are brighter for 2003. In the petrochemicals sector, slow demand and higher feedstock prices eroded margins in 2002, especially in Europe and the United States. The financial results for the first part of 2003 seem to indicate that sector profitability will not improve before 2004. (author)
Inching toward 'push-button' meshing
James Masters
2015-01-01
While "push-button" meshing remains an elusive goal, advances in 2015 have brought the technology to the point where meshes can be constructed with relative ease when appropriate surfaces are available...
Macromolecular crystallographic estructure refinement
Afonine, Pavel V.
2015-04-01
Full Text Available Model refinement is a key step in crystallographic structure determination that ensures final atomic structure of macromolecule represents measured diffraction data as good as possible. Several decades have been put into developing methods and computational tools to streamline this step. In this manuscript we provide a brief overview of major milestones of crystallographic computing and methods development pertinent to structure refinement.El refinamiento es un paso clave en el proceso de determinación de una estructura cristalográfica al garantizar que la estructura atómica de la macromolécula final represente de la mejor manera posible los datos de difracción. Han hecho falta varias décadas para poder desarrollar nuevos métodos y herramientas computacionales dirigidas a dinamizar esta etapa. En este artículo ofrecemos un breve resumen de los principales hitos en la computación cristalográfica y de los nuevos métodos relevantes para el refinamiento de estructuras.
Particle Collection Efficiency for Nylon Mesh Screens
Cena, Lorenzo G.; Ku, Bon-Ki; Peters, Thomas M.
2011-01-01
Mesh screens composed of nylon fibers leave minimal residual ash and produce no significant spectral interference when ashed for spectrometric examination. These characteristics make nylon mesh screens attractive as a collection substrate for nanoparticles. A theoretical single-fiber efficiency expression developed for wire-mesh screens was evaluated for estimating the collection efficiency of submicrometer particles for nylon mesh screens. Pressure drop across the screens, the effect of part...
Dyja, Robert; van der Zee, Kristoffer G
2016-01-01
We present an adaptive methodology for the solution of (linear and) non-linear time dependent problems that is especially tailored for massively parallel computations. The basic concept is to solve for large blocks of space-time unknowns instead of marching sequentially in time. The methodology is a combination of a computationally efficient implementation of a parallel-in-space-time finite element solver coupled with a posteriori space-time error estimates and a parallel mesh generator. This methodology enables, in principle, simultaneous adaptivity in both space and time (within the block) domains. We explore this basic concept in the context of a variety of time-steppers including $\\Theta$-schemes and Backward Differentiate Formulas. We specifically illustrate this framework with applications involving time dependent linear, quasi-linear and semi-linear diffusion equations. We focus on investigating how the coupled space-time refinement indicators for this class of problems affect spatial adaptivity. Final...
Hirt, G.; Schäfer, D.
2010-06-01
For the process design of incremental forming processes like ring rolling or stretch forging there is need for fast and accurate simulation techniques. For many applications it would be necessary to predict the microstructural evolution during the process. For this reason the FEM-software Larstran/Shape which is used for the plastomechanical simulation can be coupled with the microstructure simulation module Strucsim. This software uses phenomenological equations for the calculation of recrystallization (dynamic and static) and grain size evolution [7]. To accelerate the simulation a multi mesh method has been developed. This method uses an adapted simulation mesh with fine elements only in the locally limited contact and forming zone to achieve a reduction of the number of elements. Due to the relative movement of the tool and workpiece the adapted FE-mesh has to be remeshed regularly according to the position of the tool. To avoid loss of information caused by the use of coarse elements the multi mesh method uses a second storage mesh which represents the entire workpiece and which is discretized using only fine elements. For the update of the storage mesh the displacement vectors and changes of scalar values like temperature can be interpolated. For the application of the multi mesh method to the microstructure simulation the update algorithm for the microstructure values has to be modified. The microstructure and plastomechanical simulation have to be uncoupled and the static recrystallization and grain growth outside the forming zone have to be calculated separately. Using the multi mesh method the simulation of incremental bulk metal forming processes including a microstructural simulation can be accelerated. The acceleration factor of the simulation compared to a simulation without adaptive meshing is dependent on the reduction of elements and nodes.
2010-10-01
... 50 Wildlife and Fisheries 7 2010-10-01 2010-10-01 false Mesh size. 300.110 Section 300.110... Antarctic Marine Living Resources § 300.110 Mesh size. (a) The use of pelagic and bottom trawls having the mesh size in any part of a trawl less than indicated is prohibited for any directed fishing for the...
Markov Random Fields on Triangle Meshes
Andersen, Vedrana; Aanæs, Henrik; Bærentzen, Jakob Andreas
2010-01-01
mesh edges according to a feature detecting prior. Since we should not smooth across a sharp feature, we use edge labels to control the vertex process. In a Bayesian framework, MRF priors are combined with the likelihood function related to the mesh formation method. The output of our algorithm...... is a piecewise smooth mesh with explicit labelling of edges belonging to the sharp features....