WorldWideScience

Sample records for mesh generation tool

  1. NeuroTessMesh: A Tool for the Generation and Visualization of Neuron Meshes and Adaptive On-the-Fly Refinement

    Juan J. Garcia-Cantero

    2017-06-01

    Full Text Available Gaining a better understanding of the human brain continues to be one of the greatest challenges for science, largely because of the overwhelming complexity of the brain and the difficulty of analyzing the features and behavior of dense neural networks. Regarding analysis, 3D visualization has proven to be a useful tool for the evaluation of complex systems. However, the large number of neurons in non-trivial circuits, together with their intricate geometry, makes the visualization of a neuronal scenario an extremely challenging computational problem. Previous work in this area dealt with the generation of 3D polygonal meshes that approximated the cells’ overall anatomy but did not attempt to deal with the extremely high storage and computational cost required to manage a complex scene. This paper presents NeuroTessMesh, a tool specifically designed to cope with many of the problems associated with the visualization of neural circuits that are comprised of large numbers of cells. In addition, this method facilitates the recovery and visualization of the 3D geometry of cells included in databases, such as NeuroMorpho, and provides the tools needed to approximate missing information such as the soma’s morphology. This method takes as its only input the available compact, yet incomplete, morphological tracings of the cells as acquired by neuroscientists. It uses a multiresolution approach that combines an initial, coarse mesh generation with subsequent on-the-fly adaptive mesh refinement stages using tessellation shaders. For the coarse mesh generation, a novel approach, based on the Finite Element Method, allows approximation of the 3D shape of the soma from its incomplete description. Subsequently, the adaptive refinement process performed in the graphic card generates meshes that provide good visual quality geometries at a reasonable computational cost, both in terms of memory and rendering time. All the described techniques have been

  2. User Manual for the PROTEUS Mesh Tools

    Smith, Micheal A. [Argonne National Lab. (ANL), Argonne, IL (United States); Shemon, Emily R. [Argonne National Lab. (ANL), Argonne, IL (United States)

    2015-06-01

    This report describes the various mesh tools that are provided with the PROTEUS code giving both descriptions of the input and output. In many cases the examples are provided with a regression test of the mesh tools. The most important mesh tools for any user to consider using are the MT_MeshToMesh.x and the MT_RadialLattice.x codes. The former allows the conversion between most mesh types handled by PROTEUS while the second allows the merging of multiple (assembly) meshes into a radial structured grid. Note that the mesh generation process is recursive in nature and that each input specific for a given mesh tool (such as .axial or .merge) can be used as “mesh” input for any of the mesh tools discussed in this manual.

  3. User Manual for the PROTEUS Mesh Tools

    Smith, Micheal A.; Shemon, Emily R.

    2015-01-01

    This report describes the various mesh tools that are provided with the PROTEUS code giving both descriptions of the input and output. In many cases the examples are provided with a regression test of the mesh tools. The most important mesh tools for any user to consider using are the MT M eshToMesh.x and the MT R adialLattice.x codes. The former allows the conversion between most mesh types handled by PROTEUS while the second allows the merging of multiple (assembly) meshes into a radial structured grid. Note that the mesh generation process is recursive in nature and that each input specific for a given mesh tool (such as .axial or .merge) can be used as ''mesh'' input for any of the mesh tools discussed in this manual.

  4. Documentation for MeshKit - Reactor Geometry (&mesh) Generator

    Jain, Rajeev [Argonne National Lab. (ANL), Argonne, IL (United States); Mahadevan, Vijay [Argonne National Lab. (ANL), Argonne, IL (United States)

    2015-09-30

    This report gives documentation for using MeshKit’s Reactor Geometry (and mesh) Generator (RGG) GUI and also briefly documents other algorithms and tools available in MeshKit. RGG is a program designed to aid in modeling and meshing of complex/large hexagonal and rectilinear reactor cores. RGG uses Argonne’s SIGMA interfaces, Qt and VTK to produce an intuitive user interface. By integrating a 3D view of the reactor with the meshing tools and combining them into one user interface, RGG streamlines the task of preparing a simulation mesh and enables real-time feedback that reduces accidental scripting mistakes that could waste hours of meshing. RGG interfaces with MeshKit tools to consolidate the meshing process, meaning that going from model to mesh is as easy as a button click. This report is designed to explain RGG v 2.0 interface and provide users with the knowledge and skills to pilot RGG successfully. Brief documentation of MeshKit source code, tools and other algorithms available are also presented for developers to extend and add new algorithms to MeshKit. RGG tools work in serial and parallel and have been used to model complex reactor core models consisting of conical pins, load pads, several thousands of axially varying material properties of instrumentation pins and other interstices meshes.

  5. User Manual for the PROTEUS Mesh Tools

    Smith, Micheal A. [Argonne National Lab. (ANL), Argonne, IL (United States); Shemon, Emily R [Argonne National Lab. (ANL), Argonne, IL (United States)

    2016-09-19

    PROTEUS is built around a finite element representation of the geometry for visualization. In addition, the PROTEUS-SN solver was built to solve the even-parity transport equation on a finite element mesh provided as input. Similarly, PROTEUS-MOC and PROTEUS-NEMO were built to apply the method of characteristics on unstructured finite element meshes. Given the complexity of real world problems, experience has shown that using commercial mesh generator to create rather simple input geometries is overly complex and slow. As a consequence, significant effort has been put into place to create multiple codes that help assist in the mesh generation and manipulation. There are three input means to create a mesh in PROTEUS: UFMESH, GRID, and NEMESH. At present, the UFMESH is a simple way to generate two-dimensional Cartesian and hexagonal fuel assembly geometries. The UFmesh input allows for simple assembly mesh generation while the GRID input allows the generation of Cartesian, hexagonal, and regular triangular structured grid geometry options. The NEMESH is a way for the user to create their own mesh or convert another mesh file format into a PROTEUS input format. Given that one has an input mesh format acceptable for PROTEUS, we have constructed several tools which allow further mesh and geometry construction (i.e. mesh extrusion and merging). This report describes the various mesh tools that are provided with the PROTEUS code giving both descriptions of the input and output. In many cases the examples are provided with a regression test of the mesh tools. The most important mesh tools for any user to consider using are the MT_MeshToMesh.x and the MT_RadialLattice.x codes. The former allows the conversion between most mesh types handled by PROTEUS while the second allows the merging of multiple (assembly) meshes into a radial structured grid. Note that the mesh generation process is recursive in nature and that each input specific for a given mesh tool (such as .axial

  6. RGG: Reactor geometry (and mesh) generator

    Jain, R.; Tautges, T.

    2012-01-01

    The reactor geometry (and mesh) generator RGG takes advantage of information about repeated structures in both assembly and core lattices to simplify the creation of geometry and mesh. It is released as open source software as a part of the MeshKit mesh generation library. The methodology operates in three stages. First, assembly geometry models of various types are generated by a tool called AssyGen. Next, the assembly model or models are meshed by using MeshKit tools or the CUBIT mesh generation tool-kit, optionally based on a journal file output by AssyGen. After one or more assembly model meshes have been constructed, a tool called CoreGen uses a copy/move/merge process to arrange the model meshes into a core model. In this paper, we present the current state of tools and new features in RGG. We also discuss the parallel-enabled CoreGen, which in several cases achieves super-linear speedups since the problems fit in available RAM at higher processor counts. Several RGG applications - 1/6 VHTR model, 1/4 PWR reactor core, and a full-core model for Monju - are reported. (authors)

  7. GENERATION OF IRREGULAR HEXAGONAL MESHES

    Vlasov Aleksandr Nikolaevich

    2012-07-01

    Decomposition is performed in a constructive way and, as option, it involves meshless representation. Further, this mapping method is used to generate the calculation mesh. In this paper, the authors analyze different cases of mapping onto simply connected and bi-connected canonical domains. They represent forward and backward mapping techniques. Their potential application for generation of nonuniform meshes within the framework of the asymptotic homogenization theory is also performed to assess and project effective characteristics of heterogeneous materials (composites.

  8. A comprehensive tool for image-based generation of fetus and pregnant women mesh models for numerical dosimetry studies

    Dahdouh, S; Serrurier, A; De la Plata, J-P; Anquez, J; Angelini, E D; Bloch, I; Varsier, N; Wiart, J

    2014-01-01

    Fetal dosimetry studies require the development of accurate numerical 3D models of the pregnant woman and the fetus. This paper proposes a 3D articulated fetal growth model covering the main phases of pregnancy and a pregnant woman model combining the utero-fetal structures and a deformable non-pregnant woman body envelope. The structures of interest were automatically or semi-automatically (depending on the stage of pregnancy) segmented from a database of images and surface meshes were generated. By interpolating linearly between fetal structures, each one can be generated at any age and in any position. A method is also described to insert the utero-fetal structures in the maternal body. A validation of the fetal models is proposed, comparing a set of biometric measurements to medical reference charts. The usability of the pregnant woman model in dosimetry studies is also investigated, with respect to the influence of the abdominal fat layer. (paper)

  9. Evolutionary algorithm based optimization of hydraulic machines utilizing a state-of-the-art block coupled CFD solver and parametric geometry and mesh generation tools

    S, Kyriacou; E, Kontoleontos; S, Weissenberger; L, Mangani; E, Casartelli; I, Skouteropoulou; M, Gattringer; A, Gehrer; M, Buchmayr

    2014-03-01

    An efficient hydraulic optimization procedure, suitable for industrial use, requires an advanced optimization tool (EASY software), a fast solver (block coupled CFD) and a flexible geometry generation tool. EASY optimization software is a PCA-driven metamodel-assisted Evolutionary Algorithm (MAEA (PCA)) that can be used in both single- (SOO) and multiobjective optimization (MOO) problems. In MAEAs, low cost surrogate evaluation models are used to screen out non-promising individuals during the evolution and exclude them from the expensive, problem specific evaluation, here the solution of Navier-Stokes equations. For additional reduction of the optimization CPU cost, the PCA technique is used to identify dependences among the design variables and to exploit them in order to efficiently drive the application of the evolution operators. To further enhance the hydraulic optimization procedure, a very robust and fast Navier-Stokes solver has been developed. This incompressible CFD solver employs a pressure-based block-coupled approach, solving the governing equations simultaneously. This method, apart from being robust and fast, also provides a big gain in terms of computational cost. In order to optimize the geometry of hydraulic machines, an automatic geometry and mesh generation tool is necessary. The geometry generation tool used in this work is entirely based on b-spline curves and surfaces. In what follows, the components of the tool chain are outlined in some detail and the optimization results of hydraulic machine components are shown in order to demonstrate the performance of the presented optimization procedure.

  10. Evolutionary algorithm based optimization of hydraulic machines utilizing a state-of-the-art block coupled CFD solver and parametric geometry and mesh generation tools

    Kyriacou S; Kontoleontos E; Weissenberger S; Mangani L; Casartelli E; Skouteropoulou I; Gattringer M; Gehrer A; Buchmayr M

    2014-01-01

    An efficient hydraulic optimization procedure, suitable for industrial use, requires an advanced optimization tool (EASY software), a fast solver (block coupled CFD) and a flexible geometry generation tool. EASY optimization software is a PCA-driven metamodel-assisted Evolutionary Algorithm (MAEA (PCA)) that can be used in both single- (SOO) and multiobjective optimization (MOO) problems. In MAEAs, low cost surrogate evaluation models are used to screen out non-promising individuals during the evolution and exclude them from the expensive, problem specific evaluation, here the solution of Navier-Stokes equations. For additional reduction of the optimization CPU cost, the PCA technique is used to identify dependences among the design variables and to exploit them in order to efficiently drive the application of the evolution operators. To further enhance the hydraulic optimization procedure, a very robust and fast Navier-Stokes solver has been developed. This incompressible CFD solver employs a pressure-based block-coupled approach, solving the governing equations simultaneously. This method, apart from being robust and fast, also provides a big gain in terms of computational cost. In order to optimize the geometry of hydraulic machines, an automatic geometry and mesh generation tool is necessary. The geometry generation tool used in this work is entirely based on b-spline curves and surfaces. In what follows, the components of the tool chain are outlined in some detail and the optimization results of hydraulic machine components are shown in order to demonstrate the performance of the presented optimization procedure

  11. Automatic mesh generation with QMESH program

    Ise, Takeharu; Tsutsui, Tsuneo

    1977-05-01

    Usage of the two-dimensional self-organizing mesh generation program, QMESH, is presented together with the descriptions and the experience, as it has recently been converted and reconstructed from the NEACPL version to the FACOM. The program package consists of the QMESH code to generate quadrilaterial meshes with smoothing techniques, the QPLOT code to plot the data obtained from the QMESH on the graphic COM, and the RENUM code to renumber the meshes by using a bandwidth minimization procedure. The technique of mesh reconstructuring coupled with smoothing techniques is especially useful when one generates the meshes for computer codes based on the finite element method. Several typical examples are given for easy access to the QMESH program, which is registered in the R.B-disks of JAERI for users. (auth.)

  12. Interoperable mesh and geometry tools for advanced petascale simulations

    Diachin, L; Bauer, A; Fix, B; Kraftcheck, J; Jansen, K; Luo, X; Miller, M; Ollivier-Gooch, C; Shephard, M S; Tautges, T; Trease, H

    2007-01-01

    SciDAC applications have a demonstrated need for advanced software tools to manage the complexities associated with sophisticated geometry, mesh, and field manipulation tasks, particularly as computer architectures move toward the petascale. The Center for Interoperable Technologies for Advanced Petascale Simulations (ITAPS) will deliver interoperable and interchangeable mesh, geometry, and field manipulation services that are of direct use to SciDAC applications. The premise of our technology development goal is to provide such services as libraries that can be used with minimal intrusion into application codes. To develop these technologies, we focus on defining a common data model and data-structure neutral interfaces that unify a number of different services such as mesh generation and improvement, front tracking, adaptive mesh refinement, shape optimization, and solution transfer operations. We highlight the use of several ITAPS services in SciDAC applications

  13. MeshVoro: A Three-Dimensional Voronoi Mesh Building Tool for the TOUGH Family of Codes

    Freeman, C. M.; Boyle, K. L.; Reagan, M.; Johnson, J.; Rycroft, C.; Moridis, G. J.

    2013-09-30

    Few tools exist for creating and visualizing complex three-dimensional simulation meshes, and these have limitations that restrict their application to particular geometries and circumstances. Mesh generation needs to trend toward ever more general applications. To that end, we have developed MeshVoro, a tool that is based on the Voro (Rycroft 2009) library and is capable of generating complex threedimensional Voronoi tessellation-based (unstructured) meshes for the solution of problems of flow and transport in subsurface geologic media that are addressed by the TOUGH (Pruess et al. 1999) family of codes. MeshVoro, which includes built-in data visualization routines, is a particularly useful tool because it extends the applicability of the TOUGH family of codes by enabling the scientifically robust and relatively easy discretization of systems with challenging 3D geometries. We describe several applications of MeshVoro. We illustrate the ability of the tool to straightforwardly transform a complex geological grid into a simulation mesh that conforms to the specifications of the TOUGH family of codes. We demonstrate how MeshVoro can describe complex system geometries with a relatively small number of grid blocks, and we construct meshes for geometries that would have been practically intractable with a standard Cartesian grid approach. We also discuss the limitations and appropriate applications of this new technology.

  14. ZONE: a finite element mesh generator

    Burger, M.J.

    1976-05-01

    The ZONE computer program is a finite-element mesh generator which produces the nodes and element description of any two-dimensional geometry. The geometry is subdivided into a mesh of quadrilateral and triangular zones arranged sequentially in an ordered march through the geometry. The order of march can be chosen so that the minimum bandwidth is obtained. The node points are defined in terms of the x and y coordinates in a global rectangular coordinate system. The zones generated are quadrilaterals or triangles defined by four node points in a counterclockwise sequence. Node points defining the outside boundary are generated to describe pressure boundary conditions. The mesh that is generated can be used as input to any two-dimensional as well as any axisymmetrical structure program. The output from ZONE is essentially the input file to NAOS, HONDO, and other axisymmetric finite element programs. 14 figures

  15. Computational mesh generation for vascular structures with deformable surfaces

    Putter, S. de; Laffargue, F.; Breeuwer, M.; Vosse, F.N. van de; Gerritsen, F.A.; Philips Medical Systems, Best

    2006-01-01

    Computational blood flow and vessel wall mechanics simulations for vascular structures are becoming an important research tool for patient-specific surgical planning and intervention. An important step in the modelling process for patient-specific simulations is the creation of the computational mesh based on the segmented geometry. Most known solutions either require a large amount of manual processing or lead to a substantial difference between the segmented object and the actual computational domain. We have developed a chain of algorithms that lead to a closely related implementation of image segmentation with deformable models and 3D mesh generation. The resulting processing chain is very robust and leads both to an accurate geometrical representation of the vascular structure as well as high quality computational meshes. The chain of algorithms has been tested on a wide variety of shapes. A benchmark comparison of our mesh generation application with five other available meshing applications clearly indicates that the new approach outperforms the existing methods in the majority of cases. (orig.)

  16. Feature-Sensitive Tetrahedral Mesh Generation with Guaranteed Quality

    Wang, Jun; Yu, Zeyun

    2012-01-01

    Tetrahedral meshes are being extensively used in finite element methods (FEM). This paper proposes an algorithm to generate feature-sensitive and high-quality tetrahedral meshes from an arbitrary surface mesh model. A top-down octree subdivision is conducted on the surface mesh and a set of tetrahedra are constructed using adaptive body-centered cubic (BCC) lattices. Special treatments are given to the tetrahedra near the surface such that the quality of the resulting tetrahedral mesh is prov...

  17. Polygonal Prism Mesh in the Viscous Layers for the Polyhedral Mesh Generator, PolyGen

    Lee, Sang Yong; Park, Chan Eok; Kim, Shin Whan

    2015-01-01

    Polyhedral mesh has been known to have some benefits over the tetrahedral mesh. Efforts have been made to set up a polyhedral mesh generation system with open source programs SALOME and TetGen. The evaluation has shown that the polyhedral mesh generation system is promising. But it is necessary to extend the capability of the system to handle the viscous layers to be a generalized mesh generator. A brief review to the previous works on the mesh generation for the viscous layers will be made in section 2. Several challenging issues for the polygonal prism mesh generation will be discussed as well. The procedure to generate a polygonal prism mesh will be discussed in detail in section 3. Conclusion will be followed in section 4. A procedure to generate meshes in the viscous layers with PolyGen has been successfully designed. But more efforts have to be exercised to find the best way for the generating meshes for viscous layers. Using the extrusion direction of the STL data will the first of the trials in the near future

  18. Symbolic Block Decomposition In Hexahedral Mesh Generation

    Andrzej Adamek

    2005-01-01

    Full Text Available Hexahedral mesh generation for three-dimensional solid objects is often done in stages. Usually an object is first subdivided into simple-shaped subregions, which then are filled withhexahedral finite elements. This article presents an automatic subdividing method of polyhedron with planar faces. The subdivision is based on medial surface, axes and nodes of a solid.The main emphasis is put on creating a topology of subregions. Obtaining such a topologyinvolves defining a graph structure OMG which contains necessary information about medialsurface topology and object topology, followed by simple symbolic processing on it.

  19. Automated quadrilateral mesh generation for digital image structures

    2011-01-01

    With the development of advanced imaging technology, digital images are widely used. This paper proposes an automatic quadrilateral mesh generation algorithm for multi-colour imaged structures. It takes an original arbitrary digital image as an input for automatic quadrilateral mesh generation, this includes removing the noise, extracting and smoothing the boundary geometries between different colours, and automatic all-quad mesh generation with the above boundaries as constraints. An application example is...

  20. HypGrid2D. A 2-d mesh generator

    Soerensen, N N

    1998-03-01

    The implementation of a hyperbolic mesh generation procedure, based on an equation for orthogonality and an equation for the cell face area is described. The method is fast, robust and gives meshes with good smoothness and orthogonality. The procedure is implemented in a program called HypGrid2D. The HypGrid2D program is capable of generating C-, O- and `H`-meshes for use in connection with the EllipSys2D Navier-Stokes solver. To illustrate the capabilities of the program, some test examples are shown. First a series of C-meshes are generated around a NACA-0012 airfoil. Secondly a series of O-meshes are generated around a NACA-65-418 airfoil. Finally `H`-meshes are generated over a Gaussian hill and a linear escarpment. (au)

  1. SALOME PLATFORM and TetGen for Polyhedral Mesh Generation

    Lee, Sang Yong; Park, Chan Eok; Kim, Shin Whan [KEPCO E and C Company, Inc., Daejeon (Korea, Republic of)

    2014-05-15

    SPACE and CUPID use the unstructured mesh and they also require reliable mesh generation system. The combination of CAD system and mesh generation system is necessary to cope with a large number of cells and the complex fluid system with structural materials inside. In the past, a CAD system Pro/Engineer and mesh generator Pointwise were evaluated for this application. But, the cost of those commercial CAD and mesh generator is sometimes a great burden. Therefore, efforts have been made to set up a mesh generation system with open source programs. The evaluation of the TetGen has been made in focusing the application for the polyhedral mesh generation. In this paper, SALOME will be evaluated for the efforts in conjunction with TetGen. In section 2, review will be made on the CAD and mesh generation capability of SALOME. SALOME and TetGen codes are being integrated to construct robust polyhedral mesh generator. Edge removal on the flat surface and vertex reattachment to the solid are two challenging tasks. It is worthwhile to point out that the Python script capability of the SALOME should be fully utilized for the future investigation.

  2. AUTOMATIC MESH GENERATION OF 3-D GEOMETRIC MODELS

    刘剑飞

    2003-01-01

    In this paper the presentation of the ball-packing method is reviewed,and a scheme to generate mesh for complex 3-D geometric models is given,which consists of 4 steps:(1)create nodes in 3-D models by ball-packing method,(2)connect nodes to generate mesh by 3-D Delaunay triangulation,(3)retrieve the boundary of the model after Delaunay triangulation,(4)improve the mesh.

  3. Improved mesh generator for the POISSON Group Codes

    Gupta, R.C.

    1987-01-01

    This paper describes the improved mesh generator of the POISSON Group Codes. These improvements enable one to have full control over the way the mesh is generated and in particular the way the mesh density is distributed throughout this model. A higher mesh density in certain regions coupled with a successively lower mesh density in others keeps the accuracy of the field computation high and the requirements on the computer time and computer memory low. The mesh is generated with the help of codes AUTOMESH and LATTICE; both have gone through a major upgrade. Modifications have also been made in the POISSON part of these codes. We shall present an example of a superconducting dipole magnet to explain how to use this code. The results of field computations are found to be reliable within a few parts in a hundred thousand even in such complex geometries

  4. Image-Based Geometric Modeling and Mesh Generation

    2013-01-01

    As a new interdisciplinary research area, “image-based geometric modeling and mesh generation” integrates image processing, geometric modeling and mesh generation with finite element method (FEM) to solve problems in computational biomedicine, materials sciences and engineering. It is well known that FEM is currently well-developed and efficient, but mesh generation for complex geometries (e.g., the human body) still takes about 80% of the total analysis time and is the major obstacle to reduce the total computation time. It is mainly because none of the traditional approaches is sufficient to effectively construct finite element meshes for arbitrarily complicated domains, and generally a great deal of manual interaction is involved in mesh generation. This contributed volume, the first for such an interdisciplinary topic, collects the latest research by experts in this area. These papers cover a broad range of topics, including medical imaging, image alignment and segmentation, image-to-mesh conversion,...

  5. Moving mesh generation with a sequential approach for solving PDEs

    In moving mesh methods, physical PDEs and a mesh equation derived from equidistribution of an error metrics (so-called the monitor function) are simultaneously solved and meshes are dynamically concentrated on steep regions (Lim et al., 2001). However, the simultaneous solution procedure...... a simple and robust moving mesh algorithm in one or multidimension. In this study, we propose a sequential solution procedure including two separate parts: prediction step to obtain an approximate solution to a next time level (integration of physical PDEs) and regriding step at the next time level (mesh...... generation and solution interpolation). Convection terms, which appear in physical PDEs and a mesh equation, are discretized by a WENO (Weighted Essentially Non-Oscillatory) scheme under the consrvative form. This sequential approach is to keep the advantages of robustness and simplicity for the static...

  6. Volume Decomposition and Feature Recognition for Hexahedral Mesh Generation

    GADH,RAJIT; LU,YONG; TAUTGES,TIMOTHY J.

    1999-09-27

    Considerable progress has been made on automatic hexahedral mesh generation in recent years. Several automatic meshing algorithms have proven to be very reliable on certain classes of geometry. While it is always worth pursuing general algorithms viable on more general geometry, a combination of the well-established algorithms is ready to take on classes of complicated geometry. By partitioning the entire geometry into meshable pieces matched with appropriate meshing algorithm the original geometry becomes meshable and may achieve better mesh quality. Each meshable portion is recognized as a meshing feature. This paper, which is a part of the feature based meshing methodology, presents the work on shape recognition and volume decomposition to automatically decompose a CAD model into meshable volumes. There are four phases in this approach: (1) Feature Determination to extinct decomposition features, (2) Cutting Surfaces Generation to form the ''tailored'' cutting surfaces, (3) Body Decomposition to get the imprinted volumes; and (4) Meshing Algorithm Assignment to match volumes decomposed with appropriate meshing algorithms. The feature determination procedure is based on the CLoop feature recognition algorithm that is extended to be more general. Results are demonstrated over several parts with complicated topology and geometry.

  7. Procedure for the automatic mesh generation of innovative gear teeth

    Radicella Andrea Chiaramonte

    2016-01-01

    Full Text Available After having described gear wheels with teeth having the two sides constituted by different involutes and their importance in engineering applications, we stress the need for an efficient procedure for the automatic mesh generation of innovative gear teeth. First, we describe the procedure for the subdivision of the tooth profile in the various possible cases, then we show the method for creating the subdivision mesh, defined by two series of curves called meridians and parallels. Finally, we describe how the above procedure for automatic mesh generation is able to solve specific cases that may arise when dealing with teeth having the two sides constituted by different involutes.

  8. AUTOMATIC MESH GENERATION OF 3—D GEOMETRIC MODELS

    刘剑飞

    2003-01-01

    In this paper the presentation of the ball-packing method is reviewed, and a schemeto generate mesh for complex 3-D geometric models is given, which consists of 4 steps: (1) createnodes in 3-D models by ball-packing method, (2) connect nodes to generate mesh by 3-D Delaunaytriangulation, (3) retrieve the boundary of the model after Delaunay triangulation, (4) improve themesh.

  9. Mesh Generation via Local Bisection Refinement of Triangulated Grids

    2015-06-01

    Science and Technology Organisation DSTO–TR–3095 ABSTRACT This report provides a comprehensive implementation of an unstructured mesh generation method...and Technology Organisation 506 Lorimer St, Fishermans Bend, Victoria 3207, Australia Telephone: 1300 333 362 Facsimile: (03) 9626 7999 c© Commonwealth...their behaviour is critically linked to Maubach’s method and the data structures N and T . The top- level mesh refinement algorithm is also presented

  10. Software tools for manipulating fe mesh, virtual surgery and post-processing

    Milašinović Danko Z.

    2009-01-01

    Full Text Available This paper describes a set of software tools which we developed for the calculation of fluid flow through cardiovascular organs. Our tools work with medical data from a CT scanner, but could be used with any other 3D input data. For meshing we used a Tetgen tetrahedral mesh generator, as well as a mesh re-generator that we have developed for conversion of tetrahedral elements into bricks. After adequate meshing we used our PAKF solver for calculation of fluid flow. For human-friendly presentation of results we developed a set of post-processing software tools. With modification of 2D mesh (boundary of cardiovascular organ it is possible to do virtual surgery, so in a case of an aorta with aneurism, which we had received from University Clinical center in Heidelberg from a multi-slice 64-CT scanner, we removed the aneurism and ran calculations on both geometrical models afterwards. The main idea of this methodology is creating a system that could be used in clinics.

  11. r-Adaptive mesh generation for shell finite element analysis

    Cho, Maenghyo; Jun, Seongki

    2004-01-01

    An r-adaptive method or moving grid technique relocates a grid so that it becomes concentrated in the desired region. This concentration improves the accuracy and efficiency of finite element solutions. We apply the r-adaptive method to computational mesh of shell surfaces, which is initially regular and uniform. The r-adaptive method, given by Liao and Anderson [Appl. Anal. 44 (1992) 285], aggregate the grid in the region with a relatively high weight function without any grid-tangling. The stress error estimator is calculated in the initial uniform mesh for a weight function. However, since the r-adaptive method is a method that moves the grid, shell surface geometry error such as curvature error and mesh distortion error will increase. Therefore, to represent the exact geometry of a shell surface and to prevent surface geometric errors, we use the Naghdi's shell theory and express the shell surface by a B-spline patch. In addition, using a nine-node element, which is relatively less sensitive to mesh distortion, we try to diminish mesh distortion error in the application of an r-adaptive method. In the numerical examples, it is shown that the values of the error estimator for a cylinder, hemisphere, and torus in the overall domain can be reduced effectively by using the mesh generated by the r-adaptive method. Also, the reductions of the estimated relative errors are demonstrated in the numerical examples. In particular, a new functional is proposed to construct an adjusted mesh configuration by considering a mesh distortion measure as well as the stress error function. The proposed weight function provides a reliable mesh adaptation method after a parameter value in the weight function is properly chosen

  12. Study on boundary search method for DFM mesh generation

    Li Ri

    2012-08-01

    Full Text Available The boundary mesh of the casting model was determined by direct calculation on the triangular facets extracted from the STL file of the 3D model. Then the inner and outer grids of the model were identified by the algorithm in which we named Inner Seed Grid Method. Finally, a program to automatically generate a 3D FDM mesh was compiled. In the paper, a method named Triangle Contraction Search Method (TCSM was put forward to ensure not losing the boundary grids; while an algorithm to search inner seed grids to identify inner/outer grids of the casting model was also brought forward. Our algorithm was simple, clear and easy to construct program. Three examples for the casting mesh generation testified the validity of the program.

  13. Loft: An Automated Mesh Generator for Stiffened Shell Aerospace Vehicles

    Eldred, Lloyd B.

    2011-01-01

    Loft is an automated mesh generation code that is designed for aerospace vehicle structures. From user input, Loft generates meshes for wings, noses, tanks, fuselage sections, thrust structures, and so on. As a mesh is generated, each element is assigned properties to mark the part of the vehicle with which it is associated. This property assignment is an extremely powerful feature that enables detailed analysis tasks, such as load application and structural sizing. This report is presented in two parts. The first part is an overview of the code and its applications. The modeling approach that was used to create the finite element meshes is described. Several applications of the code are demonstrated, including a Next Generation Launch Technology (NGLT) wing-sizing study, a lunar lander stage study, a launch vehicle shroud shape study, and a two-stage-to-orbit (TSTO) orbiter. Part two of the report is the program user manual. The manual includes in-depth tutorials and a complete command reference.

  14. Aranha: a 2D mesh generator for triangular finite elements

    Fancello, E.A.; Salgado, A.C.; Feijoo, R.A.

    1990-01-01

    A method for generating unstructured meshes for linear and quadratic triangular finite elements is described in this paper. Some topics on the C language data structure used in the development of the program Aranha are also presented. The applicability for adaptive remeshing is shown and finally several examples are included to illustrate the performance of the method in irregular connected planar domains. (author)

  15. MESH2D Grid generator design and use

    Flach, G. P. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL)

    2017-10-31

    Mesh2d is a Fortran90 program originally designed to generate two-dimensional structured grids of the form [x(i),y(i,j)] where [x,y] are grid coordinates identified by indices (i,j). x-coordinates depending only on index i implies strictly vertical x-grid lines, whereas the y-grid lines can undulate. Mesh2d also assigns an integer material type to each grid cell, mtyp(i,j), in a user-specified manner. The complete grid is specified through three separate input files defining the x(i), y(i,j), and mtyp(i,j) variations. Since the original development effort, Mesh2d has been extended to more general two-dimensional structured grids of the form [x(i,j),(i,j)].

  16. Manual for automatic generation of finite element models of spiral bevel gears in mesh

    Bibel, G. D.; Reddy, S.; Kumar, A.

    1994-01-01

    The goal of this research is to develop computer programs that generate finite element models suitable for doing 3D contact analysis of faced milled spiral bevel gears in mesh. A pinion tooth and a gear tooth are created and put in mesh. There are two programs: Points.f and Pat.f to perform the analysis. Points.f is based on the equation of meshing for spiral bevel gears. It uses machine tool settings to solve for an N x M mesh of points on the four surfaces, pinion concave and convex, and gear concave and convex. Points.f creates the file POINTS.OUT, an ASCI file containing N x M points for each surface. (N is the number of node points along the length of the tooth, and M is nodes along the height.) Pat.f reads POINTS.OUT and creates the file tl.out. Tl.out is a series of PATRAN input commands. In addition to the mesh density on the tooth face, additional user specified variables are the number of finite elements through the thickness, and the number of finite elements along the tooth full fillet. A full fillet is assumed to exist for both the pinion and gear.

  17. Unstructured Mesh Movement and Viscous Mesh Generation for CFD-Based Design Optimization, Phase II

    National Aeronautics and Space Administration — The innovations proposed are twofold: 1) a robust unstructured mesh movement method able to handle isotropic (Euler), anisotropic (viscous), mixed element (hybrid)...

  18. Coupling LaGrit unstructured mesh generation and model setup with TOUGH2 flow and transport: A case study

    Sentís, Manuel Lorenzo; Gable, Carl W.

    2017-11-01

    There are many applications in science and engineering modeling where an accurate representation of a complex model geometry in the form of a mesh is important. In applications of flow and transport in subsurface porous media, this is manifest in models that must capture complex geologic stratigraphy, structure (faults, folds, erosion, deposition) and infrastructure (tunnels, boreholes, excavations). Model setup, defined as the activities of geometry definition, mesh generation (creation, optimization, modification, refine, de-refine, smooth), assigning material properties, initial conditions and boundary conditions requires specialized software tools to automate and streamline the process. In addition, some model setup tools will provide more utility if they are designed to interface with and meet the needs of a particular flow and transport software suite. A control volume discretization that uses a two point flux approximation is for example most accurate when the underlying control volumes are 2D or 3D Voronoi tessellations. In this paper we will present the coupling of LaGriT, a mesh generation and model setup software suite and TOUGH2 (Pruess et al., 1999) to model subsurface flow problems and we show an example of how LaGriT can be used as a model setup tool for the generation of a Voronoi mesh for the simulation program TOUGH2. To generate the MESH file for TOUGH2 from the LaGriT output a standalone module Lagrit2Tough2 was developed, which is presented here and will be included in a future release of LaGriT. In this paper an alternative method to generate a Voronoi mesh for TOUGH2 with LaGriT is presented and thanks to the modular and command based structure of LaGriT this method is well suited to generating a mesh for complex models.

  19. Status of the Polyhedral Mesh Generator using SALOME PLATFORM and TetGen

    Lee, Sang Yong; Park, Chan Eok; Kim, Shin Whan

    2014-01-01

    Recently developed porous body approach codes such as SPACE and CUPID require a CAD system to estimate the porosity. Since they use the unstructured mesh and they also require reliable mesh generation system. The combination of CAD system and mesh generation system is necessary to cope with a large number of cells and the complex fluid system with structural materials inside. In the past, a CAD system Pro/Engineer and mesh generator Pointwise were evaluated for this application. But, the cost of those commercial CAD and mesh generator is sometimes a great burden. Therefore, efforts have been made to set up a mesh generation system with open source programs. The evaluation of the TetGen has been made in focusing the application for the polyhedral mesh generation. In this paper, SALOME will be described for the efforts to combine TetGen with it. In section 2, brief introduction will be made on the CAD and mesh generation capability of SALOME and Tetgen. SALOME and TetGen codes are being integrated to construct robust polyhedral mesh generator. Procedures to merge boundary faces and to cut concave cells are developed to remove concave cells to get final convex polyhedral mesh. Treating the internal boundary face, i.e. non-manifold face will be the next task in the future investigation

  20. Crack growth simulation for plural crack using hexahedral mesh generation technique

    Orita, Y; Wada, Y; Kikuchi, M

    2010-01-01

    This paper describes a surface crack growth simulation using a new mesh generation technique. The generated mesh is constituted of all hexahedral elements. Hexahedral elements are suitable for an analysis of fracture mechanics parameters, i.e. stress intensity factor. The advantage of a hexahedral mesh is good accuracy of an analysis and less number of degrees of freedoms than a tetrahedral mesh. In this study, a plural crack growth simulation is computed using the hexahedral mesh and its distribution of stress intensity factor is investigated.

  1. Efficient computation of clipped Voronoi diagram for mesh generation

    Yan, Dongming

    2013-04-01

    The Voronoi diagram is a fundamental geometric structure widely used in various fields, especially in computer graphics and geometry computing. For a set of points in a compact domain (i.e. a bounded and closed 2D region or a 3D volume), some Voronoi cells of their Voronoi diagram are infinite or partially outside of the domain, but in practice only the parts of the cells inside the domain are needed, as when computing the centroidal Voronoi tessellation. Such a Voronoi diagram confined to a compact domain is called a clipped Voronoi diagram. We present an efficient algorithm to compute the clipped Voronoi diagram for a set of sites with respect to a compact 2D region or a 3D volume. We also apply the proposed method to optimal mesh generation based on the centroidal Voronoi tessellation. Crown Copyright © 2011 Published by Elsevier Ltd. All rights reserved.

  2. Efficient computation of clipped Voronoi diagram for mesh generation

    Yan, Dongming; Wang, Wen Ping; Lé vy, Bruno L.; Liu, Yang

    2013-01-01

    The Voronoi diagram is a fundamental geometric structure widely used in various fields, especially in computer graphics and geometry computing. For a set of points in a compact domain (i.e. a bounded and closed 2D region or a 3D volume), some Voronoi cells of their Voronoi diagram are infinite or partially outside of the domain, but in practice only the parts of the cells inside the domain are needed, as when computing the centroidal Voronoi tessellation. Such a Voronoi diagram confined to a compact domain is called a clipped Voronoi diagram. We present an efficient algorithm to compute the clipped Voronoi diagram for a set of sites with respect to a compact 2D region or a 3D volume. We also apply the proposed method to optimal mesh generation based on the centroidal Voronoi tessellation. Crown Copyright © 2011 Published by Elsevier Ltd. All rights reserved.

  3. Web Tools: The Second Generation

    Pascopella, Angela

    2008-01-01

    Web 2.0 tools and technologies, or second generation tools, help districts to save time and money, and eliminate the need to transfer or move files back and forth across computers. Many Web 2.0 tools help students think critically and solve problems, which falls under the 21st-century skills. The second-generation tools are growing in popularity…

  4. h-Adaptive Mesh Generation using Electric Field Intensity Value as a Criterion (in Japanese)

    Toyonaga, Kiyomi; Cingoski, Vlatko; Kaneda, Kazufumi; Yamashita, Hideo

    1994-01-01

    Finite mesh divisions are essential to obtain accurate solution of two dimensional electric field analysis. It requires the technical knowledge to generate a suitable fine mesh divisions. In electric field problem, analysts are usually interested in the electric field intensity and its distribution. In order to obtain electric field intensity with high-accuracy, we have developed and adaptive mesh generator using electric field intensity value as a criterion.

  5. Adaptive mesh generation for image registration and segmentation

    Fogtmann, Mads; Larsen, Rasmus

    2013-01-01

    measure. The method was tested on a T1 weighted MR volume of an adult brain and showed a 66% reduction in the number of mesh vertices compared to a red-subdivision strategy. The deformation capability of the mesh was tested by registration to five additional T1-weighted MR volumes....

  6. Challenges in Second-Generation Wireless Mesh Networks

    Pescapé Antonio

    2008-01-01

    Full Text Available Wireless mesh networks have the potential to provide ubiquitous high-speed Internet access at low costs. The good news is that initial deployments of WiFi meshes show the feasibility of providing ubiquitous Internet connectivity. However, their performance is far below the necessary and achievable limit. Moreover, users' subscription in the existing meshes is dismal even though the technical challenges to get connectivity are low. This paper provides an overview of the current status of mesh networks' deployment, and highlights the technical, economical, and social challenges that need to be addressed in the next years. As a proof-of-principle study, we discuss the above-mentioned challenges with reference to three real networks: (i MagNets, an operator-driven planned two-tier mesh network; (ii Berlin Freifunk network as a pure community-driven single-tier network; (iii Weimar Freifunk network, also a community-driven but two-tier network.

  7. LOOM-P: a finite element mesh generation program with on-line graphic display

    Ise, Takeharu; Yamazaki, Toshio.

    1977-06-01

    A description of the two-dimensional mesh generation program, LOOM-P, is given in detail. The program is developed newly to produce a mesh network for a reactor core geometry with the help of an automatic mesh generation routine built in it, under the control of the refresh-type graphic display. It is therefore similar to the edit program of the self-organizing mesh generator, QMESH-RENUM. Additional techniques are incorporated to improve the pattern of mesh elements by means of on-line conversational mode. The obtained mesh network is edited out as input data to the three-dimensional neutron diffusion theory code, FEM-BABEL. (auth.)

  8. ORMGEN3D, 3-D Crack Geometry FEM Mesh Generator

    Bass, B.R.; Bryson, J.W.

    1994-01-01

    1 - Description of program or function: ORMGEN3D is a finite element mesh generator for computational fracture mechanics analysis. The program automatically generates a three-dimensional finite element model for six different crack geometries. These geometries include flat plates with straight or curved surface cracks and cylinders with part-through cracks on the outer or inner surface. Mathematical or user-defined crack shapes may be considered. The curved cracks may be semicircular, semi-elliptical, or user-defined. A cladding option is available that allows for either an embedded or penetrating crack in the clad material. 2 - Method of solution: In general, one eighth or one-quarter of the structure is modelled depending on the configuration or option selected. The program generates a core of special wedge or collapsed prism elements at the crack front to introduce the appropriate stress singularity at the crack tip. The remainder of the structure is modelled with conventional 20-node iso-parametric brick elements. Element group I of the finite element model consists of an inner core of special crack tip elements surrounding the crack front enclosed by a single layer of conventional brick elements. Eight element divisions are used in a plane orthogonal to the crack front, while the number of element divisions along the arc length of the crack front is user-specified. The remaining conventional brick elements of the model constitute element group II. 3 - Restrictions on the complexity of the problem: Maxima of 5,500 nodes, 4 layers of clad elements

  9. Opfront: mesh

    2015-01-01

    Mesh generation and visualization software based on the CGAL library. Folder content: drawmesh Visualize slices of the mesh (surface/volumetric) as wireframe on top of an image (3D). drawsurf Visualize surfaces of the mesh (surface/volumetric). img2mesh Convert isosurface in image to volumetric m...... mesh (medit format). img2off Convert isosurface in image to surface mesh (off format). off2mesh Convert surface mesh (off format) to volumetric mesh (medit format). reduce Crop and resize 3D and stacks of images. data Example data to test the library on...

  10. EURCYL. A program to generate finite element meshes for pressure vessel nozzles

    De Windt, P.; Reynen, J.

    1974-12-01

    EURCYL is a program dealing with the automatic generation of finite element meshes for pressure vessel nozzles, using isoparametric elements with 8, 20 or 32 nodes. Options exist to generate BWR nozzles as well as PWR nozzles

  11. Parallel paving: An algorithm for generating distributed, adaptive, all-quadrilateral meshes on parallel computers

    Lober, R.R.; Tautges, T.J.; Vaughan, C.T.

    1997-03-01

    Paving is an automated mesh generation algorithm which produces all-quadrilateral elements. It can additionally generate these elements in varying sizes such that the resulting mesh adapts to a function distribution, such as an error function. While powerful, conventional paving is a very serial algorithm in its operation. Parallel paving is the extension of serial paving into parallel environments to perform the same meshing functions as conventional paving only on distributed, discretized models. This extension allows large, adaptive, parallel finite element simulations to take advantage of paving`s meshing capabilities for h-remap remeshing. A significantly modified version of the CUBIT mesh generation code has been developed to host the parallel paving algorithm and demonstrate its capabilities on both two dimensional and three dimensional surface geometries and compare the resulting parallel produced meshes to conventionally paved meshes for mesh quality and algorithm performance. Sandia`s {open_quotes}tiling{close_quotes} dynamic load balancing code has also been extended to work with the paving algorithm to retain parallel efficiency as subdomains undergo iterative mesh refinement.

  12. Mesh Generation and Adaption for High Reynolds Number RANS Computations, Phase I

    National Aeronautics and Space Administration — This proposal offers to provide NASA with an automatic mesh generator for the simulation of aerodynamic flows using Reynolds-Averages Navier-Stokes (RANS) models....

  13. Mesh Generation and Adaption for High Reynolds Number RANS Computations, Phase II

    National Aeronautics and Space Administration — This proposal offers to provide NASA with an automatic mesh generator for the simulation of aerodynamic flows using Reynolds-Averages Navier-Stokes (RANS) models....

  14. Generation of hybrid meshes for the simulation of petroleum reservoirs; Generation de maillages hybrides pour la simulation de reservoirs petroliers

    Balaven-Clermidy, S.

    2001-12-01

    Oil reservoir simulations study multiphase flows in porous media. These flows are described and evaluated through numerical schemes on a discretization of the reservoir domain. In this thesis, we were interested in this spatial discretization and a new kind of hybrid mesh has been proposed where the radial nature of flows in the vicinity of wells is directly taken into account in the geometry. Our modular approach described wells and their drainage area through radial circular meshes. These well meshes are inserted in a structured reservoir mesh (a Corner Point Geometry mesh) made up with hexahedral cells. Finally, in order to generate a global conforming mesh, proper connections are realized between the different kinds of meshes through unstructured transition ones. To compute these transition meshes that we want acceptable in terms of finite volume methods, an automatic method based on power diagrams has been developed. Our approach can deal with a homogeneous anisotropic medium and allows the user to insert vertical or horizontal wells as well as secondary faults in the reservoir mesh. Our work has been implemented, tested and validated in 2D and 2D1/2. It can also be extended in 3D when the geometrical constraints are simplicial ones: points, segments and triangles. (author)

  15. Molecular surface mesh generation by filtering electron density map.

    Giard, Joachim; Macq, Benoît

    2010-01-01

    Bioinformatics applied to macromolecules are now widely spread and in continuous expansion. In this context, representing external molecular surface such as the Van der Waals Surface or the Solvent Excluded Surface can be useful for several applications. We propose a fast and parameterizable algorithm giving good visual quality meshes representing molecular surfaces. It is obtained by isosurfacing a filtered electron density map. The density map is the result of the maximum of Gaussian functions placed around atom centers. This map is filtered by an ideal low-pass filter applied on the Fourier Transform of the density map. Applying the marching cubes algorithm on the inverse transform provides a mesh representation of the molecular surface.

  16. Molecular Surface Mesh Generation by Filtering Electron Density Map

    Joachim Giard

    2010-01-01

    Full Text Available Bioinformatics applied to macromolecules are now widely spread and in continuous expansion. In this context, representing external molecular surface such as the Van der Waals Surface or the Solvent Excluded Surface can be useful for several applications. We propose a fast and parameterizable algorithm giving good visual quality meshes representing molecular surfaces. It is obtained by isosurfacing a filtered electron density map. The density map is the result of the maximum of Gaussian functions placed around atom centers. This map is filtered by an ideal low-pass filter applied on the Fourier Transform of the density map. Applying the marching cubes algorithm on the inverse transform provides a mesh representation of the molecular surface.

  17. Automated hexahedral mesh generation from biomedical image data: applications in limb prosthetics.

    Zachariah, S G; Sanders, J E; Turkiyyah, G M

    1996-06-01

    A general method to generate hexahedral meshes for finite element analysis of residual limbs and similar biomedical geometries is presented. The method utilizes skeleton-based subdivision of cross-sectional domains to produce simple subdomains in which structured meshes are easily generated. Application to a below-knee residual limb and external prosthetic socket is described. The residual limb was modeled as consisting of bones, soft tissue, and skin. The prosthetic socket model comprised a socket wall with an inner liner. The geometries of these structures were defined using axial cross-sectional contour data from X-ray computed tomography, optical scanning, and mechanical surface digitization. A tubular surface representation, using B-splines to define the directrix and generator, is shown to be convenient for definition of the structure geometries. Conversion of cross-sectional data to the compact tubular surface representation is direct, and the analytical representation simplifies geometric querying and numerical optimization within the mesh generation algorithms. The element meshes remain geometrically accurate since boundary nodes are constrained to lie on the tubular surfaces. Several element meshes of increasing mesh density were generated for two residual limbs and prosthetic sockets. Convergence testing demonstrated that approximately 19 elements are required along a circumference of the residual limb surface for a simple linear elastic model. A model with the fibula absent compared with the same geometry with the fibula present showed differences suggesting higher distal stresses in the absence of the fibula. Automated hexahedral mesh generation algorithms for sliced data represent an advancement in prosthetic stress analysis since they allow rapid modeling of any given residual limb and optimization of mesh parameters.

  18. INGEN: a general-purpose mesh generator for finite element codes

    Cook, W.A.

    1979-05-01

    INGEN is a general-purpose mesh generator for two- and three-dimensional finite element codes. The basic parts of the code are surface and three-dimensional region generators that use linear-blending interpolation formulas. These generators are based on an i, j, k index scheme that is used to number nodal points, construct elements, and develop displacement and traction boundary conditions. This code can generate truss elements (2 modal points); plane stress, plane strain, and axisymmetry two-dimensional continuum elements (4 to 8 nodal points); plate elements (4 to 8 nodal points); and three-dimensional continuum elements (8 to 21 nodal points). The traction loads generated are consistent with the element generated. The expansion--contraction option is of special interest. This option makes it possible to change an existing mesh such that some regions are refined and others are made coarser than the original mesh. 9 figures

  19. New software developments for quality mesh generation and optimization from biomedical imaging data.

    Yu, Zeyun; Wang, Jun; Gao, Zhanheng; Xu, Ming; Hoshijima, Masahiko

    2014-01-01

    In this paper we present a new software toolkit for generating and optimizing surface and volumetric meshes from three-dimensional (3D) biomedical imaging data, targeted at image-based finite element analysis of some biomedical activities in a single material domain. Our toolkit includes a series of geometric processing algorithms including surface re-meshing and quality-guaranteed tetrahedral mesh generation and optimization. All methods described have been encapsulated into a user-friendly graphical interface for easy manipulation and informative visualization of biomedical images and mesh models. Numerous examples are presented to demonstrate the effectiveness and efficiency of the described methods and toolkit. Copyright © 2013 Elsevier Ireland Ltd. All rights reserved.

  20. ZONE: a finite element mesh generator. [In FORTRAN IV for CDC 7600

    Burger, M. J.

    1976-05-01

    The ZONE computer program is a finite-element mesh generator which produces the nodes and element description of any two-dimensional geometry. The geometry is subdivided into a mesh of quadrilateral and triangular zones arranged sequentially in an ordered march through the geometry. The order of march can be chosen so that the minimum bandwidth is obtained. The node points are defined in terms of the x and y coordinates in a global rectangular coordinate system. The zones generated are quadrilaterals or triangles defined by four node points in a counterclockwise sequence. Node points defining the outside boundary are generated to describe pressure boundary conditions. The mesh that is generated can be used as input to any two-dimensional as well as any axisymmetrical structure program. The output from ZONE is essentially the input file to NAOS, HONDO, and other axisymmetric finite element programs. 14 figures. (RWR)

  1. Atlas-Based Automatic Generation of Subject-Specific Finite Element Tongue Meshes.

    Bijar, Ahmad; Rohan, Pierre-Yves; Perrier, Pascal; Payan, Yohan

    2016-01-01

    Generation of subject-specific 3D finite element (FE) models requires the processing of numerous medical images in order to precisely extract geometrical information about subject-specific anatomy. This processing remains extremely challenging. To overcome this difficulty, we present an automatic atlas-based method that generates subject-specific FE meshes via a 3D registration guided by Magnetic Resonance images. The method extracts a 3D transformation by registering the atlas' volume image to the subject's one, and establishes a one-to-one correspondence between the two volumes. The 3D transformation field deforms the atlas' mesh to generate the subject-specific FE mesh. To preserve the quality of the subject-specific mesh, a diffeomorphic non-rigid registration based on B-spline free-form deformations is used, which guarantees a non-folding and one-to-one transformation. Two evaluations of the method are provided. First, a publicly available CT-database is used to assess the capability to accurately capture the complexity of each subject-specific Lung's geometry. Second, FE tongue meshes are generated for two healthy volunteers and two patients suffering from tongue cancer using MR images. It is shown that the method generates an appropriate representation of the subject-specific geometry while preserving the quality of the FE meshes for subsequent FE analysis. To demonstrate the importance of our method in a clinical context, a subject-specific mesh is used to simulate tongue's biomechanical response to the activation of an important tongue muscle, before and after cancer surgery.

  2. Mobility Models for Next Generation Wireless Networks Ad Hoc, Vehicular and Mesh Networks

    Santi, Paolo

    2012-01-01

    Mobility Models for Next Generation Wireless Networks: Ad Hoc, Vehicular and Mesh Networks provides the reader with an overview of mobility modelling, encompassing both theoretical and practical aspects related to the challenging mobility modelling task. It also: Provides up-to-date coverage of mobility models for next generation wireless networksOffers an in-depth discussion of the most representative mobility models for major next generation wireless network application scenarios, including WLAN/mesh networks, vehicular networks, wireless sensor networks, and

  3. An Approach for Patient-Specific Multi-domain Vascular Mesh Generation Featuring Spatially Varying Wall Thickness Modeling

    Raut, Samarth S.; Liu, Peng; Finol, Ender A.

    2015-01-01

    In this work, we present a computationally efficient image-derived volume mesh generation approach for vasculatures that implements spatially varying patient-specific wall thickness with a novel inward extrusion of the wall surface mesh. Multi-domain vascular meshes with arbitrary numbers, locations, and patterns of both iliac bifurcations and thrombi can be obtained without the need to specify features or landmark points as input. In addition, the mesh output is coordinate-frame independent ...

  4. An expert system for automatic mesh generation for Sn particle transport simulation in parallel environment

    Apisit, Patchimpattapong; Alireza, Haghighat; Shedlock, D.

    2003-01-01

    An expert system for generating an effective mesh distribution for the SN particle transport simulation has been developed. This expert system consists of two main parts: 1) an algorithm for generating an effective mesh distribution in a serial environment, and 2) an algorithm for inference of an effective domain decomposition strategy for parallel computing. For the first part, the algorithm prepares an effective mesh distribution considering problem physics and the spatial differencing scheme. For the second part, the algorithm determines a parallel-performance-index (PPI), which is defined as the ratio of the granularity to the degree-of-coupling. The parallel-performance-index provides expected performance of an algorithm depending on computing environment and resources. A large index indicates a high granularity algorithm with relatively low coupling among processors. This expert system has been successfully tested within the PENTRAN (Parallel Environment Neutral-Particle Transport) code system for simulating real-life shielding problems. (authors)

  5. 2D automatic body-fitted structured mesh generation using advancing extraction method

    Zhang, Yaoxin; Jia, Yafei

    2018-01-01

    This paper presents an automatic mesh generation algorithm for body-fitted structured meshes in Computational Fluids Dynamics (CFD) analysis using the Advancing Extraction Method (AEM). The method is applicable to two-dimensional domains with complex geometries, which have the hierarchical tree-like topography with extrusion-like structures (i.e., branches or tributaries) and intrusion-like structures (i.e., peninsula or dikes). With the AEM, the hierarchical levels of sub-domains can be identified, and the block boundary of each sub-domain in convex polygon shape in each level can be extracted in an advancing scheme. In this paper, several examples were used to illustrate the effectiveness and applicability of the proposed algorithm for automatic structured mesh generation, and the implementation of the method.

  6. An expert system for automatic mesh generation for Sn particle transport simulation in parallel environment

    Apisit, Patchimpattapong [Electricity Generating Authority of Thailand, Office of Corporate Planning, Bangkruai, Nonthaburi (Thailand); Alireza, Haghighat; Shedlock, D. [Florida Univ., Department of Nuclear and Radiological Engineering, Gainesville, FL (United States)

    2003-07-01

    An expert system for generating an effective mesh distribution for the SN particle transport simulation has been developed. This expert system consists of two main parts: 1) an algorithm for generating an effective mesh distribution in a serial environment, and 2) an algorithm for inference of an effective domain decomposition strategy for parallel computing. For the first part, the algorithm prepares an effective mesh distribution considering problem physics and the spatial differencing scheme. For the second part, the algorithm determines a parallel-performance-index (PPI), which is defined as the ratio of the granularity to the degree-of-coupling. The parallel-performance-index provides expected performance of an algorithm depending on computing environment and resources. A large index indicates a high granularity algorithm with relatively low coupling among processors. This expert system has been successfully tested within the PENTRAN (Parallel Environment Neutral-Particle Transport) code system for simulating real-life shielding problems. (authors)

  7. SLIC: an interactive mesh generator for finite element and finite difference application programs

    Gerhard, M.A.; Greenlaw, R.C.

    1979-01-01

    Computers with extended memory, such as the CDC STAR 100 and the CRAY 1 with mega-word capacities, are greatly enlarging the size of finite element problems which can be solved. The cost of developing and testing large meshes can be prohibitive unless one uses a computer program for mesh generation and plotting. SLIC is an interactive mesh program which builds and plots 2- and 3-D continuum meshes from interactive terminal or disc input. The user inputs coordinates for certain key points and enters commands which complete the description of the geometry. Entire surfaces and volumes are then generated from the geometric skeleton. SLIC allows the user to correct input errors and saves the corrected command list for later reuse. The mesh can be plotted on a video display at any stage of development to evaluate the work in progress. Output is in the form of an input file to a user-selected computer code. Among the available output types are ADINA, SAP4, and NIKE2D. 11 figures

  8. Transonic Airfoil Flow Simulation. Part I: Mesh Generation and Inviscid Method

    Vladimir CARDOS

    2010-06-01

    Full Text Available A calculation method for the subsonic and transonic viscous flow over airfoil using thedisplacement surface concept is described. Part I presents a mesh generation method forcomputational grid and a finite volume method for the time-dependent Euler equations. The inviscidsolution is used for the inviscid-viscous coupling procedure presented in the Part II.

  9. Medical Image Processing for Fully Integrated Subject Specific Whole Brain Mesh Generation

    Chih-Yang Hsu

    2015-05-01

    Full Text Available Currently, anatomically consistent segmentation of vascular trees acquired with magnetic resonance imaging requires the use of multiple image processing steps, which, in turn, depend on manual intervention. In effect, segmentation of vascular trees from medical images is time consuming and error prone due to the tortuous geometry and weak signal in small blood vessels. To overcome errors and accelerate the image processing time, we introduce an automatic image processing pipeline for constructing subject specific computational meshes for entire cerebral vasculature, including segmentation of ancillary structures; the grey and white matter, cerebrospinal fluid space, skull, and scalp. To demonstrate the validity of the new pipeline, we segmented the entire intracranial compartment with special attention of the angioarchitecture from magnetic resonance imaging acquired for two healthy volunteers. The raw images were processed through our pipeline for automatic segmentation and mesh generation. Due to partial volume effect and finite resolution, the computational meshes intersect with each other at respective interfaces. To eliminate anatomically inconsistent overlap, we utilized morphological operations to separate the structures with a physiologically sound gap spaces. The resulting meshes exhibit anatomically correct spatial extent and relative positions without intersections. For validation, we computed critical biometrics of the angioarchitecture, the cortical surfaces, ventricular system, and cerebrospinal fluid (CSF spaces and compared against literature values. Volumina and surface areas of the computational mesh were found to be in physiological ranges. In conclusion, we present an automatic image processing pipeline to automate the segmentation of the main intracranial compartments including a subject-specific vascular trees. These computational meshes can be used in 3D immersive visualization for diagnosis, surgery planning with haptics

  10. Mapping method for generating three-dimensional meshes: past and present

    Cook, W.A.; Oakes, W.R.

    1982-01-01

    Two transformations are derived in this paper. One is a mapping of a unit square onto a surve and the other is a mapping of a unit cube onto a three-dimensional region. Two meshing computer programs are then discussed that use these mappings. The first is INGEN, which has been used to calculate three-dimensional meshes for approximately 15 years. This meshing program uses an index scheme to number boundaries, surfaces, and regions. With such an index scheme, it is possible to control nodal points, elements, and boundary conditions. The second is ESCHER, a meshing program now being developed. Two primary considerations governing development of ESCHER are that meshes graded using quadrilaterals are required and that edge-line geometry defined by Computer-Aided Design/Computer-Aided Manufacturing (CAD/CAM) systems will be a major source of geometry definition. This program separates the processes of nodal-point connectivity generation, computation of nodal-point mapping space coordinates, and mapping of nodal points into model space

  11. Parallel octree-based hexahedral mesh generation for eulerian to lagrangian conversion.

    Staten, Matthew L.; Owen, Steven James

    2010-09-01

    Computational simulation must often be performed on domains where materials are represented as scalar quantities or volume fractions at cell centers of an octree-based grid. Common examples include bio-medical, geotechnical or shock physics calculations where interface boundaries are represented only as discrete statistical approximations. In this work, we introduce new methods for generating Lagrangian computational meshes from Eulerian-based data. We focus specifically on shock physics problems that are relevant to ASC codes such as CTH and Alegra. New procedures for generating all-hexahedral finite element meshes from volume fraction data are introduced. A new primal-contouring approach is introduced for defining a geometric domain. New methods for refinement, node smoothing, resolving non-manifold conditions and defining geometry are also introduced as well as an extension of the algorithm to handle tetrahedral meshes. We also describe new scalable MPI-based implementations of these procedures. We describe a new software module, Sculptor, which has been developed for use as an embedded component of CTH. We also describe its interface and its use within the mesh generation code, CUBIT. Several examples are shown to illustrate the capabilities of Sculptor.

  12. 3D active shape models of human brain structures: application to patient-specific mesh generation

    Ravikumar, Nishant; Castro-Mateos, Isaac; Pozo, Jose M.; Frangi, Alejandro F.; Taylor, Zeike A.

    2015-03-01

    The use of biomechanics-based numerical simulations has attracted growing interest in recent years for computer-aided diagnosis and treatment planning. With this in mind, a method for automatic mesh generation of brain structures of interest, using statistical models of shape (SSM) and appearance (SAM), for personalised computational modelling is presented. SSMs are constructed as point distribution models (PDMs) while SAMs are trained using intensity profiles sampled from a training set of T1-weighted magnetic resonance images. The brain structures of interest are, the cortical surface (cerebrum, cerebellum & brainstem), lateral ventricles and falx-cerebri membrane. Two methods for establishing correspondences across the training set of shapes are investigated and compared (based on SSM quality): the Coherent Point Drift (CPD) point-set registration method and B-spline mesh-to-mesh registration method. The MNI-305 (Montreal Neurological Institute) average brain atlas is used to generate the template mesh, which is deformed and registered to each training case, to establish correspondence over the training set of shapes. 18 healthy patients' T1-weightedMRimages form the training set used to generate the SSM and SAM. Both model-training and model-fitting are performed over multiple brain structures simultaneously. Compactness and generalisation errors of the BSpline-SSM and CPD-SSM are evaluated and used to quantitatively compare the SSMs. Leave-one-out cross validation is used to evaluate SSM quality in terms of these measures. The mesh-based SSM is found to generalise better and is more compact, relative to the CPD-based SSM. Quality of the best-fit model instance from the trained SSMs, to test cases are evaluated using the Hausdorff distance (HD) and mean absolute surface distance (MASD) metrics.

  13. 6th International Meshing Roundtable '97

    White, D.

    1997-09-01

    The goal of the 6th International Meshing Roundtable is to bring together researchers and developers from industry, academia, and government labs in a stimulating, open environment for the exchange of technical information related to the meshing process. In the pas~ the Roundtable has enjoyed significant participation born each of these groups from a wide variety of countries. The Roundtable will consist of technical presentations from contributed papers and abstracts, two invited speakers, and two invited panels of experts discussing topics related to the development and use of automatic mesh generation tools. In addition, this year we will feature a "Bring Your Best Mesh" competition and poster session to encourage discussion and participation from a wide variety of mesh generation tool users. The schedule and evening social events are designed to provide numerous opportunities for informal dialog. A proceedings will be published by Sandia National Laboratories and distributed at the Roundtable. In addition, papers of exceptionally high quaIity will be submitted to a special issue of the International Journal of Computational Geometry and Applications. Papers and one page abstracts were sought that present original results on the meshing process. Potential topics include but are got limited to: Unstructured triangular and tetrahedral mesh generation Unstructured quadrilateral and hexahedral mesh generation Automated blocking and structured mesh generation Mixed element meshing Surface mesh generation Geometry decomposition and clean-up techniques Geometry modification techniques related to meshing Adaptive mesh refinement and mesh quality control Mesh visualization Special purpose meshing algorithms for particular applications Theoretical or novel ideas with practical potential Technical presentations from industrial researchers.

  14. Automatic two- and three-dimensional mesh generation based on fuzzy knowledge processing

    Yagawa, G.; Yoshimura, S.; Soneda, N.; Nakao, K.

    1992-09-01

    This paper describes the development of a novel automatic FEM mesh generation algorithm based on the fuzzy knowledge processing technique. A number of local nodal patterns are stored in a nodal pattern database of the mesh generation system. These nodal patterns are determined a priori based on certain theories or past experience of experts of FEM analyses. For example, such human experts can determine certain nodal patterns suitable for stress concentration analyses of cracks, corners, holes and so on. Each nodal pattern possesses a membership function and a procedure of node placement according to this function. In the cases of the nodal patterns for stress concentration regions, the membership function which is utilized in the fuzzy knowledge processing has two meanings, i.e. the “closeness” of nodal location to each stress concentration field as well as “nodal density”. This is attributed to the fact that a denser nodal pattern is required near a stress concentration field. What a user has to do in a practical mesh generation process are to choose several local nodal patterns properly and to designate the maximum nodal density of each pattern. After those simple operations by the user, the system places the chosen nodal patterns automatically in an analysis domain and on its boundary, and connects them smoothly by the fuzzy knowledge processing technique. Then triangular or tetrahedral elements are generated by means of the advancing front method. The key issue of the present algorithm is an easy control of complex two- or three-dimensional nodal density distribution by means of the fuzzy knowledge processing technique. To demonstrate fundamental performances of the present algorithm, a prototype system was constructed with one of object-oriented languages, Smalltalk-80 on a 32-bit microcomputer, Macintosh II. The mesh generation of several two- and three-dimensional domains with cracks, holes and junctions was presented as examples.

  15. Algebraic mesh generation for large scale viscous-compressible aerodynamic simulation

    Smith, R.E.

    1984-01-01

    Viscous-compressible aerodynamic simulation is the numerical solution of the compressible Navier-Stokes equations and associated boundary conditions. Boundary-fitted coordinate systems are well suited for the application of finite difference techniques to the Navier-Stokes equations. An algebraic approach to boundary-fitted coordinate systems is one where an explicit functional relation describes a mesh on which a solution is obtained. This approach has the advantage of rapid-precise mesh control. The basic mathematical structure of three algebraic mesh generation techniques is described. They are transfinite interpolation, the multi-surface method, and the two-boundary technique. The Navier-Stokes equations are transformed to a computational coordinate system where boundary-fitted coordinates can be applied. Large-scale computation implies that there is a large number of mesh points in the coordinate system. Computation of viscous compressible flow using boundary-fitted coordinate systems and the application of this computational philosophy on a vector computer are presented

  16. Mesh generation and energy group condensation studies for the jaguar deterministic transport code

    Kennedy, R. A.; Watson, A. M.; Iwueke, C. I.; Edwards, E. J.

    2012-01-01

    The deterministic transport code Jaguar is introduced, and the modeling process for Jaguar is demonstrated using a two-dimensional assembly model of the Hoogenboom-Martin Performance Benchmark Problem. This single assembly model is being used to test and analyze optimal modeling methodologies and techniques for Jaguar. This paper focuses on spatial mesh generation and energy condensation techniques. In this summary, the models and processes are defined as well as thermal flux solution comparisons with the Monte Carlo code MC21. (authors)

  17. Mesh generation and energy group condensation studies for the jaguar deterministic transport code

    Kennedy, R. A.; Watson, A. M.; Iwueke, C. I.; Edwards, E. J. [Knolls Atomic Power Laboratory, Bechtel Marine Propulsion Corporation, P.O. Box 1072, Schenectady, NY 12301-1072 (United States)

    2012-07-01

    The deterministic transport code Jaguar is introduced, and the modeling process for Jaguar is demonstrated using a two-dimensional assembly model of the Hoogenboom-Martin Performance Benchmark Problem. This single assembly model is being used to test and analyze optimal modeling methodologies and techniques for Jaguar. This paper focuses on spatial mesh generation and energy condensation techniques. In this summary, the models and processes are defined as well as thermal flux solution comparisons with the Monte Carlo code MC21. (authors)

  18. Comment on "A note on generalized radial mesh generation for plasma electronic structure"

    Pain, J.-Ch.

    2011-12-01

    In a recent note, B.G. Wilson and V. Sonnad [1] proposed a very useful closed form expression for the efficient generation of analytic log-linear radial meshes. The central point of the note is an implicit equation for the parameter h, involving Lambert's function W[x]. The authors mention that they are unaware of any direct proof of this equation (they obtained it by re-summing the Taylor expansion of h[α] using high-order coefficients obtained by analytic differentiation of the implicit definition using symbolic manipulation). In the present comment, we propose a direct proof of that equation.

  19. ZONE, Finite Elements Method Quadrilateral and Triangular Mesh Generator for 2-D Axisymmetric Geometry

    Burger, M. J.

    1981-01-01

    1 - Description of problem or function: The ZONE program is a finite element mesh generator which produces the nodes and element description of any two-dimensional geometry. The geometry is divided into a mesh of quadrilateral and triangular zones defined by node points taken in a counter-clockwise sequence. The zones are arranged sequentially in an ordered march through the geometry. The order can be chosen so that the minimum bandwidth is obtained. The mesh that is generated can be used as input to any two-dimensional as well as any axisymmetrical structure program. 2 - Method of solution: The basic concept used is the definition of a two-dimensional structure by the intersection of two sets of lines which describe the geometric and material boundaries. A set of lines called meridians define the geometric and material boundaries and generally run in the same direction. Another set of linear line segments called rays which intersect the meridians are also defined at the material and geometric boundaries. The section of the structure between successive rays is called a region. The ray segment between any two consecutive ray-meridian intersections or void area in the structure is called a layer and is described as passing through, or bounding a material. The boundaries can be directly defined as a sequence of straight line segments or can be computed in terms of elliptic segments or circular arcs. A meridian or ray can also be made to follow a previously-defined meridian or ray at a fixed distance by invoking an offset option. 3 - Restrictions on the complexity of the problem: The following are limited only by a DIMENSION statement. The code currently has a maxima of: 100 coordinate points defining a meridian or ray, 40 meridians, 40 layers. There are no limits on the number of zones or nodes for any problems

  20. Systems Prototyping with Fourth Generation Tools.

    Sholtys, Phyllis

    1983-01-01

    The development of information systems using an engineering approach that uses both traditional programing techniques and fourth generation software tools is described. Fourth generation applications tools are used to quickly develop a prototype system that is revised as the user clarifies requirements. (MLW)

  1. Automatic mesh generation for finite element calculations in the case of thermal loads

    Cords, H.; Zimmermann, R.

    1975-01-01

    The presentation describes a method to generate finite element nodal point networks on the basis of isothermals and flux lines. Such a mesh provides a relatively fine partitioning at regions where pronounced temperature variations exist. In case of entirely thermal loads a net of this kind is advantageous since the refinement is provided at exactly those locations where high stress levels are expected. In the present contribution the method was employed to analyze the structural behavior of a nuclear fuel element under operating conditions. The graphite block fuel elements for high temperature reactors are of prismatic shape with a large number of parallel bores in the axial direction. Some of these bores are open at both ends and cooling is effected by helium flowing through. Blind holes contain the fuel as compacts or cartridges. The basic temperature distribution in a horizontal section of the block was obtained by the boundary point least squares method which yields analytical expressions for both temperature and thermal flux. The corresponding computer code was presented at an earlier SMiRT conference. The method is particularly useful for regular arrays of heat sources and sinks as encountered in heat exchanger problems. The generated mesh matches the requirements of a subsequent structural analysis with finite elements provided there are no other than thermal loads

  2. Optical breast shape capture and finite-element mesh generation for electrical impedance tomography

    Forsyth, J; Borsic, A; Halter, R J; Hartov, A; Paulsen, K D

    2011-01-01

    X-ray mammography is the standard for breast cancer screening. The development of alternative imaging modalities is desirable because mammograms expose patients to ionizing radiation. Electrical impedance tomography (EIT) may be used to determine tissue conductivity, a property which is an indicator of cancer presence. EIT is also a low-cost imaging solution and does not involve ionizing radiation. In breast EIT, impedance measurements are made using electrodes placed on the surface of the patient's breast. The complex conductivity of the volume of the breast is estimated by a reconstruction algorithm. EIT reconstruction is a severely ill-posed inverse problem. As a result, noisy instrumentation and incorrect modelling of the electrodes and domain shape produce significant image artefacts. In this paper, we propose a method that has the potential to reduce these errors by accurately modelling the patient breast shape. A 3D hand-held optical scanner is used to acquire the breast geometry and electrode positions. We develop methods for processing the data from the scanner and producing volume meshes accurately matching the breast surface and electrode locations, which can be used for image reconstruction. We demonstrate this method for a plaster breast phantom and a human subject. Using this approach will allow patient-specific finite-element meshes to be generated which has the potential to improve the clinical value of EIT for breast cancer diagnosis

  3. Generation of reservoir models on flexible meshes; Generation de modeles de reservoir sur maillage flexible

    Ricard, L.

    2005-12-15

    The high level geo-statistic description of the subsurface are often far too detailed for use in routine flow simulators. To make flow simulations tractable, the number of grid blocks has to be reduced: an approximation, still relevant with flow description, is necessary. In this work, we place the emphasis on the scaling procedure from the fine scale model to the multi-scale reservoir model. Two main problems appear: Near wells, faults and channels, the volume of flexible cells may be less than fine ones, so we need to solve a down-scaling problem; Far from these regions, the volume of cells are bigger than fine ones so we need to solve an up-scaling problem. In this work, research has been done on each of these three areas: down-scaling, up-scaling and fluid flow simulation. For each of these subjects, a review, some news improvements and comparative study are proposed. The proposed down-scaling method is build to be compatible with existing data integration methods. The comparative study shows that empirical methods are not enough accurate to solve the problem. Concerning the up-scaling step, the proposed approach is based on an existing method: the perturbed boundary conditions. An extension to unstructured mesh is developed for the inter-cell permeability tensor. The comparative study shows that numerical methods are not always as accurate as expected and the empirical model can be sufficient in lot of cases. A new approach to single-phase fluid flow simulation is developed. This approach can handle with full tensorial permeability fields with source or sink terms.(author)

  4. QMESH RENUM QPLOT, Mesh Generator on 2-D Bodies for Finite Element Method Analysis, with Plot Utility

    Jones, R.E.; Schkade, A.F.; Eyberger, L.R.

    1991-01-01

    1 - Description of problem or function: A set of five programs which make up a self-organising mesh generation package. QMESH generates meshes having quadrilateral elements on arbitrarily-shaped, two-dimensional (planar or axisymmetric) bodies. It is designed for use with two-dimensional finite element analysis applications. A flexible hierarchical input scheme is used to describe bodies to QMESH as collections of regions. A mesh for each region is developed independently, with the final assembly and bandwidth minimization performed by the independent program, RENUM or RENUM8. RENUM is applied when four-node elements are desired. Eight-node elements (with mid-side nodes) may be obtained with RENUM8., QPLOT and QPLOT8 are plot programs for meshes generated by the QMESH/RENUM and QMESH/RENUM8 program pairs, respectively. QPLOT and QPLOT8 automatically section the mesh into appropriately-sized sections for legible display of node and element numbers. An overall plot showing the position of the selected plot areas is produced. 2 - Method of solution: The mesh generating process for each individual region begins with the installation of an initial mesh which is a transformation of a regular grid on the unit square. The dimensions and orientation of the initial mesh may be defined by the user or, optionally, may be chosen by QMESH. Various smoothing algorithms may be applied to the initial mesh. Then, the mesh may be 'restructured' using an iterative scheme involving 'element pair restructuring', 'acute element deletion', and smoothing. In element pair restructuring, the interface side between two elements is removed and placed between two different nodes belonging to the pair of elements, provided that the change produces an overall improvement in the shapes of the two elements. In acute element deletion, an element having one diagonal much shorter than the other is deleted by collapsing the short diagonal to zero length The exact order in which restructuring, element

  5. Hybrid mesh generation for the new generation of oil reservoir simulators: 3D extension; Generation de maillage hybride pour les simulateurs de reservoir petrolier de nouvelle generation: extension 3D

    Flandrin, N.

    2005-09-15

    During the exploitation of an oil reservoir, it is important to predict the recovery of hydrocarbons and to optimize its production. A better comprehension of the physical phenomena requires to simulate 3D multiphase flows in increasingly complex geological structures. In this thesis, we are interested in this spatial discretization and we propose to extend in 3D the 2D hybrid model proposed by IFP in 1998 that allows to take directly into account in the geometry the radial characteristics of the flows. In these hybrid meshes, the wells and their drainage areas are described by structured radial circular meshes and the reservoirs are represented by structured meshes that can be a non uniform Cartesian grid or a Corner Point Geometry grids. In order to generate a global conforming mesh, unstructured transition meshes based on power diagrams and satisfying finite volume properties are used to connect the structured meshes together. Two methods have been implemented to generate these transition meshes: the first one is based on a Delaunay triangulation, the other one uses a frontal approach. Finally, some criteria are introduced to measure the quality of the transition meshes and optimization procedures are proposed to increase this quality under finite volume properties constraints. (author)

  6. Tetrahedral mesh generation of real terrain and topography effect on ERT image in Beishan region, Gansu province

    Lu Debao; Zhou Qiyou; Xiao Anlin; Song Zhen

    2014-01-01

    The paper starts from tetrahedral meshes generation of real terrain, a detailed way of tetradralization toward complicated terrain has been proposed based on comparing of advantage and disadvantage of several methods. DEM image has been used to help to generate tetrahedral mesh of research area. And then, forward soft Gmdata is used to calculate and analyze the topography effect on ERT Image with different kinds of terrain. Meanwhile, a quantitative way to define the topography effect was presented. Based on that, the method is used to eliminate the topography effect. The results show the method is effective and useful. (authors)

  7. An eFTD-VP framework for efficiently generating patient-specific anatomically detailed facial soft tissue FE mesh for craniomaxillofacial surgery simulation.

    Zhang, Xiaoyan; Kim, Daeseung; Shen, Shunyao; Yuan, Peng; Liu, Siting; Tang, Zhen; Zhang, Guangming; Zhou, Xiaobo; Gateno, Jaime; Liebschner, Michael A K; Xia, James J

    2018-04-01

    Accurate surgical planning and prediction of craniomaxillofacial surgery outcome requires simulation of soft tissue changes following osteotomy. This can only be achieved by using an anatomically detailed facial soft tissue model. The current state-of-the-art of model generation is not appropriate to clinical applications due to the time-intensive nature of manual segmentation and volumetric mesh generation. The conventional patient-specific finite element (FE) mesh generation methods are to deform a template FE mesh to match the shape of a patient based on registration. However, these methods commonly produce element distortion. Additionally, the mesh density for patients depends on that of the template model. It could not be adjusted to conduct mesh density sensitivity analysis. In this study, we propose a new framework of patient-specific facial soft tissue FE mesh generation. The goal of the developed method is to efficiently generate a high-quality patient-specific hexahedral FE mesh with adjustable mesh density while preserving the accuracy in anatomical structure correspondence. Our FE mesh is generated by eFace template deformation followed by volumetric parametrization. First, the patient-specific anatomically detailed facial soft tissue model (including skin, mucosa, and muscles) is generated by deforming an eFace template model. The adaptation of the eFace template model is achieved by using a hybrid landmark-based morphing and dense surface fitting approach followed by a thin-plate spline interpolation. Then, high-quality hexahedral mesh is constructed by using volumetric parameterization. The user can control the resolution of hexahedron mesh to best reflect clinicians' need. Our approach was validated using 30 patient models and 4 visible human datasets. The generated patient-specific FE mesh showed high surface matching accuracy, element quality, and internal structure matching accuracy. They can be directly and effectively used for clinical

  8. Fast precalculated triangular mesh algorithm for 3D binary computer-generated holograms.

    Yang, Fan; Kaczorowski, Andrzej; Wilkinson, Tim D

    2014-12-10

    A new method for constructing computer-generated holograms using a precalculated triangular mesh is presented. The speed of calculation can be increased dramatically by exploiting both the precalculated base triangle and GPU parallel computing. Unlike algorithms using point-based sources, this method can reconstruct a more vivid 3D object instead of a "hollow image." In addition, there is no need to do a fast Fourier transform for each 3D element every time. A ferroelectric liquid crystal spatial light modulator is used to display the binary hologram within our experiment and the hologram of a base right triangle is produced by utilizing just a one-step Fourier transform in the 2D case, which can be expanded to the 3D case by multiplying by a suitable Fresnel phase plane. All 3D holograms generated in this paper are based on Fresnel propagation; thus, the Fresnel plane is treated as a vital element in producing the hologram. A GeForce GTX 770 graphics card with 2 GB memory is used to achieve parallel computing.

  9. Methods for high-resolution anisotropic finite element modeling of the human head: automatic MR white matter anisotropy-adaptive mesh generation.

    Lee, Won Hee; Kim, Tae-Seong

    2012-01-01

    This study proposes an advanced finite element (FE) head modeling technique through which high-resolution FE meshes adaptive to the degree of tissue anisotropy can be generated. Our adaptive meshing scheme (called wMesh) uses MRI structural information and fractional anisotropy maps derived from diffusion tensors in the FE mesh generation process, optimally reflecting electrical properties of the human brain. We examined the characteristics of the wMeshes through various qualitative and quantitative comparisons to the conventional FE regular-sized meshes that are non-adaptive to the degree of white matter anisotropy. We investigated numerical differences in the FE forward solutions that include the electrical potential and current density generated by current sources in the brain. The quantitative difference was calculated by two statistical measures of relative difference measure (RDM) and magnification factor (MAG). The results show that the wMeshes are adaptive to the anisotropic density of the WM anisotropy, and they better reflect the density and directionality of tissue conductivity anisotropy. Our comparison results between various anisotropic regular mesh and wMesh models show that there are substantial differences in the EEG forward solutions in the brain (up to RDM=0.48 and MAG=0.63 in the electrical potential, and RDM=0.65 and MAG=0.52 in the current density). Our analysis results indicate that the wMeshes produce different forward solutions that are different from the conventional regular meshes. We present some results that the wMesh head modeling approach enhances the sensitivity and accuracy of the FE solutions at the interfaces or in the regions where the anisotropic conductivities change sharply or their directional changes are complex. The fully automatic wMesh generation technique should be useful for modeling an individual-specific and high-resolution anisotropic FE head model incorporating realistic anisotropic conductivity distributions

  10. Power generation using carbon mesh cathodes with different diffusion layers in microbial fuel cells

    Luo, Yong

    2011-11-01

    An inexpensive carbon material, carbon mesh, was examined to replace the more expensive carbon cloth usually used to make cathodes in air-cathode microbial fuel cells (MFCs). Three different diffusion layers were tested using carbon mesh: poly(dimethylsiloxane) (PDMS), polytetrafluoroethylene (PTFE), and Goretex cloth. Carbon mesh with a mixture of PDMS and carbon black as a diffusion layer produced a maximum power density of 1355 ± 62 mW m -2 (normalized to the projected cathode area), which was similar to that obtained with a carbon cloth cathode (1390 ± 72 mW m-2). Carbon mesh with a PTFE diffusion layer produced only a slightly lower (6.6%) maximum power density (1303 ± 48 mW m-2). The Coulombic efficiencies were a function of current density, with the highest value for the carbon mesh and PDMS (79%) larger than that for carbon cloth (63%). The cost of the carbon mesh cathode with PDMS/Carbon or PTFE (excluding catalyst and binder costs) is only 2.5% of the cost of the carbon cloth cathode. These results show that low cost carbon materials such as carbon mesh can be used as the cathode in an MFC without reducing the performance compared to more expensive carbon cloth. © 2011 Elsevier B.V.

  11. An Efficient Mesh Generation Method for Fractured Network System Based on Dynamic Grid Deformation

    Shuli Sun

    2013-01-01

    Full Text Available Meshing quality of the discrete model influences the accuracy, convergence, and efficiency of the solution for fractured network system in geological problem. However, modeling and meshing of such a fractured network system are usually tedious and difficult due to geometric complexity of the computational domain induced by existence and extension of fractures. The traditional meshing method to deal with fractures usually involves boundary recovery operation based on topological transformation, which relies on many complicated techniques and skills. This paper presents an alternative and efficient approach for meshing fractured network system. The method firstly presets points on fractures and then performs Delaunay triangulation to obtain preliminary mesh by point-by-point centroid insertion algorithm. Then the fractures are exactly recovered by local correction with revised dynamic grid deformation approach. Smoothing algorithm is finally applied to improve the quality of mesh. The proposed approach is efficient, easy to implement, and applicable to the cases of initial existing fractures and extension of fractures. The method is successfully applied to modeling of two- and three-dimensional discrete fractured network (DFN system in geological problems to demonstrate its effectiveness and high efficiency.

  12. An enhanced geometry-independent mesh weight window generator for MCNP

    Evans, T.M.; Hendricks, J.S.

    1997-01-01

    A new, enhanced, weight window generator suite has been developed for MCNP trademark. The new generator correctly estimates importances in either an user-specified, geometry-independent orthogonal grid or in MCNP geometric cells. The geometry-independent option alleviates the need to subdivide the MCNP cell geometry for variance reduction purposes. In addition, the new suite corrects several pathologies in the existing MCNP weight window generator. To verify the correctness of the new implementation, comparisons are performed with the analytical solution for the cell importance. Using the new generator, differences between Monte Carlo generated and analytical importances are less than 0.1%. Also, assumptions implicit in the original MCNP generator are shown to be poor in problems with high scattering media. The new generator is fully compatible with MCNP's AVATAR trademark automatic variance reduction method. The new generator applications, together with AVATAR, gives MCNP an enhanced suite of variance reduction methods. The flexibility and efficacy of this suite is demonstrated in a neutron porosity tool well-logging problem

  13. ImageParser: a tool for finite element generation from three-dimensional medical images

    Yamada T

    2004-10-01

    Full Text Available Abstract Background The finite element method (FEM is a powerful mathematical tool to simulate and visualize the mechanical deformation of tissues and organs during medical examinations or interventions. It is yet a challenge to build up an FEM mesh directly from a volumetric image partially because the regions (or structures of interest (ROIs may be irregular and fuzzy. Methods A software package, ImageParser, is developed to generate an FEM mesh from 3-D tomographic medical images. This software uses a semi-automatic method to detect ROIs from the context of image including neighboring tissues and organs, completes segmentation of different tissues, and meshes the organ into elements. Results The ImageParser is shown to build up an FEM model for simulating the mechanical responses of the breast based on 3-D CT images. The breast is compressed by two plate paddles under an overall displacement as large as 20% of the initial distance between the paddles. The strain and tangential Young's modulus distributions are specified for the biomechanical analysis of breast tissues. Conclusion The ImageParser can successfully exact the geometry of ROIs from a complex medical image and generate the FEM mesh with customer-defined segmentation information.

  14. Next generation tools for genomic data generation, distribution, and visualization

    Nix David A

    2010-09-01

    Full Text Available Abstract Background With the rapidly falling cost and availability of high throughput sequencing and microarray technologies, the bottleneck for effectively using genomic analysis in the laboratory and clinic is shifting to one of effectively managing, analyzing, and sharing genomic data. Results Here we present three open-source, platform independent, software tools for generating, analyzing, distributing, and visualizing genomic data. These include a next generation sequencing/microarray LIMS and analysis project center (GNomEx; an application for annotating and programmatically distributing genomic data using the community vetted DAS/2 data exchange protocol (GenoPub; and a standalone Java Swing application (GWrap that makes cutting edge command line analysis tools available to those who prefer graphical user interfaces. Both GNomEx and GenoPub use the rich client Flex/Flash web browser interface to interact with Java classes and a relational database on a remote server. Both employ a public-private user-group security model enabling controlled distribution of patient and unpublished data alongside public resources. As such, they function as genomic data repositories that can be accessed manually or programmatically through DAS/2-enabled client applications such as the Integrated Genome Browser. Conclusions These tools have gained wide use in our core facilities, research laboratories and clinics and are freely available for non-profit use. See http://sourceforge.net/projects/gnomex/, http://sourceforge.net/projects/genoviz/, and http://sourceforge.net/projects/useq.

  15. Next generation tools for genomic data generation, distribution, and visualization.

    Nix, David A; Di Sera, Tonya L; Dalley, Brian K; Milash, Brett A; Cundick, Robert M; Quinn, Kevin S; Courdy, Samir J

    2010-09-09

    With the rapidly falling cost and availability of high throughput sequencing and microarray technologies, the bottleneck for effectively using genomic analysis in the laboratory and clinic is shifting to one of effectively managing, analyzing, and sharing genomic data. Here we present three open-source, platform independent, software tools for generating, analyzing, distributing, and visualizing genomic data. These include a next generation sequencing/microarray LIMS and analysis project center (GNomEx); an application for annotating and programmatically distributing genomic data using the community vetted DAS/2 data exchange protocol (GenoPub); and a standalone Java Swing application (GWrap) that makes cutting edge command line analysis tools available to those who prefer graphical user interfaces. Both GNomEx and GenoPub use the rich client Flex/Flash web browser interface to interact with Java classes and a relational database on a remote server. Both employ a public-private user-group security model enabling controlled distribution of patient and unpublished data alongside public resources. As such, they function as genomic data repositories that can be accessed manually or programmatically through DAS/2-enabled client applications such as the Integrated Genome Browser. These tools have gained wide use in our core facilities, research laboratories and clinics and are freely available for non-profit use. See http://sourceforge.net/projects/gnomex/, http://sourceforge.net/projects/genoviz/, and http://sourceforge.net/projects/useq.

  16. Power generation using carbon mesh cathodes with different diffusion layers in microbial fuel cells

    Luo, Yong; Zhang, Fang; Wei, Bin; Liu, Guangli; Zhang, Renduo; Logan, Bruce E.

    2011-01-01

    to that obtained with a carbon cloth cathode (1390 ± 72 mW m-2). Carbon mesh with a PTFE diffusion layer produced only a slightly lower (6.6%) maximum power density (1303 ± 48 mW m-2). The Coulombic efficiencies were a function of current density, with the highest

  17. Three dimensional stress analysis of nozzle-to-shell intersections by the finite element method and a auto-mesh generation program

    Fujihara, Hirohiko; Ueda, Masahiro

    1975-01-01

    In the design of chemical reactors or nuclear pressure vessels it is often important to evaluate the stress distribution in nozzle-to-shell intersections. The finite element method is a powerful tool for stress analysis, but it has a defects to require troublesome work in preparing input data. Specially, the mesh data of oblique nozzles and tangential nozzles, in which stress concentration is very high, are very difficult to be prepared. The authors made a mesh generation program which can be used to any nozzle-to-shell intersections, and combining this program with a three dimensional stress analysis program by the finite element method they made the stress analysis of nozzle-to-shell intersections under internal pressure. Consequently, stresses, strains and deformations of nozzles nonsymmetrical to spherical shells and nozzles tangential to cylindrical shells were made clear and it was shown that the curvature of the inner surface of the nozzle corner was a controlling factor in reducing stress concentration. (auth.)

  18. BOT3P5.2, 3D Mesh Generator and Graphical Display of Geometry for Radiation Transport Codes, Display of Results

    Orsi, Roberto; Bidaud, Adrien

    2007-01-01

    1 - Description of program or function: BOT3P was originally conceived as a set of standard FORTRAN 77 language programs in order to give the users of the DORT and TORT deterministic transport codes some useful diagnostic tools to prepare and check their input data files. Later versions extended the possibility to produce the geometrical, material distribution and fixed neutron source data to other deterministic transport codes such as TWODANT/THREEDANT of the DANTSYS system, PARTISN and, potentially, to any transport code through BOT3P binary output files that can be easily interfaced (see, for example, the Russian two-dimensional (2D) and three-dimensional (3D) discrete ordinates neutron, photon and charged particle transport codes KASKAD-S-2.5 and KATRIN-2.0). As from Version 5.1 BOT3P contained important additions specifically addressed to radiation transport analysis for medical applications. BOT3P-5.2 contains new graphics capabilities. Some of them enable users to select space sub-domains of the total mesh grid in order to improve the zoom simulation of the geometry, both in 2D cuts and in 3D. Moreover the new BOT3P module (PDTM) may improve the interface of BOT3P geometrical models to transport analysis codes. The following programs are included in the BOT3P software package: GGDM, DDM, GGTM, DTM2, DTM3, RVARSCL, COMPARE, MKSRC, CATSM, DTET, and PDTM. The main features of these different programs are described. 2 - Methods: GGDM and GGTM work similarly from the logical point of view. Since the 3D case is more general, the following description refers to GGTM. All the co-ordinate values that characterise the geometrical scheme at the basis of the 3D transport code geometrical and material model are read, sorted and all stored if different from the neighbouring ones more than an input tolerance established by the user. These co-ordinates are always present in the fine-mesh boundary arrays independently of the mesh grid refinement options, because they

  19. Next-Generation Tools For Next-Generation Surveys

    Murray, S. G.

    2017-04-01

    The next generation of large-scale galaxy surveys, across the electromagnetic spectrum, loom on the horizon as explosively game-changing datasets, in terms of our understanding of cosmology and structure formation. We are on the brink of a torrent of data that is set to both confirm and constrain current theories to an unprecedented level, and potentially overturn many of our conceptions. One of the great challenges of this forthcoming deluge is to extract maximal scientific content from the vast array of raw data. This challenge requires not only well-understood and robust physical models, but a commensurate network of software implementations with which to efficiently apply them. The halo model, a semi-analytic treatment of cosmological spatial statistics down to nonlinear scales, provides an excellent mathematical framework for exploring the nature of dark matter. This thesis presents a next-generation toolkit based on the halo model formalism, intended to fulfil the requirements of next-generation surveys. Our toolkit comprises three tools: (i) hmf, a comprehensive and flexible calculator for halo mass functions (HMFs) within extended Press-Schechter theory, (ii) the MRP distribution for extremely efficient analytic characterisation of HMFs, and (iii) halomod, an extension of hmf which provides support for the full range of halo model components. In addition to the development and technical presentation of these tools, we apply each to the task of physical modelling. With hmf, we determine the precision of our knowledge of the HMF, due to uncertainty in our knowledge of the cosmological parameters, over the past decade of cosmic microwave background (CMB) experiments. We place rule-of-thumb uncertainties on the predicted HMF for the Planck cosmology, and find that current limits on the precision are driven by modeling uncertainties rather than those from cosmological parameters. With the MRP, we create and test a method for robustly fitting the HMF to observed

  20. Power generation using an activated carbon and metal mesh cathode in a microbial fuel cell

    Zhang, Fang

    2009-11-01

    An inexpensive activated carbon (AC) air cathode was developed as an alternative to a platinum-catalyzed electrode for oxygen reduction in a microbial fuel cell (MFC). AC was cold-pressed with a polytetrafluoroethylene (PTFE) binder to form the cathode around a Ni mesh current collector. This cathode construction avoided the need for carbon cloth or a metal catalyst, and produced a cathode with high activity for oxygen reduction at typical MFC current densities. Tests with the AC cathode produced a maximum power density of 1220 mW/m2 (normalized to cathode projected surface area; 36 W/m3 based on liquid volume) compared to 1060 mW/m2 obtained by Pt catalyzed carbon cloth cathode. The Coulombic efficiency ranged from 15% to 55%. These findings show that AC is a cost-effective material for achieving useful rates of oxygen reduction in air cathode MFCs. © 2009 Elsevier B.V. All rights reserved.

  1. Two-phase flow pattern measurements with a wire mesh sensor in a direct steam generating solar thermal collector

    Berger, Michael; Mokhtar, Marwan; Zahler, Christian; Willert, Daniel; Neuhäuser, Anton; Schleicher, Eckhard

    2017-06-01

    At Industrial Solar's test facility in Freiburg (Germany), two phase flow patterns have been measured by using a wire mesh sensor from Helmholtz Zentrum Dresden-Rossendorf (HZDR). Main purpose of the measurements was to compare observed two-phase flow patterns with expected flow patterns from models. The two-phase flow pattern is important for the design of direct steam generating solar collectors. Vibrations should be avoided in the peripheral piping, and local dry-outs or large circumferential temperature gradients should be prevented in the absorber tubes. Therefore, the choice of design for operation conditions like mass flow and steam quality are an important step in the engineering process of such a project. Results of a measurement with the wire mesh sensor are the flow pattern and the plug or slug frequency at the given operating conditions. Under the assumption of the collector power, which can be assumed from previous measurements at the same collector and adaption with sun position and incidence angle modifier, also the slip can be evaluated for a wire mesh sensor measurement. Measurements have been performed at different mass flows and pressure levels. Transient behavior has been tested for flashing, change of mass flow, and sudden changes of irradiation (cloud simulation). This paper describes the measurements and the method of evaluation. Results are shown as extruded profiles in top view and in side view. Measurement and model are compared. The tests have been performed at low steam quality, because of the limits of the test facility. Conclusions and implications for possible future measurements at larger collectors are also presented in this paper.

  2. Using bio.tools to generate and annotate workbench tool descriptions

    Hillion, Kenzo-Hugo; Kuzmin, Ivan; Khodak, Anton

    2017-01-01

    - which have been registered in the ELIXIR tools registry (https://bio.tools) - into workbench environments by generating tool description templates. ToolDog includes two modules. The first module analyses the source code of the bioinformatics software with language-specific plugins, and generates...

  3. Wire-mesh capped deposition sensors: Novel passive tool for coarse fraction flux estimation of radon thoron progeny in indoor environments

    Mayya, Y.S.; Mishra, Rosaline; Prajith, Rama; Sapra, B.K.; Kushwaha, H.S.

    2010-01-01

    Deposition-based 222 Rn and 220 Rn progeny sensors act as unique, passive tools for determining the long time-averaged progeny deposition fluxes in the environment. The use of these deposition sensors as progeny concentration monitors was demonstrated in typical indoor environments as conceptually superior alternatives to gas-based indirect monitoring methods. In the present work, the dependency of these deposition monitors on various environmental parameters is minimized by capping the deposition sensor with a suitable wire mesh. These wire-mesh capped deposition sensors measure the coarse fraction deposition flux, which is less dependent on the change in environmental parameters like ventilation rate and turbulence. The calibration of these wire-mesh capped coarse fraction progeny sensors was carried out by laboratory controlled experiments. These sensors were deployed both in indoor and in occupational environments having widely different ventilation rates. The obtained coarse fraction deposition velocities were fairly constant in these environments, which further confirmed that the signal on the wire-mesh capped sensors show the least dependency on the change in environmental parameters. This technique has the potential to serve as a passive particle sizer in the general context of nanoparticles using progeny species as surrogates. On the whole, there exists a strong case for developing a passive system that responds only to coarse fraction for providing alternative tools for dosimetry and environmental fine particle research. - Research highlights: → Wire-mesh capped deposition sensor measures the coarse fraction deposition flux → Coarse fraction deposition flux less dependent on environmental conditions → Wire-mesh capped deposition sensor as passive particle sizer

  4. Towards a personalised learning mesh: the implementation of a low overhead, multipath learning tool

    Kevin Porteous

    2011-02-01

    Full Text Available Many studies have shown that students today live in an environment of multiple, simultaneous, short-lived stimuli which they access from wherever they may be. However institutional teaching is still based on traditional, long, sequential, attended presentations. In order to bridge that gap, there have been a number of moves over the past few years to develop and integrate lecture capture into the learning environment. Often these systems are large and require a major commitment from the institution in terms of licences and infrastructure. Given the constrained financial environment for many academic institutions, these systems are not a viable option for many. The authors have extended their normal lecture capture activity in their teaching to form an integrated learning resource. The captured media is mounted into a content management system which allows the media to be repurposed along with other content to provide an integrated support tool for student enquiry and self study which better matches their unstructured social experience. This paper describes the development of the pilot system based on a minimal hardware requirement and limited post processing. The evolution of the system pilot is described and the development of the specification which then led to the live prototype is discussed. Issues that impact on the effectiveness of the prototype are covered and the strategy (based on classroom feedback for developing the prototype into a full system for deployment across a range of desktop and mobile platforms is introduced.

  5. Test Generators: Teacher's Tool or Teacher's Headache?

    Eiser, Leslie

    1988-01-01

    Discusses the advantages and disadvantages of test generation programs. Includes setting up, printing exams and "bells and whistles." Reviews eight computer packages for Apple and IBM personal computers. Compares features, costs, and usage. (CW)

  6. Induction generator models in dynamic simulation tools

    Knudsen, Hans; Akhmatov, Vladislav

    1999-01-01

    For AC network with large amount of induction generators (windmills) the paper demonstrates a significant discrepancy in the simulated voltage recovery after fault in weak networks when comparing dynamic and transient stability descriptions and the reasons of discrepancies are explained...

  7. Using bio.tools to generate and annotate workbench tool descriptions [version 1; referees: 2 approved

    Kenzo-Hugo Hillion

    2017-11-01

    Full Text Available Workbench and workflow systems such as Galaxy, Taverna, Chipster, or Common Workflow Language (CWL-based frameworks, facilitate the access to bioinformatics tools in a user-friendly, scalable and reproducible way. Still, the integration of tools in such environments remains a cumbersome, time consuming and error-prone process. A major consequence is the incomplete or outdated description of tools that are often missing important information, including parameters and metadata such as publication or links to documentation. ToolDog (Tool DescriptiOn Generator facilitates the integration of tools - which have been registered in the ELIXIR tools registry (https://bio.tools - into workbench environments by generating tool description templates. ToolDog includes two modules. The first module analyses the source code of the bioinformatics software with language-specific plugins, and generates a skeleton for a Galaxy XML or CWL tool description. The second module is dedicated to the enrichment of the generated tool description, using metadata provided by bio.tools. This last module can also be used on its own to complete or correct existing tool descriptions with missing metadata.

  8. Efficient Simplification Methods for Generating High Quality LODs of 3D Meshes

    Muhammad Hussain

    2009-01-01

    Two simplification algorithms are proposed for automatic decimation of polygonal models, and for generating their LODs. Each algorithm orders vertices according to their priority values and then removes them iteratively. For setting the priority value of each vertex, exploiting normal field of its one-ring neighborhood, we introduce a new measure of geometric fidelity that reflects well the local geometric features of the vertex. After a vertex is selected, using other measures of geometric distortion that are based on normal field deviation and distance measure, it is decided which of the edges incident on the vertex is to be collapsed for removing it. The collapsed edge is substituted with a new vertex whose position is found by minimizing the local quadric error measure. A comparison with the state-of-the-art algorithms reveals that the proposed algorithms are simple to implement, are computationally more efficient, generate LODs with better quality, and preserve salient features even after drastic simplification. The methods are useful for applications such as 3D computer games, virtual reality, where focus is on fast running time, reduced memory overhead, and high quality LODs.

  9. Design and installation of a strategically placed algae mesh barrier at OPG Pickering Nuclear Generating Station

    Marttila, D.; Patrick, P.; Gregoris, C.

    2009-01-01

    Ontario Power Generation's Pickering Nuclear has experienced a number of events in which attached algae have become entrained in the water intake costing approximately $30M over the 1995-2005 period as a result of deratings, Unit shutdowns and other operational issues. In 2005-2006 OPG and Kinectrics worked collaboratively on evaluating different potential solutions to reduce the impact of algae on the station. One of the solutions developed by Kinectrics included a strategically placed barrier net designed to regulate algae flow into the station intake. In 2006, Kinectrics designed and installed the system, the first of its kind at a Nuclear Power Plant in Canada. The system was operational by May 2007. OPG completed an effectiveness study in 2007 and concluded the barrier system had a beneficial effect on reducing algae impact on the station. (author)

  10. Design Tool for 5-20 MW Direct Drive Generators

    Leban, Krisztina Monika; Ritchie, Ewen; Argeseanu, Alin

    2014-01-01

    This paper reports on a machine design tool for large (5-10MW) direct drive electrical generator. The aim of the work is to construct a flexible calculation tool that enables the analysis of different ideas and concepts for generator design. The tool is intended for engineers that are involved...... in the design of wind turbine systems. The design tool comprises calculation modules that are kept as independent as possible from each other so that new machine geometries and types can be modelled by reusing, recombining and modifying the different modules. Choice of the most suitable candidates...

  11. BGen: A UML Behavior Network Generator Tool

    Huntsberger, Terry; Reder, Leonard J.; Balian, Harry

    2010-01-01

    BGen software was designed for autogeneration of code based on a graphical representation of a behavior network used for controlling automatic vehicles. A common format used for describing a behavior network, such as that used in the JPL-developed behavior-based control system, CARACaS ["Control Architecture for Robotic Agent Command and Sensing" (NPO-43635), NASA Tech Briefs, Vol. 32, No. 10 (October 2008), page 40] includes a graph with sensory inputs flowing through the behaviors in order to generate the signals for the actuators that drive and steer the vehicle. A computer program to translate Unified Modeling Language (UML) Freeform Implementation Diagrams into a legacy C implementation of Behavior Network has been developed in order to simplify the development of C-code for behavior-based control systems. UML is a popular standard developed by the Object Management Group (OMG) to model software architectures graphically. The C implementation of a Behavior Network is functioning as a decision tree.

  12. HeldenSurface: A CAD Tool to Generate High-Quality Surfaces, Phase I

    National Aeronautics and Space Administration — One of the primary shortcomings identified during the NASA sponsored CFD Vision 2030 Study conducted during 2012-2014 was that the generation of meshes suitable for...

  13. Generation of an artificial skin construct containing a non-degradable fiber mesh: a potential transcutaneous interface

    Cahn, Frederick [Biomedical Strategies Inc., San Diego, CA (United States); Kyriakides, Themis R [Vascular Biology and Therapeutics, Yale University, New Haven, CT 06536-9812 (United States)], E-mail: themis.kyriakides@yale.edu

    2008-09-01

    Generation of a stable interface between soft tissues and biomaterials could improve the function of transcutaneous prostheses, primarily by minimizing chronic infections. We hypothesized that inclusion of non-biodegradable biomaterials in an artificial skin substrate would improve integration of the neodermis. In the present study, we compared the biocompatibility of an experimental substrate, consisting of collagen and glycosylaminoglycans, with commercially available artificial skin of similar composition. By utilizing a mouse excisional wound model, we found that the source of collagen (bovine tendon versus hide), extent of injury and wound contraction were critical determinants of inflammation and neodermis formation. Reducing the extent of injury to underlying muscle reduced inflammation and improved remodeling; the improved conditions allowed the detection of a pro-inflammatory effect of hide-derived collagen. To eliminate the complication of wound contraction, subsequent grafts were performed in guinea pigs and showed that inclusion of carbon fibers or non-degradable sutures resulted in increased foreign body response (FBR) and altered remodeling. On the other hand, inclusion of a polyester multi-stranded mesh induced a mild FBR and allowed normal neodermis formation. Taken together, our observations suggest that non-degradable biomaterials can be embedded in an artificial skin construct without compromising its ability to induce neodermis formation.

  14. New Geometry of Worm Face Gear Drives with Conical and Cylindrical Worms: Generation, Simulation of Meshing, and Stress Analysis

    Litvin, Faydor L.; Nava, Alessandro; Fan, Qi; Fuentes, Alfonso

    2002-01-01

    New geometry of face worm gear drives with conical and cylindrical worms is proposed. The generation of the face worm-gear is based on application of a tilted head-cutter (grinding tool) instead of application of a hob applied at present. The generation of a conjugated worm is based on application of a tilted head-cutter (grinding tool) as well. The bearing contact of the gear drive is localized and is oriented longitudinally. A predesigned parabolic function of transmission errors for reduction of noise and vibration is provided. The stress analysis of the gear drive is performed using a three-dimensional finite element analysis. The contacting model is automatically generated. The developed theory is illustrated with numerical examples.

  15. Toward An Unstructured Mesh Database

    Rezaei Mahdiraji, Alireza; Baumann, Peter Peter

    2014-05-01

    Unstructured meshes are used in several application domains such as earth sciences (e.g., seismology), medicine, oceanography, cli- mate modeling, GIS as approximate representations of physical objects. Meshes subdivide a domain into smaller geometric elements (called cells) which are glued together by incidence relationships. The subdivision of a domain allows computational manipulation of complicated physical structures. For instance, seismologists model earthquakes using elastic wave propagation solvers on hexahedral meshes. The hexahedral con- tains several hundred millions of grid points and millions of hexahedral cells. Each vertex node in the hexahedrals stores a multitude of data fields. To run simulation on such meshes, one needs to iterate over all the cells, iterate over incident cells to a given cell, retrieve coordinates of cells, assign data values to cells, etc. Although meshes are used in many application domains, to the best of our knowledge there is no database vendor that support unstructured mesh features. Currently, the main tool for querying and manipulating unstructured meshes are mesh libraries, e.g., CGAL and GRAL. Mesh li- braries are dedicated libraries which includes mesh algorithms and can be run on mesh representations. The libraries do not scale with dataset size, do not have declarative query language, and need deep C++ knowledge for query implementations. Furthermore, due to high coupling between the implementations and input file structure, the implementations are less reusable and costly to maintain. A dedicated mesh database offers the following advantages: 1) declarative querying, 2) ease of maintenance, 3) hiding mesh storage structure from applications, and 4) transparent query optimization. To design a mesh database, the first challenge is to define a suitable generic data model for unstructured meshes. We proposed ImG-Complexes data model as a generic topological mesh data model which extends incidence graph model to multi

  16. SUPERIMPOSED MESH PLOTTING IN MCNP

    J. HENDRICKS

    2001-02-01

    The capability to plot superimposed meshes has been added to MCNP{trademark}. MCNP4C featured a superimposed mesh weight window generator which enabled users to set up geometries without having to subdivide geometric cells for variance reduction. The variance reduction was performed with weight windows on a rectangular or cylindrical mesh superimposed over the physical geometry. Experience with the new capability was favorable but also indicated that a number of enhancements would be very beneficial, particularly a means of visualizing the mesh and its values. The mathematics for plotting the mesh and its values is described here along with a description of other upgrades.

  17. The Exercise: An Exercise Generator Tool for the SOURCe Project

    Kakoyianni-Doa, Fryni; Tziafa, Eleni; Naskos, Athanasios

    2016-01-01

    The Exercise, an Exercise generator in the SOURCe project, is a tool that complements the properties and functionalities of the SOURCe project, which includes the search engine for the Searchable Online French-Greek parallel corpus for the UniveRsity of Cyprus (SOURCe) (Kakoyianni-Doa & Tziafa, 2013), the PENCIL (an alignment tool)…

  18. Automatic generation of 3D fine mesh geometries for the analysis of the venus-3 shielding benchmark experiment with the Tort code

    Pescarini, M.; Orsi, R.; Martinelli, T.

    2003-01-01

    In many practical radiation transport applications today the cost for solving refined, large size and complex multi-dimensional problems is not so much computing but is linked to the cumbersome effort required by an expert to prepare a detailed geometrical model, verify and validate that it is correct and represents, to a specified tolerance, the real design or facility. This situation is, in particular, relevant and frequent in reactor core criticality and shielding calculations, with three-dimensional (3D) general purpose radiation transport codes, requiring a very large number of meshes and high performance computers. The need for developing tools that make easier the task to the physicist or engineer, by reducing the time required, by facilitating through effective graphical display the verification of correctness and, finally, that help the interpretation of the results obtained, has clearly emerged. The paper shows the results of efforts in this field through detailed simulations of a complex shielding benchmark experiment. In the context of the activities proposed by the OECD/NEA Nuclear Science Committee (NSC) Task Force on Computing Radiation Dose and Modelling of Radiation-Induced Degradation of Reactor Components (TFRDD), the ENEA-Bologna Nuclear Data Centre contributed with an analysis of the VENUS-3 low-flux neutron shielding benchmark experiment (SCK/CEN-Mol, Belgium). One of the targets of the work was to test the BOT3P system, originally developed at the Nuclear Data Centre in ENEA-Bologna and actually released to OECD/NEA Data Bank for free distribution. BOT3P, ancillary system of the DORT (2D) and TORT (3D) SN codes, permits a flexible automatic generation of spatial mesh grids in Cartesian or cylindrical geometry, through combinatorial geometry algorithms, following a simplified user-friendly approach. This system demonstrated its validity also in core criticality analyses, as for example the Lewis MOX fuel benchmark, permitting to easily

  19. Efficient Data Generation and Publication as a Test Tool

    Einstein, Craig Jakob

    2017-01-01

    A tool to facilitate the generation and publication of test data was created to test the individual components of a command and control system designed to launch spacecraft. Specifically, this tool was built to ensure messages are properly passed between system components. The tool can also be used to test whether the appropriate groups have access (read/write privileges) to the correct messages. The messages passed between system components take the form of unique identifiers with associated values. These identifiers are alphanumeric strings that identify the type of message and the additional parameters that are contained within the message. The values that are passed with the message depend on the identifier. The data generation tool allows for the efficient creation and publication of these messages. A configuration file can be used to set the parameters of the tool and also specify which messages to pass.

  20. Design Tool for Direct Drive Wind Turbine Generators

    Leban, Krisztina Monika

    . A comparison of the selected machine types in view of up-scaling to 20 [MW] was performed. As an example fitness criterion, the use of active materials for the generators was considered. Based on this, suggestions for 20 [MW] generators were made. The results are discussed and future work, directions......The current work offers a comparison of the proposed machine geometries for 6 [MW] direct drive wind generator candidates with the prospective of up scaling to 20MW. The suggestions are based on a design tool especially built for this investigation. The in-built flexibility of the design tool gives...

  1. Notes on the Mesh Handler and Mesh Data Conversion

    Lee, Sang Yong; Park, Chan Eok

    2009-01-01

    At the outset of the development of the thermal-hydraulic code (THC), efforts have been made to utilize the recent technology of the computational fluid dynamics. Among many of them, the unstructured mesh approach was adopted to alleviate the restriction of the grid handling system. As a natural consequence, a mesh handler (MH) has been developed to manipulate the complex mesh data from the mesh generator. The mesh generator, Gambit, was chosen at the beginning of the development of the code. But a new mesh generator, Pointwise, was introduced to get more flexible mesh generation capability. An open source code, Paraview, was chosen as a post processor, which can handle unstructured as well as structured mesh data. Overall data processing system for THC is shown in Figure-1. There are various file formats to save the mesh data in the permanent storage media. A couple of dozen of file formats are found even in the above mentioned programs. A competent mesh handler should have the capability to import or export mesh data as many as possible formats. But, in reality, there are two aspects that make it difficult to achieve the competence. The first aspect to consider is the time and efforts to program the interface code. And the second aspect, which is even more difficult one, is the fact that many mesh data file formats are proprietary information. In this paper, some experience of the development of the format conversion programs will be presented. File formats involved are Gambit neutral format, Ansys-CFX grid file format, VTK legacy file format, Nastran format and CGNS

  2. MEAT: An Authoring Tool for Generating Adaptable Learning Resources

    Kuo, Yen-Hung; Huang, Yueh-Min

    2009-01-01

    Mobile learning (m-learning) is a new trend in the e-learning field. The learning services in m-learning environments are supported by fundamental functions, especially the content and assessment services, which need an authoring tool to rapidly generate adaptable learning resources. To fulfill the imperious demand, this study proposes an…

  3. An innovative simulation tool for waste to energy generation opportunities

    Bilal Abderezzak

    2017-03-01

    Full Text Available The new world energy policies encourage the use of renewable energy sources with clean technologies, and abandon progressively the fossil fuel dependence. Another energy generation trend called commonly the “Waste-to-Energy” solution, uses organic waste as a response for two major problems: energy generation and waste management. Thanks to the anaerobic digestion, the organic waste can provide a biogas composed essentially from Carbone dioxide (CO2 and Methane (CH4. This work aims essentially to help students, researchers and even decision makers to consider the importance of biogas generation. The proposed tool is the last version of our previous tool which is enhanced and completed. It presents the potential to produce biogas of any shortlisted kind of waste, including also some energy valorization ways. A technical economical data are introduced for eventual feasibility studies.

  4. Computed tomography landmark-based semi-automated mesh morphing and mapping techniques: generation of patient specific models of the human pelvis without segmentation.

    Salo, Zoryana; Beek, Maarten; Wright, David; Whyne, Cari Marisa

    2015-04-13

    Current methods for the development of pelvic finite element (FE) models generally are based upon specimen specific computed tomography (CT) data. This approach has traditionally required segmentation of CT data sets, which is time consuming and necessitates high levels of user intervention due to the complex pelvic anatomy. The purpose of this research was to develop and assess CT landmark-based semi-automated mesh morphing and mapping techniques to aid the generation and mechanical analysis of specimen-specific FE models of the pelvis without the need for segmentation. A specimen-specific pelvic FE model (source) was created using traditional segmentation methods and morphed onto a CT scan of a different (target) pelvis using a landmark-based method. The morphed model was then refined through mesh mapping by moving the nodes to the bone boundary. A second target model was created using traditional segmentation techniques. CT intensity based material properties were assigned to the morphed/mapped model and to the traditionally segmented target models. Models were analyzed to evaluate their geometric concurrency and strain patterns. Strains generated in a double-leg stance configuration were compared to experimental strain gauge data generated from the same target cadaver pelvis. CT landmark-based morphing and mapping techniques were efficiently applied to create a geometrically multifaceted specimen-specific pelvic FE model, which was similar to the traditionally segmented target model and better replicated the experimental strain results (R(2)=0.873). This study has shown that mesh morphing and mapping represents an efficient validated approach for pelvic FE model generation without the need for segmentation. Copyright © 2015 Elsevier Ltd. All rights reserved.

  5. Wire-mesh sensors: an experimental tool for two-phase CDF model development and code validation

    Prasser, H.M. [Forschungszentrum Rossendorf e.V., Dresden (Germany)

    2004-07-01

    Full text of publication follows:The Institute of Safety Research of the Forschungszentrum Rossendorf, Germany, has developed electrode-mesh sensors, which allow the measurement of the electrical conductivity distribution in a flow duct. This can be used either for the detection of the gaseous phase in a gas-liquid flow or for mixing studies in single phase flow, when the components have different electric conductivities. Two grids of crossing wires are placed into the flow closely behind each other. The wires of the first plane (transmitter plane) are supplied with pulses of a driving voltage in a successive order. The data acquisition is done by measuring the electrical currents arriving at the second grid (receiver wires). After the last transmitter electrode has been activated, a two-dimensional matrix is available that reflects the conductivities at crossing points of the electrodes of the two grids. Sequences of these 2D distributions are recorded with a rate of up to 10 kHz. Due to the high measuring rate each bubble is mapped in several successive instantaneous frames. This allows to obtain bubble size distributions as well as bubble-size resolved gas fraction profiles beside the visualisation and the calculation of profiles of the time-averaged void fraction. Two sensors placed behind each other can furthermore be used for bubble velocity measurements using cross-correlation techniques. Sensors with three layers of electrode grids can be used for the measurement of the velocity of individual bubbles. The sensor is widely used to study the evolution of the flow pattern in an upwards air-water flow. The experiments aim at closure equations describing forces acting on bubbles as well as coalescence and fragmentation frequencies for the implementation in CFD-codes. The largest sensor used until now has a circular measuring cross-section of about 200 mm diameter and is equipped with two grids of 64 wires. Therefore, the spatial resolution is 3 mm, the measuring

  6. Wire-mesh sensors: an experimental tool for two-phase CDF model development and code validation

    Prasser, H.M.

    2004-01-01

    Full text of publication follows:The Institute of Safety Research of the Forschungszentrum Rossendorf, Germany, has developed electrode-mesh sensors, which allow the measurement of the electrical conductivity distribution in a flow duct. This can be used either for the detection of the gaseous phase in a gas-liquid flow or for mixing studies in single phase flow, when the components have different electric conductivities. Two grids of crossing wires are placed into the flow closely behind each other. The wires of the first plane (transmitter plane) are supplied with pulses of a driving voltage in a successive order. The data acquisition is done by measuring the electrical currents arriving at the second grid (receiver wires). After the last transmitter electrode has been activated, a two-dimensional matrix is available that reflects the conductivities at crossing points of the electrodes of the two grids. Sequences of these 2D distributions are recorded with a rate of up to 10 kHz. Due to the high measuring rate each bubble is mapped in several successive instantaneous frames. This allows to obtain bubble size distributions as well as bubble-size resolved gas fraction profiles beside the visualisation and the calculation of profiles of the time-averaged void fraction. Two sensors placed behind each other can furthermore be used for bubble velocity measurements using cross-correlation techniques. Sensors with three layers of electrode grids can be used for the measurement of the velocity of individual bubbles. The sensor is widely used to study the evolution of the flow pattern in an upwards air-water flow. The experiments aim at closure equations describing forces acting on bubbles as well as coalescence and fragmentation frequencies for the implementation in CFD-codes. The largest sensor used until now has a circular measuring cross-section of about 200 mm diameter and is equipped with two grids of 64 wires. Therefore, the spatial resolution is 3 mm, the measuring

  7. An Infrastructure for UML-Based Code Generation Tools

    Wehrmeister, Marco A.; Freitas, Edison P.; Pereira, Carlos E.

    The use of Model-Driven Engineering (MDE) techniques in the domain of distributed embedded real-time systems are gain importance in order to cope with the increasing design complexity of such systems. This paper discusses an infrastructure created to build GenERTiCA, a flexible tool that supports a MDE approach, which uses aspect-oriented concepts to handle non-functional requirements from embedded and real-time systems domain. GenERTiCA generates source code from UML models, and also performs weaving of aspects, which have been specified within the UML model. Additionally, this paper discusses the Distributed Embedded Real-Time Compact Specification (DERCS), a PIM created to support UML-based code generation tools. Some heuristics to transform UML models into DERCS, which have been implemented in GenERTiCA, are also discussed.

  8. Surface meshing with curvature convergence

    Li, Huibin; Zeng, Wei; Morvan, Jean-Marie; Chen, Liming; Gu, Xianfengdavid

    2014-01-01

    Surface meshing plays a fundamental role in graphics and visualization. Many geometric processing tasks involve solving geometric PDEs on meshes. The numerical stability, convergence rates and approximation errors are largely determined by the mesh qualities. In practice, Delaunay refinement algorithms offer satisfactory solutions to high quality mesh generations. The theoretical proofs for volume based and surface based Delaunay refinement algorithms have been established, but those for conformal parameterization based ones remain wide open. This work focuses on the curvature measure convergence for the conformal parameterization based Delaunay refinement algorithms. Given a metric surface, the proposed approach triangulates its conformal uniformization domain by the planar Delaunay refinement algorithms, and produces a high quality mesh. We give explicit estimates for the Hausdorff distance, the normal deviation, and the differences in curvature measures between the surface and the mesh. In contrast to the conventional results based on volumetric Delaunay refinement, our stronger estimates are independent of the mesh structure and directly guarantee the convergence of curvature measures. Meanwhile, our result on Gaussian curvature measure is intrinsic to the Riemannian metric and independent of the embedding. In practice, our meshing algorithm is much easier to implement and much more efficient. The experimental results verified our theoretical results and demonstrated the efficiency of the meshing algorithm. © 2014 IEEE.

  9. Surface meshing with curvature convergence

    Li, Huibin

    2014-06-01

    Surface meshing plays a fundamental role in graphics and visualization. Many geometric processing tasks involve solving geometric PDEs on meshes. The numerical stability, convergence rates and approximation errors are largely determined by the mesh qualities. In practice, Delaunay refinement algorithms offer satisfactory solutions to high quality mesh generations. The theoretical proofs for volume based and surface based Delaunay refinement algorithms have been established, but those for conformal parameterization based ones remain wide open. This work focuses on the curvature measure convergence for the conformal parameterization based Delaunay refinement algorithms. Given a metric surface, the proposed approach triangulates its conformal uniformization domain by the planar Delaunay refinement algorithms, and produces a high quality mesh. We give explicit estimates for the Hausdorff distance, the normal deviation, and the differences in curvature measures between the surface and the mesh. In contrast to the conventional results based on volumetric Delaunay refinement, our stronger estimates are independent of the mesh structure and directly guarantee the convergence of curvature measures. Meanwhile, our result on Gaussian curvature measure is intrinsic to the Riemannian metric and independent of the embedding. In practice, our meshing algorithm is much easier to implement and much more efficient. The experimental results verified our theoretical results and demonstrated the efficiency of the meshing algorithm. © 2014 IEEE.

  10. Urogynecologic Surgical Mesh Implants

    ... procedures performed to treat pelvic floor disorders with surgical mesh: Transvaginal mesh to treat POP Transabdominal mesh to treat ... address safety risks Final Order for Reclassification of Surgical Mesh for Transvaginal Pelvic Organ Prolapse Repair Final Order for Effective ...

  11. Ichthyoplankton Classification Tool using Generative Adversarial Networks and Transfer Learning

    Aljaafari, Nura

    2018-04-15

    The study and the analysis of marine ecosystems is a significant part of the marine science research. These systems are valuable resources for fisheries, improving water quality and can even be used in drugs production. The investigation of ichthyoplankton inhabiting these ecosystems is also an important research field. Ichthyoplankton are fish in their early stages of life. In this stage, the fish have relatively similar shape and are small in size. The currently used way of identifying them is not optimal. Marine scientists typically study such organisms by sending a team that collects samples from the sea which is then taken to the lab for further investigation. These samples need to be studied by an expert and usually end needing a DNA sequencing. This method is time-consuming and requires a high level of experience. The recent advances in AI have helped to solve and automate several difficult tasks which motivated us to develop a classification tool for ichthyoplankton. We show that using machine learning techniques, such as generative adversarial networks combined with transfer learning solves such a problem with high accuracy. We show that using traditional machine learning algorithms fails to solve it. We also give a general framework for creating a classification tool when the dataset used for training is a limited dataset. We aim to build a user-friendly tool that can be used by any user for the classification task and we aim to give a guide to the researchers so that they can follow in creating a classification tool.

  12. MCM generator: a Java-based tool for generating medical metadata.

    Munoz, F; Hersh, W

    1998-01-01

    In a previous paper we introduced the need to implement a mechanism to facilitate the discovery of relevant Web medical documents. We maintained that the use of META tags, specifically ones that define the medical subject and resource type of a document, help towards this goal. We have now developed a tool to facilitate the generation of these tags for the authors of medical documents. Written entirely in Java, this tool makes use of the SAPHIRE server, and helps the author identify the Medical Subject Heading terms that most appropriately describe the subject of the document. Furthermore, it allows the author to generate metadata tags for the 15 elements that the Dublin Core considers as core elements in the description of a document. This paper describes the use of this tool in the cataloguing of Web and non-Web medical documents, such as images, movie, and sound files.

  13. Acoustic tool for leak inspections in steam generator divider plates

    Kowalewski, J.; Maak, P.; Colbert, M.

    1997-01-01

    Cross flow leakage through divider plates in the primary head of steam generators, may be a contributor to RIHT rise which can have a negative effect on the operation of a plant. A method to provide quick and reliable inspection of divider plate leakage can be very useful in helping make timely and effective maintenance decisions. A novel acoustic tool for performing inspections in drained steam generators during shutdowns has been developed by OH Technologies and successfully demonstrated in a field application. The technique, referred as ALIS (acoustic leak inspection system), performs a scan of the divider plate face and presents a graphic image of the leakage paths and an estimate of the total leakage area. (author)

  14. Management of complications of mesh surgery.

    Lee, Dominic; Zimmern, Philippe E

    2015-07-01

    Transvaginal placements of synthetic mid-urethral slings and vaginal meshes have largely superseded traditional tissue repairs in the current era because of presumed efficacy and ease of implant with device 'kits'. The use of synthetic material has generated novel complications including mesh extrusion, pelvic and vaginal pain and mesh contraction. In this review, our aim is to discuss the management, surgical techniques and outcomes associated with mesh removal. Recent publications have seen an increase in presentation of these mesh-related complications, and reports from multiple tertiary centers have suggested that not all patients benefit from surgical intervention. Although the true incidence of mesh complications is unknown, recent publications can serve to guide physicians and inform patients of the surgical outcomes from mesh-related complications. In addition, the literature highlights the growing need for a registry to account for a more accurate reporting of these events and to counsel patients on the risk and benefits before proceeding with mesh surgeries.

  15. Development of tools for automatic generation of PLC code

    Koutli, Maria; Rochez, Jacques

    This Master thesis was performed at CERN and more specifically in the EN-ICE-PLC section. The Thesis describes the integration of two PLC platforms, that are based on CODESYS development tool, to the CERN defined industrial framework, UNICOS. CODESYS is a development tool for PLC programming, based on IEC 61131-3 standard, and is adopted by many PLC manufacturers. The two PLC development environments are, the SoMachine from Schneider and the TwinCAT from Beckhoff. The two CODESYS compatible PLCs, should be controlled by the SCADA system of Siemens, WinCC OA. The framework includes a library of Function Blocks (objects) for the PLC programs and a software for automatic generation of the PLC code based on this library, called UAB. The integration aimed to give a solution that is shared by both PLC platforms and was based on the PLCOpen XML scheme. The developed tools were demonstrated by creating a control application for both PLC environments and testing of the behavior of the code of the library.

  16. A software platform for continuum modeling of ion channels based on unstructured mesh

    Tu, B; Bai, S Y; Xie, Y; Zhang, L B; Lu, B Z; Chen, M X

    2014-01-01

    Most traditional continuum molecular modeling adopted finite difference or finite volume methods which were based on a structured mesh (grid). Unstructured meshes were only occasionally used, but an increased number of applications emerge in molecular simulations. To facilitate the continuum modeling of biomolecular systems based on unstructured meshes, we are developing a software platform with tools which are particularly beneficial to those approaches. This work describes the software system specifically for the simulation of a typical, complex molecular procedure: ion transport through a three-dimensional channel system that consists of a protein and a membrane. The platform contains three parts: a meshing tool chain for ion channel systems, a parallel finite element solver for the Poisson–Nernst–Planck equations describing the electrodiffusion process of ion transport, and a visualization program for continuum molecular modeling. The meshing tool chain in the platform, which consists of a set of mesh generation tools, is able to generate high-quality surface and volume meshes for ion channel systems. The parallel finite element solver in our platform is based on the parallel adaptive finite element package PHG which wass developed by one of the authors [1]. As a featured component of the platform, a new visualization program, VCMM, has specifically been developed for continuum molecular modeling with an emphasis on providing useful facilities for unstructured mesh-based methods and for their output analysis and visualization. VCMM provides a graphic user interface and consists of three modules: a molecular module, a meshing module and a numerical module. A demonstration of the platform is provided with a study of two real proteins, the connexin 26 and hemolysin ion channels. (paper)

  17. World wide matching of registration metrology tools of various generations

    Laske, F.; Pudnos, A.; Mackey, L.; Tran, P.; Higuchi, M.; Enkrich, C.; Roeth, K.-D.; Schmidt, K.-H.; Adam, D.; Bender, J.

    2008-10-01

    Turn around time/cycle time is a key success criterion in the semiconductor photomask business. Therefore, global mask suppliers typically allocate work loads based on fab capability and utilization capacity. From a logistical point of view, the manufacturing location of a photomask should be transparent to the customer (mask user). Matching capability of production equipment and especially metrology tools is considered a key enabler to guarantee cross site manufacturing flexibility. Toppan, with manufacturing sites in eight countries worldwide, has an on-going program to match the registration metrology systems of all its production sites. This allows for manufacturing flexibility and risk mitigation.In cooperation with Vistec Semiconductor Systems, Toppan has recently completed a program to match the Vistec LMS IPRO systems at all production sites worldwide. Vistec has developed a new software feature which allows for significantly improved matching of LMS IPRO(x) registration metrology tools of various generations. We will report on the results of the global matching campaign of several of the leading Toppan sites.

  18. Analysis Tools for Next-Generation Hadron Spectroscopy Experiments

    Battaglieri, M.; Briscoe, B. J.; Celentano, A.; Chung, S.-U.; D'Angelo, A.; De Vita, R.; Döring, M.; Dudek, J.; Eidelman, S.; Fegan, S.; Ferretti, J.; Filippi, A.; Fox, G.; Galata, G.; García-Tecocoatzi, H.; Glazier, D. I.; Grube, B.; Hanhart, C.; Hoferichter, M.; Hughes, S. M.; Ireland, D. G.; Ketzer, B.; Klein, F. J.; Kubis, B.; Liu, B.; Masjuan, P.; Mathieu, V.; McKinnon, B.; Mitchel, R.; Nerling, F.; Paul, S.; Peláez, J. R.; Rademacker, J.; Rizzo, A.; Salgado, C.; Santopinto, E.; Sarantsev, A. V.; Sato, T.; Schlüter, T.; [Silva]da Silva, M. L. L.; Stankovic, I.; Strakovsky, I.; Szczepaniak, A.; Vassallo, A.; Walford, N. K.; Watts, D. P.; Zana, L.

    The series of workshops on New Partial-Wave Analysis Tools for Next-Generation Hadron Spectroscopy Experiments was initiated with the ATHOS 2012 meeting, which took place in Camogli, Italy, June 20-22, 2012. It was followed by ATHOS 2013 in Kloster Seeon near Munich, Germany, May 21-24, 2013. The third, ATHOS3, meeting is planned for April 13-17, 2015 at The George Washington University Virginia Science and Technology Campus, USA. The workshops focus on the development of amplitude analysis tools for meson and baryon spectroscopy, and complement other programs in hadron spectroscopy organized in the recent past including the INT-JLab Workshop on Hadron Spectroscopy in Seattle in 2009, the International Workshop on Amplitude Analysis in Hadron Spectroscopy at the ECT*-Trento in 2011, the School on Amplitude Analysis in Modern Physics in Bad Honnef in 2011, the Jefferson Lab Advanced Study Institute Summer School in 2012, and the School on Concepts of Modern Amplitude Analysis Techniques in Flecken-Zechlin near Berlin in September 2013. The aim of this document is to summarize the discussions that took place at the ATHOS 2012 and ATHOS 2013 meetings. We do not attempt a comprehensive review of the field of amplitude analysis, but offer a collection of thoughts that we hope may lay the ground for such a document.

  19. Analysis Tools for Next-Generation Hadron Spectroscopy Experiments

    Battaglieri, Marco; Briscoe, William; Celentano, Andrea; Chung, Suh-Urk; D'Angelo, Annalisa; De Vita, Rafaella; Döring, Michael; Dudek, Jozef; Eidelman, S.; Fegan, Stuart; Ferretti, J.; Filippi, A.; Fox, G.; Galata, G.; Garcia-Tecocoatzi, H.; Glazier, Derek; Grube, B.; Hanhart, C.; Hoferichter, M.; Hughes, S. M.; Ireland, David G.; Ketzer, B.; Klein, Franz J.; Kubis, B.; Liu, B.; Masjuan, P.; Mathieu, Vincent; McKinnon, Brian; Mitchel, R.; Nerling, F.; Paul, S.; Peláez, J. R.; Rademacker, J.; Rizzo, Alessandro; Salgado, Carlos; Santopinto, E.; Sarantsev, Andrey V.; Sato, Toru; Schlüter, T.; Da Silva, M. L.L.; Stankovic, I.; Strakovsky, Igor; Szczepaniak, Adam; Vassallo, A.; Walford, Natalie K.; Watts, Daniel P.

    2015-01-01

    The series of workshops on New Partial-Wave Analysis Tools for Next-Generation Hadron Spectroscopy Experiments was initiated with the ATHOS 2012 meeting, which took place in Camogli, Italy, June 20-22, 2012. It was followed by ATHOS 2013 in Kloster Seeon near Munich, Germany, May 21-24, 2013. The third, ATHOS3, meeting is planned for April 13-17, 2015 at The George Washington University Virginia Science and Technology Campus, USA. The workshops focus on the development of amplitude analysis tools for meson and baryon spectroscopy, and complement other programs in hadron spectroscopy organized in the recent past including the INT-JLab Workshop on Hadron Spectroscopy in Seattle in 2009, the International Workshop on Amplitude Analysis in Hadron Spectroscopy at the ECT*-Trento in 2011, the School on Amplitude Analysis in Modern Physics in Bad Honnef in 2011, the Jefferson Lab Advanced Study Institute Summer School in 2012, and the School on Concepts of Modern Amplitude Analysis Techniques in Flecken-Zechlin near Berlin in September 2013. The aim of this document is to summarize the discussions that took place at the ATHOS 2012 and ATHOS 2013 meetings. We do not attempt a comprehensive review of the field of amplitude analysis, but offer a collection of thoughts that we hope may lay the ground for such a document

  20. Sandia Generated Matrix Tool (SGMT) v. 1.0

    2010-03-24

    Provides a tool with which create and characterize a very large set of matrix-based visual analogy problems that have properties that are similar to Raven's Progressive Matrices (RPMs)™. The software uses the same underlying patterns found in RPMs to generate large numbers of unique matrix problems using parameters chosen by the researcher. Specifically, the software is designed so that researchers can choose the type, direction, and number of relations in a problem and then create any number of unique matrices that share the same underlying structure (e.g. changes in numerosity in a diagonal pattern) but have different surface features (e.g. shapes, colors).Raven's Progressive Matrices (RPMs) ™ are a widely-used test for assessing intelligence and reasoning ability. Since the test is non-verbal, it can be applied to many different populations and has been used all over the world. However, there are relatively few matrices in the sets developed by Raven, which limits their use in experiments requiring large numbers of stimuli. This tool creates a matrix set in a systematic way that allows researchers to have a great deal of control over the underlying structure, surface features, and difficulty of the matrix problems while providing a large set of novel matrices with which to conduct experiments.

  1. Computerized Generation and Simulation of Meshing and Contact of New Type of Novikov-Wildhaber Helical Gears

    Litvin, Faydor

    2000-01-01

    .... Such a function results in the reduction of noise and vibrations. Methods for the generation of the proposed gear tooth surfaces by grinding and hobbing are considered, and a tooth contact analysis (TCA...

  2. Development of Next Generation Multiphase Pipe Flow Prediction Tools

    Cem Sarica; Holden Zhang

    2006-05-31

    The developments of oil and gas fields in deep waters (5000 ft and more) will become more common in the future. It is inevitable that production systems will operate under multiphase flow conditions (simultaneous flow of gas, oil and water possibly along with sand, hydrates, and waxes). Multiphase flow prediction tools are essential for every phase of hydrocarbon recovery from design to operation. Recovery from deep-waters poses special challenges and requires accurate multiphase flow predictive tools for several applications, including the design and diagnostics of the production systems, separation of phases in horizontal wells, and multiphase separation (topside, seabed or bottom-hole). It is crucial for any multiphase separation technique, either at topside, seabed or bottom-hole, to know inlet conditions such as flow rates, flow patterns, and volume fractions of gas, oil and water coming into the separation devices. Therefore, the development of a new generation of multiphase flow predictive tools is needed. The overall objective of the proposed study is to develop a unified model for gas-oil-water three-phase flow in wells, flow lines, and pipelines to predict flow characteristics such as flow patterns, phase distributions, and pressure gradient encountered during petroleum production at different flow conditions (pipe diameter and inclination, fluid properties and flow rates). In the current multiphase modeling approach, flow pattern and flow behavior (pressure gradient and phase fractions) prediction modeling are separated. Thus, different models based on different physics are employed, causing inaccuracies and discontinuities. Moreover, oil and water are treated as a pseudo single phase, ignoring the distinct characteristics of both oil and water, and often resulting in inaccurate design that leads to operational problems. In this study, a new model is being developed through a theoretical and experimental study employing a revolutionary approach. The

  3. Steam Generator Analysis Tools and Modeling of Degradation Mechanisms

    Yetisir, M.; Pietralik, J.; Tapping, R.L.

    2004-01-01

    The degradation of steam generators (SGs) has a significant effect on nuclear heat transport system effectiveness and the lifetime and overall efficiency of a nuclear power plant. Hence, quantification of the effects of degradation mechanisms is an integral part of a SG degradation management strategy. Numerical analysis tools such as THIRST, a 3-dimensional (3D) thermal hydraulics code for recirculating SGs; SLUDGE, a 3D sludge prediction code; CHECWORKS a flow-accelerated corrosion prediction code for nuclear piping, PIPO-FE, a SG tube vibration code; and VIBIC and H3DMAP, 3D non-linear finite-element codes to predict SG tube fretting wear can be used to assess the impacts of various maintenance activities on SG thermal performance. These tools are also found to be invaluable at the design stage to influence the design by determining margins or by helping the designers minimize or avoid known degradation mechanisms. In this paper, the aforementioned numerical tools and their application to degradation mechanisms in CANDU recirculating SGs are described. In addition, the following degradation mechanisms are identified and their effect on SG thermal efficiency and lifetime are quantified: primary-side fouling, secondary-side fouling, fretting wear, and flow-accelerated corrosion (FAC). Primary-side tube inner diameter fouling has been a major contributor to SG thermal degradation. Using the results of thermalhydraulic analysis and field data, fouling margins are calculated. Individual effects of primary- and secondary-side fouling are separated through analyses, which allow station operators to decide what type of maintenance activity to perform and when to perform the maintenance activity. Prediction of the fretting-wear rate of tubes allows designers to decide on the number and locations of support plates and U-bend supports. The prediction of FAC rates for SG internals allows designers to select proper materials, and allows operators to adjust the SG maintenance

  4. BOT3P: a mesh generation software package for the transport analysis codes Dort, Tort, Twodant, Threedant and MCNP

    Orsi, R.

    2003-01-01

    Bot3p consists of a set of standard Fortran 77 language programs that gives the users of the deterministic transport codes Dort and Tort some useful diagnostic tools to prepare and check the geometry of their input data files for both Cartesian and cylindrical geometries including graphical display modules. Bot3p produces at the same time the geometrical and material distribution data for the deterministic transport codes Twodant and Threedant and, only in three-dimensional (3D) Cartesian geometry, for the Monte Carlo Transport Code MCNP. This makes it possible to compare directly for the same geometry the effects stemming from the use of different data libraries and solution approaches on transport analysis results. Through the use of Bot3p, radiation transport problems with complex 3D geometrical structures can be modelled easily, as a relatively small amount of engineer-time is required and refinement is achieved by changing few parameters. This tool is useful for solving very large challenging problems. (author)

  5. 3D geological and hydrogeological modeling as design tools for the Conawapa generating station

    Mann, J.; Sharif, S.; Smith, B. [KGS Group, Winnipeg, MB (Canada); Cook, G.N.; Osiowy, B.J. [Manitoba Hydro, Winnipeg, MB (Canada)

    2008-07-01

    Following the project's suspension in the early 1990s, part of Manitoba Hydro's recommitment study involved digital modeling of geological and hydrogeological data for the foundation design and analysis of the proposed Conawapa generating station in northern Manitoba. Three-dimensional geological and hydrogeological models have been developed to consolidate and improve the designer's ability to understand all of the information, and to assist in developing engineering alternatives which will improve the overall confidence of the design. The tools are also being leveraged for use in environmental studies. This paper provided an overview of the Conawapa site and 3-dimensional modeling goals. It described the geology and hydrogeology of the Conawapa site as well as the bedrock structure and Karst development. The paper also presented the central concepts of 3-dimensional modeling studies, including the flow of information from database to modeling software platforms. The construction of the Conawapa geological model was also presented, with particular reference to an overview of the MVS software; mesh design; and model buildup logic. The construction of the Conawapa hydrogeological model was discussed in terms of the finite element code FEFLOW software; conceptual model design; and initial observations of Conawapa groundwater flow modeling. It was concluded that recent advancement and application of 3-dimensional geological visualization software to engineering and environmental projects, including at the future Conawapa site using MVS and FEFLOW, have shown that complicated geological data can be organized, displayed, and analysed in a systematic way, to improve site visualization, understanding, and data relationships. 19 refs., 9 figs.

  6. Development of Next Generation Multiphase Pipe Flow Prediction Tools

    Tulsa Fluid Flow

    2008-08-31

    The developments of fields in deep waters (5000 ft and more) is a common occurrence. It is inevitable that production systems will operate under multiphase flow conditions (simultaneous flow of gas-oil-and water possibly along with sand, hydrates, and waxes). Multiphase flow prediction tools are essential for every phase of the hydrocarbon recovery from design to operation. The recovery from deep-waters poses special challenges and requires accurate multiphase flow predictive tools for several applications including the design and diagnostics of the production systems, separation of phases in horizontal wells, and multiphase separation (topside, seabed or bottom-hole). It is very crucial to any multiphase separation technique that is employed either at topside, seabed or bottom-hole to know inlet conditions such as the flow rates, flow patterns, and volume fractions of gas, oil and water coming into the separation devices. The overall objective was to develop a unified model for gas-oil-water three-phase flow in wells, flow lines, and pipelines to predict the flow characteristics such as flow patterns, phase distributions, and pressure gradient encountered during petroleum production at different flow conditions (pipe diameter and inclination, fluid properties and flow rates). The project was conducted in two periods. In Period 1 (four years), gas-oil-water flow in pipes were investigated to understand the fundamental physical mechanisms describing the interaction between the gas-oil-water phases under flowing conditions, and a unified model was developed utilizing a novel modeling approach. A gas-oil-water pipe flow database including field and laboratory data was formed in Period 2 (one year). The database was utilized in model performance demonstration. Period 1 primarily consisted of the development of a unified model and software to predict the gas-oil-water flow, and experimental studies of the gas-oil-water project, including flow behavior description and

  7. The Rack-Gear Tool Generation Modelling. Non-Analytical Method Developed in CATIA, Using the Relative Generating Trajectories Method

    Teodor, V. G.; Baroiu, N.; Susac, F.; Oancea, N.

    2016-11-01

    The modelling of a curl of surfaces associated with a pair of rolling centrodes, when it is known the profile of the rack-gear's teeth profile, by direct measuring, as a coordinate matrix, has as goal the determining of the generating quality for an imposed kinematics of the relative motion of tool regarding the blank. In this way, it is possible to determine the generating geometrical error, as a base of the total error. The generation modelling allows highlighting the potential errors of the generating tool, in order to correct its profile, previously to use the tool in machining process. A method developed in CATIA is proposed, based on a new method, namely the method of “relative generating trajectories”. They are presented the analytical foundation, as so as some application for knows models of rack-gear type tools used on Maag teething machines.

  8. Expressive Design Tools: Procedural Content Generation for Game Designers

    Smith, Gillian Margaret

    2012-01-01

    Games are shaped by the tools we use to make them and our ability to model the concepts they address. Vast improvements in computer graphics technology, processing power, storage capacity, and physics simulations have driven game design for the past forty years, leading to beautiful, spacious, detailed, and highly immersive worlds supporting games that are, for the most part, fundamentally about movement, collision, and other physics-based concepts. Designers use increasingly complex tools th...

  9. Computer-generated movies as an analytic tool

    Elliott, R.L.

    1978-01-01

    One of the problems faced by the users of large, sophisticated modeling programs at the Los Alamos Scientific Laboratory (LASL) is the analysis of the results of their calculations. One of the more productive and frequently spectacular methods is the production of computer-generated movies. An overview of the generation of computer movies at LASL is presented. The hardware, software, and generation techniques are briefly discussed

  10. Remote controller for tool moving in tube especially rotopeenig tool in steam generator

    Tocci, J.L.; Blotas, B.; Francois, D.

    1988-01-01

    To introduce at distance a tool such a rotopeening tool in a tube, the tool is attached to the extremity of a flexible cable which is enclosed in a flexible sheath. The sheath extremity is fixed on the arm of a carrier. The other extremity of the sheath is fixed in translation on a command bed. The tool introduction is made by a motoreduction gear placed on the bed translating the extremity of the sheath by pinion-rack system [fr

  11. Mesh Excision: Is Total Mesh Excision Necessary?

    Wolff, Gillian F; Winters, J Christian; Krlin, Ryan M

    2016-04-01

    Nearly 29% of women will undergo a secondary, repeat operation for pelvic organ prolapse (POP) symptom recurrence following a primary repair, as reported by Abbott et al. (Am J Obstet Gynecol 210:163.e1-163.e1, 2014). In efforts to decrease the rates of failure, graft materials have been utilized to augment transvaginal repairs. Following the success of using polypropylene mesh (PPM) for stress urinary incontinence (SUI), the use of PPM in the transvaginal repair of POP increased. However, in recent years, significant concerns have been raised about the safety of PPM mesh. Complications, some specific to mesh, such as exposures, erosion, dyspareunia, and pelvic pain, have been reported with increased frequency. In the current literature, there is not substantive evidence to suggest that PPM has intrinsic properties that warrant total mesh removal in the absence of complications. There are a number of complications that can occur after transvaginal mesh placement that do warrant surgical intervention after failure of conservative therapy. In aggregate, there are no high-quality controlled studies that clearly demonstrate that total mesh removal is consistently more likely to achieve pain reduction. In the cases of obstruction and erosion, it seems clear that definitive removal of the offending mesh is associated with resolution of symptoms in the majority of cases and reasonable practice. There are a number of complications that can occur with removal of mesh, and patients should be informed of this as they formulate a choice of treatment. We will review these considerations as we examine the clinical question of whether total versus partial removal of mesh is necessary for the resolution of complications following transvaginal mesh placement.

  12. Development of tools for automatic generation of PLC code

    Koutli, Maria; Chasapis, Georgios; Rochez, Jacques

    2014-01-01

    This Master thesis was performed at CERN and more specifically in the EN-ICE-PLC section. The Thesis describes the integration of two PLC platforms, that are based on CODESYS development tool, to the CERN defined industrial framework, UNICOS. CODESYS is a development tool for PLC programming, based on IEC 61131-3 standard, and is adopted by many PLC manufacturers. The two PLC development environments are, the SoMachine from Schneider and the TwinCAT from Beckhoff. The two CODESYS compatible P...

  13. Highly Symmetric and Congruently Tiled Meshes for Shells and Domes

    Rasheed, Muhibur; Bajaj, Chandrajit

    2016-01-01

    We describe the generation of all possible shell and dome shapes that can be uniquely meshed (tiled) using a single type of mesh face (tile), and following a single meshing (tiling) rule that governs the mesh (tile) arrangement with maximal vertex, edge and face symmetries. Such tiling arrangements or congruently tiled meshed shapes, are frequently found in chemical forms (fullerenes or Bucky balls, crystals, quasi-crystals, virus nano shells or capsids), and synthetic shapes (cages, sports domes, modern architectural facades). Congruently tiled meshes are both aesthetic and complete, as they support maximal mesh symmetries with minimal complexity and possess simple generation rules. Here, we generate congruent tilings and meshed shape layouts that satisfy these optimality conditions. Further, the congruent meshes are uniquely mappable to an almost regular 3D polyhedron (or its dual polyhedron) and which exhibits face-transitive (and edge-transitive) congruency with at most two types of vertices (each type transitive to the other). The family of all such congruently meshed polyhedra create a new class of meshed shapes, beyond the well-studied regular, semi-regular and quasi-regular classes, and their duals (platonic, Catalan and Johnson). While our new mesh class is infinite, we prove that there exists a unique mesh parametrization, where each member of the class can be represented by two integer lattice variables, and moreover efficiently constructable. PMID:27563368

  14. Ichthyoplankton Classification Tool using Generative Adversarial Networks and Transfer Learning

    Aljaafari, Nura

    2018-01-01

    . This method is time-consuming and requires a high level of experience. The recent advances in AI have helped to solve and automate several difficult tasks which motivated us to develop a classification tool for ichthyoplankton. We show that using machine

  15. Connectivity editing for quadrilateral meshes

    Peng, Chihan; Zhang, Eugene; Kobayashi, Yoshihiro; Wonka, Peter

    2011-01-01

    We propose new connectivity editing operations for quadrilateral meshes with the unique ability to explicitly control the location, orientation, type, and number of the irregular vertices (valence not equal to four) in the mesh while preserving sharp edges. We provide theoretical analysis on what editing operations are possible and impossible and introduce three fundamental operations to move and re-orient a pair of irregular vertices. We argue that our editing operations are fundamental, because they only change the quad mesh in the smallest possible region and involve the fewest irregular vertices (i.e., two). The irregular vertex movement operations are supplemented by operations for the splitting, merging, canceling, and aligning of irregular vertices. We explain how the proposed highlevel operations are realized through graph-level editing operations such as quad collapses, edge flips, and edge splits. The utility of these mesh editing operations are demonstrated by improving the connectivity of quad meshes generated from state-of-art quadrangulation techniques. © 2011 ACM.

  16. Connectivity editing for quadrilateral meshes

    Peng, Chihan

    2011-12-12

    We propose new connectivity editing operations for quadrilateral meshes with the unique ability to explicitly control the location, orientation, type, and number of the irregular vertices (valence not equal to four) in the mesh while preserving sharp edges. We provide theoretical analysis on what editing operations are possible and impossible and introduce three fundamental operations to move and re-orient a pair of irregular vertices. We argue that our editing operations are fundamental, because they only change the quad mesh in the smallest possible region and involve the fewest irregular vertices (i.e., two). The irregular vertex movement operations are supplemented by operations for the splitting, merging, canceling, and aligning of irregular vertices. We explain how the proposed highlevel operations are realized through graph-level editing operations such as quad collapses, edge flips, and edge splits. The utility of these mesh editing operations are demonstrated by improving the connectivity of quad meshes generated from state-of-art quadrangulation techniques. © 2011 ACM.

  17. Let's collocate: student generated worksheets as a motivational tool

    Simpson, Adam John

    2006-01-01

    This article discusses the process of producing collocation worksheets and the values of these worksheets as a motivational tool within a tertiary level preparatory English program. Firstly, the method by which these worksheets were produced is described, followed by an analysis of their effectiveness as a resource in terms of student motivation, personalisation, involvement in the development of the curriculum and in raising awareness of corpus linguistics and its applications.

  18. JVM: Java Visual Mapping tool for next generation sequencing read.

    Yang, Ye; Liu, Juan

    2015-01-01

    We developed a program JVM (Java Visual Mapping) for mapping next generation sequencing read to reference sequence. The program is implemented in Java and is designed to deal with millions of short read generated by sequence alignment using the Illumina sequencing technology. It employs seed index strategy and octal encoding operations for sequence alignments. JVM is useful for DNA-Seq, RNA-Seq when dealing with single-end resequencing. JVM is a desktop application, which supports reads capacity from 1 MB to 10 GB.

  19. ARBUS: A FORTRAN tool for generating tree structure diagrams

    Ferrero, C.; Zanger, M.

    1992-02-01

    The FORTRAN77 stand-alone code ARBUS has been designed to aid the user by providing a tree structure diagram generating utility for computer programs written in FORTRAN language. This report is intended to describe the main purpose and features of ARBUS and to highlight some additional applications of the code by means of practical test cases. (orig.) [de

  20. WASP as a planning tool of electrical generation systems expansion

    D'Isidoro, G.

    1984-01-01

    The ''Wien Automatic System Package'' (WASP), consists of six modules or computer programmes which assist in decision taking process in expanding an electrical generation network. A general description of this model is made and some conclusions are drawn from the data processed to this date

  1. Computer Generated Optical Illusions: A Teaching and Research Tool.

    Bailey, Bruce; Harman, Wade

    Interactive computer-generated simulations that highlight psychological principles were investigated in this study in which 33 female and 19 male undergraduate college student volunteers of median age 21 matched line and circle sizes in six variations of Ponzo's illusion. Prior to working with the illusions, data were collected based on subjects'…

  2. MCBooster: a tool for MC generation for massively parallel platforms

    Alves Junior, Antonio Augusto

    2016-01-01

    MCBooster is a header-only, C++11-compliant library for the generation of large samples of phase-space Monte Carlo events on massively parallel platforms. It was released on GitHub in the spring of 2016. The library core algorithms implement the Raubold-Lynch method; they are able to generate the full kinematics of decays with up to nine particles in the final state. The library supports the generation of sequential decays as well as the parallel evaluation of arbitrary functions over the generated events. The output of MCBooster completely accords with popular and well-tested software packages such as GENBOD (W515 from CERNLIB) and TGenPhaseSpace from the ROOT framework. MCBooster is developed on top of the Thrust library and runs on Linux systems. It deploys transparently on NVidia CUDA-enabled GPUs as well as multicore CPUs. This contribution summarizes the main features of MCBooster. A basic description of the user interface and some examples of applications are provided, along with measurements of perfor...

  3. Multiscale Toxicology - Building the Next Generation Tools for Toxicology

    Thrall, Brian D.; Minard, Kevin R.; Teeguarden, Justin G.; Waters, Katrina M.

    2012-09-01

    A Cooperative Research and Development Agreement (CRADA) was sponsored by Battelle Memorial Institute (Battelle, Columbus), to initiate a collaborative research program across multiple Department of Energy (DOE) National Laboratories aimed at developing a suite of new capabilities for predictive toxicology. Predicting the potential toxicity of emerging classes of engineered nanomaterials was chosen as one of two focusing problems for this program. PNNL’s focus toward this broader goal was to refine and apply experimental and computational tools needed to provide quantitative understanding of nanoparticle dosimetry for in vitro cell culture systems, which is necessary for comparative risk estimates for different nanomaterials or biological systems. Research conducted using lung epithelial and macrophage cell models successfully adapted magnetic particle detection and fluorescent microscopy technologies to quantify uptake of various forms of engineered nanoparticles, and provided experimental constraints and test datasets for benchmark comparison against results obtained using an in vitro computational dosimetry model, termed the ISSD model. The experimental and computational approaches developed were used to demonstrate how cell dosimetry is applied to aid in interpretation of genomic studies of nanoparticle-mediated biological responses in model cell culture systems. The combined experimental and theoretical approach provides a highly quantitative framework for evaluating relationships between biocompatibility of nanoparticles and their physical form in a controlled manner.

  4. Profiting from competition: Financial tools for electric generation companies

    Richter, Charles William, Jr.

    Regulations governing the operation of electric power systems in North America and many other areas of the world are undergoing major changes designed to promote competition. This process of change is often referred to as deregulation. Participants in deregulated electricity systems may find that their profits will greatly benefit from the implementation of successful bidding strategies. While the goal of the regulators may be to create rules which balance reliable power system operation with maximization of the total benefit to society, the goal of generation companies is to maximize their profit, i.e., return to their shareholders. The majority of the research described here is conducted from the point of view of generation companies (GENCOs) wishing to maximize their expected utility function, which is generally comprised of expected profit and risk. Strategies that help a GENCO to maximize its objective function must consider the impact of (and aid in making) operating decisions that may occur within a few seconds to multiple years. The work described here assumes an environment in which energy service companies (ESCOs) buy and GENCOs sell power via double auctions in regional commodity exchanges. Power is transported on wires owned by transmission companies (TRANSCOs) and distribution companies (DISTCOs). The proposed market framework allows participants to trade electrical energy contracts via the spot, futures, options, planning, and swap markets. An important method of studying these proposed markets and the behavior of participating agents is the field of experimental/computational economics. For much of the research reported here, the market simulator developed by Kumar and Sheble and similar simulators has been adapted to allow computerized agents to trade energy. Creating computerized agents that can react as rationally or irrationally as a human trader is a difficult problem for which we have turned to the field of artificial intelligence. Some of our

  5. Transfer of safety responsibilities to future generations: regulatory tools

    Kotra, Janet P.

    2008-01-01

    In a forward-looking local development plan, Nye County defends a series of principles like safety, equity, and societal acceptability of responsibility (safety being foremost). The Nye County community clearly advocates permanent oversight of facilities. To respond to community requirements the regulators can establish requirements and guidance to ensure that safety obligations that can reasonably be discharged are in fact carried out and that remaining obligations are transferred as responsibly as possible, so that subsequent generations have the maximum flexibility to discharge their responsibility. There are transferred burdens of cost, risk and effort and these need to be at least partially compensated for by ensuring a subsequent transfer of information, resources and continuity of education, skills and research. The US regulatory requirements for disposal in a geological repository set out obligations in terms of land-ownership and control, records maintenance, performance confirmation, post-closure monitoring, monuments and markers, archives and records preservation and post-closure oversight. For the future the Nye County is proposing that there would be a co-ordinated involvement of the county in planning, development, operation and long term monitoring of the repository. They want to encourage the development of a live-work community for the repository workers so that they will be engaged in the local community as well as working at the facility

  6. Multiscale Toxicology- Building the Next Generation Tools for Toxicology

    Retterer, S. T. [ORNL; Holsapple, M. P. [Battelle Memorial Institute

    2013-10-31

    A Cooperative Research and Development Agreement (CRADA) was established between Battelle Memorial Institute (BMI), Pacific Northwest National Laboratory (PNNL), Oak Ridge National Laboratory (ORNL), Brookhaven National Laboratory (BNL), Lawrence Livermore National Laboratory (LLNL) with the goal of combining the analytical and synthetic strengths of the National Laboratories with BMI's expertise in basic and translational medical research to develop a collaborative pipeline and suite of high throughput and imaging technologies that could be used to provide a more comprehensive understanding of material and drug toxicology in humans. The Multi-Scale Toxicity Initiative (MSTI), consisting of the team members above, was established to coordinate cellular scale, high-throughput in vitro testing, computational modeling and whole animal in vivo toxicology studies between MSTI team members. Development of a common, well-characterized set of materials for testing was identified as a crucial need for the initiative. Two research tracks were established by BMI during the course of the CRADA. The first research track focused on the development of tools and techniques for understanding the toxicity of nanomaterials, specifically inorganic nanoparticles (NPs). ORNL"s work focused primarily on the synthesis, functionalization and characterization of a common set of NPs for dissemination to the participating laboratories. These particles were synthesized to retain the same surface characteristics and size, but to allow visualization using the variety of imaging technologies present across the team. Characterization included the quantitative analysis of physical and chemical properties of the materials as well as the preliminary assessment of NP toxicity using commercially available toxicity screens and emerging optical imaging strategies. Additional efforts examined the development of high-throughput microfluidic and imaging assays for measuring NP uptake, localization, and

  7. PC graphics generation and management tool for real-time applications

    Truong, Long V.

    1992-01-01

    A graphics tool was designed and developed for easy generation and management of personal computer graphics. It also provides methods and 'run-time' software for many common artificial intelligence (AI) or expert system (ES) applications.

  8. Current situation of transvaginal mesh repair for pelvic organ prolapse.

    Zhu, Lan; Zhang, Lei

    2014-09-01

    Surgical mesh is a metallic or polymeric screen intended to be implanted to reinforce soft tissue or bone where weakness exists. Surgical mesh has been used since the 1950s to repair abdominal hernias. In the 1970s, gynecologists began using surgical mesh products to indicate the repair of pelvic organ prolapse (POP), and in the 1990s, gynecologists began using surgical mesh for POP. Then the U.S. Food and Drug Administration (FDA) approved the first surgical mesh product specifically for use in POP. Surgical mesh materials can be divided into several categories. Most surgical mesh devices cleared for POP procedures are composed of non-absorbable synthetic polypropylene. Mesh can be placed in the anterior vaginal wall to aid in the correction of cystocele (anterior repair), in the posterior vaginal wall to aid in correction of rectocele (posterior repair), or attached to the top of the vagina to correct uterine prolapse or vaginal apical prolapse (apical repair). Over the past decades, surgical mesh products for transvaginal POP repair became incorporated into "kits" that included tools to aid in the delivery and insertion of the mesh. Surgical mesh kits continue to evolve, adding new insertion tools, tissue fixation anchors, surgical techniques, and ab- sorbable and biological materials. This procedure has been performed popularly. It was also performed increased in China. But this new technique met some trouble recently and let shake in urogynecology.

  9. GeneratorSE: A Sizing Tool for Variable-Speed Wind Turbine Generators

    Sethuraman, Latha [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Dykes, Katherine L [National Renewable Energy Laboratory (NREL), Golden, CO (United States)

    2017-09-22

    This report documents a set of analytical models employed by the optimization algorithms within the GeneratorSE framework. The initial values and boundary conditions employed for the generation of the various designs and initial estimates for basic design dimensions, masses, and efficiency for the four different models of generators are presented and compared with empirical data collected from previous studies and some existing commercial turbines. These models include designs applicable for variable-speed, high-torque application featuring direct-drive synchronous generators and low-torque application featuring induction generators. In all of the four models presented, the main focus of optimization is electromagnetic design with the exception of permanent-magnet and wire-wound synchronous generators, wherein the structural design is also optimized. Thermal design is accommodated in GeneratorSE as a secondary attribute by limiting the winding current densities to acceptable limits. A preliminary validation of electromagnetic design was carried out by comparing the optimized magnetic loading against those predicted by numerical simulation in FEMM4.2, a finite-element software for analyzing electromagnetic and thermal physics problems for electrical machines. For direct-drive synchronous generators, the analytical models for the structural design are validated by static structural analysis in ANSYS.

  10. T3i: A Tool for Generating and Querying Test Suites for Java

    Prasetya, I.S.W.B.

    2015-01-01

    T3i is an automated unit-testing tool to test Java classes. To expose interactions T3i generates test-cases in the form of sequences of calls to the methods of the target class. What separates it from other testing tools is that it treats test suites as first class objects and allows users to e.g.

  11. How to model wireless mesh networks topology

    Sanni, M L; Hashim, A A; Anwar, F; Ali, S; Ahmed, G S M

    2013-01-01

    The specification of network connectivity model or topology is the beginning of design and analysis in Computer Network researches. Wireless Mesh Networks is an autonomic network that is dynamically self-organised, self-configured while the mesh nodes establish automatic connectivity with the adjacent nodes in the relay network of wireless backbone routers. Researches in Wireless Mesh Networks range from node deployment to internetworking issues with sensor, Internet and cellular networks. These researches require modelling of relationships and interactions among nodes including technical characteristics of the links while satisfying the architectural requirements of the physical network. However, the existing topology generators model geographic topologies which constitute different architectures, thus may not be suitable in Wireless Mesh Networks scenarios. The existing methods of topology generation are explored, analysed and parameters for their characterisation are identified. Furthermore, an algorithm for the design of Wireless Mesh Networks topology based on square grid model is proposed in this paper. The performance of the topology generated is also evaluated. This research is particularly important in the generation of a close-to-real topology for ensuring relevance of design to the intended network and validity of results obtained in Wireless Mesh Networks researches

  12. Hernia Surgical Mesh Implants

    ... knitted mesh or non-knitted sheet forms. The synthetic materials used can be absorbable, non-absorbable or a combination of absorbable and non-absorbable materials. Animal-derived mesh are made of animal tissue, such as intestine or skin, that has been processed and disinfected to be ...

  13. PetriCode: A Tool for Template-Based Code Generation from CPN Models

    Simonsen, Kent Inge

    2014-01-01

    Code generation is an important part of model driven methodologies. In this paper, we present PetriCode, a software tool for generating protocol software from a subclass of Coloured Petri Nets (CPNs). The CPN subclass is comprised of hierarchical CPN models describing a protocol system at different...

  14. RETRANS - A tool to verify the functional equivalence of automatically generated source code with its specification

    Miedl, H.

    1998-01-01

    Following the competent technical standards (e.g. IEC 880) it is necessary to verify each step in the development process of safety critical software. This holds also for the verification of automatically generated source code. To avoid human errors during this verification step and to limit the cost effort a tool should be used which is developed independently from the development of the code generator. For this purpose ISTec has developed the tool RETRANS which demonstrates the functional equivalence of automatically generated source code with its underlying specification. (author)

  15. 22nd International Meshing Roundtable

    Staten, Matthew

    2014-01-01

    This volume contains the articles presented at the 22nd International Meshing Roundtable (IMR) organized, in part, by Sandia National Laboratories and was held on Oct 13-16, 2013 in Orlando, Florida, USA.  The first IMR was held in 1992, and the conference series has been held annually since.  Each year the IMR brings together researchers, developers, and application experts in a variety of disciplines, from all over the world, to present and discuss ideas on mesh generation and related topics.  The technical papers in this volume present theoretical and novel ideas and algorithms with practical potential, as well as technical applications in science and engineering, geometric modeling, computer graphics and visualization.

  16. 21st International Meshing Roundtable

    Weill, Jean-Christophe

    2013-01-01

    This volume contains the articles presented at the 21st International Meshing Roundtable (IMR) organized, in part, by Sandia National Laboratories and was held on October 7–10, 2012 in San Jose, CA, USA. The first IMR was held in 1992, and the conference series has been held annually since. Each year the IMR brings together researchers, developers, and application experts in a variety of disciplines, from all over the world, to present and discuss ideas on mesh generation and related topics. The technical papers in this volume present theoretical and novel ideas and algorithms with practical potential, as well as technical applications in science and engineering, geometric modeling, computer graphics, and visualization.

  17. Automated planning target volume generation: an evaluation pitting a computer-based tool against human experts

    Ketting, Case H.; Austin-Seymour, Mary; Kalet, Ira; Jacky, Jon; Kromhout-Schiro, Sharon; Hummel, Sharon; Unger, Jonathan; Fagan, Lawrence M.; Griffin, Tom

    1997-01-01

    Purpose: Software tools are seeing increased use in three-dimensional treatment planning. However, the development of these tools frequently omits careful evaluation before placing them in clinical use. This study demonstrates the application of a rigorous evaluation methodology using blinded peer review to an automated software tool that produces ICRU-50 planning target volumes (PTVs). Methods and Materials: Seven physicians from three different institutions involved in three-dimensional treatment planning participated in the evaluation. Four physicians drew partial PTVs on nine test cases, consisting of four nasopharynx and five lung primaries. Using the same information provided to the human experts, the computer tool generated PTVs for comparison. The remaining three physicians, designated evaluators, individually reviewed the PTVs for acceptability. To exclude bias, the evaluators were blinded to the source (human or computer) of the PTVs they reviewed. Their scorings of the PTVs were statistically examined to determine if the computer tool performed as well as the human experts. Results: The computer tool was as successful as the human experts in generating PTVs. Failures were primarily attributable to insufficient margins around the clinical target volume and to encroachment upon critical structures. In a qualitative analysis, the human and computer experts displayed similar types and distributions of errors. Conclusions: Rigorous evaluation of computer-based radiotherapy tools requires comparison to current practice and can reveal areas for improvement before the tool enters clinical practice

  18. Capacity Analysis of Wireless Mesh Networks

    M. I. Gumel

    2012-06-01

    Full Text Available The next generation wireless networks experienced a great development with emergence of wireless mesh networks (WMNs, which can be regarded as a realistic solution that provides wireless broadband access. The limited available bandwidth makes capacity analysis of the network very essential. While the network offers broadband wireless access to community and enterprise users, the problems that limit the network capacity must be addressed to exploit the optimum network performance. The wireless mesh network capacity analysis shows that the throughput of each mesh node degrades in order of l/n with increasing number of nodes (n in a linear topology. The degradation is found to be higher in a fully mesh network as a result of increase in interference and MAC layer contention in the network.

  19. MeSH Now: automatic MeSH indexing at PubMed scale via learning to rank.

    Mao, Yuqing; Lu, Zhiyong

    2017-04-17

    MeSH indexing is the task of assigning relevant MeSH terms based on a manual reading of scholarly publications by human indexers. The task is highly important for improving literature retrieval and many other scientific investigations in biomedical research. Unfortunately, given its manual nature, the process of MeSH indexing is both time-consuming (new articles are not immediately indexed until 2 or 3 months later) and costly (approximately ten dollars per article). In response, automatic indexing by computers has been previously proposed and attempted but remains challenging. In order to advance the state of the art in automatic MeSH indexing, a community-wide shared task called BioASQ was recently organized. We propose MeSH Now, an integrated approach that first uses multiple strategies to generate a combined list of candidate MeSH terms for a target article. Through a novel learning-to-rank framework, MeSH Now then ranks the list of candidate terms based on their relevance to the target article. Finally, MeSH Now selects the highest-ranked MeSH terms via a post-processing module. We assessed MeSH Now on two separate benchmarking datasets using traditional precision, recall and F 1 -score metrics. In both evaluations, MeSH Now consistently achieved over 0.60 in F-score, ranging from 0.610 to 0.612. Furthermore, additional experiments show that MeSH Now can be optimized by parallel computing in order to process MEDLINE documents on a large scale. We conclude that MeSH Now is a robust approach with state-of-the-art performance for automatic MeSH indexing and that MeSH Now is capable of processing PubMed scale documents within a reasonable time frame. http://www.ncbi.nlm.nih.gov/CBBresearch/Lu/Demo/MeSHNow/ .

  20. Geometrically Consistent Mesh Modification

    Bonito, A.

    2010-01-01

    A new paradigm of adaptivity is to execute refinement, coarsening, and smoothing of meshes on manifolds with incomplete information about their geometry and yet preserve position and curvature accuracy. We refer to this collectively as geometrically consistent (GC) mesh modification. We discuss the concept of discrete GC, show the failure of naive approaches, and propose and analyze a simple algorithm that is GC and accuracy preserving. © 2010 Society for Industrial and Applied Mathematics.

  1. Adaptive and dynamic meshing methods for numerical simulations

    Acikgoz, Nazmiye

    For the numerical simulation of many problems of engineering interest, it is desirable to have an automated mesh adaption tool capable of producing high quality meshes with an affordably low number of mesh points. This is important especially for problems, which are characterized by anisotropic features of the solution and require mesh clustering in the direction of high gradients. Another significant issue in meshing emerges in the area of unsteady simulations with moving boundaries or interfaces, where the motion of the boundary has to be accommodated by deforming the computational grid. Similarly, there exist problems where current mesh needs to be adapted to get more accurate solutions because either the high gradient regions are initially predicted inaccurately or they change location throughout the simulation. To solve these problems, we propose three novel procedures. For this purpose, in the first part of this work, we present an optimization procedure for three-dimensional anisotropic tetrahedral grids based on metric-driven h-adaptation. The desired anisotropy in the grid is dictated by a metric that defines the size, shape, and orientation of the grid elements throughout the computational domain. Through the use of topological and geometrical operators, the mesh is iteratively adapted until the final mesh minimizes a given objective function. In this work, the objective function measures the distance between the metric of each simplex and a target metric, which can be either user-defined (a-priori) or the result of a-posteriori error analysis. During the adaptation process, one tries to decrease the metric-based objective function until the final mesh is compliant with the target within a given tolerance. However, in regions such as corners and complex face intersections, the compliance condition was found to be very difficult or sometimes impossible to satisfy. In order to address this issue, we propose an optimization process based on an ad

  2. The C language auto-generation of reactor trip logic caused by steam generator water level using CASE tools

    Kim, Jang Yeol; Lee, Jang Soo

    1999-01-01

    The purpose is to produce a model of nuclear reactor trip logic caused by the steam generator water level of Wolsung 2/3/4 unit through an activity chart and a statechart and to produce C language automatically using statechart-based formalism and statemate MAGNUM toolset suggested by David Harel Formalism. It was worth attempting auto-generation of C language through we manually made Software Requirement specification(SRS) for safety-critical software using statechart-based formalism. Most of the phase of the software life-cycle except the software requirement specification of an analysis phase were generated automatically by Computer Aided Software Engineering(CASE) tools. It was verified that automatically produced C language has high productivity, portability, and quality through the simulation. (Author). 6 refs., 6 figs

  3. CAPAClTYANALYSIS OF WIRELESS MESH NET\\VORKS

    The limited available bandwidth makes capacity analysis of the network very essential. ... Wireless mesh networks can also be employed for wide variety ofapplications such ... wireless mesh networks using OPNET (Optimized Network Engineering Tool) Modeller 1-J..5. The .... /bps using I I Mbps data rate and 12000 bits.

  4. An Approach to Quad Meshing Based On Cross Valued Maps and the Ginzburg-Landau Theory

    Viertel, Ryan [Univ. of Utah, Salt Lake City, UT (United States). Dept. of Mathematics; Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Osting, Braxton [Univ. of Utah, Salt Lake City, UT (United States). Dept. of Mathematics

    2017-08-01

    A generalization of vector fields, referred to as N-direction fields or cross fields when N=4, has been recently introduced and studied for geometry processing, with applications in quadrilateral (quad) meshing, texture mapping, and parameterization. We make the observation that cross field design for two-dimensional quad meshing is related to the well-known Ginzburg-Landau problem from mathematical physics. This identification yields a variety of theoretical tools for efficiently computing boundary-aligned quad meshes, with provable guarantees on the resulting mesh, for example, the number of mesh defects and bounds on the defect locations. The procedure for generating the quad mesh is to (i) find a complex-valued "representation" field that minimizes the Dirichlet energy subject to a boundary constraint, (ii) convert the representation field into a boundary-aligned, smooth cross field, (iii) use separatrices of the cross field to partition the domain into four sided regions, and (iv) mesh each of these four-sided regions using standard techniques. Under certain assumptions on the geometry of the domain, we prove that this procedure can be used to produce a cross field whose separatrices partition the domain into four sided regions. To solve the energy minimization problem for the representation field, we use an extension of the Merriman-Bence-Osher (MBO) threshold dynamics method, originally conceived as an algorithm to simulate motion by mean curvature, to minimize the Ginzburg-Landau energy for the optimal representation field. Lastly, we demonstrate the method on a variety of test domains.

  5. Coarse mesh code development

    Lieberoth, J.

    1975-06-15

    The numerical solution of the neutron diffusion equation plays a very important role in the analysis of nuclear reactors. A wide variety of numerical procedures has been proposed, at which most of the frequently used numerical methods are fundamentally based on the finite- difference approximation where the partial derivatives are approximated by the finite difference. For complex geometries, typical of the practical reactor problems, the computational accuracy of the finite-difference method is seriously affected by the size of the mesh width relative to the neutron diffusion length and by the heterogeneity of the medium. Thus, a very large number of mesh points are generally required to obtain a reasonably accurate approximate solution of the multi-dimensional diffusion equation. Since the computation time is approximately proportional to the number of mesh points, a detailed multidimensional analysis, based on the conventional finite-difference method, is still expensive even with modern large-scale computers. Accordingly, there is a strong incentive to develop alternatives that can reduce the number of mesh-points and still retain accuracy. One of the promising alternatives is the finite element method, which consists of the expansion of the neutron flux by piecewise polynomials. One of the advantages of this procedure is its flexibility in selecting the locations of the mesh points and the degree of the expansion polynomial. The small number of mesh points of the coarse grid enables to store the results of several of the least outer iterations and to calculate well extrapolated values of them by comfortable formalisms. This holds especially if only one energy distribution of fission neutrons is assumed for all fission processes in the reactor, because the whole information of an outer iteration is contained in a field of fission rates which has the size of all mesh points of the coarse grid.

  6. Case-Based Pedagogy Using Student-Generated Vignettes: A Pre-Service Intercultural Awareness Tool

    Cournoyer, Amy

    2010-01-01

    This qualitative study investigated the effectiveness of case-based pedagogy as an instructional tool aimed at increasing cultural awareness and competence in the preparation of 18 pre-service and in-service students enrolled in an Intercultural Education course. Each participant generated a vignette based on an instructional challenge identified…

  7. NUMERICAL STUDY OF ELECTROMAGNETIC WAVES GENERATED BY A PROTOTYPE DIELECTRIC LOGGING TOOL

    To understand the electromagnetic waves generated by a prototype dielectric logging tool, a numerical study was conducted using both the finite-difference, time-domain method and a frequency- wavenumber method. When the propagation velocity in the borehole was greater than th...

  8. Evaluation of Computer Tools for Idea Generation and Team Formation in Project-Based Learning

    Ardaiz-Villanueva, Oscar; Nicuesa-Chacon, Xabier; Brene-Artazcoz, Oscar; Sanz de Acedo Lizarraga, Maria Luisa; Sanz de Acedo Baquedano, Maria Teresa

    2011-01-01

    The main objective of this research was to validate the effectiveness of Wikideas and Creativity Connector tools to stimulate the generation of ideas and originality by university students organized into groups according to their indexes of creativity and affinity. Another goal of the study was to evaluate the classroom climate created by these…

  9. Gear cutting tools fundamentals of design and computation

    Radzevich, Stephen P

    2010-01-01

    Presents the DG/K-based method of surface generation, a novel and practical mathematical method for designing gear cutting tools with optimal parameters. This book proposes a scientific classification for the various kinds of the gear machining meshes, discussing optimal designs of gear cutting tools.

  10. Dry metal forming of high alloy steel using laser generated aluminum bronze tools

    Freiße Hannes

    2015-01-01

    Full Text Available Regarding the optimization of forming technology in economic and environmental aspects, avoiding lubricants is an approach to realize the vision of a new green technology. The resulting direct contact between the tool and the sheet in non-lubricated deep drawing causes higher stress and depends mainly on the material combination. The tribological system in dry sliding has to be assessed by means on the one hand of the resulting friction coefficient and on the other hand of the wear of the tool and sheet material. The potential to generate tailored tribological systems for dry metal forming could be shown within the investigations by using different material combinations and by applying different laser cladding process parameters. Furthermore, the feasibility of additive manufacturing of a deep drawing tool was demonstrated. The tool was successfully applied to form circular cups in a dry metal forming process.

  11. Trajectory Assessment and Modification Tools for Next Generation Air Traffic Management Operations

    Brasil, Connie; Lee, Paul; Mainini, Matthew; Lee, Homola; Lee, Hwasoo; Prevot, Thomas; Smith, Nancy

    2011-01-01

    This paper reviews three Next Generation Air Transportation System (NextGen) based high fidelity air traffic control human-in-the-loop (HITL) simulations, with a focus on the expected requirement of enhanced automated trajectory assessment and modification tools to support future air traffic flow management (ATFM) planning positions. The simulations were conducted at the National Aeronautics and Space Administration (NASA) Ames Research Centers Airspace Operations Laboratory (AOL) in 2009 and 2010. The test airspace for all three simulations assumed the mid-term NextGenEn-Route high altitude environment utilizing high altitude sectors from the Kansas City and Memphis Air Route Traffic Control Centers. Trajectory assessment, modification and coordination decision support tools were developed at the AOL in order to perform future ATFM tasks. Overall tool usage results and user acceptability ratings were collected across three areas of NextGen operatoins to evaluate the tools. In addition to the usefulness and usability feedback, feasibility issues, benefits, and future requirements were also addressed. Overall, the tool sets were rated very useful and usable, and many elements of the tools received high scores and were used frequently and successfully. Tool utilization results in all three HITLs showed both user and system benefits including better airspace throughput, reduced controller workload, and highly effective communication protocols in both full Data Comm and mixed-equipage environments.

  12. Adaptive scallop height tool path generation for robot-based incremental sheet metal forming

    Seim, Patrick; Möllensiep, Dennis; Störkle, Denis Daniel; Thyssen, Lars; Kuhlenkötter, Bernd

    2016-10-01

    Incremental sheet metal forming is an emerging process for the production of individualized products or prototypes in low batch sizes and with short times to market. In these processes, the desired shape is produced by the incremental inward motion of the workpiece-independent forming tool in depth direction and its movement along the contour in lateral direction. Based on this shape production, the tool path generation is a key factor on e.g. the resulting geometric accuracy, the resulting surface quality, and the working time. This paper presents an innovative tool path generation based on a commercial milling CAM package considering the surface quality and working time. This approach offers the ability to define a specific scallop height as an indicator of the surface quality for specific faces of a component. Moreover, it decreases the required working time for the production of the entire component compared to the use of a commercial software package without this adaptive approach. Different forming experiments have been performed to verify the newly developed tool path generation. Mainly, this approach serves to solve the existing conflict of combining the working time and the surface quality within the process of incremental sheet metal forming.

  13. Ultrasonic wall thickness gauging for ferritic steam generator tubing as an in-service inspection tool

    Haesen, W.M.J.; Tromp, Th.J.

    1980-01-01

    In-service inspection of LWR steam generators is more or less a standard routine operation. The situation can be very different for LMFBRs. For the SNR 300 (Kalkar Power Station) the situation is different because the steam generators have ferritic tubing. The tube walls are comparatively thick, 2 to 4.5 mm. During inservice examinations the steam generators will be drained on both sides, however on the sodium side a sodium film will be present. Furthermore the SNR 300 will have two types of steam generator. A straight tube design and a helical coil design will be used. Both types consist of a evaporator and superheater. The steam generators are of course not radioactive. It is obvious that in this case the eddy current (EC) technique is not an enviable inservice inspection tool. Basically EC is a surface flaw detection technique. Only the saturation magnetisation method will improve the EC technique sufficiently for ferritic material. However the 'in bore examination' with the saturation technique was, in case of the SNR 300 steam generator tubing, considered impossible since the inner diameters are fairly small. Furthermore sodium traces may influence the EC method. Although multifrequency methods can solve this problem, EC is not considered as a useful tool for examining ferritic tubing. Another method is to employ the 'stray flux' method which is under development with the TNO organization in Holland. The EC and stray flux method do have one drawback, these methods do not detect gradual changes in wall thickness. Ultrasonic examinations will be used in the SNR 300 as the main inspection tool for the steam generators. In this paper the reasons why ultrasonic examination was selected are explained. The results of the development work on this subject are discussed

  14. Design tools for computer-generated display of information to operators

    O'Brien, J.F.; Cain, D.G.; Sun, B.K.H.

    1985-01-01

    More and more computers are being used to process and display information to operators who control nuclear power plants. Implementation of computer-generated displays in power plant control rooms represents a considerable design challenge for industry designers. Over the last several years, the EPRI has conducted research aimed at providing industry designers tools to meet this new design challenge. These tools provide guidance in defining more 'intelligent' information for plant control and in developing effective displays to communicate this information to the operators. (orig./HP)

  15. A software tool for simulation of surfaces generated by ball nose end milling

    Bissacco, Giuliano

    2004-01-01

    , for prediction of surface topography of ball nose end milled surfaces, was developed. Such software tool is based on a simplified model of the ideal tool motion and neglects the effects due to run-out, static and dynamic deflections and error motions, but has the merit of generating in output a file in a format...... readable by a surface processor software (SPIP [2]), for calculation of a number of surface roughness parameters. In the next paragraph a description of the basic features of ball nose end milled surfaces is given, while in paragraph 3 the model is described....

  16. Plant management tools tested with a small-scale distributed generation laboratory

    Ferrari, Mario L.; Traverso, Alberto; Pascenti, Matteo; Massardo, Aristide F.

    2014-01-01

    Highlights: • Thermal grid innovative layouts. • Experimental rig for distributed generation. • Real-time management tool. • Experimental results for plant management. • Comparison with results from an optimization complete software. - Abstract: Optimization of power generation with smart grids is an important issue for extensive sustainable development of distributed generation. Since an experimental approach is essential for implementing validated optimization software, the TPG research team of the University of Genoa has installed a laboratory facility for carrying out studies on polygeneration grids. The facility consists of two co-generation prime movers based on conventional technology: a 100 kWe gas turbine (mGT) and a 20 kWe internal combustion engine (ICE). The rig high flexibility allows the possibility of integration with renewable-source based devices, such as biomass-fed boilers and solar panels. Special attention was devoted to thermal distribution grid design. To ensure the possibility of application in medium-large districts, composed of several buildings including energy users, generators or both, an innovative layout based on two ring pipes was examined. Thermal storage devices were also included in order to have a complete hardware platform suitable for assessing the performance of different management tools. The test presented in this paper was carried out with both the mGT and the ICE connected to this innovative thermal grid, while users were emulated by means of fan coolers controlled by inverters. During this test the plant is controlled by a real-time model capable of calculating a machine performance ranking, which is necessary in order to split power demands between the prime movers (marginal cost decrease objective). A complete optimization tool devised by TPG (ECoMP program) was also used in order to obtain theoretical results considering the same machines and load values. The data obtained with ECoMP were compared with the

  17. Forecasting municipal solid waste generation using prognostic tools and regression analysis.

    Ghinea, Cristina; Drăgoi, Elena Niculina; Comăniţă, Elena-Diana; Gavrilescu, Marius; Câmpean, Teofil; Curteanu, Silvia; Gavrilescu, Maria

    2016-11-01

    For an adequate planning of waste management systems the accurate forecast of waste generation is an essential step, since various factors can affect waste trends. The application of predictive and prognosis models are useful tools, as reliable support for decision making processes. In this paper some indicators such as: number of residents, population age, urban life expectancy, total municipal solid waste were used as input variables in prognostic models in order to predict the amount of solid waste fractions. We applied Waste Prognostic Tool, regression analysis and time series analysis to forecast municipal solid waste generation and composition by considering the Iasi Romania case study. Regression equations were determined for six solid waste fractions (paper, plastic, metal, glass, biodegradable and other waste). Accuracy Measures were calculated and the results showed that S-curve trend model is the most suitable for municipal solid waste (MSW) prediction. Copyright © 2016 Elsevier Ltd. All rights reserved.

  18. Reliability Oriented Design Tool For the New Generation of Grid Connected PV-Inverters

    Sintamarean, Nicolae Cristian; Blaabjerg, Frede; Wang, Huai

    2015-01-01

    is achieved and is further used as an input to the lifetime model. The proposed reliability-oriented design tool is used to study the impact of mission profile (MP) variation and device degradation (aging) in the PV inverter lifetime. The obtained results indicate that the MP of the field where the PV...... inverter is operating has an important impact (up to 70%) on the converter lifetime expectation, and it should be considered in the design stage to better optimize the converter design margin. In order to have correct lifetime estimation, it is crucial to consider also the device degradation feedback (in......This paper introduces a reliability-oriented design tool for a new generation of grid-connected photovoltaic (PV) inverters. The proposed design tool consists of a real field mission profile (RFMP) model (for two operating regions: USA and Denmark), a PV panel model, a grid-connected PV inverter...

  19. Analysis of the Requirements Generation Process for the Logistics Analysis and Wargame Support Tool

    2017-06-01

    impact everything from strategic logistic operations down to the energy demands at the company level. It also looks at the force structure of the...this requirement. 34. The system shall determine the efficiency of the logistics network with respect to an estimated cost of fuel used to deliver...REQUIREMENTS GENERATION PROCESS FOR THE LOGISTICS ANALYSIS AND WARGAME SUPPORT TOOL by Jonathan M. Swan June 2017 Thesis Advisor

  20. Computer Tool for Automatically Generated 3D Illustration in Real Time from Archaeological Scanned Pieces

    Luis López; Germán Arroyo; Domingo Martín

    2012-01-01

    The graphical documentation process of archaeological pieces requires the active involvement of a professional artist to recreate beautiful illustrations using a wide variety of expressive techniques. Frequently, the artist’s work is limited by the inconvenience of working only with the photographs of the pieces he is going to illustrate. This paper presents a software tool that allows the easy generation of illustrations in real time from 3D scanned models. The developed interface allows the...

  1. Method and Tool for Design Process Navigation and Automatic Generation of Simulation Models for Manufacturing Systems

    Nakano, Masaru; Kubota, Fumiko; Inamori, Yutaka; Mitsuyuki, Keiji

    Manufacturing system designers should concentrate on designing and planning manufacturing systems instead of spending their efforts on creating the simulation models to verify the design. This paper proposes a method and its tool to navigate the designers through the engineering process and generate the simulation model automatically from the design results. The design agent also supports collaborative design projects among different companies or divisions with distributed engineering and distributed simulation techniques. The idea was implemented and applied to a factory planning process.

  2. AQME: A forensic mitochondrial DNA analysis tool for next-generation sequencing data.

    Sturk-Andreaggi, Kimberly; Peck, Michelle A; Boysen, Cecilie; Dekker, Patrick; McMahon, Timothy P; Marshall, Charla K

    2017-11-01

    The feasibility of generating mitochondrial DNA (mtDNA) data has expanded considerably with the advent of next-generation sequencing (NGS), specifically in the generation of entire mtDNA genome (mitogenome) sequences. However, the analysis of these data has emerged as the greatest challenge to implementation in forensics. To address this need, a custom toolkit for use in the CLC Genomics Workbench (QIAGEN, Hilden, Germany) was developed through a collaborative effort between the Armed Forces Medical Examiner System - Armed Forces DNA Identification Laboratory (AFMES-AFDIL) and QIAGEN Bioinformatics. The AFDIL-QIAGEN mtDNA Expert, or AQME, generates an editable mtDNA profile that employs forensic conventions and includes the interpretation range required for mtDNA data reporting. AQME also integrates an mtDNA haplogroup estimate into the analysis workflow, which provides the analyst with phylogenetic nomenclature guidance and a profile quality check without the use of an external tool. Supplemental AQME outputs such as nucleotide-per-position metrics, configurable export files, and an audit trail are produced to assist the analyst during review. AQME is applied to standard CLC outputs and thus can be incorporated into any mtDNA bioinformatics pipeline within CLC regardless of sample type, library preparation or NGS platform. An evaluation of AQME was performed to demonstrate its functionality and reliability for the analysis of mitogenome NGS data. The study analyzed Illumina mitogenome data from 21 samples (including associated controls) of varying quality and sample preparations with the AQME toolkit. A total of 211 tool edits were automatically applied to 130 of the 698 total variants reported in an effort to adhere to forensic nomenclature. Although additional manual edits were required for three samples, supplemental tools such as mtDNA haplogroup estimation assisted in identifying and guiding these necessary modifications to the AQME-generated profile. Along

  3. Developing next-generation telehealth tools and technologies: patients, systems, and data perspectives.

    Ackerman, Michael J; Filart, Rosemarie; Burgess, Lawrence P; Lee, Insup; Poropatich, Ronald K

    2010-01-01

    The major goals of telemedicine today are to develop next-generation telehealth tools and technologies to enhance healthcare delivery to medically underserved populations using telecommunication technology, to increase access to medical specialty services while decreasing healthcare costs, and to provide training of healthcare providers, clinical trainees, and students in health-related fields. Key drivers for these tools and technologies are the need and interest to collaborate among telehealth stakeholders, including patients, patient communities, research funders, researchers, healthcare services providers, professional societies, industry, healthcare management/economists, and healthcare policy makers. In the development, marketing, adoption, and implementation of these tools and technologies, communication, training, cultural sensitivity, and end-user customization are critical pieces to the process. Next-generation tools and technologies are vehicles toward personalized medicine, extending the telemedicine model to include cell phones and Internet-based telecommunications tools for remote and home health management with video assessment, remote bedside monitoring, and patient-specific care tools with event logs, patient electronic profile, and physician note-writing capability. Telehealth is ultimately a system of systems in scale and complexity. To cover the full spectrum of dynamic and evolving needs of end-users, we must appreciate system complexity as telehealth moves toward increasing functionality, integration, interoperability, outreach, and quality of service. Toward that end, our group addressed three overarching questions: (1) What are the high-impact topics? (2) What are the barriers to progress? and (3) What roles can the National Institutes of Health and its various institutes and centers play in fostering the future development of telehealth?

  4. Polyhedral meshing as an innovative approach to computational domain discretization of a cyclone in a fluidized bed CLC unit

    Sosnowski Marcin

    2017-01-01

    Full Text Available Chemical Looping Combustion (CLC is a technology that allows the separation of CO2, which is generated by the combustion of fossil fuels. The majority of process designs currently under investigation are systems of coupled fluidized beds. Advances in the development of power generation system using CLC cannot be introduced without using numerical modelling as a research tool. The primary and critical activity in numerical modelling is the computational domain discretization. It influences the numerical diffusion as well as convergence of the model and therefore the overall accuracy of the obtained results. Hence an innovative approach of computational domain discretization using polyhedral (POLY mesh is proposed in the paper. This method reduces both the numerical diffusion of the mesh as well as the time cost of preparing the model for subsequent calculation. The major advantage of POLY mesh is that each individual cell has many neighbours, so gradients can be much better approximated in comparison to commonly-used tetrahedral (TET mesh. POLYs are also less sensitive to stretching than TETs which results in better numerical stability of the model. Therefore detailed comparison of numerical modelling results concerning subsection of CLC system using tetrahedral and polyhedral mesh is covered in the paper.

  5. A Planning Tool for Estimating Waste Generated by a Radiological Incident and Subsequent Decontamination Efforts - 13569

    Boe, Timothy; Lemieux, Paul; Schultheisz, Daniel; Peake, Tom; Hayes, Colin

    2013-01-01

    Management of debris and waste from a wide-area radiological incident would probably constitute a significant percentage of the total remediation cost and effort. The U.S. Environmental Protection Agency's (EPA's) Waste Estimation Support Tool (WEST) is a unique planning tool for estimating the potential volume and radioactivity levels of waste generated by a radiological incident and subsequent decontamination efforts. The WEST was developed to support planners and decision makers by generating a first-order estimate of the quantity and characteristics of waste resulting from a radiological incident. The tool then allows the user to evaluate the impact of various decontamination/demolition strategies on the waste types and volumes generated. WEST consists of a suite of standalone applications and Esri R ArcGIS R scripts for rapidly estimating waste inventories and levels of radioactivity generated from a radiological contamination incident as a function of user-defined decontamination and demolition approaches. WEST accepts Geographic Information System (GIS) shape-files defining contaminated areas and extent of contamination. Building stock information, including square footage, building counts, and building composition estimates are then generated using the Federal Emergency Management Agency's (FEMA's) Hazus R -MH software. WEST then identifies outdoor surfaces based on the application of pattern recognition to overhead aerial imagery. The results from the GIS calculations are then fed into a Microsoft Excel R 2007 spreadsheet with a custom graphical user interface where the user can examine the impact of various decontamination/demolition scenarios on the quantity, characteristics, and residual radioactivity of the resulting waste streams. (authors)

  6. A Planning Tool for Estimating Waste Generated by a Radiological Incident and Subsequent Decontamination Efforts - 13569

    Boe, Timothy [Oak Ridge Institute for Science and Education, Research Triangle Park, NC 27711 (United States); Lemieux, Paul [U.S. Environmental Protection Agency, Research Triangle Park, NC 27711 (United States); Schultheisz, Daniel; Peake, Tom [U.S. Environmental Protection Agency, Washington, DC 20460 (United States); Hayes, Colin [Eastern Research Group, Inc, Morrisville, NC 26560 (United States)

    2013-07-01

    Management of debris and waste from a wide-area radiological incident would probably constitute a significant percentage of the total remediation cost and effort. The U.S. Environmental Protection Agency's (EPA's) Waste Estimation Support Tool (WEST) is a unique planning tool for estimating the potential volume and radioactivity levels of waste generated by a radiological incident and subsequent decontamination efforts. The WEST was developed to support planners and decision makers by generating a first-order estimate of the quantity and characteristics of waste resulting from a radiological incident. The tool then allows the user to evaluate the impact of various decontamination/demolition strategies on the waste types and volumes generated. WEST consists of a suite of standalone applications and Esri{sup R} ArcGIS{sup R} scripts for rapidly estimating waste inventories and levels of radioactivity generated from a radiological contamination incident as a function of user-defined decontamination and demolition approaches. WEST accepts Geographic Information System (GIS) shape-files defining contaminated areas and extent of contamination. Building stock information, including square footage, building counts, and building composition estimates are then generated using the Federal Emergency Management Agency's (FEMA's) Hazus{sup R}-MH software. WEST then identifies outdoor surfaces based on the application of pattern recognition to overhead aerial imagery. The results from the GIS calculations are then fed into a Microsoft Excel{sup R} 2007 spreadsheet with a custom graphical user interface where the user can examine the impact of various decontamination/demolition scenarios on the quantity, characteristics, and residual radioactivity of the resulting waste streams. (authors)

  7. Automatic generation of bioinformatics tools for predicting protein-ligand binding sites.

    Komiyama, Yusuke; Banno, Masaki; Ueki, Kokoro; Saad, Gul; Shimizu, Kentaro

    2016-03-15

    Predictive tools that model protein-ligand binding on demand are needed to promote ligand research in an innovative drug-design environment. However, it takes considerable time and effort to develop predictive tools that can be applied to individual ligands. An automated production pipeline that can rapidly and efficiently develop user-friendly protein-ligand binding predictive tools would be useful. We developed a system for automatically generating protein-ligand binding predictions. Implementation of this system in a pipeline of Semantic Web technique-based web tools will allow users to specify a ligand and receive the tool within 0.5-1 day. We demonstrated high prediction accuracy for three machine learning algorithms and eight ligands. The source code and web application are freely available for download at http://utprot.net They are implemented in Python and supported on Linux. shimizu@bi.a.u-tokyo.ac.jp Supplementary data are available at Bioinformatics online. © The Author 2015. Published by Oxford University Press.

  8. System-level tools and reconfigurable computing for next-generation HWIL systems

    Stark, Derek; McAulay, Derek; Cantle, Allan J.; Devlin, Malachy

    2001-08-01

    Previous work has been presented on the creation of computing architectures called DIME, which addressed the particular computing demands of hardware in the loop systems. These demands include low latency, high data rates and interfacing. While it is essential to have a capable platform for handling and processing of the data streams, the tools must also complement this so that a system's engineer is able to construct their final system. The paper will present the work in the area of integration of system level design tools, such as MATLAB and SIMULINK, with a reconfigurable computing platform. This will demonstrate how algorithms can be implemented and simulated in a familiar rapid application development environment before they are automatically transposed for downloading directly to the computing platform. This complements the established control tools, which handle the configuration and control of the processing systems leading to a tool suite for system development and implementation. As the development tools have evolved the core-processing platform has also been enhanced. These improved platforms are based on dynamically reconfigurable computing, utilizing FPGA technologies, and parallel processing methods that more than double the performance and data bandwidth capabilities. This offers support for the processing of images in Infrared Scene Projectors with 1024 X 1024 resolutions at 400 Hz frame rates. The processing elements will be using the latest generation of FPGAs, which implies that the presented systems will be rated in terms of Tera (1012) operations per second.

  9. Design and Implementation of a Cloud Computing Adoption Decision Tool: Generating a Cloud Road

    Bildosola, Iñaki; Río-Belver, Rosa; Cilleruelo, Ernesto; Garechana, Gaizka

    2015-01-01

    Migrating to cloud computing is one of the current enterprise challenges. This technology provides a new paradigm based on “on-demand payment” for information and communication technologies. In this sense, the small and medium enterprise is supposed to be the most interested, since initial investments are avoided and the technology allows gradual implementation. However, even if the characteristics and capacities have been widely discussed, entry into the cloud is still lacking in terms of practical, real frameworks. This paper aims at filling this gap, presenting a real tool already implemented and tested, which can be used as a cloud computing adoption decision tool. This tool uses diagnosis based on specific questions to gather the required information and subsequently provide the user with valuable information to deploy the business within the cloud, specifically in the form of Software as a Service (SaaS) solutions. This information allows the decision makers to generate their particular Cloud Road. A pilot study has been carried out with enterprises at a local level with a two-fold objective: to ascertain the degree of knowledge on cloud computing and to identify the most interesting business areas and their related tools for this technology. As expected, the results show high interest and low knowledge on this subject and the tool presented aims to readdress this mismatch, insofar as possible. PMID:26230400

  10. Design and Implementation of a Cloud Computing Adoption Decision Tool: Generating a Cloud Road.

    Iñaki Bildosola

    Full Text Available Migrating to cloud computing is one of the current enterprise challenges. This technology provides a new paradigm based on "on-demand payment" for information and communication technologies. In this sense, the small and medium enterprise is supposed to be the most interested, since initial investments are avoided and the technology allows gradual implementation. However, even if the characteristics and capacities have been widely discussed, entry into the cloud is still lacking in terms of practical, real frameworks. This paper aims at filling this gap, presenting a real tool already implemented and tested, which can be used as a cloud computing adoption decision tool. This tool uses diagnosis based on specific questions to gather the required information and subsequently provide the user with valuable information to deploy the business within the cloud, specifically in the form of Software as a Service (SaaS solutions. This information allows the decision makers to generate their particular Cloud Road. A pilot study has been carried out with enterprises at a local level with a two-fold objective: to ascertain the degree of knowledge on cloud computing and to identify the most interesting business areas and their related tools for this technology. As expected, the results show high interest and low knowledge on this subject and the tool presented aims to readdress this mismatch, insofar as possible.

  11. Design and Implementation of a Cloud Computing Adoption Decision Tool: Generating a Cloud Road.

    Bildosola, Iñaki; Río-Belver, Rosa; Cilleruelo, Ernesto; Garechana, Gaizka

    2015-01-01

    Migrating to cloud computing is one of the current enterprise challenges. This technology provides a new paradigm based on "on-demand payment" for information and communication technologies. In this sense, the small and medium enterprise is supposed to be the most interested, since initial investments are avoided and the technology allows gradual implementation. However, even if the characteristics and capacities have been widely discussed, entry into the cloud is still lacking in terms of practical, real frameworks. This paper aims at filling this gap, presenting a real tool already implemented and tested, which can be used as a cloud computing adoption decision tool. This tool uses diagnosis based on specific questions to gather the required information and subsequently provide the user with valuable information to deploy the business within the cloud, specifically in the form of Software as a Service (SaaS) solutions. This information allows the decision makers to generate their particular Cloud Road. A pilot study has been carried out with enterprises at a local level with a two-fold objective: to ascertain the degree of knowledge on cloud computing and to identify the most interesting business areas and their related tools for this technology. As expected, the results show high interest and low knowledge on this subject and the tool presented aims to readdress this mismatch, insofar as possible.

  12. To mesh or not to mesh: a review of pelvic organ reconstructive surgery

    Dällenbach, Patrick

    2015-01-01

    Pelvic organ prolapse (POP) is a major health issue with a lifetime risk of undergoing at least one surgical intervention estimated at close to 10%. In the 1990s, the risk of reoperation after primary standard vaginal procedure was estimated to be as high as 30% to 50%. In order to reduce the risk of relapse, gynecological surgeons started to use mesh implants in pelvic organ reconstructive surgery with the emergence of new complications. Recent studies have nevertheless shown that the risk of POP recurrence requiring reoperation is lower than previously estimated, being closer to 10% rather than 30%. The development of mesh surgery – actively promoted by the marketing industry – was tremendous during the past decade, and preceded any studies supporting its benefit for our patients. Randomized trials comparing the use of mesh to native tissue repair in POP surgery have now shown better anatomical but similar functional outcomes, and meshes are associated with more complications, in particular for transvaginal mesh implants. POP is not a life-threatening condition, but a functional problem that impairs quality of life for women. The old adage “primum non nocere” is particularly appropriate when dealing with this condition which requires no treatment when asymptomatic. It is currently admitted that a certain degree of POP is physiological with aging when situated above the landmark of the hymen. Treatment should be individualized and the use of mesh needs to be selective and appropriate. Mesh implants are probably an important tool in pelvic reconstructive surgery, but the ideal implant has yet to be found. The indications for its use still require caution and discernment. This review explores the reasons behind the introduction of mesh augmentation in POP surgery, and aims to clarify the risks, benefits, and the recognized indications for its use. PMID:25848324

  13. A generative tool for building health applications driven by ISO 13606 archetypes.

    Menárguez-Tortosa, Marcos; Martínez-Costa, Catalina; Fernández-Breis, Jesualdo Tomás

    2012-10-01

    The use of Electronic Healthcare Records (EHR) standards in the development of healthcare applications is crucial for achieving the semantic interoperability of clinical information. Advanced EHR standards make use of the dual model architecture, which provides a solution for clinical interoperability based on the separation of the information and knowledge. However, the impact of such standards is biased by the limited availability of tools that facilitate their usage and practical implementation. In this paper, we present an approach for the automatic generation of clinical applications for the ISO 13606 EHR standard, which is based on the dual model architecture. This generator has been generically designed, so it can be easily adapted to other dual model standards and can generate applications for multiple technological platforms. Such good properties are based on the combination of standards for the representation of generic user interfaces and model-driven engineering techniques.

  14. A Reconfigurable Mesh-Ring Topology for Bluetooth Sensor Networks

    Ben-Yi Wang

    2018-05-01

    Full Text Available In this paper, a Reconfigurable Mesh-Ring (RMR algorithm is proposed for Bluetooth sensor networks. The algorithm is designed in three stages to determine the optimal configuration of the mesh-ring network. Firstly, a designated root advertises and discovers its neighboring nodes. Secondly, a scatternet criterion is built to compute the minimum number of piconets and distributes the connection information for piconet and scatternet. Finally, a peak-search method is designed to determine the optimal mesh-ring configuration for various sizes of networks. To maximize the network capacity, the research problem is formulated by determining the best connectivity of available mesh links. During the formation and maintenance phases, three possible configurations (including piconet, scatternet, and hybrid are examined to determine the optimal placement of mesh links. The peak-search method is a systematic approach, and is implemented by three functional blocks: the topology formation block generates the mesh-ring topology, the routing efficiency block computes the routing performance, and the optimum decision block introduces a decision-making criterion to determine the optimum number of mesh links. Simulation results demonstrate that the optimal mesh-ring configuration can be determined and that the scatternet case achieves better overall performance than the other two configurations. The RMR topology also outperforms the conventional ring-based and cluster-based mesh methods in terms of throughput performance for Bluetooth configurable networks.

  15. Combining spray nozzle simulators with meshes: characterization of rainfall intensity and drop properties

    Carvalho, Sílvia C. P.; de Lima, João L. M. P.; de Lima, M. Isabel P.

    2013-04-01

    Rainfall simulators can be a powerful tool to increase our understanding of hydrological and geomorphological processes. Nevertheless, rainfall simulators' design and operation might be rather demanding, for achieving specific rainfall intensity distributions and drop characteristics. The pressurized simulators have some advantages over the non-pressurized simulators: drops do not rely on gravity to reach terminal velocity, but are sprayed out under pressure; pressurized simulators also yield a broad range of drop sizes in comparison with drop-formers simulators. The main purpose of this study was to explore in the laboratory the potential of combining spray nozzle simulators with meshes in order to change rainfall characteristics (rainfall intensity and diameters and fall speed of drops). Different types of spray nozzles were tested, such as single full-cone and multiple full-cone nozzles. The impact of the meshes on the simulated rain was studied by testing different materials (i.e. plastic and steel meshes), square apertures and wire thicknesses, and different vertical distances between the nozzle and the meshes underneath. The diameter and fall speed of the rain drops were measured using a Laser Precipitation Monitor (Thies Clima). The rainfall intensity range and coefficients of uniformity of the sprays and the drop size distribution, fall speed and kinetic energy were analysed. Results show that when meshes intercept drop trajectories the spatial distribution of rainfall intensity and the drop size distribution are affected. As the spray nozzles generate typically small drop sizes and narrow drop size distributions, meshes can be used to promote the formation of bigger drops and random their landing positions.

  16. Mathematics and computational methods development in U.S. department of energy-sponsored research (nuclear energy research initiative and nuclear engineering education research). 4. Development of an Expert System for Generation of an Effective Mesh Distribution for the SN Method

    Patchimpattapong, Apisit; Haghighat, Alireza

    2001-01-01

    The discrete ordinates (S N ) method is widely used to obtain numerical solutions of the transport equation. The method calls for discretization of spatial, energy, and angular variables. To generate an 'effective' spatial mesh distribution, one has to consider various factors including particle mean free path (mfp), material and source discontinuities, and problem objectives. This becomes more complicated if we consider the effect of numerics such as differencing schemes, parallel processing strategies, and computation resources. As a result, one may often over/under-mesh depending upon limitations on accuracy, computing resources, and time allotted. To overcome the foregoing issues, we are developing an expert system for input preparation of the discrete ordinates (S N ) method. This project is a part of an ongoing project sponsored by Nuclear Engineering Education Research. Our expert system consists of two parts: (a) an algorithm for generation of a mesh distribution for a serial calculation and (b) an algorithm for extension to parallel computing, which accounts for parallelization parameters including granularity, load balancing, parallel algorithms, and possible architectural issues. Thus far, we have developed a stand-alone algorithm for generation of an 'effective' mesh distribution for a serial calculation. The algorithm has been successfully tested with the Parallel Environment Neutral-Particle Transport (PENTRAN) code system. In this paper, we discuss the structure of our algorithm and present its use for simulating the VENUS-3 experimental facility. To date, we have developed and tested part 1 of this system. This part comprises of four steps: creation of a geometric model and coarse meshes, calculation of un-collided flux, selection of differencing schemes, and generation of fine-mesh distribution. For the un-collided flux calculation, we have developed a parallel code called PENFC. It is capable of calculating un-collided and first-collision fluxes

  17. An automated approach for solution based mesh adaptation to enhance numerical accuracy for a given number of grid cells

    Lucas, P.; Van Zuijlen, A.H.; Bijl, H.

    2009-01-01

    Mesh adaptation is a fairly established tool to obtain numerically accurate solutions for flow problems. Computational efficiency is, however, not always guaranteed for the adaptation strategies found in literature. Typically excessive mesh growth diminishes the potential efficiency gain. This

  18. Grid generation methods

    Liseikin, Vladimir D

    2017-01-01

    This new edition provides a description of current developments relating to grid methods, grid codes, and their applications to actual problems. Grid generation methods are indispensable for the numerical solution of differential equations. Adaptive grid-mapping techniques, in particular, are the main focus and represent a promising tool to deal with systems with singularities. This 3rd edition includes three new chapters on numerical implementations (10), control of grid properties (11), and applications to mechanical, fluid, and plasma related problems (13). Also the other chapters have been updated including new topics, such as curvatures of discrete surfaces (3). Concise descriptions of hybrid mesh generation, drag and sweeping methods, parallel algorithms for mesh generation have been included too. This new edition addresses a broad range of readers: students, researchers, and practitioners in applied mathematics, mechanics, engineering, physics and other areas of applications.

  19. Conforming to interface structured adaptive mesh refinement: 3D algorithm and implementation

    Nagarajan, Anand; Soghrati, Soheil

    2018-03-01

    A new non-iterative mesh generation algorithm named conforming to interface structured adaptive mesh refinement (CISAMR) is introduced for creating 3D finite element models of problems with complex geometries. CISAMR transforms a structured mesh composed of tetrahedral elements into a conforming mesh with low element aspect ratios. The construction of the mesh begins with the structured adaptive mesh refinement of elements in the vicinity of material interfaces. An r-adaptivity algorithm is then employed to relocate selected nodes of nonconforming elements, followed by face-swapping a small fraction of them to eliminate tetrahedrons with high aspect ratios. The final conforming mesh is constructed by sub-tetrahedralizing remaining nonconforming elements, as well as tetrahedrons with hanging nodes. In addition to studying the convergence and analyzing element-wise errors in meshes generated using CISAMR, several example problems are presented to show the ability of this method for modeling 3D problems with intricate morphologies.

  20. Energy mesh optimization for multi-level calculation schemes

    Mosca, P.; Taofiki, A.; Bellier, P.; Prevost, A.

    2011-01-01

    The industrial calculations of third generation nuclear reactors are based on sophisticated strategies of homogenization and collapsing at different spatial and energetic levels. An important issue to ensure the quality of these calculation models is the choice of the collapsing energy mesh. In this work, we show a new approach to generate optimized energy meshes starting from the SHEM 281-group library. The optimization model is applied on 1D cylindrical cells and consists of finding an energy mesh which minimizes the errors between two successive collision probability calculations. The former is realized over the fine SHEM mesh with Livolant-Jeanpierre self-shielded cross sections and the latter is performed with collapsed cross sections over the energy mesh being optimized. The optimization is done by the particle swarm algorithm implemented in the code AEMC and multigroup flux solutions are obtained from standard APOLLO2 solvers. By this new approach, a set of new optimized meshes which encompass from 10 to 50 groups has been defined for PWR and BWR calculations. This set will allow users to adapt the energy detail of the solution to the complexity of the calculation (assembly, multi-assembly, two-dimensional whole core). Some preliminary verifications, in which the accuracy of the new meshes is measured compared to a direct 281-group calculation, show that the 30-group optimized mesh offers a good compromise between simulation time and accuracy for a standard 17 x 17 UO 2 assembly with and without control rods. (author)

  1. Wireless mesh networks.

    Wang, Xinheng

    2008-01-01

    Wireless telemedicine using GSM and GPRS technologies can only provide low bandwidth connections, which makes it difficult to transmit images and video. Satellite or 3G wireless transmission provides greater bandwidth, but the running costs are high. Wireless networks (WLANs) appear promising, since they can supply high bandwidth at low cost. However, the WLAN technology has limitations, such as coverage. A new wireless networking technology named the wireless mesh network (WMN) overcomes some of the limitations of the WLAN. A WMN combines the characteristics of both a WLAN and ad hoc networks, thus forming an intelligent, large scale and broadband wireless network. These features are attractive for telemedicine and telecare because of the ability to provide data, voice and video communications over a large area. One successful wireless telemedicine project which uses wireless mesh technology is the Emergency Room Link (ER-LINK) in Tucson, Arizona, USA. There are three key characteristics of a WMN: self-organization, including self-management and self-healing; dynamic changes in network topology; and scalability. What we may now see is a shift from mobile communication and satellite systems for wireless telemedicine to the use of wireless networks based on mesh technology, since the latter are very attractive in terms of cost, reliability and speed.

  2. Isotropic 2D quadrangle meshing with size and orientation control

    Pellenard, Bertrand; Alliez, Pierre; Morvan, Jean-Marie

    2011-01-01

    We propose an approach for automatically generating isotropic 2D quadrangle meshes from arbitrary domains with a fine control over sizing and orientation of the elements. At the heart of our algorithm is an optimization procedure that, from a coarse

  3. Mutation based treatment recommendations from next generation sequencing data: a comparison of web tools.

    Patel, Jaymin M; Knopf, Joshua; Reiner, Eric; Bossuyt, Veerle; Epstein, Lianne; DiGiovanna, Michael; Chung, Gina; Silber, Andrea; Sanft, Tara; Hofstatter, Erin; Mougalian, Sarah; Abu-Khalaf, Maysa; Platt, James; Shi, Weiwei; Gershkovich, Peter; Hatzis, Christos; Pusztai, Lajos

    2016-04-19

    Interpretation of complex cancer genome data, generated by tumor target profiling platforms, is key for the success of personalized cancer therapy. How to draw therapeutic conclusions from tumor profiling results is not standardized and may vary among commercial and academically-affiliated recommendation tools. We performed targeted sequencing of 315 genes from 75 metastatic breast cancer biopsies using the FoundationOne assay. Results were run through 4 different web tools including the Drug-Gene Interaction Database (DGidb), My Cancer Genome (MCG), Personalized Cancer Therapy (PCT), and cBioPortal, for drug and clinical trial recommendations. These recommendations were compared amongst each other and to those provided by FoundationOne. The identification of a gene as targetable varied across the different recommendation sources. Only 33% of cases had 4 or more sources recommend the same drug for at least one of the usually several altered genes found in tumor biopsies. These results indicate further development and standardization of broadly applicable software tools that assist in our therapeutic interpretation of genomic data is needed. Existing algorithms for data acquisition, integration and interpretation will likely need to incorporate artificial intelligence tools to improve both content and real-time status.

  4. ModelMage: a tool for automatic model generation, selection and management.

    Flöttmann, Max; Schaber, Jörg; Hoops, Stephan; Klipp, Edda; Mendes, Pedro

    2008-01-01

    Mathematical modeling of biological systems usually involves implementing, simulating, and discriminating several candidate models that represent alternative hypotheses. Generating and managing these candidate models is a tedious and difficult task and can easily lead to errors. ModelMage is a tool that facilitates management of candidate models. It is designed for the easy and rapid development, generation, simulation, and discrimination of candidate models. The main idea of the program is to automatically create a defined set of model alternatives from a single master model. The user provides only one SBML-model and a set of directives from which the candidate models are created by leaving out species, modifiers or reactions. After generating models the software can automatically fit all these models to the data and provides a ranking for model selection, in case data is available. In contrast to other model generation programs, ModelMage aims at generating only a limited set of models that the user can precisely define. ModelMage uses COPASI as a simulation and optimization engine. Thus, all simulation and optimization features of COPASI are readily incorporated. ModelMage can be downloaded from http://sysbio.molgen.mpg.de/modelmage and is distributed as free software.

  5. NASCENT: an automatic protein interaction network generation tool for non-model organisms.

    Banky, Daniel; Ordog, Rafael; Grolmusz, Vince

    2009-04-24

    Large quantity of reliable protein interaction data are available for model organisms in public depositories (e.g., MINT, DIP, HPRD, INTERACT). Most data correspond to experiments with the proteins of Saccharomyces cerevisiae, Drosophila melanogaster, Homo sapiens, Caenorhabditis elegans, Escherichia coli and Mus musculus. For other important organisms the data availability is poor or non-existent. Here we present NASCENT, a completely automatic web-based tool and also a downloadable Java program, capable of modeling and generating protein interaction networks even for non-model organisms. The tool performs protein interaction network modeling through gene-name mapping, and outputs the resulting network in graphical form and also in computer-readable graph-forms, directly applicable by popular network modeling software. http://nascent.pitgroup.org.

  6. The Creative task Creator: a tool for the generation of customized, Web-based creativity tasks.

    Pretz, Jean E; Link, John A

    2008-11-01

    This article presents a Web-based tool for the creation of divergent-thinking and open-ended creativity tasks. A Java program generates HTML forms with PHP scripting that run an Alternate Uses Task and/or open-ended response items. Researchers may specify their own instructions, objects, and time limits, or use default settings. Participants can also be prompted to select their best responses to the Alternate Uses Task (Silvia et al., 2008). Minimal programming knowledge is required. The program runs on any server, and responses are recorded in a standard MySQL database. Responses can be scored using the consensual assessment technique (Amabile, 1996) or Torrance's (1998) traditional scoring method. Adoption of this Web-based tool should facilitate creativity research across cultures and access to eminent creators. The Creative Task Creator may be downloaded from the Psychonomic Society's Archive of Norms, Stimuli, and Data, www.psychonomic.org/archive.

  7. ODG: Omics database generator - a tool for generating, querying, and analyzing multi-omics comparative databases to facilitate biological understanding.

    Guhlin, Joseph; Silverstein, Kevin A T; Zhou, Peng; Tiffin, Peter; Young, Nevin D

    2017-08-10

    Rapid generation of omics data in recent years have resulted in vast amounts of disconnected datasets without systemic integration and knowledge building, while individual groups have made customized, annotated datasets available on the web with few ways to link them to in-lab datasets. With so many research groups generating their own data, the ability to relate it to the larger genomic and comparative genomic context is becoming increasingly crucial to make full use of the data. The Omics Database Generator (ODG) allows users to create customized databases that utilize published genomics data integrated with experimental data which can be queried using a flexible graph database. When provided with omics and experimental data, ODG will create a comparative, multi-dimensional graph database. ODG can import definitions and annotations from other sources such as InterProScan, the Gene Ontology, ENZYME, UniPathway, and others. This annotation data can be especially useful for studying new or understudied species for which transcripts have only been predicted, and rapidly give additional layers of annotation to predicted genes. In better studied species, ODG can perform syntenic annotation translations or rapidly identify characteristics of a set of genes or nucleotide locations, such as hits from an association study. ODG provides a web-based user-interface for configuring the data import and for querying the database. Queries can also be run from the command-line and the database can be queried directly through programming language hooks available for most languages. ODG supports most common genomic formats as well as generic, easy to use tab-separated value format for user-provided annotations. ODG is a user-friendly database generation and query tool that adapts to the supplied data to produce a comparative genomic database or multi-layered annotation database. ODG provides rapid comparative genomic annotation and is therefore particularly useful for non-model or

  8. Customised 3D Printing: An Innovative Training Tool for the Next Generation of Orbital Surgeons.

    Scawn, Richard L; Foster, Alex; Lee, Bradford W; Kikkawa, Don O; Korn, Bobby S

    2015-01-01

    Additive manufacturing or 3D printing is the process by which three dimensional data fields are translated into real-life physical representations. 3D printers create physical printouts using heated plastics in a layered fashion resulting in a three-dimensional object. We present a technique for creating customised, inexpensive 3D orbit models for use in orbital surgical training using 3D printing technology. These models allow trainee surgeons to perform 'wet-lab' orbital decompressions and simulate upcoming surgeries on orbital models that replicate a patient's bony anatomy. We believe this represents an innovative training tool for the next generation of orbital surgeons.

  9. Synthetic biology in mammalian cells: Next generation research tools and therapeutics

    Lienert, Florian; Lohmueller, Jason J; Garg, Abhishek; Silver, Pamela A

    2014-01-01

    Recent progress in DNA manipulation and gene circuit engineering has greatly improved our ability to programme and probe mammalian cell behaviour. These advances have led to a new generation of synthetic biology research tools and potential therapeutic applications. Programmable DNA-binding domains and RNA regulators are leading to unprecedented control of gene expression and elucidation of gene function. Rebuilding complex biological circuits such as T cell receptor signalling in isolation from their natural context has deepened our understanding of network motifs and signalling pathways. Synthetic biology is also leading to innovative therapeutic interventions based on cell-based therapies, protein drugs, vaccines and gene therapies. PMID:24434884

  10. A web based tool for storing and visualising data generated within a smart home.

    McDonald, H A; Nugent, C D; Moore, G; Finlay, D D; Hallberg, J

    2011-01-01

    There is a growing need to re-assess the current approaches available to researchers for storing and managing heterogeneous data generated within a smart home environment. In our current work we have developed the homeML Application; a web based tool to support researchers engaged in the area of smart home research as they perform experiments. Within this paper the homeML Application is presented which includes the fundamental components of the homeML Repository and the homeML Toolkit. Results from a usability study conducted by 10 computer science researchers are presented; the initial results of which have been positive.

  11. From 4D Medical Images (CT, MRI, and Ultrasound to 4D Structured Mesh Models of the Left Ventricular Endocardium for Patient-Specific Simulations

    Federico Canè

    2018-01-01

    Full Text Available With cardiovascular disease (CVD remaining the primary cause of death worldwide, early detection of CVDs becomes essential. The intracardiac flow is an important component of ventricular function, motion kinetics, wash-out of ventricular chambers, and ventricular energetics. Coupling between Computational Fluid Dynamics (CFD simulations and medical images can play a fundamental role in terms of patient-specific diagnostic tools. From a technical perspective, CFD simulations with moving boundaries could easily lead to negative volumes errors and the sudden failure of the simulation. The generation of high-quality 4D meshes (3D in space + time with 1-to-1 vertex becomes essential to perform a CFD simulation with moving boundaries. In this context, we developed a semiautomatic morphing tool able to create 4D high-quality structured meshes starting from a segmented 4D dataset. To prove the versatility and efficiency, the method was tested on three different 4D datasets (Ultrasound, MRI, and CT by evaluating the quality and accuracy of the resulting 4D meshes. Furthermore, an estimation of some physiological quantities is accomplished for the 4D CT reconstruction. Future research will aim at extending the region of interest, further automation of the meshing algorithm, and generating structured hexahedral mesh models both for the blood and myocardial volume.

  12. To mesh or not to mesh: a review of pelvic organ reconstructive surgery

    Dällenbach P

    2015-04-01

    Full Text Available Patrick Dällenbach Department of Gynecology and Obstetrics, Division of Gynecology, Urogynecology Unit, Geneva University Hospitals, Geneva, Switzerland Abstract: Pelvic organ prolapse (POP is a major health issue with a lifetime risk of undergoing at least one surgical intervention estimated at close to 10%. In the 1990s, the risk of reoperation after primary standard vaginal procedure was estimated to be as high as 30% to 50%. In order to reduce the risk of relapse, gynecological surgeons started to use mesh implants in pelvic organ reconstructive surgery with the emergence of new complications. Recent studies have nevertheless shown that the risk of POP recurrence requiring reoperation is lower than previously estimated, being closer to 10% rather than 30%. The development of mesh surgery – actively promoted by the marketing industry – was tremendous during the past decade, and preceded any studies supporting its benefit for our patients. Randomized trials comparing the use of mesh to native tissue repair in POP surgery have now shown better anatomical but similar functional outcomes, and meshes are associated with more complications, in particular for transvaginal mesh implants. POP is not a life-threatening condition, but a functional problem that impairs quality of life for women. The old adage “primum non nocere” is particularly appropriate when dealing with this condition which requires no treatment when asymptomatic. It is currently admitted that a certain degree of POP is physiological with aging when situated above the landmark of the hymen. Treatment should be individualized and the use of mesh needs to be selective and appropriate. Mesh implants are probably an important tool in pelvic reconstructive surgery, but the ideal implant has yet to be found. The indications for its use still require caution and discernment. This review explores the reasons behind the introduction of mesh augmentation in POP surgery, and aims to

  13. Mesh erosion after abdominal sacrocolpopexy.

    Kohli, N; Walsh, P M; Roat, T W; Karram, M M

    1998-12-01

    To report our experience with erosion of permanent suture or mesh material after abdominal sacrocolpopexy. A retrospective chart review was performed to identify patients who underwent sacrocolpopexy by the same surgeon over 8 years. Demographic data, operative notes, hospital records, and office charts were reviewed after sacrocolpopexy. Patients with erosion of either suture or mesh were treated initially with conservative therapy followed by surgical intervention as required. Fifty-seven patients underwent sacrocolpopexy using synthetic mesh during the study period. The mean (range) postoperative follow-up was 19.9 (1.3-50) months. Seven patients (12%) had erosions after abdominal sacrocolpopexy with two suture erosions and five mesh erosions. Patients with suture erosion were asymptomatic compared with patients with mesh erosion, who presented with vaginal bleeding or discharge. The mean (+/-standard deviation) time to erosion was 14.0+/-7.7 (range 4-24) months. Both patients with suture erosion were treated conservatively with estrogen cream. All five patients with mesh erosion required transvaginal removal of the mesh. Mesh erosion can follow abdominal sacrocolpopexy over a long time, and usually presents as vaginal bleeding or discharge. Although patients with suture erosion can be managed successfully with conservative treatment, patients with mesh erosion require surgical intervention. Transvaginal removal of the mesh with vaginal advancement appears to be an effective treatment in patients failing conservative management.

  14. A Survey of Solver-Related Geometry and Meshing Issues

    Masters, James; Daniel, Derick; Gudenkauf, Jared; Hine, David; Sideroff, Chris

    2016-01-01

    There is a concern in the computational fluid dynamics community that mesh generation is a significant bottleneck in the CFD workflow. This is one of several papers that will help set the stage for a moderated panel discussion addressing this issue. Although certain general "rules of thumb" and a priori mesh metrics can be used to ensure that some base level of mesh quality is achieved, inadequate consideration is often given to the type of solver or particular flow regime on which the mesh will be utilized. This paper explores how an analyst may want to think differently about a mesh based on considerations such as if a flow is compressible vs. incompressible or hypersonic vs. subsonic or if the solver is node-centered vs. cell-centered. This paper is a high-level investigation intended to provide general insight into how considering the nature of the solver or flow when performing mesh generation has the potential to increase the accuracy and/or robustness of the solution and drive the mesh generation process to a state where it is no longer a hindrance to the analysis process.

  15. Spot markets vs. long-term contracts - modelling tools for regional electricity generating utilities

    Grohnheit, P.E.

    1999-01-01

    A properly organised market for electricity requires that some information will be available for all market participants. Also a range of generally available modelling tools are necessary. This paper describes a set of simple models based on published data for analyses of the long-term revenues of regional utilities with combined heat and power generation (CHP), who will operate a competitive international electricity market and a local heat market. The future revenues from trade on the spot market is analysed using a load curve model, in which marginal costs are calculated on the basis of short-term costs of the available units and chronological hourly variations in the demands for electricity and heat. Assumptions on prices, marginal costs and electricity generation by the different types of generating units are studied for selected types of local electricity generators. The long-term revenue requirements to be met by long-term contracts are analysed using a traditional techno-economic optimisation model focusing on technology choice and competition among technologies over 20.30 years. A possible conclusion from this discussion is that it is important for the economic and environmental efficiency of the electricity market that local or regional generators of CHP, who are able to react on price signals, do not conclude long-term contracts that include fixed time-of-day tariff for sale of electricity. Optimisation results for a CHP region (represented by the structure of the Danish electricity and CHP market in 1995) also indicates that a market for CO 2 tradable permits is unlikely to attract major non-fossil fuel technologies for electricity generation, e.g. wind power. (au)

  16. SPRINT: A Tool to Generate Concurrent Transaction-Level Models from Sequential Code

    Richard Stahl

    2007-01-01

    Full Text Available A high-level concurrent model such as a SystemC transaction-level model can provide early feedback during the exploration of implementation alternatives for state-of-the-art signal processing applications like video codecs on a multiprocessor platform. However, the creation of such a model starting from sequential code is a time-consuming and error-prone task. It is typically done only once, if at all, for a given design. This lack of exploration of the design space often leads to a suboptimal implementation. To support our systematic C-based design flow, we have developed a tool to generate a concurrent SystemC transaction-level model for user-selected task boundaries. Using this tool, different parallelization alternatives have been evaluated during the design of an MPEG-4 simple profile encoder and an embedded zero-tree coder. Generation plus evaluation of an alternative was possible in less than six minutes. This is fast enough to allow extensive exploration of the design space.

  17. Difference frequency generation spectroscopy as a vibrational optical activity measurement tool.

    Cheon, Sangheon; Cho, Minhaeng

    2009-03-19

    Vibrational optical activity (VOA) of chiral molecules in condensed phases can be studied by using vibrational circular dichroism and Raman optical activity measurement techniques. Recently, IR-vis sum frequency generation has shown to be an alternative VOA measurement method. Such a three-wave-mixing method employing a polarization modulation technique can be a potentially useful VOA measurement tool. Here, a theoretical description of difference frequency generation (DFG) employing circularly polarized visible radiations is presented. Frequency scanning to obtain a VOA-DFG spectrum is achieved by controlling the difference between the two electronically nonresonant incident radiation frequencies. If the two incident beams are linearly polarized and their polarization directions are perpendicular to each other, one can selectively measure the all-electric-dipole-allowed chiral component of the DFG susceptibility. In addition, by using circularly polarized beams and taking the DFG difference intensity signal, which is defined as the difference between left and right circularly polarized DFG signals, additional chiral susceptibility components originating from the electric quadrupole transition can be measured. The DFG as a novel VOA measurement technique for solution samples containing chiral molecules will therefore be a useful coherent spectroscopic tool for determining absolute configuration of chiral molecules in condensed phases.

  18. Polyhedral meshing in numerical analysis of conjugate heat transfer

    Sosnowski, Marcin; Krzywanski, Jaroslaw; Grabowska, Karolina; Gnatowska, Renata

    2018-06-01

    Computational methods have been widely applied in conjugate heat transfer analysis. The very first and crucial step in such research is the meshing process which consists in dividing the analysed geometry into numerous small control volumes (cells). In Computational Fluid Dynamics (CFD) applications it is desirable to use the hexahedral cells as the resulting mesh is characterized by low numerical diffusion. Unfortunately generating such mesh can be a very time-consuming task and in case of complicated geometry - it may not be possible to generate cells of good quality. Therefore tetrahedral cells have been implemented into commercial pre-processors. Their advantage is the ease of its generation even in case of very complex geometry. On the other hand tetrahedrons cannot be stretched excessively without decreasing the mesh quality factor, so significantly larger number of cells has to be used in comparison to hexahedral mesh in order to achieve a reasonable accuracy. Moreover the numerical diffusion of tetrahedral elements is significantly higher. Therefore the polyhedral cells are proposed within the paper in order to combine the advantages of hexahedrons (low numerical diffusion resulting in accurate solution) and tetrahedrons (rapid semi-automatic generation) as well as to overcome the disadvantages of both the above mentioned mesh types. The major benefit of polyhedral mesh is that each individual cell has many neighbours, so gradients can be well approximated. Polyhedrons are also less sensitive to stretching than tetrahedrons which results in better mesh quality leading to improved numerical stability of the model. In addition, numerical diffusion is reduced due to mass exchange over numerous faces. This leads to a more accurate solution achieved with a lower cell count. Therefore detailed comparison of numerical modelling results concerning conjugate heat transfer using tetrahedral and polyhedral meshes is presented in the paper.

  19. THEORY OF GENERATIONS AS A TOOL FOR ANALYSIS, FORMATION AND DEVELOPMENT OF LABOUR POTENTIAL

    I. M. Gurova

    2016-01-01

    Full Text Available The modern Strauss–Howe generational theory created at the intersection of economic, sociological, historical and psychological Sciences. She focuses primarily on difference of attitudes of generations, due to the specific social environment, corresponding to a certain period of time. Such a perspective on issues related to the human factor in the economy, has recently attracted interest not only from researchers, but also finds application for solving practical problems in some fields of business.Subject / theme. The article is devoted to one of the urgent contemporary socio-economic problems – the issue of the formation and development of labor potential. In this context, the theory of generations is offered by the authors as a tool for the study of qualitative parameters of human resources and the planning of future work opportunities in our country. In particular, the article considers the main aspects of the classical version of the Strauss–Howe generational theory and its Russian adaptation. Statistics describing the General demographics and working population of Russia from the point of view of generational groups. On this basis, a forecast is made of the labor potential of the country in the long term structure, review and compare the basic values and business characteristics of its constituent generations. Problem areas for which use of generational approach is rational are revealed and the corresponding recommendations are made.Objectives. The purpose of this article is justification of the possible application of provisions of modern theories of generations to identify and predict the dynamics of qualitative characteristics of the domestic workforce, as well as the prospects of its use in order to control the formation and development of labor potential.Methods. Methodological basis of the presented work make comparative and economic-statistical and socio-cultural methods of analysis.Results. In the framework of this article

  20. Investigation of next-generation sequencing data of Klebsiella pneumoniae using web-based tools.

    Brhelova, Eva; Antonova, Mariya; Pardy, Filip; Kocmanova, Iva; Mayer, Jiri; Racil, Zdenek; Lengerova, Martina

    2017-11-01

    Rapid identification and characterization of multidrug-resistant Klebsiella pneumoniae strains is necessary due to the increasing frequency of severe infections in patients. The decreasing cost of next-generation sequencing enables us to obtain a comprehensive overview of genetic information in one step. The aim of this study is to demonstrate and evaluate the utility and scope of the application of web-based databases to next-generation sequenced (NGS) data. The whole genomes of 11 clinical Klebsiella pneumoniae isolates were sequenced using Illumina MiSeq. Selected web-based tools were used to identify a variety of genetic characteristics, such as acquired antimicrobial resistance genes, multilocus sequence types, plasmid replicons, and identify virulence factors, such as virulence genes, cps clusters, urease-nickel clusters and efflux systems. Using web-based tools hosted by the Center for Genomic Epidemiology, we detected resistance to 8 main antimicrobial groups with at least 11 acquired resistance genes. The isolates were divided into eight sequence types (ST11, 23, 37, 323, 433, 495 and 562, and a new one, ST1646). All of the isolates carried replicons of large plasmids. Capsular types, virulence factors and genes coding AcrAB and OqxAB efflux pumps were detected using BIGSdb-Kp, whereas the selected virulence genes, identified in almost all of the isolates, were detected using CLC Genomic Workbench software. Applying appropriate web-based online tools to NGS data enables the rapid extraction of comprehensive information that can be used for more efficient diagnosis and treatment of patients, while data processing is free of charge, easy and time-efficient.

  1. Streaming simplification of tetrahedral meshes.

    Vo, Huy T; Callahan, Steven P; Lindstrom, Peter; Pascucci, Valerio; Silva, Cláudio T

    2007-01-01

    Unstructured tetrahedral meshes are commonly used in scientific computing to represent scalar, vector, and tensor fields in three dimensions. Visualization of these meshes can be difficult to perform interactively due to their size and complexity. By reducing the size of the data, we can accomplish real-time visualization necessary for scientific analysis. We propose a two-step approach for streaming simplification of large tetrahedral meshes. Our algorithm arranges the data on disk in a streaming, I/O-efficient format that allows coherent access to the tetrahedral cells. A quadric-based simplification is sequentially performed on small portions of the mesh in-core. Our output is a coherent streaming mesh which facilitates future processing. Our technique is fast, produces high quality approximations, and operates out-of-core to process meshes too large for main memory.

  2. The THIRST chemistry module as a tool to determine optimal steam generator corrosion control strategies

    Heppner, K.; Laroche, S.; Pietralik, J.

    2006-01-01

    As part of a coordinated program, AECL is developing a set of tools to aid with the prediction and management of steam generator performance. Although stress corrosion cracking (of Alloy 800) has not been detected in any operating steam generator, for life management it is necessary to develop mechanistic models to predict the conditions under which stress corrosion cracking is plausible. Therefore, constant extension rate tests were carried out for Alloy 800 under various steam generator crevice chemistry conditions at applied potentials. These tests were designed to evaluate the stress corrosion cracking susceptibility of Alloy 800 under CANDU( steam generator operating conditions. Based on the experimental results, the recommended electrochemical corrosion potential/pH zone for Alloy 800 determined by electrochemical polarization measurements was verified with the respect of stress corrosion cracking susceptibility. The effects of lead contamination on the stress corrosion cracking susceptibility of Alloy 800 tubing were also evaluated. The experimental results from constant extension rate tests obtained under applied potentials suggest that Alloy 800 has good performance inside much of a previously recommended electrochemical corrosion potential/pH zone determined by electrochemical analysis. Alloy 800 is not susceptible to stress corrosion cracking under normal CANDU steam generator operating conditions. However, Alloy 800 may be susceptible to stress corrosion cracking under near-neutral crevice chemistry conditions in the presence of oxidants. In addition, stress corrosion cracking susceptibility is increased by lead contamination. This observation suggests that the previously defined electrochemical corrosion potential limit under near-neutral crevice conditions could be modified to minimize stress corrosion cracking of Alloy 800. The test results from this work also suggest that the pH dependency of the stress corrosion cracking susceptibility of Alloy 800

  3. A Monte Carlo based decision-support tool for assessing generation portfolios in future carbon constrained electricity industries

    Vithayasrichareon, Peerapat; MacGill, Iain F.

    2012-01-01

    This paper presents a novel decision-support tool for assessing future generation portfolios in an increasingly uncertain electricity industry. The tool combines optimal generation mix concepts with Monte Carlo simulation and portfolio analysis techniques to determine expected overall industry costs, associated cost uncertainty, and expected CO 2 emissions for different generation portfolio mixes. The tool can incorporate complex and correlated probability distributions for estimated future fossil-fuel costs, carbon prices, plant investment costs, and demand, including price elasticity impacts. The intent of this tool is to facilitate risk-weighted generation investment and associated policy decision-making given uncertainties facing the electricity industry. Applications of this tool are demonstrated through a case study of an electricity industry with coal, CCGT, and OCGT facing future uncertainties. Results highlight some significant generation investment challenges, including the impacts of uncertain and correlated carbon and fossil-fuel prices, the role of future demand changes in response to electricity prices, and the impact of construction cost uncertainties on capital intensive generation. The tool can incorporate virtually any type of input probability distribution, and support sophisticated risk assessments of different portfolios, including downside economic risks. It can also assess portfolios against multi-criterion objectives such as greenhouse emissions as well as overall industry costs. - Highlights: ► Present a decision support tool to assist generation investment and policy making under uncertainty. ► Generation portfolios are assessed based on their expected costs, risks, and CO 2 emissions. ► There is tradeoff among expected cost, risks, and CO 2 emissions of generation portfolios. ► Investment challenges include economic impact of uncertainties and the effect of price elasticity. ► CO 2 emissions reduction depends on the mix of

  4. System diagnostic builder: a rule-generation tool for expert systems that do intelligent data evaluation

    Nieten, Joseph L.; Burke, Roger

    1993-03-01

    The system diagnostic builder (SDB) is an automated knowledge acquisition tool using state- of-the-art artificial intelligence (AI) technologies. The SDB uses an inductive machine learning technique to generate rules from data sets that are classified by a subject matter expert (SME). Thus, data is captured from the subject system, classified by an expert, and used to drive the rule generation process. These rule-bases are used to represent the observable behavior of the subject system, and to represent knowledge about this system. The rule-bases can be used in any knowledge based system which monitors or controls a physical system or simulation. The SDB has demonstrated the utility of using inductive machine learning technology to generate reliable knowledge bases. In fact, we have discovered that the knowledge captured by the SDB can be used in any number of applications. For example, the knowledge bases captured from the SMS can be used as black box simulations by intelligent computer aided training devices. We can also use the SDB to construct knowledge bases for the process control industry, such as chemical production, or oil and gas production. These knowledge bases can be used in automated advisory systems to ensure safety, productivity, and consistency.

  5. AUTOMATIC WINDING GENERATION USING MATRIX REPRESENTATION - ANFRACTUS TOOL 1.0

    Daoud Ouamara

    2018-02-01

    Full Text Available This paper describes an original approach dealing with AC/DC winding design in electrical machines. A research software called “ANFRACTUS Tool 1.0”, allowing automatic generation of all windings in multi-phases electrical machines, has been developed using the matrix representation. Unlike existent methods, where the aim is to synthesize a winding with higher performances, the proposed method provides the opportunity to choose between all doable windings. The specificity of this approach is based on the fact that it take only the slots, phases and layers number as input parameters. The poles number is not requested to run the generation process. Windings generation by matrix representation may be applied for any number of slots, phases and layers. The software do not deal with the manner that coils are connected but just the emplacement of coils in each slot with its current sense. The waveform and the harmonic spectrum of the total magnetomotive force (MMF are given as result.

  6. Combining SLBL routine with landslide-generated tsunami model for a quick hazard assessment tool

    Franz, Martin; Rudaz, Benjamin; Jaboyedoff, Michel; Podladchikov, Yury

    2016-04-01

    Regions with steep topography are potentially subject to landslide-induced tsunami, because of the proximity between lakes, rivers, sea shores and potential instabilities. The concentration of the population and infrastructures on the water body shores and downstream valleys could lead to catastrophic consequences. In order to assess comprehensively this phenomenon together with the induced risks, we have developed a tool which allows the construction of the landslide geometry, and which is able to simulate its propagation, the generation and the propagation of the wave and eventually the spread on the shores or the associated downstream flow. The tool is developed in the Matlab© environment, with a graphical user interface (GUI) to select the parameters in a user-friendly manner. The whole process is done in three steps implying different methods. Firstly, the geometry of the sliding mass is constructed using the Sloping Local Base Level (SLBL) concept. Secondly, the propagation of this volume is performed using a model based on viscous flow equations. Finally, the wave generation and its propagation are simulated using the shallow water equations stabilized by the Lax-Friedrichs scheme. The transition between wet and dry bed is performed by the combination of the two latter sets of equations. The intensity map is based on the criterion of flooding in Switzerland provided by the OFEG and results from the multiplication of the velocity and the depth obtained during the simulation. The tool can be used for hazard assessment in the case of well-known landslides, where the SLBL routine can be constrained and checked for realistic construction of the geometrical model. In less-known cases, various failure plane geometries can be automatically built between given range and thus a multi-scenario approach is used. In any case, less-known parameters such as the landslide velocity, its run-out distance, etc. can also be set to vary within given ranges, leading to multi

  7. Evaluation of Semi-Automatic Metadata Generation Tools: A Survey of the Current State of the Art

    Jung-ran Park

    2015-09-01

    Full Text Available Assessment of the current landscape of semi-automatic metadata generation tools is particularly important considering the rapid development of digital repositories and the recent explosion of big data. Utilization of (semiautomatic metadata generation is critical in addressing these environmental changes and may be unavoidable in the future considering the costly and complex operation of manual metadata creation. To address such needs, this study examines the range of semi-automatic metadata generation tools (n=39 while providing an analysis of their techniques, features, and functions. The study focuses on open-source tools that can be readily utilized in libraries and other memory institutions.  The challenges and current barriers to implementation of these tools were identified. The greatest area of difficulty lies in the fact that  the piecemeal development of most semi-automatic generation tools only addresses part of the issue of semi-automatic metadata generation, providing solutions to one or a few metadata elements but not the full range elements.  This indicates that significant local efforts will be required to integrate the various tools into a coherent set of a working whole.  Suggestions toward such efforts are presented for future developments that may assist information professionals with incorporation of semi-automatic tools within their daily workflows.

  8. The APEX Quantitative Proteomics Tool: Generating protein quantitation estimates from LC-MS/MS proteomics results

    Saeed Alexander I

    2008-12-01

    Full Text Available Abstract Background Mass spectrometry (MS based label-free protein quantitation has mainly focused on analysis of ion peak heights and peptide spectral counts. Most analyses of tandem mass spectrometry (MS/MS data begin with an enzymatic digestion of a complex protein mixture to generate smaller peptides that can be separated and identified by an MS/MS instrument. Peptide spectral counting techniques attempt to quantify protein abundance by counting the number of detected tryptic peptides and their corresponding MS spectra. However, spectral counting is confounded by the fact that peptide physicochemical properties severely affect MS detection resulting in each peptide having a different detection probability. Lu et al. (2007 described a modified spectral counting technique, Absolute Protein Expression (APEX, which improves on basic spectral counting methods by including a correction factor for each protein (called Oi value that accounts for variable peptide detection by MS techniques. The technique uses machine learning classification to derive peptide detection probabilities that are used to predict the number of tryptic peptides expected to be detected for one molecule of a particular protein (Oi. This predicted spectral count is compared to the protein's observed MS total spectral count during APEX computation of protein abundances. Results The APEX Quantitative Proteomics Tool, introduced here, is a free open source Java application that supports the APEX protein quantitation technique. The APEX tool uses data from standard tandem mass spectrometry proteomics experiments and provides computational support for APEX protein abundance quantitation through a set of graphical user interfaces that partition thparameter controls for the various processing tasks. The tool also provides a Z-score analysis for identification of significant differential protein expression, a utility to assess APEX classifier performance via cross validation, and a

  9. Discrete Surface Evolution and Mesh Deformation for Aircraft Icing Applications

    Thompson, David; Tong, Xiaoling; Arnoldus, Qiuhan; Collins, Eric; McLaurin, David; Luke, Edward; Bidwell, Colin S.

    2013-01-01

    Robust, automated mesh generation for problems with deforming geometries, such as ice accreting on aerodynamic surfaces, remains a challenging problem. Here we describe a technique to deform a discrete surface as it evolves due to the accretion of ice. The surface evolution algorithm is based on a smoothed, face-offsetting approach. We also describe a fast algebraic technique to propagate the computed surface deformations into the surrounding volume mesh while maintaining geometric mesh quality. Preliminary results presented here demonstrate the ecacy of the approach for a sphere with a prescribed accretion rate, a rime ice accretion, and a more complex glaze ice accretion.

  10. Isotropic 2D quadrangle meshing with size and orientation control

    Pellenard, Bertrand

    2011-12-01

    We propose an approach for automatically generating isotropic 2D quadrangle meshes from arbitrary domains with a fine control over sizing and orientation of the elements. At the heart of our algorithm is an optimization procedure that, from a coarse initial tiling of the 2D domain, enforces each of the desirable mesh quality criteria (size, shape, orientation, degree, regularity) one at a time, in an order designed not to undo previous enhancements. Our experiments demonstrate how well our resulting quadrangle meshes conform to a wide range of input sizing and orientation fields.

  11. Numerical convergence of discrete exterior calculus on arbitrary surface meshes

    Mohamed, Mamdouh S.

    2018-02-13

    Discrete exterior calculus (DEC) is a structure-preserving numerical framework for partial differential equations solution, particularly suitable for simplicial meshes. A longstanding and widespread assumption has been that DEC requires special (Delaunay) triangulations, which complicated the mesh generation process especially for curved surfaces. This paper presents numerical evidence demonstrating that this restriction is unnecessary. Convergence experiments are carried out for various physical problems using both Delaunay and non-Delaunay triangulations. Signed diagonal definition for the key DEC operator (Hodge star) is adopted. The errors converge as expected for all considered meshes and experiments. This relieves the DEC paradigm from unnecessary triangulation limitation.

  12. Mobile Phones Democratize and Cultivate Next-Generation Imaging, Diagnostics and Measurement Tools

    Ozcan, Aydogan

    2014-01-01

    In this article, I discuss some of the emerging applications and the future opportunities and challenges created by the use of mobile phones and their embedded components for the development of next-generation imaging, sensing, diagnostics and measurement tools. The massive volume of mobile phone users, which has now reached ~7 billion, drives the rapid improvements of the hardware, software and high-end imaging and sensing technologies embedded in our phones, transforming the mobile phone into a cost-effective and yet extremely powerful platform to run e.g., biomedical tests and perform scientific measurements that would normally require advanced laboratory instruments. This rapidly evolving and continuing trend will help us transform how medicine, engineering and sciences are practiced and taught globally. PMID:24647550

  13. Computer Tool for Automatically Generated 3D Illustration in Real Time from Archaeological Scanned Pieces

    Luis López

    2012-11-01

    Full Text Available The graphical documentation process of archaeological pieces requires the active involvement of a professional artist to recreate beautiful illustrations using a wide variety of expressive techniques. Frequently, the artist’s work is limited by the inconvenience of working only with the photographs of the pieces he is going to illustrate. This paper presents a software tool that allows the easy generation of illustrations in real time from 3D scanned models. The developed interface allows the user to simulate very elaborate artistic styles through the creation of diagrams by using the available virtual lights. The software processes the diagrams to render an illustration from any given angle or position. Among the available virtual lights, there are well known techniques as silhouettes enhancement, hatching or toon shading.

  14. A powerful transgenic tool for fate mapping and functional analysis of newly generated neurons

    Vogt Weisenhorn Daniela M

    2010-12-01

    Full Text Available Abstract Background Lack of appropriate tools and techniques to study fate and functional integration of newly generated neurons has so far hindered understanding of neurogenesis' relevance under physiological and pathological conditions. Current analyses are either dependent on mitotic labeling, for example BrdU-incorporation or retroviral infection, or on the detection of transient immature neuronal markers. Here, we report a transgenic mouse model (DCX-CreERT2 for time-resolved fate analysis of newly generated neurons. This model is based on the expression of a tamoxifen-inducible Cre recombinase under the control of a doublecortin (DCX promoter, which is specific for immature neuronal cells in the CNS. Results In the DCX-CreERT2 transgenic mice, expression of CreERT2 was restricted to DCX+ cells. In the CNS of transgenic embryos and adult DCX-CreERT2 mice, tamoxifen administration caused the transient translocation of CreERT2 to the nucleus, allowing for the recombination of loxP-flanked sequences. In our system, tamoxifen administration at E14.5 resulted in reporter gene activation throughout the developing CNS of transgenic embryos. In the adult CNS, neurogenic regions were the primary sites of tamoxifen-induced reporter gene activation. In addition, reporter expression could also be detected outside of neurogenic regions in cells physiologically expressing DCX (e.g. piriform cortex, corpus callosum, hypothalamus. Four weeks after recombination, the vast majority of reporter-expressing cells were found to co-express NeuN, revealing the neuronal fate of DCX+ cells upon maturation. Conclusions This first validation demonstrates that our new DCX-CreERT2 transgenic mouse model constitutes a powerful tool to investigate neurogenesis, migration and their long-term fate of neuronal precursors. Moreover, it allows for a targeted activation or deletion of specific genes in neuronal precursors and will thereby contribute to unravel the molecular

  15. Trident: A Universal Tool for Generating Synthetic Absorption Spectra from Astrophysical Simulations

    Hummels, Cameron B.; Smith, Britton D.; Silvia, Devin W.

    2017-09-01

    Hydrodynamical simulations are increasingly able to accurately model physical systems on stellar, galactic, and cosmological scales; however, the utility of these simulations is often limited by our ability to directly compare them with the data sets produced by observers: spectra, photometry, etc. To address this problem, we have created trident, a Python-based open-source tool for post-processing hydrodynamical simulations to produce synthetic absorption spectra and related data. trident can (I) create absorption-line spectra for any trajectory through a simulated data set mimicking both background quasar and down-the-barrel configurations; (II) reproduce the spectral characteristics of common instruments like the Cosmic Origins Spectrograph; (III) operate across the ultraviolet, optical, and infrared using customizable absorption-line lists; (IV) trace simulated physical structures directly to spectral features; (v) approximate the presence of ion species absent from the simulation outputs; (VI) generate column density maps for any ion; and (vii) provide support for all major astrophysical hydrodynamical codes. trident was originally developed to aid in the interpretation of observations of the circumgalactic medium and intergalactic medium, but it remains a general tool applicable in other contexts.

  16. Trident: A Universal Tool for Generating Synthetic Absorption Spectra from Astrophysical Simulations

    Hummels, Cameron B. [TAPIR, California Institute of Technology, Pasadena, CA 91125 (United States); Smith, Britton D. [San Diego Supercomputer Center, University of California, San Diego, CA 92121 (United States); Silvia, Devin W. [Department of Physics and Astronomy, Michigan State University, East Lansing, MI 48824 (United States)

    2017-09-20

    Hydrodynamical simulations are increasingly able to accurately model physical systems on stellar, galactic, and cosmological scales; however, the utility of these simulations is often limited by our ability to directly compare them with the data sets produced by observers: spectra, photometry, etc. To address this problem, we have created trident, a Python-based open-source tool for post-processing hydrodynamical simulations to produce synthetic absorption spectra and related data. trident can (i) create absorption-line spectra for any trajectory through a simulated data set mimicking both background quasar and down-the-barrel configurations; (ii) reproduce the spectral characteristics of common instruments like the Cosmic Origins Spectrograph; (iii) operate across the ultraviolet, optical, and infrared using customizable absorption-line lists; (iv) trace simulated physical structures directly to spectral features; (v) approximate the presence of ion species absent from the simulation outputs; (vi) generate column density maps for any ion; and (vii) provide support for all major astrophysical hydrodynamical codes. trident was originally developed to aid in the interpretation of observations of the circumgalactic medium and intergalactic medium, but it remains a general tool applicable in other contexts.

  17. User-friendly Tool for Power Flow Analysis and Distributed Generation Optimisation in Radial Distribution Networks

    M. F. Akorede

    2017-06-01

    Full Text Available The intent of power distribution companies (DISCOs is to deliver electric power to their customers in an efficient and reliable manner – with minimal energy loss cost. One major way to minimise power loss on a given power system is to install distributed generation (DG units on the distribution networks. However, to maximise benefits, it is highly crucial for a DISCO to ensure that these DG units are of optimal size and sited in the best locations on the network. This paper gives an overview of a software package developed in this study, called Power System Analysis and DG Optimisation Tool (PFADOT. The main purpose of the graphical user interface-based package is to guide a DISCO in finding the optimal size and location for DG placement in radial distribution networks. The package, which is also suitable for load flow analysis, employs the GUI feature of MATLAB. Three objective functions are formulated into a single optimisation problem and solved with fuzzy genetic algorithm to simultaneously obtain DG optimal size and location. The accuracy and reliability of the developed tool was validated using several radial test systems, and the results obtained are evaluated against the existing similar package cited in the literature, which are impressive and computationally efficient.

  18. Next Generation Electromagnetic Pump Analysis Tools (PLM DOC-0005-2188). Final Report

    Stregy, Seth [GE Hitachi Nuclear Energy Americas LLC, Wilmington, NC (United States); Dasilva, Ana [GE Hitachi Nuclear Energy Americas LLC, Wilmington, NC (United States); Yilmaz, Serkan [GE Hitachi Nuclear Energy Americas LLC, Wilmington, NC (United States); Saha, Pradip [GE Hitachi Nuclear Energy Americas LLC, Wilmington, NC (United States); Loewen, Eric [GE Hitachi Nuclear Energy Americas LLC, Wilmington, NC (United States)

    2015-10-29

    This report provides the broad historical review of EM Pump development and details of MATRIX development under this project. This report summarizes the efforts made to modernize the legacy performance models used in previous EM Pump designs and the improvements made to the analysis tools. This report provides information on Tasks 1, 3, and 4 of the entire project. The research for Task 4 builds upon Task 1: Update EM Pump Databank and Task 3: Modernize the Existing EM Pump Analysis Model, which are summarized within this report. Where research for Task 2: Insulation Materials Development and Evaluation identified parameters applicable to the analysis model with Task 4, the analysis code was updated, and analyses were made for additional materials. The important design variables for the manufacture and operation of an EM Pump that the model improvement can evaluate are: space constraints; voltage capability of insulation system; maximum flux density through iron; flow rate and outlet pressure; efficiency and manufacturability. The development of the next-generation EM Pump analysis tools during this two-year program provides information in three broad areas: Status of analysis model development; Improvements made to older simulations; and Comparison to experimental data.

  19. A practical comparison of de novo genome assembly software tools for next-generation sequencing technologies.

    Wenyu Zhang

    Full Text Available The advent of next-generation sequencing technologies is accompanied with the development of many whole-genome sequence assembly methods and software, especially for de novo fragment assembly. Due to the poor knowledge about the applicability and performance of these software tools, choosing a befitting assembler becomes a tough task. Here, we provide the information of adaptivity for each program, then above all, compare the performance of eight distinct tools against eight groups of simulated datasets from Solexa sequencing platform. Considering the computational time, maximum random access memory (RAM occupancy, assembly accuracy and integrity, our study indicate that string-based assemblers, overlap-layout-consensus (OLC assemblers are well-suited for very short reads and longer reads of small genomes respectively. For large datasets of more than hundred millions of short reads, De Bruijn graph-based assemblers would be more appropriate. In terms of software implementation, string-based assemblers are superior to graph-based ones, of which SOAPdenovo is complex for the creation of configuration file. Our comparison study will assist researchers in selecting a well-suited assembler and offer essential information for the improvement of existing assemblers or the developing of novel assemblers.

  20. Recombineering strategies for developing next generation BAC transgenic tools for optogenetics and beyond.

    Ting, Jonathan T; Feng, Guoping

    2014-01-01

    The development and application of diverse BAC transgenic rodent lines has enabled rapid progress for precise molecular targeting of genetically-defined cell types in the mammalian central nervous system. These transgenic tools have played a central role in the optogenetic revolution in neuroscience. Indeed, an overwhelming proportion of studies in this field have made use of BAC transgenic Cre driver lines to achieve targeted expression of optogenetic probes in the brain. In addition, several BAC transgenic mouse lines have been established for direct cell-type specific expression of Channelrhodopsin-2 (ChR2). While the benefits of these new tools largely outweigh any accompanying challenges, many available BAC transgenic lines may suffer from confounds due in part to increased gene dosage of one or more "extra" genes contained within the large BAC DNA sequences. Here we discuss this under-appreciated issue and propose strategies for developing the next generation of BAC transgenic lines that are devoid of extra genes. Furthermore, we provide evidence that these strategies are simple, reproducible, and do not disrupt the intended cell-type specific transgene expression patterns for several distinct BAC clones. These strategies may be widely implemented for improved BAC transgenesis across diverse disciplines.

  1. Trident: A Universal Tool for Generating Synthetic Absorption Spectra from Astrophysical Simulations

    Hummels, Cameron B.; Smith, Britton D.; Silvia, Devin W.

    2017-01-01

    Hydrodynamical simulations are increasingly able to accurately model physical systems on stellar, galactic, and cosmological scales; however, the utility of these simulations is often limited by our ability to directly compare them with the data sets produced by observers: spectra, photometry, etc. To address this problem, we have created trident, a Python-based open-source tool for post-processing hydrodynamical simulations to produce synthetic absorption spectra and related data. trident can (i) create absorption-line spectra for any trajectory through a simulated data set mimicking both background quasar and down-the-barrel configurations; (ii) reproduce the spectral characteristics of common instruments like the Cosmic Origins Spectrograph; (iii) operate across the ultraviolet, optical, and infrared using customizable absorption-line lists; (iv) trace simulated physical structures directly to spectral features; (v) approximate the presence of ion species absent from the simulation outputs; (vi) generate column density maps for any ion; and (vii) provide support for all major astrophysical hydrodynamical codes. trident was originally developed to aid in the interpretation of observations of the circumgalactic medium and intergalactic medium, but it remains a general tool applicable in other contexts.

  2. Development of the On-line Acoustic Leak Detection Tool for the SFR Steam Generator Protection

    Kim, Tae-Joon; Jeong, Ji-Young; Kim, Jong-Man; Kim, Byung-Ho; Kim, Seong-O

    2007-01-01

    The successful detection of a water/steam into a sodium leak in the SFR SG (steam generator) at an early phase of a leak origin depends on the fast response and sensitivity of a leak detection system. This intention of an acoustic leak detection system is stipulated by a key impossibility of a fast detecting of an intermediate leak by the present nominal systems such as the hydrogen meter. Subject of this study is to introduce the detection performance of an on-line acoustic leak detection tool discriminated by a back-propagation neural network with a preprocessing of the 1/m Octave band analysis, and to introduce the status of an on-line development being developed with the acoustic leak detection tool(S/W) in KAERI. For a performance test, it was used with the acoustic signals for a sodium-water reaction from the injected steam into water experiments in KAERI, the acoustic signals injected from the water into the sodium obtained in IPPE, and the background noise of the PFR superheater

  3. LCFM - LIVING COLOR FRAME MAKER: PC GRAPHICS GENERATION AND MANAGEMENT TOOL FOR REAL-TIME APPLICATIONS

    Truong, L. V.

    1994-01-01

    Computer graphics are often applied for better understanding and interpretation of data under observation. These graphics become more complicated when animation is required during "run-time", as found in many typical modern artificial intelligence and expert systems. Living Color Frame Maker is a solution to many of these real-time graphics problems. Living Color Frame Maker (LCFM) is a graphics generation and management tool for IBM or IBM compatible personal computers. To eliminate graphics programming, the graphic designer can use LCFM to generate computer graphics frames. The graphical frames are then saved as text files, in a readable and disclosed format, which can be easily accessed and manipulated by user programs for a wide range of "real-time" visual information applications. For example, LCFM can be implemented in a frame-based expert system for visual aids in management of systems. For monitoring, diagnosis, and/or controlling purposes, circuit or systems diagrams can be brought to "life" by using designated video colors and intensities to symbolize the status of hardware components (via real-time feedback from sensors). Thus status of the system itself can be displayed. The Living Color Frame Maker is user friendly with graphical interfaces, and provides on-line help instructions. All options are executed using mouse commands and are displayed on a single menu for fast and easy operation. LCFM is written in C++ using the Borland C++ 2.0 compiler for IBM PC series computers and compatible computers running MS-DOS. The program requires a mouse and an EGA/VGA display. A minimum of 77K of RAM is also required for execution. The documentation is provided in electronic form on the distribution medium in WordPerfect format. A sample MS-DOS executable is provided on the distribution medium. The standard distribution medium for this program is one 5.25 inch 360K MS-DOS format diskette. The contents of the diskette are compressed using the PKWARE archiving tools

  4. Towards a New Generation of Time-Series Visualization Tools in the ESA Heliophysics Science Archives

    Perez, H.; Martinez, B.; Cook, J. P.; Herment, D.; Fernandez, M.; De Teodoro, P.; Arnaud, M.; Middleton, H. R.; Osuna, P.; Arviset, C.

    2017-12-01

    During the last decades a varied set of Heliophysics missions have allowed the scientific community to gain a better knowledge on the solar atmosphere and activity. The remote sensing images of missions such as SOHO have paved the ground for Helio-based spatial data visualization software such as JHelioViewer/Helioviewer. On the other hand, the huge amount of in-situ measurements provided by other missions such as Cluster provide a wide base for plot visualization software whose reach is still far from being fully exploited. The Heliophysics Science Archives within the ESAC Science Data Center (ESDC) already provide a first generation of tools for time-series visualization focusing on each mission's needs: visualization of quicklook plots, cross-calibration time series, pre-generated/on-demand multi-plot stacks (Cluster), basic plot zoom in/out options (Ulysses) and easy navigation through the plots in time (Ulysses, Cluster, ISS-Solaces). However, as the needs evolve and the scientists involved in new missions require to plot multi-variable data, heat maps stacks interactive synchronization and axis variable selection among other improvements. The new Heliophysics archives (such as Solar Orbiter) and the evolution of existing ones (Cluster) intend to address these new challenges. This paper provides an overview of the different approaches for visualizing time-series followed within the ESA Heliophysics Archives and their foreseen evolution.

  5. Generation of Look-Up Tables for Dynamic Job Shop Scheduling Decision Support Tool

    Oktaviandri, Muchamad; Hassan, Adnan; Mohd Shaharoun, Awaluddin

    2016-02-01

    Majority of existing scheduling techniques are based on static demand and deterministic processing time, while most job shop scheduling problem are concerned with dynamic demand and stochastic processing time. As a consequence, the solutions obtained from the traditional scheduling technique are ineffective wherever changes occur to the system. Therefore, this research intends to develop a decision support tool (DST) based on promising artificial intelligent that is able to accommodate the dynamics that regularly occur in job shop scheduling problem. The DST was designed through three phases, i.e. (i) the look-up table generation, (ii) inverse model development and (iii) integration of DST components. This paper reports the generation of look-up tables for various scenarios as a part in development of the DST. A discrete event simulation model was used to compare the performance among SPT, EDD, FCFS, S/OPN and Slack rules; the best performances measures (mean flow time, mean tardiness and mean lateness) and the job order requirement (inter-arrival time, due dates tightness and setup time ratio) which were compiled into look-up tables. The well-known 6/6/J/Cmax Problem from Muth and Thompson (1963) was used as a case study. In the future, the performance measure of various scheduling scenarios and the job order requirement will be mapped using ANN inverse model.

  6. Leveraging the power of mesh

    Glass, H. [Cellnet, Alpharetta, GA (United States)

    2006-07-01

    Mesh network applications are used by utilities for metering, demand response, and mobile workforce management. This presentation provided an overview of a multi-dimensional mesh application designed to offer improved scalability and higher throughput in advanced metering infrastructure (AMI) systems. Mesh applications can be used in AMI for load balancing and forecasting, as well as for distribution and transmission planning. New revenue opportunities can be realized through the application's ability to improve notification and monitoring services, and customer service communications. Mesh network security features include data encryption, data fragmentation and the automatic re-routing of data. In order to use mesh network applications, networks must have sufficient bandwidth and provide flexibility at the endpoint layer to support multiple devices from multiple vendors, as well as support multiple protocols. It was concluded that smart meters will not enable energy response solutions without an underlying AMI that is reliable, scalable and self-healing. .refs., tabs., figs.

  7. Form-finding with polyhedral meshes made simple

    Tang, Chengcheng; Sun, Xiang; Gomes, Maria Alexandra; Wallner, Johannes; Pottmann, Helmut

    2014-01-01

    We solve the form-finding problem for polyhedral meshes in a way which combines form, function and fabrication; taking care of user-specified constraints like boundary interpolation, planarity of faces, statics, panel size and shape, enclosed volume, and last, but not least, cost. Our main application is the interactive modeling of meshes for architectural and industrial design. Our approach can be described as guided exploration of the constraint space whose algebraic structure is simplified by introducing auxiliary variables and ensuring that constraints are at most quadratic. Computationally, we perform a projection onto the constraint space which is biased towards low values of an energy which expresses desirable "soft" properties like fairness. We have created a tool which elegantly handles difficult tasks, such as taking boundary-alignment of polyhedral meshes into account, planarization, fairing under planarity side conditions, handling hybrid meshes, and extending the treatment of static equilibrium to shapes which possess overhanging parts.

  8. Form-finding with polyhedral meshes made simple

    Tang, Chengcheng

    2014-07-27

    We solve the form-finding problem for polyhedral meshes in a way which combines form, function and fabrication; taking care of user-specified constraints like boundary interpolation, planarity of faces, statics, panel size and shape, enclosed volume, and last, but not least, cost. Our main application is the interactive modeling of meshes for architectural and industrial design. Our approach can be described as guided exploration of the constraint space whose algebraic structure is simplified by introducing auxiliary variables and ensuring that constraints are at most quadratic. Computationally, we perform a projection onto the constraint space which is biased towards low values of an energy which expresses desirable "soft" properties like fairness. We have created a tool which elegantly handles difficult tasks, such as taking boundary-alignment of polyhedral meshes into account, planarization, fairing under planarity side conditions, handling hybrid meshes, and extending the treatment of static equilibrium to shapes which possess overhanging parts.

  9. Finite element simulation of impact response of wire mesh screens

    Wang Caizheng

    2015-01-01

    Full Text Available In this paper, the response of wire mesh screens to low velocity impact with blunt objects is investigated using finite element (FE simulation. The woven wire mesh is modelled with homogeneous shell elements with equivalent smeared mechanical properties. The mechanical behaviour of the woven wire mesh was determined experimentally with tensile tests on steel wire mesh coupons to generate the data for the smeared shell material used in the FE. The effects of impacts with a low mass (4 kg and a large mass (40 kg providing the same impact energy are studied. The joint between the wire mesh screen and the aluminium frame surrounding it is modelled using contact elements with friction between the corresponding elements. Damage to the screen of different types compromising its structural integrity, such as mesh separation and pulling out from the surrounding frame is modelled. The FE simulation is validated with results of impact tests conducted on woven steel wire screen meshes.

  10. Dynamic Mesh Adaptation for Front Evolution Using Discontinuous Galerkin Based Weighted Condition Number Mesh Relaxation

    Greene, Patrick T. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Schofield, Samuel P. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Nourgaliev, Robert [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2016-06-21

    A new mesh smoothing method designed to cluster mesh cells near a dynamically evolving interface is presented. The method is based on weighted condition number mesh relaxation with the weight function being computed from a level set representation of the interface. The weight function is expressed as a Taylor series based discontinuous Galerkin projection, which makes the computation of the derivatives of the weight function needed during the condition number optimization process a trivial matter. For cases when a level set is not available, a fast method for generating a low-order level set from discrete cell-centered elds, such as a volume fraction or index function, is provided. Results show that the low-order level set works equally well for the weight function as the actual level set. Meshes generated for a number of interface geometries are presented, including cases with multiple level sets. Dynamic cases for moving interfaces are presented to demonstrate the method's potential usefulness to arbitrary Lagrangian Eulerian (ALE) methods.

  11. CityZoom UP (Urban Pollution): a computational tool for the fast generation and setup of urban scenarios for CFD and dispersion modelling simulation

    Grazziotin, Pablo Colossi

    2016-01-01

    This research presents the development of CityZoom UP, the first attempt to extend existing urban planning software in order to assist in modelling urban scenarios and setting up simulation parameters for Gaussian dispersion and CFD models. Based on the previous capabilities and graphic user interfaces of CityZoom to model and validate urban scenarios based on Master Plan regulations, new graphic user interfaces, automatic mesh generation and data conversion algorithms have been created to se...

  12. Next generation human skin constructs as advanced tools for drug development.

    Abaci, H E; Guo, Zongyou; Doucet, Yanne; Jacków, Joanna; Christiano, Angela

    2017-11-01

    Many diseases, as well as side effects of drugs, manifest themselves through skin symptoms. Skin is a complex tissue that hosts various specialized cell types and performs many roles including physical barrier, immune and sensory functions. Therefore, modeling skin in vitro presents technical challenges for tissue engineering. Since the first attempts at engineering human epidermis in 1970s, there has been a growing interest in generating full-thickness skin constructs mimicking physiological functions by incorporating various skin components, such as vasculature and melanocytes for pigmentation. Development of biomimetic in vitro human skin models with these physiological functions provides a new tool for drug discovery, disease modeling, regenerative medicine and basic research for skin biology. This goal, however, has long been delayed by the limited availability of different cell types, the challenges in establishing co-culture conditions, and the ability to recapitulate the 3D anatomy of the skin. Recent breakthroughs in induced pluripotent stem cell (iPSC) technology and microfabrication techniques such as 3D-printing have allowed for building more reliable and complex in vitro skin models for pharmaceutical screening. In this review, we focus on the current developments and prevailing challenges in generating skin constructs with vasculature, skin appendages such as hair follicles, pigmentation, immune response, innervation, and hypodermis. Furthermore, we discuss the promising advances that iPSC technology offers in order to generate in vitro models of genetic skin diseases, such as epidermolysis bullosa and psoriasis. We also discuss how future integration of the next generation human skin constructs onto microfluidic platforms along with other tissues could revolutionize the early stages of drug development by creating reliable evaluation of patient-specific effects of pharmaceutical agents. Impact statement Skin is a complex tissue that hosts various

  13. Streaming Compression of Hexahedral Meshes

    Isenburg, M; Courbet, C

    2010-02-03

    We describe a method for streaming compression of hexahedral meshes. Given an interleaved stream of vertices and hexahedral our coder incrementally compresses the mesh in the presented order. Our coder is extremely memory efficient when the input stream documents when vertices are referenced for the last time (i.e. when it contains topological finalization tags). Our coder then continuously releases and reuses data structures that no longer contribute to compressing the remainder of the stream. This means in practice that our coder has only a small fraction of the whole mesh in memory at any time. We can therefore compress very large meshes - even meshes that do not file in memory. Compared to traditional, non-streaming approaches that load the entire mesh and globally reorder it during compression, our algorithm trades a less compact compressed representation for significant gains in speed, memory, and I/O efficiency. For example, on the 456k hexahedra 'blade' mesh, our coder is twice as fast and uses 88 times less memory (only 3.1 MB) with the compressed file increasing about 3% in size. We also present the first scheme for predictive compression of properties associated with hexahedral cells.

  14. 802.11s Wireless Mesh Network Visualization Application

    Mauldin, James Alexander

    2014-01-01

    Results of past experimentation at NASA Johnson Space Center showed that the IEEE 802.11s standard has better performance than the widely implemented alternative protocol B.A.T.M.A.N (Better Approach to Mobile Ad hoc Networking). 802.11s is now formally incorporated into the Wi- Fi 802.11-2012 standard, which specifies a hybrid wireless mesh networking protocol (HWMP). In order to quickly analyze changes to the routing algorithm and to support optimizing the mesh network behavior for our intended application a visualization tool was developed by modifying and integrating open source tools.

  15. A novel three-dimensional mesh deformation method based on sphere relaxation

    Zhou, Xuan; Li, Shuixiang

    2015-01-01

    In our previous work (2013) [19], we developed a disk relaxation based mesh deformation method for two-dimensional mesh deformation. In this paper, the idea of the disk relaxation is extended to the sphere relaxation for three-dimensional meshes with large deformations. We develop a node based pre-displacement procedure to apply initial movements on nodes according to their layer indices. Afterwards, the nodes are moved locally by the improved sphere relaxation algorithm to transfer boundary deformations and increase the mesh quality. A three-dimensional mesh smoothing method is also adopted to prevent the occurrence of the negative volume of elements, and further improve the mesh quality. Numerical applications in three-dimension including the wing rotation, bending beam and morphing aircraft are carried out. The results demonstrate that the sphere relaxation based approach generates the deformed mesh with high quality, especially regarding complex boundaries and large deformations

  16. A novel three-dimensional mesh deformation method based on sphere relaxation

    Zhou, Xuan [Department of Mechanics & Engineering Science, College of Engineering, Peking University, Beijing, 100871 (China); Institute of Applied Physics and Computational Mathematics, Beijing, 100094 (China); Li, Shuixiang, E-mail: lsx@pku.edu.cn [Department of Mechanics & Engineering Science, College of Engineering, Peking University, Beijing, 100871 (China)

    2015-10-01

    In our previous work (2013) [19], we developed a disk relaxation based mesh deformation method for two-dimensional mesh deformation. In this paper, the idea of the disk relaxation is extended to the sphere relaxation for three-dimensional meshes with large deformations. We develop a node based pre-displacement procedure to apply initial movements on nodes according to their layer indices. Afterwards, the nodes are moved locally by the improved sphere relaxation algorithm to transfer boundary deformations and increase the mesh quality. A three-dimensional mesh smoothing method is also adopted to prevent the occurrence of the negative volume of elements, and further improve the mesh quality. Numerical applications in three-dimension including the wing rotation, bending beam and morphing aircraft are carried out. The results demonstrate that the sphere relaxation based approach generates the deformed mesh with high quality, especially regarding complex boundaries and large deformations.

  17. Mesh Adaptation and Shape Optimization on Unstructured Meshes, Phase I

    National Aeronautics and Space Administration — In this SBIR CRM proposes to implement the entropy adjoint method for solution adaptive mesh refinement into the Loci/CHEM unstructured flow solver. The scheme will...

  18. The Environmental Scenario Generator (ESG: a distributed environmental data archive analysis tool

    E A Kihn

    2006-01-01

    Full Text Available The Environmental Scenario Generator (ESG is a network distributed software system designed to allow a user to interact with archives of environmental data for the purpose of scenario extraction, data analysis and integration with existing models that require environmental input. The ESG uses fuzzy-logic based search tools to allow a user to look for specific environmental scenarios in vast archives by specifying the search in human linguistic terms. For example, the user can specify a scenario such as a "cloud free week" or "high winds and low pressure" and then search relevant archives available across the network to get a list of matching events. The ESG hooks to existing archives of data by providing a simple communication framework and an efficient data model for exchanging data. Once data has been delivered by the distributed archives in the ESG data model, it can easily be accessed by the visualization, integration and analysis components to meet specific user requests. The ESG implementation provides a framework which can be taken as a pattern applicable to other distributed archive systems.

  19. Mersiline mesh in premaxillary augmentation.

    Foda, Hossam M T

    2005-01-01

    Premaxillary retrusion may distort the aesthetic appearance of the columella, lip, and nasal tip. This defect is characteristically seen in, but not limited to, patients with cleft lip nasal deformity. This study investigated 60 patients presenting with premaxillary deficiencies in which Mersiline mesh was used to augment the premaxilla. All the cases had surgery using the external rhinoplasty technique. Two methods of augmentation with Mersiline mesh were used: the Mersiline roll technique, for the cases with central symmetric deficiencies, and the Mersiline packing technique, for the cases with asymmetric deficiencies. Premaxillary augmentation with Mersiline mesh proved to be simple technically, easy to perform, and not associated with any complications. Periodic follow-up evaluation for a mean period of 32 months (range, 12-98 months) showed that an adequate degree of premaxillary augmentation was maintained with no clinically detectable resorption of the mesh implant.

  20. Influence of mesh non-orthogonality on numerical simulation of buoyant jet flows

    Ishigaki, Masahiro; Abe, Satoshi; Sibamoto, Yasuteru; Yonomoto, Taisuke

    2017-01-01

    Highlights: • Influence of mesh non-orthogonality on numerical solution of buoyant jet flows. • Buoyant jet flows are simulated with hexahedral and prismatic meshes. • Jet instability with prismatic meshes may be overestimated compared to that with hexahedral meshes. • Modified solvers that can reduce the influence of mesh non-orthogonality and reduce computation time are proposed. - Abstract: In the present research, we discuss the influence of mesh non-orthogonality on numerical solution of a type of buoyant flow. Buoyant jet flows are simulated numerically with hexahedral and prismatic mesh elements in an open source Computational Fluid Dynamics (CFD) code called “OpenFOAM”. Buoyant jet instability obtained with the prismatic meshes may be overestimated compared to that obtained with the hexahedral meshes when non-orthogonal correction is not applied in the code. Although the non-orthogonal correction method can improve the instability generated by mesh non-orthogonality, it may increase computation time required to reach a convergent solution. Thus, we propose modified solvers that can reduce the influence of mesh non-orthogonality and reduce the computation time compared to the existing solvers in OpenFOAM. It is demonstrated that calculations for a buoyant jet with a large temperature difference are performed faster by the modified solver.

  1. Influence of mesh non-orthogonality on numerical simulation of buoyant jet flows

    Ishigaki, Masahiro, E-mail: ishigaki.masahiro@jaea.go.jp; Abe, Satoshi; Sibamoto, Yasuteru; Yonomoto, Taisuke

    2017-04-01

    Highlights: • Influence of mesh non-orthogonality on numerical solution of buoyant jet flows. • Buoyant jet flows are simulated with hexahedral and prismatic meshes. • Jet instability with prismatic meshes may be overestimated compared to that with hexahedral meshes. • Modified solvers that can reduce the influence of mesh non-orthogonality and reduce computation time are proposed. - Abstract: In the present research, we discuss the influence of mesh non-orthogonality on numerical solution of a type of buoyant flow. Buoyant jet flows are simulated numerically with hexahedral and prismatic mesh elements in an open source Computational Fluid Dynamics (CFD) code called “OpenFOAM”. Buoyant jet instability obtained with the prismatic meshes may be overestimated compared to that obtained with the hexahedral meshes when non-orthogonal correction is not applied in the code. Although the non-orthogonal correction method can improve the instability generated by mesh non-orthogonality, it may increase computation time required to reach a convergent solution. Thus, we propose modified solvers that can reduce the influence of mesh non-orthogonality and reduce the computation time compared to the existing solvers in OpenFOAM. It is demonstrated that calculations for a buoyant jet with a large temperature difference are performed faster by the modified solver.

  2. Field-aligned mesh joinery

    Cignoni, Paolo; Pietroni, Nico; Malomo, Luigi

    2014-01-01

    Mesh joinery is an innovative method to produce illustrative shape approximations suitable for fabrication. Mesh joinery is capable of producing complex fabricable structures in an efficient and visually pleasing manner. We represent an input geometry as a set of planar pieces arranged to compose a rigid structure, by exploiting an efficient slit mechanism. Since slices are planar, to fabricate them a standard 2D cutting system is enough. We automatically arrange slices according to a smooth ...

  3. Topological patterns of mesh textures in serpentinites

    Miyazawa, M.; Suzuki, A.; Shimizu, H.; Okamoto, A.; Hiraoka, Y.; Obayashi, I.; Tsuji, T.; Ito, T.

    2017-12-01

    Serpentinization is a hydration process that forms serpentine minerals and magnetite within the oceanic lithosphere. Microfractures crosscut these minerals during the reactions, and the structures look like mesh textures. It has been known that the patterns of microfractures and the system evolutions are affected by the hydration reaction and fluid transport in fractures and within matrices. This study aims at quantifying the topological patterns of the mesh textures and understanding possible conditions of fluid transport and reaction during serpentinization in the oceanic lithosphere. Two-dimensional simulation by the distinct element method (DEM) generates fracture patterns due to serpentinization. The microfracture patterns are evaluated by persistent homology, which measures features of connected components of a topological space and encodes multi-scale topological features in the persistence diagrams. The persistence diagrams of the different mesh textures are evaluated by principal component analysis to bring out the strong patterns of persistence diagrams. This approach help extract feature values of fracture patterns from high-dimensional and complex datasets.

  4. Improved Mesh_Based Image Morphing ‎

    Mohammed Abdullah Taha

    2017-11-01

    Full Text Available Image morphing is a multi-step process that generates a sequence of transitions between two images. The thought is to get a ₔgrouping of middle pictures which, when ₔassembled with the first pictures would represent the change from one picture to the other.  The process of morphing requires time and attention to detail in order to get good results. Morphing image requires at least two processes warping and cross dissolve. Warping is the process of geometric transformation of images. The cross dissolve is the process interpolation of color of eachₔ pixel from the first image value to theₔ corresponding second imageₔ value over the time. Image morphing techniques differ from in the approach of image warping procedure. This work presents a survey of different techniques to construct morphing images by review the different warping techniques. One of the predominant approaches of warping process is mesh warping which suffers from some problems including ghosting. This work proposed and implements an improved mesh warping technique to construct morphing images. The results show that the proposed approach can overcome the problems of the traditional mesh technique

  5. Computational fluid dynamics and particle image velocimetry assisted design tools for a new generation of trochoidal gear pumps

    M Garcia-Vilchez

    2015-06-01

    Full Text Available Trochoidal gear pumps produce significant flow pulsations that result in pressure pulsations, which interact with the system where they are connected, shortening the life of both the pump and circuit components. The complicated aspects of the operation of a gerotor pump make computational fluid dynamics the proper tool for modelling and simulating its flow characteristics. A three-dimensional model with deforming mesh computational fluid dynamics is presented, including the effects of the manufacturing tolerance and the leakage inside the pump. A new boundary condition is created for the simulation of the solid contact in the interteeth radial clearance. The experimental study of the pump is carried out by means of time-resolved particle image velocimetry, and results are qualitatively evaluated, thanks to the numerical simulation results. Time-resolved particle image velocimetry is developed in order to adapt it to the gerotor pump, and it is proved to be a feasible alternative to obtain the instantaneous flow of the pump in a direct mode, which would allow the determination of geometries that minimize the non-desired flow pulsations. Thus, a new methodology involving computational fluid dynamics and time-resolved particle image velocimetry is presented, which allows the obtaining of the instantaneous flow of the pump in a direct mode without altering its behaviour significantly.

  6. The influences of mesh subdivision on nonlinear fracture analysis for surface cracked structures

    Shimakawa, T.

    1991-01-01

    The leak-before-break (LBB) concept can be expected to be applied not only to safety assessment, but also to the rationalization of nuclear power plants. The development of a method to evaluate fracture characteristics is required to establish this concept. The finite element method (FEM) is one of the most useful tools for this evaluation. However, the influence of various factors on the solution is not well understood and the reliability has not been fully verified. In this study, elastic-plastic 3D analyses are performed for two kinds of surface cracked structure, and the influence of mesh design is discussed. The first problem is surface crack growth in a carbon steel plate subjected to tension loading. A crack extension analysis is performed under a generation phase simulation using the crack release technique. Numerical instability of the J-integral solution is observed when the number of elements in the thickness direction of the ligament is reduced to three. The influence of mesh design in the ligament on the solution is discussed. The second problem is a circumferential part-through crack in a carbon steel pipe subjected to a bending moment. Two kinds of mesh design are employed, and a comparison between two sets of results shows that the number of elements on the crack surface also affects the solution as well as the number of elements in the ligament. (author)

  7. Finite element meshing approached as a global minimization process

    WITKOWSKI,WALTER R.; JUNG,JOSEPH; DOHRMANN,CLARK R.; LEUNG,VITUS J.

    2000-03-01

    The ability to generate a suitable finite element mesh in an automatic fashion is becoming the key to being able to automate the entire engineering analysis process. However, placing an all-hexahedron mesh in a general three-dimensional body continues to be an elusive goal. The approach investigated in this research is fundamentally different from any other that is known of by the authors. A physical analogy viewpoint is used to formulate the actual meshing problem which constructs a global mathematical description of the problem. The analogy used was that of minimizing the electrical potential of a system charged particles within a charged domain. The particles in the presented analogy represent duals to mesh elements (i.e., quads or hexes). Particle movement is governed by a mathematical functional which accounts for inter-particles repulsive, attractive and alignment forces. This functional is minimized to find the optimal location and orientation of each particle. After the particles are connected a mesh can be easily resolved. The mathematical description for this problem is as easy to formulate in three-dimensions as it is in two- or one-dimensions. The meshing algorithm was developed within CoMeT. It can solve the two-dimensional meshing problem for convex and concave geometries in a purely automated fashion. Investigation of the robustness of the technique has shown a success rate of approximately 99% for the two-dimensional geometries tested. Run times to mesh a 100 element complex geometry were typically in the 10 minute range. Efficiency of the technique is still an issue that needs to be addressed. Performance is an issue that is critical for most engineers generating meshes. It was not for this project. The primary focus of this work was to investigate and evaluate a meshing algorithm/philosophy with efficiency issues being secondary. The algorithm was also extended to mesh three-dimensional geometries. Unfortunately, only simple geometries were tested

  8. Method and system for mesh network embedded devices

    Wang, Ray (Inventor)

    2009-01-01

    A method and system for managing mesh network devices. A mesh network device with integrated features creates an N-way mesh network with a full mesh network topology or a partial mesh network topology.

  9. Mesh versus non-mesh repair of ventral abdominal hernias

    Jawaid, M.A.; Talpur, A.H.

    2008-01-01

    To investigate the relative effectiveness of mesh and suture repair of ventral abdominal hernias in terms of clinical outcome, quality of life and rate of recurrence in both the techniques. This is a retrospective descriptive analysis of 236 patients with mesh and non-mesh repair of primary ventral hernias performed between January 2000 to December 2004 at Surgery Department, Liaquat University of Medical and Health Sciences, Jamshoro. The record sheets of the patients were analyzed and data retrieved to compare the results of both techniques for short-term and long-term results. The data retrieved is statistically analyzed on SPSS version 11. There were 43 (18.22%) males and 193 (81.77%) females with a mean age of 51.79 years and a range of 59 (81-22). Para-umbilical hernia was the commonest of ventral hernia and accounted for 49.8% (n=118) of the total study population followed by incisional hernia comprising 24% (n=57) of the total number. There was a significant difference in the recurrent rate at 3 years interval with 23/101 (22.77%) recurrences in suture-repaired subjects compared to 10/135 (7.40%) in mesh repair group. Chronic pain lasting up to 1-2 years was noted in 14 patients with suture repair. Wound infection is comparatively more common (8.14%) in mesh group. The other variables such as operative and postoperative complications, total hospital stay and quality of life is also discussed. Mesh repair of ventral hernia is much superior to non-mesh suture repair in terms of recurrence and overall outcome. (author)

  10. Tools for Generating Useful Time-series Data from PhenoCam Images

    Milliman, T. E.; Friedl, M. A.; Frolking, S.; Hufkens, K.; Klosterman, S.; Richardson, A. D.; Toomey, M. P.

    2012-12-01

    The PhenoCam project (http://phenocam.unh.edu/) is tasked with acquiring, processing, and archiving digital repeat photography to be used for scientific studies of vegetation phenological processes. Over the past 5 years the PhenoCam project has collected over 2 million time series images for a total over 700 GB of image data. Several papers have been published describing derived "vegetation indices" (such as green-chromatic-coordinate or gcc) which can be compared to standard measures such as NDVI or EVI. Imagery from our archive is available for download but converting series of images for a particular camera into useful scientific data, while simple in principle, is complicated by a variety of factors. Cameras are often exposed to harsh weather conditions (high wind, rain, ice, snow pile up), which result in images where the field of view (FOV) is partially obscured or completely blocked for periods of time. The FOV can also change for other reasons (mount failures, tower maintenance, etc.) Some of the relatively inexpensive cameras that are being used can also temporarily lose color balance or exposure controls resulting in loss of imagery. All these factors negatively influence the automated analysis of the image time series making this a non-trivial task. Here we discuss the challenges of processing PhenoCam image time-series for vegetation monitoring and the associated data management tasks. We describe our current processing framework and a simple standardized output format for the resulting time-series data. The time-series data in this format will be generated for specific "regions of interest" (ROI's) for each of the cameras in the PhenoCam network. This standardized output (which will be updated daily) can be considered 'the pulse' of a particular camera and will provide a default phenological dynamic for said camera. The time-series data can also be viewed as a higher level product which can be used to generate "vegetation indices", like gcc, for

  11. Evaluation of machine learning tools for inspection of steam generator tube structures using pulsed eddy current

    Buck, J. A.; Underhill, P. R.; Morelli, J.; Krause, T. W.

    2017-02-01

    Degradation of nuclear steam generator (SG) tubes and support structures can result in a loss of reactor efficiency. Regular in-service inspection, by conventional eddy current testing (ECT), permits detection of cracks, measurement of wall loss, and identification of other SG tube degradation modes. However, ECT is challenged by overlapping degradation modes such as might occur for SG tube fretting accompanied by tube off-set within a corroding ferromagnetic support structure. Pulsed eddy current (PEC) is an emerging technology examined here for inspection of Alloy-800 SG tubes and associated carbon steel drilled support structures. Support structure hole size was varied to simulate uniform corrosion, while SG tube was off-set relative to hole axis. PEC measurements were performed using a single driver with an 8 pick-up coil configuration in the presence of flat-bottom rectangular frets as an overlapping degradation mode. A modified principal component analysis (MPCA) was performed on the time-voltage data in order to reduce data dimensionality. The MPCA scores were then used to train a support vector machine (SVM) that simultaneously targeted four independent parameters associated with; support structure hole size, tube off-centering in two dimensions and fret depth. The support vector machine was trained, tested, and validated on experimental data. Results were compared with a previously developed artificial neural network (ANN) trained on the same data. Estimates of tube position showed comparable results between the two machine learning tools. However, the ANN produced better estimates of hole inner diameter and fret depth. The better results from ANN analysis was attributed to challenges associated with the SVM when non-constant variance is present in the data.

  12. Systems Prototyping with Fourth Generation Tools: One Answer to the Productivity Puzzle? AIR 1983 Annual Forum Paper.

    Sholtys, Phyllis A.

    The development of information systems using an engineering approach employing both traditional programming techniques and nonprocedural languages is described. A fourth generation application tool is used to develop a prototype system that is revised and expanded as the user clarifies individual requirements. When fully defined, a combination of…

  13. Next-Generation Ion Thruster Design Tool to Support Future Space Missions, Phase II

    National Aeronautics and Space Administration — Computational tools that accurately predict the performance of electric propulsion devices are highly desirable and beneficial to NASA and the broader electric...

  14. What can next generation sequencing do for you? Next generation sequencing as a valuable tool in plant research

    Bräutigam, Andrea; Gowik, Udo

    2010-01-01

    Next generation sequencing (NGS) technologies have opened fascinating opportunities for the analysis of plants with and without a sequenced genome on a genomic scale. During the last few years, NGS methods have become widely available and cost effective. They can be applied to a wide variety of biological questions, from the sequencing of complete eukaryotic genomes and transcriptomes, to the genome-scale analysis of DNA-protein interactions. In this review, we focus on the use of NGS for pla...

  15. 3D Mesh Compression and Transmission for Mobile Robotic Applications

    Bailin Yang

    2016-01-01

    Full Text Available Mobile robots are useful for environment exploration and rescue operations. In such applications, it is crucial to accurately analyse and represent an environment, providing appropriate inputs for motion planning in order to support robot navigation and operations. 2D mapping methods are simple but cannot handle multilevel or multistory environments. To address this problem, 3D mapping methods generate structural 3D representations of the robot operating environment and its objects by 3D mesh reconstruction. However, they face the challenge of efficiently transmitting those 3D representations to system modules for 3D mapping, motion planning, and robot operation visualization. This paper proposes a quality-driven mesh compression and transmission method to address this. Our method is efficient, as it compresses a mesh by quantizing its transformed vertices without the need to spend time constructing an a-priori structure over the mesh. A visual distortion function is developed to govern the level of quantization, allowing mesh transmission to be controlled under different network conditions or time constraints. Our experiments demonstrate how the visual quality of a mesh can be manipulated by the visual distortion function.

  16. Unbiased Sampling and Meshing of Isosurfaces

    Yan, Dongming

    2014-05-07

    In this paper, we present a new technique to generate unbiased samples on isosurfaces. An isosurface, F(x,y,z) = c , of a function, F , is implicitly defined by trilinear interpolation of background grid points. The key idea of our approach is that of treating the isosurface within a grid cell as a graph (height) function in one of the three coordinate axis directions, restricted to where the slope is not too high, and integrating / sampling from each of these three. We use this unbiased sampling algorithm for applications in Monte Carlo integration, Poisson-disk sampling, and isosurface meshing.

  17. Shadowfax: Moving mesh hydrodynamical integration code

    Vandenbroucke, Bert

    2016-05-01

    Shadowfax simulates galaxy evolution. Written in object-oriented modular C++, it evolves a mixture of gas, subject to the laws of hydrodynamics and gravity, and any collisionless fluid only subject to gravity, such as cold dark matter or stars. For the hydrodynamical integration, it makes use of a (co-) moving Lagrangian mesh. The code has a 2D and 3D version, contains utility programs to generate initial conditions and visualize simulation snapshots, and its input/output is compatible with a number of other simulation codes, e.g. Gadget2 (ascl:0003.001) and GIZMO (ascl:1410.003).

  18. Unbiased Sampling and Meshing of Isosurfaces

    Yan, Dongming; Wallner, Johannes; Wonka, Peter

    2014-01-01

    In this paper, we present a new technique to generate unbiased samples on isosurfaces. An isosurface, F(x,y,z) = c , of a function, F , is implicitly defined by trilinear interpolation of background grid points. The key idea of our approach is that of treating the isosurface within a grid cell as a graph (height) function in one of the three coordinate axis directions, restricted to where the slope is not too high, and integrating / sampling from each of these three. We use this unbiased sampling algorithm for applications in Monte Carlo integration, Poisson-disk sampling, and isosurface meshing.

  19. Comparison of various spring analogy related mesh deformation techniques in two-dimensional airfoil design optimization

    Yang, Y.; Özgen, S.

    2017-06-01

    During the last few decades, CFD (Computational Fluid Dynamics) has developed greatly and has become a more reliable tool for the conceptual phase of aircraft design. This tool is generally combined with an optimization algorithm. In the optimization phase, the need for regenerating the computational mesh might become cumbersome, especially when the number of design parameters is high. For this reason, several mesh generation and deformation techniques have been developed in the past decades. One of the most widely used techniques is the Spring Analogy. There are numerous spring analogy related techniques reported in the literature: linear spring analogy, torsional spring analogy, semitorsional spring analogy, and ball vertex spring analogy. This paper gives the explanation of linear spring analogy method and angle inclusion in the spring analogy method. In the latter case, two di¨erent solution methods are proposed. The best feasible method will later be used for two-dimensional (2D) Airfoil Design Optimization with objective function being to minimize sectional drag for a required lift coe©cient at di¨erent speeds. Design variables used in the optimization include camber and thickness distribution of the airfoil. SU2 CFD is chosen as the §ow solver during the optimization procedure. The optimization is done by using Phoenix ModelCenter Optimization Tool.

  20. Statistical relation between particle contaminations in ultra pure water and defects generated by process tools

    Wali, F.; Knotter, D. Martin; Wortelboer, Ronald; Mud, Auke

    2007-01-01

    Ultra pure water supplied inside the Fab is used in different tools at different stages of processing. Data of the particles measured in ultra pure water was compared with the defect density on wafers processed on these tools and a statistical relation is found Keywords— Yield, defect density,

  1. Next generation interactive tool as a backbone for universal access to electricity

    Moner-Girona, Magda; Puig, Daniel; Mulugetta, Yacob

    2018-01-01

    Energy planning in rural areas and in developing countries most often relies on the outputs of specialised analytical tools, of which only a handful have been developed. Over the years these tools have been upgraded, and the newest among them take into consideration, to a greater or lesser extent...

  2. Development of Next Generation Synthetic Biology Tools for Use in Streptomyces venezuelae

    Phelan, Ryan M.; Sachs, Daniel; Petkiewicz, Shayne J.

    2017-01-01

    precludes rapid and predictable metabolic engineering that is possible in hosts such as Escherichia coli or Saccharomyces cerevisiae. In an effort to improve genetic tools for Streptomyces venezuelae, we developed a suite of standardized, orthogonal integration vectors and an improved method to monitor...... expression system. These tools advance S. venezuelae to be a practical host for future metabolic engineering efforts....

  3. Mesh-based parallel code coupling interface

    Wolf, K.; Steckel, B. (eds.) [GMD - Forschungszentrum Informationstechnik GmbH, St. Augustin (DE). Inst. fuer Algorithmen und Wissenschaftliches Rechnen (SCAI)

    2001-04-01

    MpCCI (mesh-based parallel code coupling interface) is an interface for multidisciplinary simulations. It provides industrial end-users as well as commercial code-owners with the facility to combine different simulation tools in one environment. Thereby new solutions for multidisciplinary problems will be created. This opens new application dimensions for existent simulation tools. This Book of Abstracts gives a short overview about ongoing activities in industry and research - all presented at the 2{sup nd} MpCCI User Forum in February 2001 at GMD Sankt Augustin. (orig.) [German] MpCCI (mesh-based parallel code coupling interface) definiert eine Schnittstelle fuer multidisziplinaere Simulationsanwendungen. Sowohl industriellen Anwender als auch kommerziellen Softwarehersteller wird mit MpCCI die Moeglichkeit gegeben, Simulationswerkzeuge unterschiedlicher Disziplinen miteinander zu koppeln. Dadurch entstehen neue Loesungen fuer multidisziplinaere Problemstellungen und fuer etablierte Simulationswerkzeuge ergeben sich neue Anwendungsfelder. Dieses Book of Abstracts bietet einen Ueberblick ueber zur Zeit laufende Arbeiten in der Industrie und in der Forschung, praesentiert auf dem 2{sup nd} MpCCI User Forum im Februar 2001 an der GMD Sankt Augustin. (orig.)

  4. SED-ML web tools: generate, modify and export standard-compliant simulation studies.

    Bergmann, Frank T; Nickerson, David; Waltemath, Dagmar; Scharm, Martin

    2017-04-15

    The Simulation Experiment Description Markup Language (SED-ML) is a standardized format for exchanging simulation studies independently of software tools. We present the SED-ML Web Tools, an online application for creating, editing, simulating and validating SED-ML documents. The Web Tools implement all current SED-ML specifications and, thus, support complex modifications and co-simulation of models in SBML and CellML formats. Ultimately, the Web Tools lower the bar on working with SED-ML documents and help users create valid simulation descriptions. http://sysbioapps.dyndns.org/SED-ML_Web_Tools/ . fbergman@caltech.edu . © The Author 2017. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com

  5. Comparison of new generation low-complexity flood inundation mapping tools with a hydrodynamic model

    Afshari, Shahab; Tavakoly, Ahmad A.; Rajib, Mohammad Adnan; Zheng, Xing; Follum, Michael L.; Omranian, Ehsan; Fekete, Balázs M.

    2018-01-01

    The objective of this study is to compare two new generation low-complexity tools, AutoRoute and Height Above the Nearest Drainage (HAND), with a two-dimensional hydrodynamic model (Hydrologic Engineering Center-River Analysis System, HEC-RAS 2D). The assessment was conducted on two hydrologically different and geographically distant test-cases in the United States, including the 16,900 km2 Cedar River (CR) watershed in Iowa and a 62 km2 domain along the Black Warrior River (BWR) in Alabama. For BWR, twelve different configurations were set up for each of the models, including four different terrain setups (e.g. with and without channel bathymetry and a levee), and three flooding conditions representing moderate to extreme hazards at 10-, 100-, and 500-year return periods. For the CR watershed, models were compared with a simplistic terrain setup (without bathymetry and any form of hydraulic controls) and one flooding condition (100-year return period). Input streamflow forcing data representing these hypothetical events were constructed by applying a new fusion approach on National Water Model outputs. Simulated inundation extent and depth from AutoRoute, HAND, and HEC-RAS 2D were compared with one another and with the corresponding FEMA reference estimates. Irrespective of the configurations, the low-complexity models were able to produce inundation extents similar to HEC-RAS 2D, with AutoRoute showing slightly higher accuracy than the HAND model. Among four terrain setups, the one including both levee and channel bathymetry showed lowest fitness score on the spatial agreement of inundation extent, due to the weak physical representation of low-complexity models compared to a hydrodynamic model. For inundation depth, the low-complexity models showed an overestimating tendency, especially in the deeper segments of the channel. Based on such reasonably good prediction skills, low-complexity flood models can be considered as a suitable alternative for fast

  6. Cache-Oblivious Mesh Layouts

    Yoon, S; Lindstrom, P; Pascucci, V; Manocha, D

    2005-01-01

    We present a novel method for computing cache-oblivious layouts of large meshes that improve the performance of interactive visualization and geometric processing algorithms. Given that the mesh is accessed in a reasonably coherent manner, we assume no particular data access patterns or cache parameters of the memory hierarchy involved in the computation. Furthermore, our formulation extends directly to computing layouts of multi-resolution and bounding volume hierarchies of large meshes. We develop a simple and practical cache-oblivious metric for estimating cache misses. Computing a coherent mesh layout is reduced to a combinatorial optimization problem. We designed and implemented an out-of-core multilevel minimization algorithm and tested its performance on unstructured meshes composed of tens to hundreds of millions of triangles. Our layouts can significantly reduce the number of cache misses. We have observed 2-20 times speedups in view-dependent rendering, collision detection, and isocontour extraction without any modification of the algorithms or runtime applications

  7. deepTools2: a next generation web server for deep-sequencing data analysis.

    Ramírez, Fidel; Ryan, Devon P; Grüning, Björn; Bhardwaj, Vivek; Kilpert, Fabian; Richter, Andreas S; Heyne, Steffen; Dündar, Friederike; Manke, Thomas

    2016-07-08

    We present an update to our Galaxy-based web server for processing and visualizing deeply sequenced data. Its core tool set, deepTools, allows users to perform complete bioinformatic workflows ranging from quality controls and normalizations of aligned reads to integrative analyses, including clustering and visualization approaches. Since we first described our deepTools Galaxy server in 2014, we have implemented new solutions for many requests from the community and our users. Here, we introduce significant enhancements and new tools to further improve data visualization and interpretation. deepTools continue to be open to all users and freely available as a web service at deeptools.ie-freiburg.mpg.de The new deepTools2 suite can be easily deployed within any Galaxy framework via the toolshed repository, and we also provide source code for command line usage under Linux and Mac OS X. A public and documented API for access to deepTools functionality is also available. © The Author(s) 2016. Published by Oxford University Press on behalf of Nucleic Acids Research.

  8. Tetrahedral meshing via maximal Poisson-disk sampling

    Guo, Jianwei

    2016-02-15

    In this paper, we propose a simple yet effective method to generate 3D-conforming tetrahedral meshes from closed 2-manifold surfaces. Our approach is inspired by recent work on maximal Poisson-disk sampling (MPS), which can generate well-distributed point sets in arbitrary domains. We first perform MPS on the boundary of the input domain, we then sample the interior of the domain, and we finally extract the tetrahedral mesh from the samples by using 3D Delaunay or regular triangulation for uniform or adaptive sampling, respectively. We also propose an efficient optimization strategy to protect the domain boundaries and to remove slivers to improve the meshing quality. We present various experimental results to illustrate the efficiency and the robustness of our proposed approach. We demonstrate that the performance and quality (e.g., minimal dihedral angle) of our approach are superior to current state-of-the-art optimization-based approaches.

  9. Automatic, unstructured mesh optimization for simulation and assessment of tide- and surge-driven hydrodynamics in a longitudinal estuary: St. Johns River

    Bacopoulos, Peter

    2018-05-01

    A localized truncation error analysis with complex derivatives (LTEA+CD) is applied recursively with advanced circulation (ADCIRC) simulations of tides and storm surge for finite element mesh optimization. Mesh optimization is demonstrated with two iterations of LTEA+CD for tidal simulation in the lower 200 km of the St. Johns River, located in northeast Florida, and achieves more than an over 50% decrease in the number of mesh nodes, relating to a twofold increase in efficiency, at a zero cost to model accuracy. The recursively generated meshes using LTEA+CD lead to successive reductions in the global cumulative truncation error associated with the model mesh. Tides are simulated with root mean square error (RMSE) of 0.09-0.21 m and index of agreement (IA) values generally in the 80s and 90s percentage ranges. Tidal currents are simulated with RMSE of 0.09-0.23 m s-1 and IA values of 97% and greater. Storm tide due to Hurricane Matthew 2016 is simulated with RMSE of 0.09-0.33 m and IA values of 75-96%. Analysis of the LTEA+CD results shows the M2 constituent to dominate the node spacing requirement in the St. Johns River, with the M4 and M6 overtides and the STEADY constituent contributing some. Friction is the predominant physical factor influencing the target element size distribution, especially along the main river stem, while frequency (inertia) and Coriolis (rotation) are supplementary contributing factors. The combination of interior- and boundary-type computational molecules, providing near-full coverage of the model domain, renders LTEA+CD an attractive mesh generation/optimization tool for complex coastal and estuarine domains. The mesh optimization procedure using LTEA+CD is automatic and extensible to other finite element-based numerical models. Discussion is provided on the scope of LTEA+CD, the starting point (mesh) of the procedure, the user-specified scaling of the LTEA+CD results, and the iteration (termination) of LTEA+CD for mesh optimization.

  10. TimeMesh – A Serious Game for European Citizenship

    Ricardo Baptista

    2013-03-01

    Full Text Available Serious games are games where the entertainment aspect is not the most relevant motivation or objective. TimeMesh is an online, multi-language, multiplayer, collaborative and social game platform for sharing and acquiring knowledge of the history of European regions. As such it is a serious game with educational characteristics. This article evaluates the use of TimeMesh with students of 13 and 14 years-old. It shows that this game is already a significant learning tool about European citizenship.

  11. Dynamic mesh optimization based on the spring analogy

    Schmidt Jonas

    2014-01-01

    Full Text Available We present an implementation of the spring analogy for three dimensional meshes in OpenFOAM. All parameters of the spring system are treated as fields that can either be pre-defined by the user, or updated at each time step according to specified geometrical regions or diffusion equations. The purpose of the method is to provide a pre-processing tool for mesh optimization. We study three simple test cases, a deformed block, an airfoil and a hill, and we analyze the evolution of skewness, non-orthogonality and aspect ratio during the approach of dynamic equilibrium.

  12. Development of Next Generation Synthetic Biology Tools for Use in Streptomyces venezuelae

    Phelan, Ryan M. [Joint BioEnergy Inst. (JBEI), Emeryville, CA (United States); Univ. of California, Berkeley, CA (United States). QB3 Inst.; Sachs, Daniel [Joint BioEnergy Inst. (JBEI), Emeryville, CA (United States); Petkiewicz, Shayne J. [Joint BioEnergy Inst. (JBEI), Emeryville, CA (United States); Barajas, Jesus F. [Joint BioEnergy Inst. (JBEI), Emeryville, CA (United States); Blake-Hedges, Jacquelyn M. [Univ. of California, Berkeley, CA (United States). Dept. of Chemistry; Thompson, Mitchell G. [Univ. of California, Berkeley, CA (United States). Dept. of Plant & Microbial Biology; Reider Apel, Amanda [Joint BioEnergy Inst. (JBEI), Emeryville, CA (United States); Rasor, Blake J. [Miami Univ., Oxford, Ohio (United States). Dept. of Biology; Katz, Leonard [Univ. of California, Berkeley, CA (United States). QB3 Inst.; Keasling, Jay D. [Joint BioEnergy Inst. (JBEI), Emeryville, CA (United States); Univ. of California, Berkeley, CA (United States). QB3 Inst.; Univ. of California, Berkeley, CA (United States). Dept. of Chemical and Biomolecular Engineering and Department of Bioengineering; Technical Univ. of Denmark, Kogle Alle (Denmark). Novo Nordisk Foundation Center for Biosustainability

    2016-09-07

    Streptomyces have a rich history as producers of important natural products and this genus of bacteria has recently garnered attention for its potential applications in the broader context of synthetic biology. However, the dearth of genetic tools available to control and monitor protein production precludes rapid and predictable metabolic engineering that is possible in hosts such as Escherichia coli or Saccharomyces cerevisiae. In an effort to improve genetic tools for Streptomyces venezuelae, we developed a suite of standardized, orthogonal integration vectors and an improved method to monitor protein production in this host. These tools were applied to characterize heterologous promoters and various attB chromosomal integration sites. A final study leveraged the characterized toolset to demonstrate its use in producing the biofuel precursor bisabolene using a chromosomally integrated expression system. In conclusion, these tools advance S. venezuelae to be a practical host for future metabolic engineering efforts.

  13. A SURVEY on WIRELESS MESH NETWORKS, ROUTING METRICS and PROTOCOLS

    Safak DURUKAN ODABASI

    2013-01-01

    Full Text Available Today, Internet has become an indispensable part of our daily lives. It has a growing user community in many fields from banking transactions to online entertainment. It will be very efficient for users, as the next generation internet access becomes wireless like frequently used services such as cellular phones. But for providing this, a new network is needed to be designed or an existing network must be improved as well as making changes on infrastructure. At this point, mesh network infrastructure arises and offers more sophisticated internet access with less need. The most important advantage of mesh networks is the capability of working without infrastructure. Mesh networks are an additional access technology more than being a renewed one in the next generation wireless networks called 4G. In this study, wireless mesh networks and example applications are mentioned. Base architecture and design factors are emphasized, current routing protocols that are used on wireless mesh networks and routing metrics on which these protocols are based, are explained. Finally, the performance effects of these protocols and metrics on different network topologies are referred.

  14. THM-GTRF: New Spider meshes, New Hydra-TH runs

    Bakosi, Jozsef [Los Alamos National Laboratory; Christon, Mark A. [Los Alamos National Laboratory; Francois, Marianne M. [Los Alamos National Laboratory; Lowrie, Robert B. [Los Alamos National Laboratory; Nourgaliev, Robert [Los Alamos National Laboratory

    2012-06-20

    Progress is reported on computational capabilities for the grid-to-rod-fretting (GTRF) problem of pressurized water reactors. Numeca's Hexpress/Hybrid mesh generator is demonstrated as an excellent alternative to generating computational meshes for complex flow geometries, such as in GTRF. Mesh assessment is carried out using standard industrial computational fluid dynamics practices. Hydra-TH, a simulation code developed at LANL for reactor thermal-hydraulics, is demonstrated on hybrid meshes, containing different element types. A series of new Hydra-TH calculations has been carried out collecting turbulence statistics. Preliminary results on the newly generated meshes are discussed; full analysis will be documented in the L3 milestone, THM.CFD.P5.05, Sept. 2012.

  15. A Linear-Elasticity Solver for Higher-Order Space-Time Mesh Deformation

    Diosady, Laslo T.; Murman, Scott M.

    2018-01-01

    A linear-elasticity approach is presented for the generation of meshes appropriate for a higher-order space-time discontinuous finite-element method. The equations of linear-elasticity are discretized using a higher-order, spatially-continuous, finite-element method. Given an initial finite-element mesh, and a specified boundary displacement, we solve for the mesh displacements to obtain a higher-order curvilinear mesh. Alternatively, for moving-domain problems we use the linear-elasticity approach to solve for a temporally discontinuous mesh velocity on each time-slab and recover a continuous mesh deformation by integrating the velocity. The applicability of this methodology is presented for several benchmark test cases.

  16. Evaluation on machined surface of hardened stainless steel generated by hard turning using coated carbide tools with wiper geometry

    Noordin, M.Y.; Kurniawan, D.; Sharif, S.

    2007-01-01

    Hard turning has been explored to be the finish machining operation for parts made of hardened steel. Its feasibility is determined partially by the quality of the resulting machined surface. This study evaluates the surface integrity of martensitic stainless steel (48 HRC) resulting from hard turning using coated carbide tool with wiper geometry at various cutting speed and feed and compares to that obtained using coated carbide tool with conventional geometry. The wiper coated carbide tool is able to produce machined surface which is of finer finish (Ra is finer than 0.4 μm at most cutting parameters) and yet is similarly inducing only minor microstructural alteration compared to its conventional counterpart. From the view of the chip morphology where continuous type of chip is desired rather than sawtooth chip type, the wiper tool generates continuous chip at almost similar range of cutting parameters compared to the case when using conventional tool. Additionally, the use of wiper tool also induces the preferred compressive residual stress at the machined surface. (author)

  17. A Study on Tooling and Its Effect on Heat Generation and Mechanical Properties of Welded Joints in Friction Stir Welding

    Tikader, Sujoy; Biswas, Pankaj; Puri, Asit Baran

    2018-04-01

    Friction stir welding (FSW) has been the most attracting solid state welding process as it serves numerous advantages like good mechanical, metallurgical properties etc. Non weldable aluminium alloys like 5XXX, 7XXX series can be simply joined by this process. In this present study a mathematical model has been developed and experiments were successfully performed to evaluate mechanical properties of FSW on similar aluminium alloys i.e. AA1100 for different process parameters and mainly two kind of tool geometry (straight cylindrical and conical or cylindrical tapered shaped pin with flat shoulder). Tensile strength and micro hardness for different process parameters are reported of the welded plate sample. It was noticed that in FSW of similar alloy with tool made of SS-310 tool steel, friction is the major contributor for the heat generation. It was seen that tool geometry, tool rotational speed, plunging force by the tool and traverse speed have significant effect on tensile strength and hardness of friction stir welded joints.

  18. BEAT: A Web-Based Boolean Expression Fault-Based Test Case Generation Tool

    Chen, T. Y.; Grant, D. D.; Lau, M. F.; Ng, S. P.; Vasa, V. R.

    2006-01-01

    BEAT is a Web-based system that generates fault-based test cases from Boolean expressions. It is based on the integration of our several fault-based test case selection strategies. The generated test cases are considered to be fault-based, because they are aiming at the detection of particular faults. For example, when the Boolean expression is in…

  19. BEST: Next-Generation Biomedical Entity Search Tool for Knowledge Discovery from Biomedical Literature.

    Sunwon Lee

    Full Text Available As the volume of publications rapidly increases, searching for relevant information from the literature becomes more challenging. To complement standard search engines such as PubMed, it is desirable to have an advanced search tool that directly returns relevant biomedical entities such as targets, drugs, and mutations rather than a long list of articles. Some existing tools submit a query to PubMed and process retrieved abstracts to extract information at query time, resulting in a slow response time and limited coverage of only a fraction of the PubMed corpus. Other tools preprocess the PubMed corpus to speed up the response time; however, they are not constantly updated, and thus produce outdated results. Further, most existing tools cannot process sophisticated queries such as searches for mutations that co-occur with query terms in the literature. To address these problems, we introduce BEST, a biomedical entity search tool. BEST returns, as a result, a list of 10 different types of biomedical entities including genes, diseases, drugs, targets, transcription factors, miRNAs, and mutations that are relevant to a user's query. To the best of our knowledge, BEST is the only system that processes free text queries and returns up-to-date results in real time including mutation information in the results. BEST is freely accessible at http://best.korea.ac.kr.

  20. Different types of mesh fixation for laparoscopic repair of inguinal hernia: A protocol for systematic review and network meta-analysis with randomized controlled trials.

    Wei, Kongyuan; Lu, Cuncun; Ge, Long; Pan, Bei; Yang, Huan; Tian, Jinhui; Cao, Nong

    2018-04-01

    Laparoscopic inguinal hernia repair has become a valid option for repair of an inguinal hernia. Due to there are several types of mesh fixation for laparoscopic repair of inguinal hernia. The study aims to assess and compare the efficacy of different types of mesh fixation for laparoscopic repair of inguinal hernia using network meta-analysis. We will systematically search PubMed, EMBASE the Cochrane library, and Chinese Biomedical Literature Database from their inception to March 2018. Randomized controlled trials (RCTs) that compared the effect of different types of mesh fixation for laparoscopic inguinal hernia repair will be included. The primary outcomes are chronic groin pain, incidence risk of hernia recurrence, and complications. Risk of bias assessment of the included RCTs will be conducted using to Cochrane risk of bias tool. A network meta-analysis will be performed using WinBUGS 1.4.3 software and the result figures will be generated using R x64 3.1.2 software and STATA V.12.0 software. Grading of Recommendations Assessment, Development and Evaluation (GRADE) will be used to assess the quality of evidence. The results of this study will be published in a peer-reviewed journal. Our study will generate evidence of laparoscopic repair of mesh fixation for adult patients with inguinal hernia and provide suggestions for clinical practice or guideline.

  1. ArcCN-Runoff: An ArcGIS tool for generating curve number and runoff maps

    Zhan, X.; Huang, M.-L.

    2004-01-01

    The development and the application of ArcCN-Runoff tool, an extension of ESRI@ ArcGIS software, are reported. This tool can be applied to determine curve numbers and to calculate runoff or infiltration for a rainfall event in a watershed. Implementation of GIS techniques such as dissolving, intersecting, and a curve-number reference table improve efficiency. Technical processing time may be reduced from days, if not weeks, to hours for producing spatially varied curve number and runoff maps. An application example for a watershed in Lyon County and Osage County, Kansas, USA, is presented. ?? 2004 Elsevier Ltd. All rights reserved.

  2. Virtual tool mark generation for efficient striation analysis in forensic science

    Ekstrand, Laura [Iowa State Univ., Ames, IA (United States)

    2012-01-01

    In 2009, a National Academy of Sciences report called for investigation into the scienti c basis behind tool mark comparisons (National Academy of Sciences, 2009). Answering this call, Chumbley et al. (2010) attempted to prove or disprove the hypothesis that tool marks are unique to a single tool. They developed a statistical algorithm that could, in most cases, discern matching and non-matching tool marks made at di erent angles by sequentially numbered screwdriver tips. Moreover, in the cases where the algorithm misinterpreted a pair of marks, an experienced forensics examiner could discern the correct outcome. While this research served to con rm the basic assumptions behind tool mark analysis, it also suggested that statistical analysis software could help to reduce the examiner's workload. This led to a new tool mark analysis approach, introduced in this thesis, that relies on 3D scans of screwdriver tip and marked plate surfaces at the micrometer scale from an optical microscope. These scans are carefully cleaned to remove noise from the data acquisition process and assigned a coordinate system that mathematically de nes angles and twists in a natural way. The marking process is then simulated by using a 3D graphics software package to impart rotations to the tip and take the projection of the tip's geometry in the direction of tool travel. The edge of this projection, retrieved from the 3D graphics software, becomes a virtual tool mark. Using this method, virtual marks are made at increments of 5 and compared to a scan of the evidence mark. The previously developed statistical package from Chumbley et al. (2010) performs the comparison, comparing the similarity of the geometry of both marks to the similarity that would occur due to random chance. The resulting statistical measure of the likelihood of the match informs the examiner of the angle of the best matching virtual mark, allowing the examiner to focus his/her mark analysis on a smaller range of angles

  3. miRanalyzer: a microRNA detection and analysis tool for next-generation sequencing experiments.

    Hackenberg, Michael; Sturm, Martin; Langenberger, David; Falcón-Pérez, Juan Manuel; Aransay, Ana M

    2009-07-01

    Next-generation sequencing allows now the sequencing of small RNA molecules and the estimation of their expression levels. Consequently, there will be a high demand of bioinformatics tools to cope with the several gigabytes of sequence data generated in each single deep-sequencing experiment. Given this scene, we developed miRanalyzer, a web server tool for the analysis of deep-sequencing experiments for small RNAs. The web server tool requires a simple input file containing a list of unique reads and its copy numbers (expression levels). Using these data, miRanalyzer (i) detects all known microRNA sequences annotated in miRBase, (ii) finds all perfect matches against other libraries of transcribed sequences and (iii) predicts new microRNAs. The prediction of new microRNAs is an especially important point as there are many species with very few known microRNAs. Therefore, we implemented a highly accurate machine learning algorithm for the prediction of new microRNAs that reaches AUC values of 97.9% and recall values of up to 75% on unseen data. The web tool summarizes all the described steps in a single output page, which provides a comprehensive overview of the analysis, adding links to more detailed output pages for each analysis module. miRanalyzer is available at http://web.bioinformatics.cicbiogune.es/microRNA/.

  4. Multigrid for refined triangle meshes

    Shapira, Yair

    1997-02-01

    A two-level preconditioning method for the solution of (locally) refined finite element schemes using triangle meshes is introduced. In the isotropic SPD case, it is shown that the condition number of the preconditioned stiffness matrix is bounded uniformly for all sufficiently regular triangulations. This is also verified numerically for an isotropic diffusion problem with highly discontinuous coefficients.

  5. User-friendly tool for power flow analysis and distributed generation ...

    The intent of power distribution companies (DISCOs) is to deliver electric power to their ... One major way to minimise power loss on a given power system is to install ... The accuracy and reliability of the developed tool was validated using ...

  6. A tool for automatic generation of RTL-level VHDL description of RNS FIR filters

    Re, Andrea Del; Nannarelli, Alberto; Re, Marco

    2004-01-01

    Although digital filters based on the Residue Number System (RNS) show high performance and low power dissipation, RNS filters are not widely used in DSP systems, because of the complexity of the algorithms involved. We present a tool to design RNS FIR filters which hides the RNS algorithms to th...

  7. Making Sense of Conceptual Tools in Student-Generated Cases: Student Teachers' Problem-Solving Processes

    Jahreie, Cecilie Flo

    2010-01-01

    This article examines the way student teachers make sense of conceptual tools when writing cases. In order to understand the problem-solving process, an analysis of the interactions is conducted. The findings show that transforming practical experiences into theoretical reflection is not a straightforward matter. To be able to elaborate on the…

  8. MAKER2: an annotation pipeline and genome-database management tool for second-generation genome projects.

    Holt, Carson; Yandell, Mark

    2011-12-22

    Second-generation sequencing technologies are precipitating major shifts with regards to what kinds of genomes are being sequenced and how they are annotated. While the first generation of genome projects focused on well-studied model organisms, many of today's projects involve exotic organisms whose genomes are largely terra incognita. This complicates their annotation, because unlike first-generation projects, there are no pre-existing 'gold-standard' gene-models with which to train gene-finders. Improvements in genome assembly and the wide availability of mRNA-seq data are also creating opportunities to update and re-annotate previously published genome annotations. Today's genome projects are thus in need of new genome annotation tools that can meet the challenges and opportunities presented by second-generation sequencing technologies. We present MAKER2, a genome annotation and data management tool designed for second-generation genome projects. MAKER2 is a multi-threaded, parallelized application that can process second-generation datasets of virtually any size. We show that MAKER2 can produce accurate annotations for novel genomes where training-data are limited, of low quality or even non-existent. MAKER2 also provides an easy means to use mRNA-seq data to improve annotation quality; and it can use these data to update legacy annotations, significantly improving their quality. We also show that MAKER2 can evaluate the quality of genome annotations, and identify and prioritize problematic annotations for manual review. MAKER2 is the first annotation engine specifically designed for second-generation genome projects. MAKER2 scales to datasets of any size, requires little in the way of training data, and can use mRNA-seq data to improve annotation quality. It can also update and manage legacy genome annotation datasets.

  9. Creating and Exploring Huge Parameter Spaces: Interactive Evolution as a Tool for Sound Generation

    Dahlstedt, Palle

    2001-01-01

    In this paper, a program is presented that applies interactive evolution to sound generation, i.e., preferred individuals are repeatedly selected from a population of genetically bred sound objects, created with various synthesis and pattern generation algorithms. This simplifies aural exploration...... applications. It is also shown how this technique can be used to simplify sound design in standard hardware synthesizers, a task normally avoided by most musicians, due to the required amount of technical understanding....

  10. THEORY OF GENERATIONS AS A TOOL FOR ANALYSIS, FORMATION AND DEVELOPMENT OF LABOUR POTENTIAL

    I. M. Gurova; S. Sh. Evdokimova

    2016-01-01

    The modern Strauss–Howe generational theory created at the intersection of economic, sociological, historical and psychological Sciences. She focuses primarily on difference of attitudes of generations, due to the specific social environment, corresponding to a certain period of time. Such a perspective on issues related to the human factor in the economy, has recently attracted interest not only from researchers, but also finds application for solving practical problems in some fields of bus...

  11. Reactor physics verification of the MCNP6 unstructured mesh capability

    Burke, T. P.; Kiedrowski, B. C.; Martz, R. L.; Martin, W. R.

    2013-01-01

    The Monte Carlo software package MCNP6 has the ability to transport particles on unstructured meshes generated from the Computed-Aided Engineering software Abaqus. Verification is performed using benchmarks with features relevant to reactor physics - Big Ten and the C5G7 computational benchmark. Various meshing strategies are tested and results are compared to reference solutions. Computational performance results are also given. The conclusions show MCNP6 is capable of producing accurate calculations for reactor physics geometries and the computational requirements for small lattice benchmarks are reasonable on modern computing platforms. (authors)

  12. Reactor physics verification of the MCNP6 unstructured mesh capability

    Burke, T. P. [Department of Nuclear Engineering and Radiological Sciences, University of Michigan, 2355 Bonisteel Boulevard, Ann Arbor, MI 48109 (United States); Kiedrowski, B. C.; Martz, R. L. [X-Computational Physics Division, Monte Carlo Codes Group, Los Alamos National Laboratory, P.O. Box 1663, Los Alamos, NM 87545 (United States); Martin, W. R. [Department of Nuclear Engineering and Radiological Sciences, University of Michigan, 2355 Bonisteel Boulevard, Ann Arbor, MI 48109 (United States)

    2013-07-01

    The Monte Carlo software package MCNP6 has the ability to transport particles on unstructured meshes generated from the Computed-Aided Engineering software Abaqus. Verification is performed using benchmarks with features relevant to reactor physics - Big Ten and the C5G7 computational benchmark. Various meshing strategies are tested and results are compared to reference solutions. Computational performance results are also given. The conclusions show MCNP6 is capable of producing accurate calculations for reactor physics geometries and the computational requirements for small lattice benchmarks are reasonable on modern computing platforms. (authors)

  13. Resterilized Polypropylene Mesh for Inguinal Hernia Repair

    2018-04-19

    Apr 19, 2018 ... Conclusion: The use of sterilized polypropylene mesh for the repair of inguinal ... and nonabsorbable materials to reduce the tissue–mesh. INTRODUCTION ... which we have been practicing in our center since we introduced ...

  14. Identification of leaky steam generators by iodine mapping technique and development of tools for cutting of tubes of steam generators of Indian PHWRS

    Subba Rao, D.

    2006-01-01

    Kakrapar Atomic Power Station (2X220 MWe) located in Mandvi Taluka of Surat District in the state of Gujarat is the fifth Nuclear Power Station of the country. It has got an excellent record in the field of operation, safety, public awareness and emergency preparedness. KAPS Unit -1 achieved first criticality in Sep-1992 and was declared for commercial operation in may-1993. KAPS Unit -2 achieved first criticality in Jan-1995 and was declared for commercial operation in Sep-1995. So far station has generated about 30 billion units.Unit-1 achieved 98.4% and was graded as a world's No.1 in year 2002 amongst all CANDU type reactors. KAPS Unit -1 has made another record of operating continuously for more than 300 days in Indian PHWR s operating history. This paper mainly deals with the Indian PHWRs Steam Generators (SG) tube leaks, leaky steam generator identification by Iodine mapping, and development of special tool for cutting, removal and plugging of leaky tubes. These Steam Generators are designed by M/s Kraft Werke Union (KWU) of Siemens Group, West Germany, and Manufactured by M/s ENSA, SPAIN for Unit- 1 and by M/s MAN-GHH, Germany for Unit- 2. First time in October-2002 one of the Steam Generators of Unit-1 developed tube leak. To identify leaky Steam Generator, KAPS has established a method of Iodine mapping. With that the leaky SG was identified in very short time and corrective actions were taken immediately. Total three tube leaks (two in SG-4 of Unit-1 and one in SG-1 of unit-2) were experienced in both Units'. Following observations were made on SG tubes failure: All failures were in cold leg side; All Failures / deterioration locations were in front of main feed water nozzle; All Failures / deterioration locations were observed to be just above tube support plate (TSP) number 4 or 5; Deterioration ( i.e. wall thinning) observed from OD side and these tubes were adjacent to failed tubes; In all the three incidents, failed / deteriorated tubes were

  15. Identifying problems and generating recommendations for enhancing complex systems: applying the abstraction hierarchy framework as an analytical tool.

    Xu, Wei

    2007-12-01

    This study adopts J. Rasmussen's (1985) abstraction hierarchy (AH) framework as an analytical tool to identify problems and pinpoint opportunities to enhance complex systems. The process of identifying problems and generating recommendations for complex systems using conventional methods is usually conducted based on incompletely defined work requirements. As the complexity of systems rises, the sheer mass of data generated from these methods becomes unwieldy to manage in a coherent, systematic form for analysis. There is little known work on adopting a broader perspective to fill these gaps. AH was used to analyze an aircraft-automation system in order to further identify breakdowns in pilot-automation interactions. Four steps follow: developing an AH model for the system, mapping the data generated by various methods onto the AH, identifying problems based on the mapped data, and presenting recommendations. The breakdowns lay primarily with automation operations that were more goal directed. Identified root causes include incomplete knowledge content and ineffective knowledge structure in pilots' mental models, lack of effective higher-order functional domain information displayed in the interface, and lack of sufficient automation procedures for pilots to effectively cope with unfamiliar situations. The AH is a valuable analytical tool to systematically identify problems and suggest opportunities for enhancing complex systems. It helps further examine the automation awareness problems and identify improvement areas from a work domain perspective. Applications include the identification of problems and generation of recommendations for complex systems as well as specific recommendations regarding pilot training, flight deck interfaces, and automation procedures.

  16. Kinetic solvers with adaptive mesh in phase space

    Arslanbekov, Robert R.; Kolobov, Vladimir I.; Frolova, Anna A.

    2013-12-01

    An adaptive mesh in phase space (AMPS) methodology has been developed for solving multidimensional kinetic equations by the discrete velocity method. A Cartesian mesh for both configuration (r) and velocity (v) spaces is produced using a “tree of trees” (ToT) data structure. The r mesh is automatically generated around embedded boundaries, and is dynamically adapted to local solution properties. The v mesh is created on-the-fly in each r cell. Mappings between neighboring v-space trees is implemented for the advection operator in r space. We have developed algorithms for solving the full Boltzmann and linear Boltzmann equations with AMPS. Several recent innovations were used to calculate the discrete Boltzmann collision integral with dynamically adaptive v mesh: the importance sampling, multipoint projection, and variance reduction methods. We have developed an efficient algorithm for calculating the linear Boltzmann collision integral for elastic and inelastic collisions of hot light particles in a Lorentz gas. Our AMPS technique has been demonstrated for simulations of hypersonic rarefied gas flows, ion and electron kinetics in weakly ionized plasma, radiation and light-particle transport through thin films, and electron streaming in semiconductors. We have shown that AMPS allows minimizing the number of cells in phase space to reduce the computational cost and memory usage for solving challenging kinetic problems.

  17. GOLLUM: a next-generation simulation tool for electron, thermal and spin transport

    Ferrer, J; García-Suárez, V M; Rodríguez-Ferradás, R; Lambert, C J; Manrique, D Zs; Visontai, D; Grace, I; Bailey, S W D; Gillemot, K; Sadeghi, Hatef; Algharagholy, L A; Oroszlany, L

    2014-01-01

    We have developed an efficient simulation tool ‘GOLLUM’ for the computation of electrical, spin and thermal transport characteristics of complex nanostructures. The new multi-scale, multi-terminal tool addresses a number of new challenges and functionalities that have emerged in nanoscale-scale transport over the past few years. To illustrate the flexibility and functionality of GOLLUM, we present a range of demonstrator calculations encompassing charge, spin and thermal transport, corrections to density functional theory such as local density approximation +U (LDA+U) and spectral adjustments, transport in the presence of non-collinear magnetism, the quantum Hall effect, Kondo and Coulomb blockade effects, finite-voltage transport, multi-terminal transport, quantum pumps, superconducting nanostructures, environmental effects, and pulling curves and conductance histograms for mechanically-controlled break-junction experiments. (paper)

  18. SEQ-POINTER: Next generation, planetary spacecraft remote sensing science observation design tool

    Boyer, Jeffrey S.

    1994-11-01

    Since Mariner, NASA-JPL planetary missions have been supported by ground software to plan and design remote sensing science observations. The software used by the science and sequence designers to plan and design observations has evolved with mission and technological advances. The original program, PEGASIS (Mariners 4, 6, and 7), was re-engineered as POGASIS (Mariner 9, Viking, and Mariner 10), and again later as POINTER (Voyager and Galileo). Each of these programs were developed under technological, political, and fiscal constraints which limited their adaptability to other missions and spacecraft designs. Implementation of a multi-mission tool, SEQ POINTER, under the auspices of the JPL Multimission Operations Systems Office (MOSO) is in progress. This version has been designed to address the limitations experienced on previous versions as they were being adapted to a new mission and spacecraft. The tool has been modularly designed with subroutine interface structures to support interchangeable celestial body and spacecraft definition models. The computational and graphics modules have also been designed to interface with data collected from previous spacecraft, or on-going observations, which describe the surface of each target body. These enhancements make SEQ POINTER a candidate for low-cost mission usage, when a remote sensing science observation design capability is required. The current and planned capabilities of the tool will be discussed. The presentation will also include a 5-10 minute video presentation demonstrating the capabilities of a proto-Cassini Project version that was adapted to test the tool. The work described in this abstract was performed by the Jet Propulsion Laboratory, California Institute of Technology, under contract to the National Aeronautics and Space Administration.

  19. Generation of orientation tools for automated zebrafish screening assays using desktop 3D printing

    Wittbrodt, Jonas N.; Liebel, Urban; Gehrig, Jochen

    2014-01-01

    Background The zebrafish has been established as the main vertebrate model system for whole organism screening applications. However, the lack of consistent positioning of zebrafish embryos within wells of microtiter plates remains an obstacle for the comparative analysis of images acquired in automated screening assays. While technical solutions to the orientation problem exist, dissemination is often hindered by the lack of simple and inexpensive ways of distributing and duplicating tools. ...

  20. Integrated approach for fusion multi-physics coupled analyses based on hybrid CAD and mesh geometries

    Qiu, Yuefeng, E-mail: yuefeng.qiu@kit.edu; Lu, Lei; Fischer, Ulrich

    2015-10-15

    Highlights: • Integrated approach for neutronics, thermal and structural analyses was developed. • MCNP5/6, TRIPOLI-4 were coupled with CFX, Fluent and ANSYS Workbench. • A novel meshing approach has been proposed for describing MC geometry. - Abstract: Coupled multi-physics analyses on fusion reactor devices require high-fidelity neutronic models, and flexible, accurate data exchanging between various calculation codes. An integrated coupling approach has been developed to enable the conversion of CAD, mesh, or hybrid geometries for Monte Carlo (MC) codes MCNP5/6, TRIPOLI-4, and translation of nuclear heating data for CFD codes Fluent, CFX and structural mechanical software ANSYS Workbench. The coupling approach has been implemented based on SALOME platform with CAD modeling, mesh generation and data visualization capabilities. A novel meshing approach has been developed for generating suitable meshes for MC geometry descriptions. The coupling approach has been concluded to be reliable and efficient after verification calculations of several application cases.

  1. Poly-generation planning: useful lessons from models and decision support tools

    Rong, Aiying; Lahdelma, Risto; Grunow, Martin

    2009-01-01

    of energy-efficient production technologies has experienced cogeneration and tri-generation and now is moving towards poly-generation. All these aspects have added new dimension in energy planning. The liberalized energy market requires techniques for planning under uncertainty. The growing environmental...... awareness calls for explicit handling of the impacts of energy generation on environment. Advanced production technologies require more sophisticated models for planning. The energy sector is one of the core application areas for operations research, decision sciences and intelligent techniques......Increasing environmental concerns and the trends towards deregulation of energy markets have become an integral part of energy policy planning. Consequently, the requirement for environmentally sound energy production technologies has gained much ground in the energy business. The development...

  2. Poly-generation planning: useful lessons from models and decision support tools

    Rong, Aiying; Lahdelma, Risto; Grunow, Martin

    2009-01-01

    Increasing environmental concerns and the trends towards deregulation of energy markets have become an integral part of energy policy planning. Consequently, the requirement for environmentally sound energy production technologies has gained much ground in the energy business. The development...... of energy-efficient production technologies has experienced cogeneration and tri-generation and now is moving towards poly-generation. All these aspects have added new dimension in energy planning. The liberalized energy market requires techniques for planning under uncertainty. The growing environmental...... awareness calls for explicit handling of the impacts of energy generation on environment. Advanced production technologies require more sophisticated models for planning. The energy sector is one of the core application areas for operations research, decision sciences and intelligent techniques...

  3. Software module for geometric product modeling and NC tool path generation

    Sidorenko, Sofija; Dukovski, Vladimir

    2003-01-01

    The intelligent CAD/CAM system named VIRTUAL MANUFACTURE is created. It is consisted of four intelligent software modules: the module for virtual NC machine creation, the module for geometric product modeling and automatic NC path generation, the module for virtual NC machining and the module for virtual product evaluation. In this paper the second intelligent software module is presented. This module enables feature-based product modeling carried out via automatic saving of the designed product geometric features as knowledge data. The knowledge data are afterwards applied for automatic NC program generation for the designed product NC machining. (Author)

  4. Practical implementation of tetrahedral mesh reconstruction in emission tomography

    Boutchko, R.; Sitek, A.; Gullberg, G. T.

    2013-05-01

    This paper presents a practical implementation of image reconstruction on tetrahedral meshes optimized for emission computed tomography with parallel beam geometry. Tetrahedral mesh built on a point cloud is a convenient image representation method, intrinsically three-dimensional and with a multi-level resolution property. Image intensities are defined at the mesh nodes and linearly interpolated inside each tetrahedron. For the given mesh geometry, the intensities can be computed directly from tomographic projections using iterative reconstruction algorithms with a system matrix calculated using an exact analytical formula. The mesh geometry is optimized for a specific patient using a two stage process. First, a noisy image is reconstructed on a finely-spaced uniform cloud. Then, the geometry of the representation is adaptively transformed through boundary-preserving node motion and elimination. Nodes are removed in constant intensity regions, merged along the boundaries, and moved in the direction of the mean local intensity gradient in order to provide higher node density in the boundary regions. Attenuation correction and detector geometric response are included in the system matrix. Once the mesh geometry is optimized, it is used to generate the final system matrix for ML-EM reconstruction of node intensities and for visualization of the reconstructed images. In dynamic PET or SPECT imaging, the system matrix generation procedure is performed using a quasi-static sinogram, generated by summing projection data from multiple time frames. This system matrix is then used to reconstruct the individual time frame projections. Performance of the new method is evaluated by reconstructing simulated projections of the NCAT phantom and the method is then applied to dynamic SPECT phantom and patient studies and to a dynamic microPET rat study. Tetrahedral mesh-based images are compared to the standard voxel-based reconstruction for both high and low signal-to-noise ratio

  5. Practical implementation of tetrahedral mesh reconstruction in emission tomography

    Boutchko, R; Gullberg, G T; Sitek, A

    2013-01-01

    This paper presents a practical implementation of image reconstruction on tetrahedral meshes optimized for emission computed tomography with parallel beam geometry. Tetrahedral mesh built on a point cloud is a convenient image representation method, intrinsically three-dimensional and with a multi-level resolution property. Image intensities are defined at the mesh nodes and linearly interpolated inside each tetrahedron. For the given mesh geometry, the intensities can be computed directly from tomographic projections using iterative reconstruction algorithms with a system matrix calculated using an exact analytical formula. The mesh geometry is optimized for a specific patient using a two stage process. First, a noisy image is reconstructed on a finely-spaced uniform cloud. Then, the geometry of the representation is adaptively transformed through boundary-preserving node motion and elimination. Nodes are removed in constant intensity regions, merged along the boundaries, and moved in the direction of the mean local intensity gradient in order to provide higher node density in the boundary regions. Attenuation correction and detector geometric response are included in the system matrix. Once the mesh geometry is optimized, it is used to generate the final system matrix for ML-EM reconstruction of node intensities and for visualization of the reconstructed images. In dynamic PET or SPECT imaging, the system matrix generation procedure is performed using a quasi-static sinogram, generated by summing projection data from multiple time frames. This system matrix is then used to reconstruct the individual time frame projections. Performance of the new method is evaluated by reconstructing simulated projections of the NCAT phantom and the method is then applied to dynamic SPECT phantom and patient studies and to a dynamic microPET rat study. Tetrahedral mesh-based images are compared to the standard voxel-based reconstruction for both high and low signal-to-noise ratio

  6. Clustering of Cochlear Oscillations in Frequency Plateaus as a Tool to Investigate SOAE Generation

    Epp, Bastian; Wit, Hero; van Dijk, Pim

    2016-01-01

    of coupled oscillators (OAM) [7] are also found in a transmission line model (TLM) which is able to generate realistic SOAEs [2] and if these frequency plateaus can be used to explain the formation of SOAEs. The simulations showed a clustering of oscillators along the simulated basilar membrane Both, the OAM...

  7. Development of the tool for generating ORIGEN2 library based on JENDL-3.2 for FBR

    Ohkawachi, Yasushi; Fukushima, Manabu

    1999-05-01

    ORIGEN2 is one of the most widely-used burnup analysis code in the world. This code has one-grouped cross section libraries compiled for various types of reactors. However, these libraries have some problems. One is that these libraries were developed from old nuclear data libraries (ENDF/B-IV,V) and the other is that core and fuel designs from which these libraries are generated do not match the current analysis. In order to solve the problems, analysis tool is developed for generating ORIGEN2 library from JENDL-3.2 considering multi-energy neutron spectrum. And eight new libraries are prepared using this tool for analysis of sodium-cooled FBR. These new libraries are prepared for eight kinds of cores in total. Seven of them are made by changing core size (small core - large core), fuel type (oxide, nitride, metal) and Pu vector as a parameter. The eighth one is a Pu burner core. Burnup calculation using both new and original libraries, shows large difference in buildup or depletion numbers of nuclides among the libraries. It is estimated that the analysis result is greatly influenced by the neutron spectrum which is used in collapse of cross section. By using this tool or new libraries, it seems to improve evaluation accuracy of buildup or depletion numbers of nuclides in transmutation research on FBR fuel cycle. (author)

  8. A complete CFD tool for flooding forecasting

    Nguyen, V.T.; Eberl, H.

    2004-01-01

    Every year, flooding does not only cause property damage of billions of dollars, but also threats to millions of human life around the world. The ability to accurately predict the extreme flooding in urban areas is of obvious importance in order to reduce flooding risks and to improve public safety. In this paper, a complete computational tool is presented that includes pre-processing, meshing, calculating and post-processing modules. The pre-processing procedure is used to interpolate the geometry of the river and floodplains where the data can not be obtained directly from measurements. The meshing procedure is implemented by a triangle mesh generator. The computational procedure is based on a Finite Element Method to discretize the two-dimensional depth-averaged equations for shallow water flow. The post-processing procedure, finally, is interfaced with Geographic Information Systems (GIS), which can serve as a tool for monitoring and as an early warning system. The numerical model is verified and calibrated through many practical projects of flood protection for rivers in Germany. The numerical results show a very good agreement with data from the field survey, as well as data from past flood events. Thus the numerical model can be used as an important tool for flood prediction. (author)

  9. Voltammetry at micro-mesh electrodes

    Wadhawan Jay D.

    2003-01-01

    Full Text Available The voltammetry at three micro-mesh electrodes is explored. It is found that at sufficiently short experimental durations, the micro-mesh working electrode first behaves as an ensemble of microband electrodes, then follows the behaviour anticipated for an array of diffusion-independent micro-ring electrodes of the same perimeter as individual grid-squares within the mesh. During prolonged electrolysis, the micro-mesh electrode follows that behaviour anticipated theoretically for a cubically-packed partially-blocked electrode. Application of the micro-mesh electrode for the electrochemical determination of carbon dioxide in DMSO electrolyte solutions is further illustrated.

  10. Environmental epigenetics: A promising venue for developing next-generation pollution biomonitoring tools in marine invertebrates.

    Suarez-Ulloa, Victoria; Gonzalez-Romero, Rodrigo; Eirin-Lopez, Jose M

    2015-09-15

    Environmental epigenetics investigates the cause-effect relationships between specific environmental factors and the subsequent epigenetic modifications triggering adaptive responses in the cell. Given the dynamic and potentially reversible nature of the different types of epigenetic marks, environmental epigenetics constitutes a promising venue for developing fast and sensible biomonitoring programs. Indeed, several epigenetic biomarkers have been successfully developed and applied in traditional model organisms (e.g., human and mouse). Nevertheless, the lack of epigenetic knowledge in other ecologically and environmentally relevant organisms has hampered the application of these tools in a broader range of ecosystems, most notably in the marine environment. Fortunately, that scenario is now changing thanks to the growing availability of complete reference genome sequences along with the development of high-throughput DNA sequencing and bioinformatic methods. Altogether, these resources make the epigenetic study of marine organisms (and more specifically marine invertebrates) a reality. By building on this knowledge, the present work provides a timely perspective highlighting the extraordinary potential of environmental epigenetic analyses as a promising source of rapid and sensible tools for pollution biomonitoring, using marine invertebrates as sentinel organisms. This strategy represents an innovative, groundbreaking approach, improving the conservation and management of natural resources in the oceans. Copyright © 2015 Elsevier Ltd. All rights reserved.

  11. Protein engineering for metabolic engineering: current and next-generation tools

    Marcheschi, Ryan J.; Gronenberg, Luisa S.; Liao, James C.

    2014-01-01

    Protein engineering in the context of metabolic engineering is increasingly important to the field of industrial biotechnology. As the demand for biologically-produced food, fuels, chemicals, food additives, and pharmaceuticals continues to grow, the ability to design and modify proteins to accomplish new functions will be required to meet the high productivity demands for the metabolism of engineered organisms. This article reviews advances of selecting, modeling, and engineering proteins to improve or alter their activity. Some of the methods have only recently been developed for general use and are just beginning to find greater application in the metabolic engineering community. We also discuss methods of generating random and targeted diversity in proteins to generate mutant libraries for analysis. Recent uses of these techniques to alter cofactor use, produce non-natural amino acids, alcohols, and carboxylic acids, and alter organism phenotypes are presented and discussed as examples of the successful engineering of proteins for metabolic engineering purposes. PMID:23589443

  12. A versatile generator of nanoparticle aerosols. A novel tool in environmental and occupational exposure assessment.

    Clemente, Alberto; Lobera, M Pilar; Balas, Francisco; Santamaria, Jesus

    2018-06-01

    The increasing presence of nanotechnology on the market entails a growing probability of finding ENMs in the environment. Nanoparticles aerosols are a yet unknown risk for human and environmental exposure that may normally occur at any point during the nanomaterial lifecycle. There is a research gap in standardized methods to assess the exposure to airborne nanoparticles in different environments. The controllable generation of nanoparticle aerosols has long been a challenging objective for researchers and industries dealing with airborne nanoparticles. In this work, a versatile system to generate nanoparticulate aerosols has been designed. The system allows the production of both i) instantaneous nanoparticle clouds and ii) continuous nanoparticle streams with quasi-stable values of particle concentration and size distribution. This novel device uses a compressed-air pressure pulse to disperse the target material into either the testing environment (instantaneous cloud formation) or a secondary chamber, from which a continuous aerosol stream can be drawn, with a tunable nanoparticle concentration. The system is robust, highly versatile and easy to operate, enabling reproducible generation of aerosols from a variety of sources. The system has been verified with four dry nanomaterials: TiO 2 , ZnO, CuO and CNT bundles. Copyright © 2017 Elsevier B.V. All rights reserved.

  13. Understanding the generative capacity of analogies as a tool for explanation

    Wong, E. David

    1993-12-01

    Psychological studies have typically portrayed analogical reasoning as a process of schema transfer from a familiar domain of understanding to a problem situation. These studies have usually examined analogical reasoning in contexts where (a) individuals possess or have been provided with appropriate, problem-specific schema, (b) the nature of the problem and/or solution is fairly well defined, and (c) ideal analogies are provided or suggested by an outside source. This study examines analogical reasoning in contexts where understanding is generated from loosely organized, incomplete prior knowledge rather than transferred from a well-structured domain of understanding. In addition, participants were asked to create, apply, and modify their own analogies - as opposed to applying a given analogy - as a heuristic for constructing, evaluating, and modifying their explanations for a particular scientific phenomena. The results provide empirical support for the generative properties of analogies; that is, analogies can stimulate new inferences and insight. Furthermore, under specific conditions, individuals can productively harness the generative capacity of their own analogies to advance their conceptual understanding of scientific phenomena.

  14. H-Morph: An indirect approach to advancing front hex meshing

    OWEN,STEVEN J.; SAIGAL,SUNIL

    2000-05-30

    H-Morph is a new automatic algorithm for the generation of a hexahedral-dominant finite element mesh for arbitrary volumes. The H-Morph method starts with an initial tetrahedral mesh and systematically transforms and combines tetrahedral into hexahedra. It uses an advancing front technique where the initial front consists of a set of prescribed quadrilateral surface facets. Fronts are individually processed by recovering each of the six quadrilateral faces of a hexahedron from the tetrahedral mesh. Recovery techniques similar to those used in boundary constrained Delaunay mesh generation are used. Tetrahedral internal to the six hexahedral faces are then removed and a hexahedron is formed. At any time during the H-Morph procedure a valid mixed hexahedral-tetrahedral mesh is in existence within the volume. The procedure continues until no tetrahedral remain within the volume, or tetrahedral remain which cannot be transformed or combined into valid hexahedral elements. Any remaining tetrahedral are typically towards the interior of the volume, generally a less critical region for analysis. Transition from tetrahedral to hexahedra in the final mesh is accomplished through pyramid shaped elements. Advantages of the proposed method include its ability to conform to an existing quadrilateral surface mesh, its ability to mesh without the need to decompose or recognize special classes of geometry, and its characteristic well-aligned layers of elements parallel to the boundary. Example test cases are presented on a variety of models.

  15. Adaptive Mesh Refinement in CTH

    Crawford, David

    1999-01-01

    This paper reports progress on implementing a new capability of adaptive mesh refinement into the Eulerian multimaterial shock- physics code CTH. The adaptivity is block-based with refinement and unrefinement occurring in an isotropic 2:1 manner. The code is designed to run on serial, multiprocessor and massive parallel platforms. An approximate factor of three in memory and performance improvements over comparable resolution non-adaptive calculations has-been demonstrated for a number of problems

  16. ADRIA: A summary of different tools for motor and generator field service activities in the international market

    Krell, M.; Wuehler, H.

    2012-01-01

    AREVA Drive In Application (ADRIA) – is an innovative program for use at sites all over the world. ADRIA is a full package for medium voltage motor and generator services growth in the long-term, which presents a wide range of applications for nuclear- , thermal- , renewable hydro power plants and all industrial plants. ADRIA support our customers with an optimized product in MV motor management systems e.g., dismantling, maintenance, repair, tools, containers with full equipped working tools or measure equipment. Are you planning to open a site? You will need special equipment and very special diagnostic methods for electrical analysis for preventive maintenance scheduling and planning; special support from experts, ADRIA is a good, unitized, and flexible solution which can be adapted to different customer requirements. • Technical modules show low service and maintenance costs due to the life time of electrical machines. • Measurement module – a solution for the actual condition of the MV motors and generators (vibration, partial discharge, infrared technology, cable analysis, motor – current – analysis) • Documentation module – for all projects in joining with quality management • Tools module – support for all activities in service (working tools, office / store containers) • Man power module – for all service in the field (Experts with world wide experience) • For pre-maintenance we use, for example, infrared technology. Experts monitor different cycles and find out which parts are creating problems. The maintenance strategy will be clear, shorter and more effective for the next outage. ADRIA offers great advantages for the global maintenance market providing time, efficiency and preventive scheduling, efforts and saving of costs. This new technology method has been successfully applied in various motor refurbishment projects in German, Spain or Brazil nuclear power plants since 2010. Further follow-up orders in German and

  17. Integrated Simulation Development and Decision Support Tool-Set for Utility Market and Distributed Solar Power Generation Electricore, Inc.

    Daye, Tony [Green Power Labs (GPL), San Diego, CA (United States)

    2013-09-30

    This project will enable utilities to develop long-term strategic plans that integrate high levels of renewable energy generation, and to better plan power system operations under high renewable penetration. The program developed forecast data streams for decision support and effective integration of centralized and distributed solar power generation in utility operations. This toolset focused on real time simulation of distributed power generation within utility grids with the emphasis on potential applications in day ahead (market) and real time (reliability) utility operations. The project team developed and demonstrated methodologies for quantifying the impact of distributed solar generation on core utility operations, identified protocols for internal data communication requirements, and worked with utility personnel to adapt the new distributed generation (DG) forecasts seamlessly within existing Load and Generation procedures through a sophisticated DMS. This project supported the objectives of the SunShot Initiative and SUNRISE by enabling core utility operations to enhance their simulation capability to analyze and prepare for the impacts of high penetrations of solar on the power grid. The impact of high penetration solar PV on utility operations is not only limited to control centers, but across many core operations. Benefits of an enhanced DMS using state-of-the-art solar forecast data were demonstrated within this project and have had an immediate direct operational cost savings for Energy Marketing for Day Ahead generation commitments, Real Time Operations, Load Forecasting (at an aggregate system level for Day Ahead), Demand Response, Long term Planning (asset management), Distribution Operations, and core ancillary services as required for balancing and reliability. This provided power system operators with the necessary tools and processes to operate the grid in a reliable manner under high renewable penetration.

  18. BPFlexTemplate: A Business Process template generation tool based on similarity and flexibility

    Latifa Ilahi

    2017-01-01

    Full Text Available In large organizations with multiple organizational units, process variants emerge due to many aspects, including local management policies, resources or socio-technical limitations. Organizations then struggle to improve a business process which has no longer a single process model to redesign, implement and adjust. In this paper, we propose an approach to tackle these two challenges: decrease the proliferation of process variants in these organizations, and foresee, at the same time, the need of having flexible business processes that allow for a certain degree of adjustment. To validate our approach, we first conducted case studies where we collected six real-world business process variants from two organizational units of the same healthcare organization. We then proposed an algorithm to derive a template process model from all the variants, which includes common and flexible process elements. We implemented our approach in a software tool called BPFlexTemplate, and tested it with the elicited variants.

  19. PIPEMicroDB: microsatellite database and primer generation tool for pigeonpea genome.

    Sarika; Arora, Vasu; Iquebal, M A; Rai, Anil; Kumar, Dinesh

    2013-01-01

    Molecular markers play a significant role for crop improvement in desirable characteristics, such as high yield, resistance to disease and others that will benefit the crop in long term. Pigeonpea (Cajanus cajan L.) is the recently sequenced legume by global consortium led by ICRISAT (Hyderabad, India) and been analysed for gene prediction, synteny maps, markers, etc. We present PIgeonPEa Microsatellite DataBase (PIPEMicroDB) with an automated primer designing tool for pigeonpea genome, based on chromosome wise as well as location wise search of primers. Total of 123 387 Short Tandem Repeats (STRs) were extracted from pigeonpea genome, available in public domain using MIcroSAtellite tool (MISA). The database is an online relational database based on 'three-tier architecture' that catalogues information of microsatellites in MySQL and user-friendly interface is developed using PHP. Search for STRs may be customized by limiting their location on chromosome as well as number of markers in that range. This is a novel approach and is not been implemented in any of the existing marker database. This database has been further appended with Primer3 for primer designing of selected markers with left and right flankings of size up to 500 bp. This will enable researchers to select markers of choice at desired interval over the chromosome. Furthermore, one can use individual STRs of a targeted region over chromosome to narrow down location of gene of interest or linked Quantitative Trait Loci (QTLs). Although it is an in silico approach, markers' search based on characteristics and location of STRs is expected to be beneficial for researchers. Database URL: http://cabindb.iasri.res.in/pigeonpea/

  20. The CRISPRdb database and tools to display CRISPRs and to generate dictionaries of spacers and repeats

    Vergnaud Gilles

    2007-05-01

    Full Text Available Abstract Background In Archeae and Bacteria, the repeated elements called CRISPRs for "clustered regularly interspaced short palindromic repeats" are believed to participate in the defence against viruses. Short sequences called spacers are stored in-between repeated elements. In the current model, motifs comprising spacers and repeats may target an invading DNA and lead to its degradation through a proposed mechanism similar to RNA interference. Analysis of intra-species polymorphism shows that new motifs (one spacer and one repeated element are added in a polarised fashion. Although their principal characteristics have been described, a lot remains to be discovered on the way CRISPRs are created and evolve. As new genome sequences become available it appears necessary to develop automated scanning tools to make available CRISPRs related information and to facilitate additional investigations. Description We have produced a program, CRISPRFinder, which identifies CRISPRs and extracts the repeated and unique sequences. Using this software, a database is constructed which is automatically updated monthly from newly released genome sequences. Additional tools were created to allow the alignment of flanking sequences in search for similarities between different loci and to build dictionaries of unique sequences. To date, almost six hundred CRISPRs have been identified in 475 published genomes. Two Archeae out of thirty-seven and about half of Bacteria do not possess a CRISPR. Fine analysis of repeated sequences strongly supports the current view that new motifs are added at one end of the CRISPR adjacent to the putative promoter. Conclusion It is hoped that availability of a public database, regularly updated and which can be queried on the web will help in further dissecting and understanding CRISPR structure and flanking sequences evolution. Subsequent analyses of the intra-species CRISPR polymorphism will be facilitated by CRISPRFinder and the

  1. Topology and boundary shape optimization as an integrated design tool

    Bendsoe, Martin Philip; Rodrigues, Helder Carrico

    1990-01-01

    The optimal topology of a two dimensional linear elastic body can be computed by regarding the body as a domain of the plane with a high density of material. Such an optimal topology can then be used as the basis for a shape optimization method that computes the optimal form of the boundary curves of the body. This results in an efficient and reliable design tool, which can be implemented via common FEM mesh generator and CAD type input-output facilities.

  2. Mesh Intercomparisons of Fog Water Collected Yield Insight Into the Nature of Fog-Drip Collection Mechanisms

    Fernandez, D.; Torregrosa, A.; Weiss-Penzias, P. S.; Oliphant, A. J.; Dodge, C.; Bowman, M.; Wilson, S.; Mairs, A. A.; Gravelle, M.; Barkley, T.

    2016-12-01

    At multiple sites across central CA, several passive fog water collectors have been deployed for the past 3 years. All of the sites employ standard Raschel polypropylene mesh as the fog collection medium and five of them also integrated a novel polypropylene mesh of German manufacture with a 3-dimensional internal structure. Additionally, six metal mesh manufactured by McMaster-Carr of various hole sizing were coated with a POSS-PEMA substance at the Massachusetts Institute of Technology and deployed in parallel with the Raschel mesh at six distinct locations. Finally, fluorine-free versions of the POSS-PEMA substance were generated by NBD Nanotechnology and coated on a much finer mesh substrate. Three of those and one control (uncoated mesh) were deployed at one of the fog collection sites for one season, along with a standard Raschel mesh. Preliminary results from one intercomparison from just one pair of mesh over two seasons seem to reveal a wind speed and also, possibly, a droplet-size dependence on the fog collection efficiency for the mesh. This study will continue to intercompare the various mesh in conjunction with the wind speed and direction data. If a collection efficiency dependence on mesh size or coating is confirmed, it may point to interesting and relevant mechanisms for fog droplet capture and collection hitherto unobserved in field conditions.

  3. TESS: A RELATIVISTIC HYDRODYNAMICS CODE ON A MOVING VORONOI MESH

    Duffell, Paul C.; MacFadyen, Andrew I.

    2011-01-01

    We have generalized a method for the numerical solution of hyperbolic systems of equations using a dynamic Voronoi tessellation of the computational domain. The Voronoi tessellation is used to generate moving computational meshes for the solution of multidimensional systems of conservation laws in finite-volume form. The mesh-generating points are free to move with arbitrary velocity, with the choice of zero velocity resulting in an Eulerian formulation. Moving the points at the local fluid velocity makes the formulation effectively Lagrangian. We have written the TESS code to solve the equations of compressible hydrodynamics and magnetohydrodynamics for both relativistic and non-relativistic fluids on a dynamic Voronoi mesh. When run in Lagrangian mode, TESS is significantly less diffusive than fixed mesh codes and thus preserves contact discontinuities to high precision while also accurately capturing strong shock waves. TESS is written for Cartesian, spherical, and cylindrical coordinates and is modular so that auxiliary physics solvers are readily integrated into the TESS framework and so that this can be readily adapted to solve general systems of equations. We present results from a series of test problems to demonstrate the performance of TESS and to highlight some of the advantages of the dynamic tessellation method for solving challenging problems in astrophysical fluid dynamics.

  4. System Diagnostic Builder - A rule generation tool for expert systems that do intelligent data evaluation. [applied to Shuttle Mission Simulator

    Nieten, Joseph; Burke, Roger

    1993-01-01

    Consideration is given to the System Diagnostic Builder (SDB), an automated knowledge acquisition tool using state-of-the-art AI technologies. The SDB employs an inductive machine learning technique to generate rules from data sets that are classified by a subject matter expert. Thus, data are captured from the subject system, classified, and used to drive the rule generation process. These rule bases are used to represent the observable behavior of the subject system, and to represent knowledge about this system. The knowledge bases captured from the Shuttle Mission Simulator can be used as black box simulations by the Intelligent Computer Aided Training devices. The SDB can also be used to construct knowledge bases for the process control industry, such as chemical production or oil and gas production.

  5. Model of Random Polygon Particles for Concrete and Mesh Automatic Subdivision

    2001-01-01

    In order to study the constitutive behavior of concrete in mesoscopic level, a new method is proposed in this paper. This method uses random polygon particles to simulate full grading broken aggregates of concrete. Based on computational geometry, we carry out the automatic generation of the triangle finite element mesh for the model of random polygon particles of concrete. The finite element mesh generated in this paper is also applicable to many other numerical methods.

  6. A validation of a ray-tracing tool used to generate bi-directional scattering distribution functions for complex fenestration systems

    McNeil, A.; Jonsson, C.J.; Appelfeld, David

    2013-01-01

    , or daylighting systems. However, such tools require users to provide bi-directional scattering distribution function (BSDF) data that describe the solar-optical performance of the CFS. A free, open-source Radiance tool genBSDF enables users to generate BSDF data for arbitrary CFS. Prior to genBSDF, BSDF data.......We explain the basis and use of the genBSDF tool and validate the tool by comparing results for four different cases to BSDF data produced via alternate methods. This validation demonstrates that BSDFs created with genBSDF are comparable to BSDFs generated analytically using TracePro and by measurement...

  7. Practical tools to implement massive parallel pyrosequencing of PCR products in next generation molecular diagnostics.

    Kim De Leeneer

    Full Text Available Despite improvements in terms of sequence quality and price per basepair, Sanger sequencing remains restricted to screening of individual disease genes. The development of massively parallel sequencing (MPS technologies heralded an era in which molecular diagnostics for multigenic disorders becomes reality. Here, we outline different PCR amplification based strategies for the screening of a multitude of genes in a patient cohort. We performed a thorough evaluation in terms of set-up, coverage and sequencing variants on the data of 10 GS-FLX experiments (over 200 patients. Crucially, we determined the actual coverage that is required for reliable diagnostic results using MPS, and provide a tool to calculate the number of patients that can be screened in a single run. Finally, we provide an overview of factors contributing to false negative or false positive mutation calls and suggest ways to maximize sensitivity and specificity, both important in a routine setting. By describing practical strategies for screening of multigenic disorders in a multitude of samples and providing answers to questions about minimum required coverage, the number of patients that can be screened in a single run and the factors that may affect sensitivity and specificity we hope to facilitate the implementation of MPS technology in molecular diagnostics.

  8. Decision-support tool for assessing future nuclear reactor generation portfolios

    Jain, Shashi; Roelofs, Ferry; Oosterlee, Cornelis W.

    2014-01-01

    Capital costs, fuel, operation and maintenance (O and M) costs, and electricity prices play a key role in the economics of nuclear power plants. Often standardized reactor designs are required to be locally adapted, which often impacts the project plans and the supply chain. It then becomes difficult to ascertain how these changes will eventually reflect in costs, which makes the capital costs component of nuclear power plants uncertain. Different nuclear reactor types compete economically by having either lower and less uncertain construction costs, increased efficiencies, lower and less uncertain fuel cycles and O and M costs etc. The decision making process related to nuclear power plants requires a holistic approach that takes into account the key economic factors and their uncertainties. We here present a decision-support tool that satisfactorily takes into account the major uncertainties in the cost elements of a nuclear power plant, to provide an optimal portfolio of nuclear reactors. The portfolio so obtained, under our model assumptions and the constraints considered, maximizes the combined returns for a given level of risk or uncertainty. These decisions are made using a combination of real option theory and mean–variance portfolio optimization. - Highlights: • Decisions to continue or abandon the construction of NPPs • Mean–variance portfolio of nuclear reactors • Sensitivity study of mean–variance portfolio of nuclear reactors

  9. Streptomyces venezuelae TX-TL - a next generation cell-free synthetic biology tool.

    Moore, Simon J; Lai, Hung-En; Needham, Hannah; Polizzi, Karen M; Freemont, Paul S

    2017-04-01

    Streptomyces venezuelae is a promising chassis in synthetic biology for fine chemical and secondary metabolite pathway engineering. The potential of S. venezuelae could be further realized by expanding its capability with the introduction of its own in vitro transcription-translation (TX-TL) system. TX-TL is a fast and expanding technology for bottom-up design of complex gene expression tools, biosensors and protein manufacturing. Herein, we introduce a S. venezuelae TX-TL platform by reporting a streamlined protocol for cell-extract preparation, demonstrating high-yield synthesis of a codon-optimized sfGFP reporter and the prototyping of a synthetic tetracycline-inducible promoter in S. venezuelae TX-TL based on the tetO-TetR repressor system. The aim of this system is to provide a host for the homologous production of exotic enzymes from Actinobacteria secondary metabolism in vitro. As an example, the authors demonstrate the soluble synthesis of a selection of enzymes (12-70 kDa) from the Streptomyces rimosus oxytetracycline pathway. © 2017 The Authors. Biotechnology Journal published by WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  10. Computational protein design-the next generation tool to expand synthetic biology applications.

    Gainza-Cirauqui, Pablo; Correia, Bruno Emanuel

    2018-05-02

    One powerful approach to engineer synthetic biology pathways is the assembly of proteins sourced from one or more natural organisms. However, synthetic pathways often require custom functions or biophysical properties not displayed by natural proteins, limitations that could be overcome through modern protein engineering techniques. Structure-based computational protein design is a powerful tool to engineer new functional capabilities in proteins, and it is beginning to have a profound impact in synthetic biology. Here, we review efforts to increase the capabilities of synthetic biology using computational protein design. We focus primarily on computationally designed proteins not only validated in vitro, but also shown to modulate different activities in living cells. Efforts made to validate computational designs in cells can illustrate both the challenges and opportunities in the intersection of protein design and synthetic biology. We also highlight protein design approaches, which although not validated as conveyors of new cellular function in situ, may have rapid and innovative applications in synthetic biology. We foresee that in the near-future, computational protein design will vastly expand the functional capabilities of synthetic cells. Copyright © 2018. Published by Elsevier Ltd.

  11. DynamiX, numerical tool for design of next-generation x-ray telescopes.

    Chauvin, Maxime; Roques, Jean-Pierre

    2010-07-20

    We present a new code aimed at the simulation of grazing-incidence x-ray telescopes subject to deformations and demonstrate its ability with two test cases: the Simbol-X and the International X-ray Observatory (IXO) missions. The code, based on Monte Carlo ray tracing, computes the full photon trajectories up to the detector plane, accounting for the x-ray interactions and for the telescope motion and deformation. The simulation produces images and spectra for any telescope configuration using Wolter I mirrors and semiconductor detectors. This numerical tool allows us to study the telescope performance in terms of angular resolution, effective area, and detector efficiency, accounting for the telescope behavior. We have implemented an image reconstruction method based on the measurement of the detector drifts by an optical sensor metrology. Using an accurate metrology, this method allows us to recover the loss of angular resolution induced by the telescope instability. In the framework of the Simbol-X mission, this code was used to study the impacts of the parameters on the telescope performance. In this paper we present detailed performance analysis of Simbol-X, taking into account the satellite motions and the image reconstruction. To illustrate the versatility of the code, we present an additional performance analysis with a particular configuration of IXO.

  12. DynamiX, numerical tool for design of next-generation x-ray telescopes

    Chauvin, Maxime; Roques, Jean-Pierre

    2010-01-01

    We present a new code aimed at the simulation of grazing-incidence x-ray telescopes subject to deformations and demonstrate its ability with two test cases: the Simbol-X and the International X-ray Observatory (IXO) missions. The code, based on Monte Carlo ray tracing, computes the full photon trajectories up to the detector plane, accounting for the x-ray interactions and for the telescope motion and deformation. The simulation produces images and spectra for any telescope configuration using Wolter I mirrors and semiconductor detectors. This numerical tool allows us to study the telescope performance in terms of angular resolution, effective area, and detector efficiency, accounting for the telescope behavior. We have implemented an image reconstruction method based on the measurement of the detector drifts by an optical sensor metrology. Using an accurate metrology, this method allows us to recover the loss of angular resolution induced by the telescope instability. In the framework of the Simbol-X mission, this code was used to study the impacts of the parameters on the telescope performance. In this paper we present detailed performance analysis of Simbol-X, taking into account the satellite motions and the image reconstruction. To illustrate the versatility of the code, we present an additional performance analysis with a particular configuration of IXO.

  13. DynamiX, numerical tool for design of next-generation x-ray telescopes

    Chauvin, Maxime; Roques, Jean-Pierre

    2010-07-20

    We present a new code aimed at the simulation of grazing-incidence x-ray telescopes subject to deformations and demonstrate its ability with two test cases: the Simbol-X and the International X-ray Observatory (IXO) missions. The code, based on Monte Carlo ray tracing, computes the full photon trajectories up to the detector plane, accounting for the x-ray interactions and for the telescope motion and deformation. The simulation produces images and spectra for any telescope configuration using Wolter I mirrors and semiconductor detectors. This numerical tool allows us to study the telescope performance in terms of angular resolution, effective area, and detector efficiency, accounting for the telescope behavior. We have implemented an image reconstruction method based on the measurement of the detector drifts by an optical sensor metrology. Using an accurate metrology, this method allows us to recover the loss of angular resolution induced by the telescope instability. In the framework of the Simbol-X mission, this code was used to study the impacts of the parameters on the telescope performance. In this paper we present detailed performance analysis of Simbol-X, taking into account the satellite motions and the image reconstruction. To illustrate the versatility of the code, we present an additional performance analysis with a particular configuration of IXO.

  14. Understanding the mobile internet to develop the next generation of online medical teaching tools.

    Desai, Tejas; Christiano, Cynthia; Ferris, Maria

    2011-01-01

    Healthcare providers (HCPs) use online medical information for self-directed learning and patient care. Recently, the mobile internet has emerged as a new platform for accessing medical information as it allows mobile devices to access online information in a manner compatible with their restricted storage. We investigated mobile internet usage parameters to direct the future development of mobile internet teaching websites. Nephrology On-Demand Mobile (NOD(M)) (http://www.nephrologyondemand.org) was made accessible to all mobile devices. From February 1 to December 31, 2010, HCP use of NOD(M) was tracked using code inserted into the root files. Nephrology On-Demand received 15,258 visits, of which approximately 10% were made to NOD(M), with the majority coming from the USA. Most access to NOD(M) was through the Apple iOS family of devices and cellular connections were the most frequently used. These findings provide a basis for the future development of mobile nephrology and medical teaching tools.

  15. The Characteristics and Generating Mechanism of Large Precipitates in Ti-Containing H13 Tool Steel

    Xie, You; Cheng, Guoguang; Chen, Lie; Zhang, Yandong; Yan, Qingzhong

    2017-02-01

    The characteristics of large precipitates in H13 tool steel with 0.015wt% Ti were studied. The result shows that three types of phases larger than 1 μm exist in the as-cast ingot, that is, (Ti, V) (C, N) type phase, (V, Mo, Cr)C type phase and sulfide. (Ti, V) (C, N) type phase could be further classified as the homogeneous Ti-rich one and the Ti-V-rich one in which Ti/V ratio gradually changes. (V, Mo, Cr)C type phase contains the V-rich one and the Mo-Cr-rich one. The compositional characteristics in all of them have little relation with the cutting position or cooling rate. The precipitating process could be well described through calculation by Thermo-Calc software. During solidification, the primary phase (Ti, V)(C, N) first starts to precipitate in the form of Ti-rich carbonitride. With the development of solidification, the ratio of Ti decreases and that of V increases. Then the primary phase Ti-V-rich (Ti, V)(C, N) and V-rich (V, Mo, Cr)C appears successively. Mo-Cr-rich (V, Mo, Cr)C phase does not precipitate until the solidification process reaches to the end. Sulfide precipitates before (V, Mo, Cr)C type phase and it could act as the nucleus of (V, Mo, Cr)C.

  16. New tool to assemble repetitive regions using next-generation sequencing data

    Kuśmirek, Wiktor; Nowak, Robert M.; Neumann, Łukasz

    2017-08-01

    The next generation sequencing techniques produce a large amount of sequencing data. Some part of the genome are composed of repetitive DNA sequences, which are very problematic for the existing genome assemblers. We propose a modification of the algorithm for a DNA assembly, which uses the relative frequency of reads to properly reconstruct repetitive sequences. The new approach was implemented and tested, as a demonstration of the capability of our software we present some results for model organisms. The new implementation, using a three-layer software architecture was selected, where the presentation layer, data processing layer, and data storage layer were kept separate. Source code as well as demo application with web interface and the additional data are available at project web-page: http://dnaasm.sourceforge.net.

  17. Confinement of electron beams by mesh arrays in a relativistic klystron amplifier

    Wang Pingshan; Gu Binlin

    1998-01-01

    Theoretical and experimental results of intense beam confinement by conducting meshes in a relativistic klystron amplifier (RKA) are presented. Electron motions in a steady intense electron beam confined by conducting meshes are analyzed with an approximate space charge field distribution. And the conditions for steady beam transportation are discussed. Experimental results of a long distance (60 cm) transportation of an intense beam (400 kV, 2.5 kA) generated by a linear induction accelerator are presented. Experimental results of modulated beam transportation confined by the mesh array are presented also. The results show that the focusing ability of the conducting meshes is not very sensitive to the beam energy. And the meshes can be used effectively in a RKA to replace the magnetic field system

  18. A Mechanistic Study of Wetting Superhydrophobic Porous 3D Meshes

    Yohe, Stefan T.; Freedman, Jonathan D.; Falde, Eric J.; Colson, Yolonda L.; Grinstaff, Mark W.

    2014-01-01

    Superhydrophobic, porous, 3D materials composed of poly( ε -caprolactone) (PCL) and the hydrophobic polymer dopant poly(glycerol monostearate-co- ε -caprolactone) (PGC-C18) are fabricated using the electrospinning technique. These 3D materials are distinct from 2D superhydrophobic surfaces, with maintenance of air at the surface as well as within the bulk of the material. These superhydrophobic materials float in water, and when held underwater and pressed, an air bubble is released and will rise to the surface. By changing the PGC-C18 doping concentration in the meshes and/or the fiber size from the micro- to nanoscale, the long-term stability of the entrapped air layer is controlled. The rate of water infiltration into the meshes, and the resulting displacement of the entrapped air, is quantitatively measured using X-ray computed tomography. The properties of the meshes are further probed using surfactants and solvents of different surface tensions. Finally, the application of hydraulic pressure is used to quantify the breakthrough pressure to wet the meshes. The tools for fabrication and analysis of these superhydrophobic materials as well as the ability to control the robustness of the entrapped air layer are highly desirable for a number of existing and emerging applications. PMID:25309305

  19. Tool for generation of seismic floor response spectra for secondary system design

    Cardoso, Tarcisio F.; Almeida, Andreia A. Diniz de

    2009-01-01

    The spectral analysis is still a valuable method to the seismic structure design, especially when one focalizes the topics of secondary systems in large industrial installations, as nuclear power plants. Two aspects of this situation add their arguments to recommend the use of this kind of analysis: the random character of the excitation and the multiplicity and the variability of the secondary systems. The first aspect can be managed if one assumes the site seismicity represented by a power spectrum density function of the ground acceleration, and then, by the systematic resolution of a first passage problem, to develop a uniformly probable response spectrum. The second one suggests also a probabilistic approach to the response spectrum in order to be representative all over the extensive group of systems with different characteristics, which can be enrolled in a plant. The present paper proposes a computational tool to achieve in-structure floor response spectra for secondary system design, which includes a probabilistic approach and considers coupling effects between primary and inelastic secondary systems. The analysis is performed in the frequency domain, with SASSI2000 system. A set of auxiliary programs are developed to consider three-dimensional models and their responses to a generic base excitation, acting in 3 orthogonal directions. The ground excitation is transferred to a secondary system SDOF model conveniently attached to the primary system. Then, a uniformly probable coupled response spectrum is obtained using a first passage analysis. In this work, the ExeSASSI program is created to manage SASSI2000 several modules and a set of auxiliary programs created to perform the probabilistic analyses. (author)

  20. An automated graphics tool for comparative genomics: the Coulson plot generator.

    Field, Helen I; Coulson, Richard M R; Field, Mark C

    2013-04-27

    Comparative analysis is an essential component to biology. When applied to genomics for example, analysis may require comparisons between the predicted presence and absence of genes in a group of genomes under consideration. Frequently, genes can be grouped into small categories based on functional criteria, for example membership of a multimeric complex, participation in a metabolic or signaling pathway or shared sequence features and/or paralogy. These patterns of retention and loss are highly informative for the prediction of function, and hence possible biological context, and can provide great insights into the evolutionary history of cellular functions. However, representation of such information in a standard spreadsheet is a poor visual means from which to extract patterns within a dataset. We devised the Coulson Plot, a new graphical representation that exploits a matrix of pie charts to display comparative genomics data. Each pie is used to describe a complex or process from a separate taxon, and is divided into sectors corresponding to the number of proteins (subunits) in a complex/process. The predicted presence or absence of proteins in each complex are delineated by occupancy of a given sector; this format is visually highly accessible and makes pattern recognition rapid and reliable. A key to the identity of each subunit, plus hierarchical naming of taxa and coloring are included. A java-based application, the Coulson plot generator (CPG) automates graphic production, with a tab or comma-delineated text file as input and generating an editable portable document format or svg file. CPG software may be used to rapidly convert spreadsheet data to a graphical matrix pie chart format. The representation essentially retains all of the information from the spreadsheet but presents a graphically rich format making comparisons and identification of patterns significantly clearer. While the Coulson plot format is highly useful in comparative genomics, its

  1. Laparoscopic appendicectomy for suspected mesh-induced appendicitis after laparoscopic transabdominal preperitoneal polypropylene mesh inguinal herniorraphy

    Jennings Jason

    2010-01-01

    Full Text Available Laparoscopic inguinal herniorraphy via a transabdominal preperitoneal (TAPP approach using Polypropylene Mesh (Mesh and staples is an accepted technique. Mesh induces a localised inflammatory response that may extend to, and involve, adjacent abdominal and pelvic viscera such as the appendix. We present an interesting case of suspected Mesh-induced appendicitis treated successfully with laparoscopic appendicectomy, without Mesh removal, in an elderly gentleman who presented with symptoms and signs of acute appendicitis 18 months after laparoscopic inguinal hernia repair. Possible mechanisms for Mesh-induced appendicitis are briefly discussed.

  2. Applying Ancestry and Sex Computation as a Quality Control Tool in Targeted Next-Generation Sequencing.

    Mathias, Patrick C; Turner, Emily H; Scroggins, Sheena M; Salipante, Stephen J; Hoffman, Noah G; Pritchard, Colin C; Shirts, Brian H

    2016-03-01

    To apply techniques for ancestry and sex computation from next-generation sequencing (NGS) data as an approach to confirm sample identity and detect sample processing errors. We combined a principal component analysis method with k-nearest neighbors classification to compute the ancestry of patients undergoing NGS testing. By combining this calculation with X chromosome copy number data, we determined the sex and ancestry of patients for comparison with self-report. We also modeled the sensitivity of this technique in detecting sample processing errors. We applied this technique to 859 patient samples with reliable self-report data. Our k-nearest neighbors ancestry screen had an accuracy of 98.7% for patients reporting a single ancestry. Visual inspection of principal component plots was consistent with self-report in 99.6% of single-ancestry and mixed-ancestry patients. Our model demonstrates that approximately two-thirds of potential sample swaps could be detected in our patient population using this technique. Patient ancestry can be estimated from NGS data incidentally sequenced in targeted panels, enabling an inexpensive quality control method when coupled with patient self-report. © American Society for Clinical Pathology, 2016. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.

  3. Design-a-wetland: a tool for generating and assessing constructed wetland designs for wastewater treatment

    Casaril, Carolina J.

    2007-01-01

    Full text: Full text: The hydrological cycle is a key cycle affected by current and predicted climate change. Wetlands are one of the key ecosystems within the hydrological cycle and could contribute significantly in facing the challenges of climate change, such as water shortage. The impact of wetlands on greenhouse gas emissions is much debated and, conversely, the impact of climate change on wetlands also raises many questions. There have been many attempts to harness and integrate the natural capacities of wetlands into constructed systems. These systems are especially designed for multiple purposes. They can be used for wastewater treatment and reuse, and have the potential to increase sustainability by changing land and water use practices. This project generates a 'Design-A-Wetland' prototype model, designed to facilitate decision-making in the creation of constructed wetlands. Constructed wetlands are specifically tailored to their end use; water treatment fish and fowl habitat, flood buffer zones, or sequestration of greenhouse gases. This project attempts to answer the following questions: Can a single integrated decision model be created for the design and assessment of artificial wetlands, provided either entry or exit standards are known and specified?; Can the elements of a system of interfacing the model with public consultation be specified?; The project identifies model schematics and lays the groundwork for modelling suited to the wide variety of inputs required for decision making

  4. Method and system for progressive mesh storage and reconstruction using wavelet-encoded height fields

    Baxes, Gregory A. (Inventor); Linger, Timothy C. (Inventor)

    2011-01-01

    Systems and methods are provided for progressive mesh storage and reconstruction using wavelet-encoded height fields. A method for progressive mesh storage includes reading raster height field data, and processing the raster height field data with a discrete wavelet transform to generate wavelet-encoded height fields. In another embodiment, a method for progressive mesh storage includes reading texture map data, and processing the texture map data with a discrete wavelet transform to generate wavelet-encoded texture map fields. A method for reconstructing a progressive mesh from wavelet-encoded height field data includes determining terrain blocks, and a level of detail required for each terrain block, based upon a viewpoint. Triangle strip constructs are generated from vertices of the terrain blocks, and an image is rendered utilizing the triangle strip constructs. Software products that implement these methods are provided.

  5. Numerical Tools for Multicomponent, Multiphase, Reactive Processes: Flow of CO{sub 2} in Porous Medium

    Khattri, Sanjay Kumar

    2006-07-01

    The thesis is concerned with numerically simulating multicomponent, multiphase, reactive transport in heterogeneous porous medium. Such processes are ubiquitous, for example, deposition of green house gases, flow of hydrocarbons and groundwater remediation. Understanding such processes is important from social and economic point of view. For the success of geological sequestration, an accurate estimation of migration patterns of green-house gases is essential. Due to an ever increasing computer power, computational mathematics has become an important tool for predicting dynamics of porous media fluids. Numerical and mathematical modelling of processes in a domain requires grid generation in the domain, discretization of the continuum equations on the generated grid, solution of the formed linear or nonlinear system of discrete equations and finally visualization of the results. The thesis is composed of three chapters and eight papers. Chapter 2 presents two techniques for generating structured quadrilateral and hexahedral meshes. These techniques are called algebraic and elliptic methods. Algebraic techniques are by far the most simple and computationally efficient method for grid generation. Transfinite interpolation operators are a kind of algebraic grid generation technique. In this chapter, many transfinite interpolation operators for grid generation are derived from 1D projection operators. In this chapter, some important properties of hexahedral elements are also mentioned. These properties are useful in discretization of partial differential equations on hexahedral mesh, improving quality of the hexahedral mesh, mesh generation and visualization. Chapter 3 is about CO{sub 2} flow in porous media. In this chapter, we present the mathematical models and their discretization for capturing major physical processes associated with CO{sub 2} deposition in geological formations. Some important simulations of practical applications in 2D and 3D are presented

  6. Numerical Tools for Multicomponent, Multiphase, Reactive Processes: Flow of CO{sub 2} in Porous Medium

    Khattri, Sanjay Kumar

    2006-07-01

    The thesis is concerned with numerically simulating multicomponent, multiphase, reactive transport in heterogeneous porous medium. Such processes are ubiquitous, for example, deposition of green house gases, flow of hydrocarbons and groundwater remediation. Understanding such processes is important from social and economic point of view. For the success of geological sequestration, an accurate estimation of migration patterns of green-house gases is essential. Due to an ever increasing computer power, computational mathematics has become an important tool for predicting dynamics of porous media fluids. Numerical and mathematical modelling of processes in a domain requires grid generation in the domain, discretization of the continuum equations on the generated grid, solution of the formed linear or nonlinear system of discrete equations and finally visualization of the results. The thesis is composed of three chapters and eight papers. Chapter 2 presents two techniques for generating structured quadrilateral and hexahedral meshes. These techniques are called algebraic and elliptic methods. Algebraic techniques are by far the most simple and computationally efficient method for grid generation. Transfinite interpolation operators are a kind of algebraic grid generation technique. In this chapter, many transfinite interpolation operators for grid generation are derived from 1D projection operators. In this chapter, some important properties of hexahedral elements are also mentioned. These properties are useful in discretization of partial differential equations on hexahedral mesh, improving quality of the hexahedral mesh, mesh generation and visualization. Chapter 3 is about CO{sub 2} flow in porous media. In this chapter, we present the mathematical models and their discretization for capturing major physical processes associated with CO{sub 2} deposition in geological formations. Some important simulations of practical applications in 2D and 3D are presented

  7. Short term post-operative morphing of sacrocolpopexy mesh measured by magnetic resonance imaging.

    Sindhwani, Nikhil; Callewaert, Geertje; Deprest, Thomas; Housmans, Susanne; Van Beckevoort, Dirk; Deprest, Jan

    2018-04-01

    Sacrocolpopexy (SC) involves suspension of the vaginal vault or cervix to the sacrum using a mesh. Following insertion, the meshes have been observed to have undergone dimensional changes. To quantify dimensional changes of meshes following implantation and characterize their morphology in-vivo. 24 patients underwent SC using PolyVinyliDeneFluoride mesh loaded with Fe 3 O 4 particles. Tailored anterior and posterior mesh flaps were sutured to the respective vaginal walls, uniting at the apex. The posterior flap continued to the sacrum and was attached there. Meshes were visualized on magnetic resonance (MR) imaging at 12 [3-12] (median [range]) months postoperatively and 3D models of the mesh were generated. Dynamic MR sequences were acquired during valsalva to record mesh mobility. The area of the vagina effectively supported by the mesh (Effective Support Area (ESA)) was calculated. The 3D models' wall thickness map was analyzed to identify the locations of mesh folding. Intraclass correlation (ICC) was calculated to test the reliability of the methods. To measure the laxity and flatness of the mesh, the curvature and the ellipticity of the sacral flap were calculated. The ESA calculation methodology had ICC = 0.97. A reduction of 75.49 [61.55-78.67] % (median [IQR]) in area, 47.64 [38.07-59.81] % in anterior flap, and of 23.95 [10.96-27.21] % in the posterior flap was measured. The mesh appeared thicker near its attachment at the sacral promontory (n = 19) and near the vaginal apex (n = 22). The laxity of the mesh was 1.13 [1.10-1.16] and 60.55 [49.76-76.25] % of the sacral flap was flat. We could not reliably measure mesh mobility (ICC = 0.16). A methodology for complete 3D characterization of SC meshes using MR images was presented. After implantation, the supported area is much lower than what is prepared prior to implantation. We propose this happened during the surgery itself. Copyright © 2018 Elsevier Ltd. All rights reserved.

  8. Assessment of the anti-biofouling potentials of a copper iodide-doped nylon mesh.

    Sato, Tetsuya; Fujimori, Yoshie; Nakayama, Tsuruo; Gotoh, Yasuo; Sunaga, Yoshihiko; Nemoto, Michiko; Matsunaga, Tadashi; Tanaka, Tsuyoshi

    2012-08-01

    We propose a copper iodide (CuI)-doped nylon mesh prepared using polyiodide ions as a precursor toward anti-biofouling polymer textile. The CuI-doped nylon mesh was subjected to the prevention of biofouling in marine environments. The attachment of the marine organisms was markedly inhibited on the CuI-doped nylon mesh surface until 249 days. Scanning electron microscopy-energy dispersive X-ray analysis indicated that copper compounds were maintained in the nylon mesh after the field experiment, although copper content in the nylon mesh was reduced. Therefore, the copper ions slowly dissolved from nylon mesh will contribute to the long-term prevention of biofouling. Furthermore, electron spin resonance analysis revealed the generation of reactive oxygen species (ROS) from CuI-doped nylon mesh after the field experiment. One of the possibilities for toxic action of copper ions will be the direct effect of Cu+ -induced ROS on biofilm forming on nylon mesh surface. The proposed polymer textile can be applied to fishing and aquafarming nets, mooring rope for ship, or silt fence to restrict polluted water in marine environments.

  9. Evaluation of mechanical properties in metal wire mesh supported selective catalytic reduction (SCR) catalyst structures

    Rajath, S.; Siddaraju, C.; Nandakishora, Y.; Roy, Sukumar

    2018-04-01

    The objective of this research is to evaluate certain specific mechanical properties of certain stainless steel wire mesh supported Selective catalytic reduction catalysts structures wherein the physical properties of the metal wire mesh and also its surface treatments played vital role thereby influencing the mechanical properties. As the adhesion between the stainless steel wire mesh and the catalyst material determines the bond strength and the erosion resistance of catalyst structures, surface modifications of the metal- wire mesh structure in order to facilitate the interface bonding is therefore very important to realize enhanced level of mechanical properties. One way to enhance such adhesion properties, the stainless steel wire mesh is treated with the various acids, i.e., chromic acid, phosphoric acid including certain mineral acids and combination of all those in various molar ratios that could generate surface active groups on metal surface that promotes good interface structure between the metal- wire mesh and metal oxide-based catalyst material and then the stainless steel wire mesh is dipped in the glass powder slurry containing some amount of organic binder. As a result of which the said catalyst material adheres to the metal-wire mesh surface more effectively that improves the erosion profile of supported catalysts structure including bond strength.

  10. Sierra toolkit computational mesh conceptual model

    Baur, David G.; Edwards, Harold Carter; Cochran, William K.; Williams, Alan B.; Sjaardema, Gregory D.

    2010-01-01

    The Sierra Toolkit computational mesh is a software library intended to support massively parallel multi-physics computations on dynamically changing unstructured meshes. This domain of intended use is inherently complex due to distributed memory parallelism, parallel scalability, heterogeneity of physics, heterogeneous discretization of an unstructured mesh, and runtime adaptation of the mesh. Management of this inherent complexity begins with a conceptual analysis and modeling of this domain of intended use; i.e., development of a domain model. The Sierra Toolkit computational mesh software library is designed and implemented based upon this domain model. Software developers using, maintaining, or extending the Sierra Toolkit computational mesh library must be familiar with the concepts/domain model presented in this report.

  11. Drishti: a volume exploration and presentation tool

    Limaye, Ajay

    2012-10-01

    Among several rendering techniques for volumetric data, direct volume rendering is a powerful visualization tool for a wide variety of applications. This paper describes the major features of hardware based volume exploration and presentation tool - Drishti. The word, Drishti, stands for vision or insight in Sanskrit, an ancient Indian language. Drishti is a cross-platform open-source volume rendering system that delivers high quality, state of the art renderings. The features in Drishti include, though not limited to, production quality rendering, volume sculpting, multi-resolution zooming, transfer function blending, profile generation, measurement tools, mesh generation, stereo/anaglyph/crosseye renderings. Ultimately, Drishti provides an intuitive and powerful interface for choreographing animations.

  12. Astronomy as a Tool for Training the Next Generation Technical Workforce

    Romero, V.; Walsh, G.; Ryan, W.; Ryan, E.

    A major challenge for today's institutes of higher learning is training the next generation of scientists, engineers, and optical specialists to be proficient in the latest technologies they will encounter when they enter the workforce. Although research facilities can offer excellent hands-on instructional opportunities, integrating such experiential learning into academic coursework without disrupting normal operations at such facilities can be difficult. Also, motivating entry level students to increase their skill levels by undertaking and successfully completing difficult coursework can require more creative instructional approaches, including fostering a fun, non-threatening environment for enhancing basic abilities. Astronomy is a universally appealing subject area, and can be very effective as a foundation for cultivating advanced competencies. We report on a project underway at the New Mexico Institute of Mining and Technology (NM Tech), a science and engineering school in Socorro, NM, to incorporate a state-of-the-art optical telescope and laboratory experiments into an entry-level course in basic engineering. Students enrolled in an explosive engineering course were given a topical problem in Planetary Astronomy: they were asked to develop a method to energetically mitigate a potentially hazardous impact between our planet and a Near-Earth asteroid to occur sometime in the future. They were first exposed to basic engineering training in the areas of fracture and material response to failure under different environmental conditions through lectures and traditional laboratory exercises. The students were then given access to NM Tech's Magdalena Ridge Observatory's (MRO) 2.4-meter telescope to collect physical characterization data, (specifically shape information) on two potentially hazardous asteroids (one roughly spherical, the other an elongated ellipsoid). Finally, the students used NM Tech's Energetic Materials Research and Testing Center (EMRTC) to

  13. Generation of a Skeleton Corpus of Digital Objects for the Validation and Evaluation of Format Identification Tools and Signatures

    Ross Spencer

    2013-06-01

    Full Text Available To preserve digital information it is vital that the format of that information can be identified, in-perpetuity. This is the major focus of research within the field of Digital Preservation. The National Archives of the UK called for the Digital Preservation and Digital Curation communities to develop a test corpus of digital objects to help further develop tools to aid this purpose. Following that call, an attempt has been made to develop the suite.This paper initially outlines a methodology to generate a skeleton corpus using simple user-generated digital objects. It then explores the lessons learnt in the generation of a corpus using scripting language techniques from the file format signatures described in The National Archives PRONOM technical registry. It will also discuss the use of the digital signature for this purpose, the benefits of developing a test corpus using this technique. Finally, this paper will outline a methodology for future research before exploring how the community can best make use of the output of this project and how this project needs to be taken forward to completion.

  14. Anisotropic evaluation of synthetic surgical meshes.

    Saberski, E R; Orenstein, S B; Novitsky, Y W

    2011-02-01

    The material properties of meshes used in hernia repair contribute to the overall mechanical behavior of the repair. The anisotropic potential of synthetic meshes, representing a difference in material properties (e.g., elasticity) in different material axes, is not well defined to date. Haphazard orientation of anisotropic mesh material can contribute to inconsistent surgical outcomes. We aimed to characterize and compare anisotropic properties of commonly used synthetic meshes. Six different polypropylene (Trelex(®), ProLite™, Ultrapro™), polyester (Parietex™), and PTFE-based (Dualmesh(®), Infinit) synthetic meshes were selected. Longitudinal and transverse axes were defined for each mesh, and samples were cut in each axis orientation. Samples underwent uniaxial tensile testing, from which the elastic modulus (E) in each axis was determined. The degree of anisotropy (λ) was calculated as a logarithmic expression of the ratio between the elastic modulus in each axis. Five of six meshes displayed significant anisotropic behavior. Ultrapro™ and Infinit exhibited approximately 12- and 20-fold differences between perpendicular axes, respectively. Trelex(®), ProLite™, and Parietex™ were 2.3-2.4 times. Dualmesh(®) was the least anisotropic mesh, without marked difference between the axes. Anisotropy of synthetic meshes has been underappreciated. In this study, we found striking differences between elastic properties of perpendicular axes for most commonly used synthetic meshes. Indiscriminate orientation of anisotropic mesh may adversely affect hernia repairs. Proper labeling of all implants by manufacturers should be mandatory. Understanding the specific anisotropic behavior of synthetic meshes should allow surgeons to employ rational implant orientation to maximize outcomes of hernia repair.

  15. Parallel Performance Optimizations on Unstructured Mesh-based Simulations

    Sarje, Abhinav; Song, Sukhyun; Jacobsen, Douglas; Huck, Kevin; Hollingsworth, Jeffrey; Malony, Allen; Williams, Samuel; Oliker, Leonid

    2015-01-01

    © The Authors. Published by Elsevier B.V. This paper addresses two key parallelization challenges the unstructured mesh-based ocean modeling code, MPAS-Ocean, which uses a mesh based on Voronoi tessellations: (1) load imbalance across processes, and (2) unstructured data access patterns, that inhibit intra- and inter-node performance. Our work analyzes the load imbalance due to naive partitioning of the mesh, and develops methods to generate mesh partitioning with better load balance and reduced communication. Furthermore, we present methods that minimize both inter- and intranode data movement and maximize data reuse. Our techniques include predictive ordering of data elements for higher cache efficiency, as well as communication reduction approaches. We present detailed performance data when running on thousands of cores using the Cray XC30 supercomputer and show that our optimization strategies can exceed the original performance by over 2×. Additionally, many of these solutions can be broadly applied to a wide variety of unstructured grid-based computations.

  16. Unstructured mesh adaptivity for urban flooding modelling

    Hu, R.; Fang, F.; Salinas, P.; Pain, C. C.

    2018-05-01

    Over the past few decades, urban floods have been gaining more attention due to their increase in frequency. To provide reliable flooding predictions in urban areas, various numerical models have been developed to perform high-resolution flood simulations. However, the use of high-resolution meshes across the whole computational domain causes a high computational burden. In this paper, a 2D control-volume and finite-element flood model using adaptive unstructured mesh technology has been developed. This adaptive unstructured mesh technique enables meshes to be adapted optimally in time and space in response to the evolving flow features, thus providing sufficient mesh resolution where and when it is required. It has the advantage of capturing the details of local flows and wetting and drying front while reducing the computational cost. Complex topographic features are represented accurately during the flooding process. For example, the high-resolution meshes around the buildings and steep regions are placed when the flooding water reaches these regions. In this work a flooding event that happened in 2002 in Glasgow, Scotland, United Kingdom has been simulated to demonstrate the capability of the adaptive unstructured mesh flooding model. The simulations have been performed using both fixed and adaptive unstructured meshes, and then results have been compared with those published 2D and 3D results. The presented method shows that the 2D adaptive mesh model provides accurate results while having a low computational cost.

  17. Adaptive hybrid mesh refinement for multiphysics applications

    Khamayseh, Ahmed; Almeida, Valmor de

    2007-01-01

    The accuracy and convergence of computational solutions of mesh-based methods is strongly dependent on the quality of the mesh used. We have developed methods for optimizing meshes that are comprised of elements of arbitrary polygonal and polyhedral type. We present in this research the development of r-h hybrid adaptive meshing technology tailored to application areas relevant to multi-physics modeling and simulation. Solution-based adaptation methods are used to reposition mesh nodes (r-adaptation) or to refine the mesh cells (h-adaptation) to minimize solution error. The numerical methods perform either the r-adaptive mesh optimization or the h-adaptive mesh refinement method on the initial isotropic or anisotropic meshes to equidistribute weighted geometric and/or solution error function. We have successfully introduced r-h adaptivity to a least-squares method with spherical harmonics basis functions for the solution of the spherical shallow atmosphere model used in climate modeling. In addition, application of this technology also covers a wide range of disciplines in computational sciences, most notably, time-dependent multi-physics, multi-scale modeling and simulation

  18. Meshes optimized for discrete exterior calculus (DEC).

    Mousley, Sarah C. [Univ. of Illinois, Urbana-Champaign, IL (United States); Deakin, Michael [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Knupp, Patrick [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Mitchell, Scott A. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2017-12-01

    We study the optimization of an energy function used by the meshing community to measure and improve mesh quality. This energy is non-traditional because it is dependent on both the primal triangulation and its dual Voronoi (power) diagram. The energy is a measure of the mesh's quality for usage in Discrete Exterior Calculus (DEC), a method for numerically solving PDEs. In DEC, the PDE domain is triangulated and this mesh is used to obtain discrete approximations of the continuous operators in the PDE. The energy of a mesh gives an upper bound on the error of the discrete diagonal approximation of the Hodge star operator. In practice, one begins with an initial mesh and then makes adjustments to produce a mesh of lower energy. However, we have discovered several shortcomings in directly optimizing this energy, e.g. its non-convexity, and we show that the search for an optimized mesh may lead to mesh inversion (malformed triangles). We propose a new energy function to address some of these issues.

  19. Transrectal Mesh Erosion Requiring Bowel Resection.

    Kemp, Marta Maria; Slim, Karem; Rabischong, Benoît; Bourdel, Nicolas; Canis, Michel; Botchorishvili, Revaz

    To report a case of a transrectal mesh erosion as complication of laparoscopic promontofixation with mesh repair, necessitating bowel resection and subsequent surgical interventions. Sacrocolpopexy has become a standard procedure for vaginal vault prolapse [1], and the laparoscopic approach has gained popularity owing to more rapid recovery and less morbidity [2,3]. Mesh erosion is a well-known complication of surgical treatment for prolapse as reported in several negative evaluations, including a report from the US Food and Drug Administration in 2011 [4]. Mesh complications are more common after surgeries via the vaginal approach [5]; nonetheless, the incidence of vaginal mesh erosion after laparoscopic procedures is as high as 9% [6]. The incidence of transrectal mesh exposure after laparoscopic ventral rectopexy is roughly 1% [7]. The diagnosis may be delayed because of its rarity and variable presentation. In addition, polyester meshes, such as the mesh used in this case, carry a higher risk of exposure [8]. A 57-year-old woman experiencing genital prolapse, with the cervix classified as +3 according to the Pelvic Organ Prolapse Quantification system, underwent laparoscopic standard sacrocolpopexy using polyester mesh. Subtotal hysterectomy and bilateral adnexectomy were performed concomitantly. A 3-year follow-up consultation demonstrated no signs or symptoms of erosion of any type. At 7 years after the surgery, however, the patient presented with rectal discharge, diagnosed as infectious rectocolitis with the isolation of Clostridium difficile. She underwent a total of 5 repair surgeries in a period of 4 months, including transrectal resection of exposed mesh, laparoscopic ablation of mesh with digestive resection, exploratory laparoscopy with abscess drainage, and exploratory laparoscopy with ablation of residual mesh and transverse colostomy. She recovered well after the last intervention, exhibiting no signs of vaginal or rectal fistula and no recurrence

  20. Parallel adaptive simulations on unstructured meshes

    Shephard, M S; Jansen, K E; Sahni, O; Diachin, L A

    2007-01-01

    This paper discusses methods being developed by the ITAPS center to support the execution of parallel adaptive simulations on unstructured meshes. The paper first outlines the ITAPS approach to the development of interoperable mesh, geometry and field services to support the needs of SciDAC application in these areas. The paper then demonstrates the ability of unstructured adaptive meshing methods built on such interoperable services to effectively solve important physics problems. Attention is then focused on ITAPs' developing ability to solve adaptive unstructured mesh problems on massively parallel computers

  1. Generation of an ABCG2GFPn-puro transgenic line - A tool to study ABCG2 expression in mice

    Orford, Michael; Mean, Richard; Lapathitis, George; Genethliou, Nicholas; Panayiotou, Elena; Panayi, Helen; Malas, Stavros

    2009-01-01

    The ATP-binding cassette (ABC) transporter 2 (ABCG2) is expressed by stem cells in many organs and in stem cells of solid tumors. These cells are isolated based on the side population (SP) phenotype, a Hoechst 3342 dye efflux property believed to be conferred by ABCG2. Because of the limitations of this approach we generated transgenic mice that express Nuclear GFP (GFPn) coupled to the Puromycin-resistance gene, under the control of ABCG2 promoter/enhancer sequences. We show that ABCG2 is expressed in neural progenitors of the developing forebrain and spinal cord and in embryonic and adult endothelial cells of the brain. Using the neurosphere assay, we isolated tripotent ABCG2-expressing neural stem cells from embryonic mouse brain. This transgenic line is a powerful tool for studying the expression of ABCG2 in many tissues and for performing functional studies in different experimental settings.

  2. Generation of an ABCG2{sup GFPn-puro} transgenic line - A tool to study ABCG2 expression in mice

    Orford, Michael; Mean, Richard; Lapathitis, George; Genethliou, Nicholas; Panayiotou, Elena; Panayi, Helen [The Cyprus Institute of Neurology and Genetics, Airport Avenue, No. 6, Agios Dometios 2370, Nicosia (Cyprus); Malas, Stavros, E-mail: smalas@cing.ac.cy [The Cyprus Institute of Neurology and Genetics, Airport Avenue, No. 6, Agios Dometios 2370, Nicosia (Cyprus); Department of Biological Sciences, University of Cyprus, P.O. Box 20537, 1678 Nicosia (Cyprus)

    2009-06-26

    The ATP-binding cassette (ABC) transporter 2 (ABCG2) is expressed by stem cells in many organs and in stem cells of solid tumors. These cells are isolated based on the side population (SP) phenotype, a Hoechst 3342 dye efflux property believed to be conferred by ABCG2. Because of the limitations of this approach we generated transgenic mice that express Nuclear GFP (GFPn) coupled to the Puromycin-resistance gene, under the control of ABCG2 promoter/enhancer sequences. We show that ABCG2 is expressed in neural progenitors of the developing forebrain and spinal cord and in embryonic and adult endothelial cells of the brain. Using the neurosphere assay, we isolated tripotent ABCG2-expressing neural stem cells from embryonic mouse brain. This transgenic line is a powerful tool for studying the expression of ABCG2 in many tissues and for performing functional studies in different experimental settings.

  3. Development of new generation software tools for simulation of electron beam formation in novel high power gyrotrons

    Sabchevski, S.; Zhelyazkov, I.; Benova, E.; Atanassov, V.; Dankov, P.; Thumm, M.; Dammertz, G.; Piosczyk, B.; Illy, S.; Tran, M. Q.; Alberti, S.; Hogge, J.-Ph

    2006-07-01

    Computer aided design (CAD) based on numerical experiments performed by using adequate physical models and efficient simulation codes is an indispensable tool for development, investigation, and optimization of gyrotrons used as radiation sources for electron cyclotron resonance heating (ECRH) of fusion plasmas. In this paper, we review briefly the state-of-the-art in the field of modelling and simulation of intense, relativistic, helical electron beams formed in the electron-optical systems (EOS) of powerful gyrotrons. We discuss both the limitations of the known computer codes and the requirements for increasing their capabilities for solution of various design problems that are being envisaged in the development of the next generation gyrotrons for ECRH. Moreover, we present the concept followed by us in an attempt to unite the advantages of the modern programming techniques with self-consistent, first-principles 3D physical models in the creation of a new highly efficient and versatile software package for simulation of powerful gyrotrons.

  4. Solving kinetic equations with adaptive mesh in phase space for rarefied gas dynamics and plasma physics (Invited)

    Kolobov, Vladimir; Arslanbekov, Robert; Frolova, Anna

    2014-01-01

    The paper describes an Adaptive Mesh in Phase Space (AMPS) technique for solving kinetic equations with deterministic mesh-based methods. The AMPS technique allows automatic generation of adaptive Cartesian mesh in both physical and velocity spaces using a Tree-of-Trees data structure. We illustrate advantages of AMPS for simulations of rarefied gas dynamics and electron kinetics on low temperature plasmas. In particular, we consider formation of the velocity distribution functions in hypersonic flows, particle kinetics near oscillating boundaries, and electron kinetics in a radio-frequency sheath. AMPS provide substantial savings in computational cost and increased efficiency of the mesh-based kinetic solvers

  5. Solving kinetic equations with adaptive mesh in phase space for rarefied gas dynamics and plasma physics (Invited)

    Kolobov, Vladimir [CFD Research Corporation, Huntsville, AL 35805, USA and The University of Alabama in Huntsville, Huntsville, AL 35805 (United States); Arslanbekov, Robert [CFD Research Corporation, Huntsville, AL 35805 (United States); Frolova, Anna [Computing Center of the Russian Academy of Sciences, Moscow, 119333 (Russian Federation)

    2014-12-09

    The paper describes an Adaptive Mesh in Phase Space (AMPS) technique for solving kinetic equations with deterministic mesh-based methods. The AMPS technique allows automatic generation of adaptive Cartesian mesh in both physical and velocity spaces using a Tree-of-Trees data structure. We illustrate advantages of AMPS for simulations of rarefied gas dynamics and electron kinetics on low temperature plasmas. In particular, we consider formation of the velocity distribution functions in hypersonic flows, particle kinetics near oscillating boundaries, and electron kinetics in a radio-frequency sheath. AMPS provide substantial savings in computational cost and increased efficiency of the mesh-based kinetic solvers.

  6. A conceptual framework for technology-enabled and technology-dependent user behavior toward device mesh and mesh app

    Ming-Hsiung Hsiao

    2018-06-01

    Full Text Available The device mesh and mesh app revealed by Gartner as the future strategic technology trend are able to predict people's need from their historic data, then provides the needed services or service innovation to support their activity engagement. However, many theories have identified that it is the motivation, rather than technology, that drives people to engage in activities or tasks. For this reason, this study builds a conceptual framework by integrating the extant logic and theories to explore how future technology would generate benefits for people. It integrates task-technology fit (TTF model and motivation theory (mainly expectancy-value theory to explain such technology user behavior. It also points out the difference between technology-enabled and technology-dependent user behavior and concludes that too much emphasis on the role of technology with too little attention on motivation would distort technology user behavior, and the role of technology as well. Keywords: Device mesh, Mesh app, Expectancy-value theory, Task-technology fit (TTF, Technology-enabled user, Technology-dependent user

  7. Next-generation phage display: integrating and comparing available molecular tools to enable cost-effective high-throughput analysis.

    Emmanuel Dias-Neto

    2009-12-01

    Full Text Available Combinatorial phage display has been used in the last 20 years in the identification of protein-ligands and protein-protein interactions, uncovering relevant molecular recognition events. Rate-limiting steps of combinatorial phage display library selection are (i the counting of transducing units and (ii the sequencing of the encoded displayed ligands. Here, we adapted emerging genomic technologies to minimize such challenges.We gained efficiency by applying in tandem real-time PCR for rapid quantification to enable bacteria-free phage display library screening, and added phage DNA next-generation sequencing for large-scale ligand analysis, reporting a fully integrated set of high-throughput quantitative and analytical tools. The approach is far less labor-intensive and allows rigorous quantification; for medical applications, including selections in patients, it also represents an advance for quantitative distribution analysis and ligand identification of hundreds of thousands of targeted particles from patient-derived biopsy or autopsy in a longer timeframe post library administration. Additional advantages over current methods include increased sensitivity, less variability, enhanced linearity, scalability, and accuracy at much lower cost. Sequences obtained by qPhage plus pyrosequencing were similar to a dataset produced from conventional Sanger-sequenced transducing-units (TU, with no biases due to GC content, codon usage, and amino acid or peptide frequency. These tools allow phage display selection and ligand analysis at >1,000-fold faster rate, and reduce costs approximately 250-fold for generating 10(6 ligand sequences.Our analyses demonstrates that whereas this approach correlates with the traditional colony-counting, it is also capable of a much larger sampling, allowing a faster, less expensive, more accurate and consistent analysis of phage enrichment. Overall, qPhage plus pyrosequencing is superior to TU-counting plus Sanger

  8. Tensile Behaviour of Welded Wire Mesh and Hexagonal Metal Mesh for Ferrocement Application

    Tanawade, A. G.; Modhera, C. D.

    2017-08-01

    Tension tests were conducted on welded mesh and hexagonal Metal mesh. Welded Mesh is available in the market in different sizes. The two types are analysed viz. Ø 2.3 mm and Ø 2.7 mm welded mesh, having opening size 31.75 mm × 31.75 mm and 25.4 mm × 25.4 mm respectively. Tensile strength test was performed on samples of welded mesh in three different orientations namely 0°, 30° and 45° degrees with the loading axis and hexagonal Metal mesh of Ø 0.7 mm, having opening 19.05 × 19.05 mm. Experimental tests were conducted on samples of these meshes. The objective of this study was to investigate the behaviour of the welded mesh and hexagonal Metal mesh. The result shows that the tension load carrying capacity of welded mesh of Ø 2.7 mm of 0° orientation is good as compared to Ø2.3 mm mesh and ductility of hexagonal Metal mesh is good in behaviour.

  9. Mesh optimization for microbial fuel cell cathodes constructed around stainless steel mesh current collectors

    Zhang, Fang; Merrill, Matthew D.; Tokash, Justin C.; Saito, Tomonori; Cheng, Shaoan; Hickner, Michael A.; Logan, Bruce E.

    2011-01-01

    that the mesh properties of these cathodes can significantly affect performance. Cathodes made from the coarsest mesh (30-mesh) achieved the highest maximum power of 1616 ± 25 mW m-2 (normalized to cathode projected surface area; 47.1 ± 0.7 W m-3 based on liquid

  10. Intravesical midurethral sling mesh erosion secondary to transvaginal mesh reconstructive surgery

    Sukanda Bin Jaili

    2015-05-01

    Conclusion: Repeated vaginal reconstructive surgery may jeopardize a primary mesh or sling, and pose a high risk of mesh erosion, which may be delayed for several years. Removal of the mesh erosion and bladder repair are feasible pervaginally with good outcome.

  11. GeneAnalytics: An Integrative Gene Set Analysis Tool for Next Generation Sequencing, RNAseq and Microarray Data.

    Ben-Ari Fuchs, Shani; Lieder, Iris; Stelzer, Gil; Mazor, Yaron; Buzhor, Ella; Kaplan, Sergey; Bogoch, Yoel; Plaschkes, Inbar; Shitrit, Alina; Rappaport, Noa; Kohn, Asher; Edgar, Ron; Shenhav, Liraz; Safran, Marilyn; Lancet, Doron; Guan-Golan, Yaron; Warshawsky, David; Shtrichman, Ronit

    2016-03-01

    Postgenomics data are produced in large volumes by life sciences and clinical applications of novel omics diagnostics and therapeutics for precision medicine. To move from "data-to-knowledge-to-innovation," a crucial missing step in the current era is, however, our limited understanding of biological and clinical contexts associated with data. Prominent among the emerging remedies to this challenge are the gene set enrichment tools. This study reports on GeneAnalytics™ ( geneanalytics.genecards.org ), a comprehensive and easy-to-apply gene set analysis tool for rapid contextualization of expression patterns and functional signatures embedded in the postgenomics Big Data domains, such as Next Generation Sequencing (NGS), RNAseq, and microarray experiments. GeneAnalytics' differentiating features include in-depth evidence-based scoring algorithms, an intuitive user interface and proprietary unified data. GeneAnalytics employs the LifeMap Science's GeneCards suite, including the GeneCards®--the human gene database; the MalaCards-the human diseases database; and the PathCards--the biological pathways database. Expression-based analysis in GeneAnalytics relies on the LifeMap Discovery®--the embryonic development and stem cells database, which includes manually curated expression data for normal and diseased tissues, enabling advanced matching algorithm for gene-tissue association. This assists in evaluating differentiation protocols and discovering biomarkers for tissues and cells. Results are directly linked to gene, disease, or cell "cards" in the GeneCards suite. Future developments aim to enhance the GeneAnalytics algorithm as well as visualizations, employing varied graphical display items. Such attributes make GeneAnalytics a broadly applicable postgenomics data analyses and interpretation tool for translation of data to knowledge-based innovation in various Big Data fields such as precision medicine, ecogenomics, nutrigenomics, pharmacogenomics, vaccinomics

  12. Online newspapers as an effective tool to generate interest in reading for students of Youth and Adult Education

    Daniel Luis Notari

    2017-10-01

    Full Text Available One of the main challenges to improve the quality of Brazilian education is to promote reading and an understanding of what is read. This situation is more complex in Youth and Adult Education, which has focused on students who exhibited academic failure in mainstream classes or had not engaged in reading and writing for many years. In addition to adolescents, adults and elderly individuals have the same interests when they are in school; however, the identification of a way to engage these different audiences is a challenge. Thus, we propose reading newspapers online to generate an interest in reading, as well as determining how to select updated texts that are able to draw the attention of students. The findings indicated 100% of the students were satisfied with the activity. Nevertheless, when the times of reading performed in the Educational Computer Laboratory and classroom were compared, the evaluation that employed computer tools was far superior to the classroom. We conclude that although online newspaper reading is quite simple and easy to perform, it is an important tool capable of stimulating and developing the taste and habit of reading and is effective for different audiences.

  13. Epidemiology of transmissible diseases: Array hybridization and next generation sequencing as universal nucleic acid-mediated typing tools.

    Michael Dunne, W; Pouseele, Hannes; Monecke, Stefan; Ehricht, Ralf; van Belkum, Alex

    2017-09-21

    The magnitude of interest in the epidemiology of transmissible human diseases is reflected in the vast number of tools and methods developed recently with the expressed purpose to characterize and track evolutionary changes that occur in agents of these diseases over time. Within the past decade a new suite of such tools has become available with the emergence of the so-called "omics" technologies. Among these, two are exponents of the ongoing genomic revolution. Firstly, high-density nucleic acid probe arrays have been proposed and developed using various chemical and physical approaches. Via hybridization-mediated detection of entire genes or genetic polymorphisms in such genes and intergenic regions these so called "DNA chips" have been successfully applied for distinguishing very closely related microbial species and strains. Second and even more phenomenal, next generation sequencing (NGS) has facilitated the assessment of the complete nucleotide sequence of entire microbial genomes. This technology currently provides the most detailed level of bacterial genotyping and hence allows for the resolution of microbial spread and short-term evolution in minute detail. We will here review the very recent history of these two technologies, sketch their usefulness in the elucidation of the spread and epidemiology of mostly hospital-acquired infections and discuss future developments. Copyright © 2017 Elsevier B.V. All rights reserved.

  14. A new generation of tools for search, recovery and quality evaluation of World Wide Web medical resources.

    Aguillo, I

    2000-01-01

    Although the Internet is already a valuable information resource in medicine, there are important challenges to be faced before physicians and general users will have extensive access to this information. As a result of a research effort to compile a health-related Internet directory, new tools and strategies have been developed to solve key problems derived from the explosive growth of medical information on the Net and the great concern over the quality of such critical information. The current Internet search engines lack some important capabilities. We suggest using second generation tools (client-side based) able to deal with large quantities of data and to increase the usability of the records recovered. We tested the capabilities of these programs to solve health-related information problems, recognising six groups according to the kind of topics addressed: Z39.50 clients, downloaders, multisearchers, tracing agents, indexers and mappers. The evaluation of the quality of health information available on the Internet could require a large amount of human effort. A possible solution may be to use quantitative indicators based on the hypertext visibility of the Web sites. The cybermetric measures are valid for quality evaluation if they are derived from indirect peer review by experts with Web pages citing the site. The hypertext links acting as citations need to be extracted from a controlled sample of quality super-sites.

  15. Analyzing Item Generation with Natural Language Processing Tools for the "TOEIC"® Listening Test. Research Report. ETS RR-17-52

    Yoon, Su-Youn; Lee, Chong Min; Houghton, Patrick; Lopez, Melissa; Sakano, Jennifer; Loukina, Anastasia; Krovetz, Bob; Lu, Chi; Madani, Nitin

    2017-01-01

    In this study, we developed assistive tools and resources to support TOEIC® Listening test item generation. There has recently been an increased need for a large pool of items for these tests. This need has, in turn, inspired efforts to increase the efficiency of item generation while maintaining the quality of the created items. We aimed to…

  16. Quadrilateral mesh fitting that preserves sharp features based on multi-normals for Laplacian energy

    Yusuke Imai

    2014-04-01

    Full Text Available Because the cost of performance testing using actual products is expensive, manufacturers use lower-cost computer-aided design simulations for this function. In this paper, we propose using hexahedral meshes, which are more accurate than tetrahedral meshes, for finite element analysis. We propose automatic hexahedral mesh generation with sharp features to precisely represent the corresponding features of a target shape. Our hexahedral mesh is generated using a voxel-based algorithm. In our previous works, we fit the surface of the voxels to the target surface using Laplacian energy minimization. We used normal vectors in the fitting to preserve sharp features. However, this method could not represent concave sharp features precisely. In this proposal, we improve our previous Laplacian energy minimization by adding a term that depends on multi-normal vectors instead of using normal vectors. Furthermore, we accentuate a convex/concave surface subset to represent concave sharp features.

  17. Computational performance of Free Mesh Method applied to continuum mechanics problems

    YAGAWA, Genki

    2011-01-01

    The free mesh method (FMM) is a kind of the meshless methods intended for particle-like finite element analysis of problems that are difficult to handle using global mesh generation, or a node-based finite element method that employs a local mesh generation technique and a node-by-node algorithm. The aim of the present paper is to review some unique numerical solutions of fluid and solid mechanics by employing FMM as well as the Enriched Free Mesh Method (EFMM), which is a new version of FMM, including compressible flow and sounding mechanism in air-reed instruments as applications to fluid mechanics, and automatic remeshing for slow crack growth, dynamic behavior of solid as well as large-scale Eigen-frequency of engine block as applications to solid mechanics. PMID:21558753

  18. The utilization of mesh meteorological data maps for agricultural activity in hilly and mountainous area

    Ueyama, H.

    2008-01-01

    Hilly and mountainous areas occupy approximately 70% of Japan, and the area of farmland in these regions is decreasing; these areas are defined as those from the outer plains to the mountains. The development of strategies for the revitalization of local agriculture in hilly and mountainous areas is therefore a significant problem in Japan. Systematic agriculture is efficient in hilly and mountainous areas, and distribution maps are effective planning tools for evaluating the meteorological conditions for individual farms in those areas where farms are small and interspersed. Public agricultural research centers in each prefecture of Japan have developed mesh meteorological data maps with some kilometers grid cell resolutions for local agriculture, and have been made many studies using mesh meteorological data maps. However, critical variations exist between estimated mesh data and actual meteorological condition within the area of each grid cell. To address this problem, methods of estimating air temperature and solar radiation on a 50 m mesh (latitude 1.5 sec x longitude 2.25 sec) were developed. While many studies with mesh meteorological data maps have been made, numbers of concrete examples of utility for agricultural activity in hilly and mountainous areas have been few. This paper presents therefore some studies for utilization facilitated of mesh meteorological data maps in hilly and mountainous areas. And furthermore, it is proposed some guides to utilize mesh meteorological data maps for the purpose of revitalizing an agricultural activity in hilly and mountainous area with concrete examples

  19. Phenotype Instance Verification and Evaluation Tool (PIVET): A Scaled Phenotype Evidence Generation Framework Using Web-Based Medical Literature

    Ke, Junyuan; Ho, Joyce C; Ghosh, Joydeep; Wallace, Byron C

    2018-01-01

    Background Researchers are developing methods to automatically extract clinically relevant and useful patient characteristics from raw healthcare datasets. These characteristics, often capturing essential properties of patients with common medical conditions, are called computational phenotypes. Being generated by automated or semiautomated, data-driven methods, such potential phenotypes need to be validated as clinically meaningful (or not) before they are acceptable for use in decision making. Objective The objective of this study was to present Phenotype Instance Verification and Evaluation Tool (PIVET), a framework that uses co-occurrence analysis on an online corpus of publically available medical journal articles to build clinical relevance evidence sets for user-supplied phenotypes. PIVET adopts a conceptual framework similar to the pioneering prototype tool PheKnow-Cloud that was developed for the phenotype validation task. PIVET completely refactors each part of the PheKnow-Cloud pipeline to deliver vast improvements in speed without sacrificing the quality of the insights PheKnow-Cloud achieved. Methods PIVET leverages indexing in NoSQL databases to efficiently generate evidence sets. Specifically, PIVET uses a succinct representation of the phenotypes that corresponds to the index on the corpus database and an optimized co-occurrence algorithm inspired by the Aho-Corasick algorithm. We compare PIVET’s phenotype representation with PheKnow-Cloud’s by using PheKnow-Cloud’s experimental setup. In PIVET’s framework, we also introduce a statistical model trained on domain expert–verified phenotypes to automatically classify phenotypes as clinically relevant or not. Additionally, we show how the classification model can be used to examine user-supplied phenotypes in an online, rather than batch, manner. Results PIVET maintains the discriminative power of PheKnow-Cloud in terms of identifying clinically relevant phenotypes for the same corpus with

  20. Phenotype Instance Verification and Evaluation Tool (PIVET): A Scaled Phenotype Evidence Generation Framework Using Web-Based Medical Literature.

    Henderson, Jette; Ke, Junyuan; Ho, Joyce C; Ghosh, Joydeep; Wallace, Byron C

    2018-05-04

    Researchers are developing methods to automatically extract clinically relevant and useful patient characteristics from raw healthcare datasets. These characteristics, often capturing essential properties of patients with common medical conditions, are called computational phenotypes. Being generated by automated or semiautomated, data-driven methods, such potential phenotypes need to be validated as clinically meaningful (or not) before they are acceptable for use in decision making. The objective of this study was to present Phenotype Instance Verification and Evaluation Tool (PIVET), a framework that uses co-occurrence analysis on an online corpus of publically available medical journal articles to build clinical relevance evidence sets for user-supplied phenotypes. PIVET adopts a conceptual framework similar to the pioneering prototype tool PheKnow-Cloud that was developed for the phenotype validation task. PIVET completely refactors each part of the PheKnow-Cloud pipeline to deliver vast improvements in speed without sacrificing the quality of the insights PheKnow-Cloud achieved. PIVET leverages indexing in NoSQL databases to efficiently generate evidence sets. Specifically, PIVET uses a succinct representation of the phenotypes that corresponds to the index on the corpus database and an optimized co-occurrence algorithm inspired by the Aho-Corasick algorithm. We compare PIVET's phenotype representation with PheKnow-Cloud's by using PheKnow-Cloud's experimental setup. In PIVET's framework, we also introduce a statistical model trained on domain expert-verified phenotypes to automatically classify phenotypes as clinically relevant or not. Additionally, we show how the classification model can be used to examine user-supplied phenotypes in an online, rather than batch, manner. PIVET maintains the discriminative power of PheKnow-Cloud in terms of identifying clinically relevant phenotypes for the same corpus with which PheKnow-Cloud was originally developed, but

  1. SARAPAN-A simulated-annealing-based tool to generate random patterned-channel-age in CANDU fuel management analyses

    Kastanya, Doddy [Safety and Licensing Department, Candesco Division of Kinectrics Inc., Toronto (Canada)

    2017-02-15

    In any reactor physics analysis, the instantaneous power distribution in the core can be calculated when the actual bundle-wise burnup distribution is known. Considering the fact that CANDU (Canada Deuterium Uranium) utilizes on-power refueling to compensate for the reduction of reactivity due to fuel burnup, in the CANDU fuel management analysis, snapshots of power and burnup distributions can be obtained by simulating and tracking the reactor operation over an extended period using various tools such as the *SIMULATE module of the Reactor Fueling Simulation Program (RFSP) code. However, for some studies, such as an evaluation of a conceptual design of a next-generation CANDU reactor, the preferred approach to obtain a snapshot of the power distribution in the core is based on the patterned-channel-age model implemented in the *INSTANTAN module of the RFSP code. The objective of this approach is to obtain a representative snapshot of core conditions quickly. At present, such patterns could be generated by using a program called RANDIS, which is implemented within the *INSTANTAN module. In this work, we present an alternative approach to derive the patterned-channel-age model where a simulated-annealing-based algorithm is used to find such patterns, which produce reasonable power distributions.

  2. SARAPAN—A Simulated-Annealing-Based Tool to Generate Random Patterned-Channel-Age in CANDU Fuel Management Analyses

    Doddy Kastanya

    2017-02-01

    Full Text Available In any reactor physics analysis, the instantaneous power distribution in the core can be calculated when the actual bundle-wise burnup distribution is known. Considering the fact that CANDU (Canada Deuterium Uranium utilizes on-power refueling to compensate for the reduction of reactivity due to fuel burnup, in the CANDU fuel management analysis, snapshots of power and burnup distributions can be obtained by simulating and tracking the reactor operation over an extended period using various tools such as the *SIMULATE module of the Reactor Fueling Simulation Program (RFSP code. However, for some studies, such as an evaluation of a conceptual design of a next-generation CANDU reactor, the preferred approach to obtain a snapshot of the power distribution in the core is based on the patterned-channel-age model implemented in the *INSTANTAN module of the RFSP code. The objective of this approach is to obtain a representative snapshot of core conditions quickly. At present, such patterns could be generated by using a program called RANDIS, which is implemented within the *INSTANTAN module. In this work, we present an alternative approach to derive the patterned-channel-age model where a simulated-annealing-based algorithm is used to find such patterns, which produce reasonable power distributions.

  3. Philippine case study: comparison of health impacts, costs and benefits of different generating technologies using IAEA's tools/methodologies

    Mundo, Mary Rose Q.; Arriola, Herminio

    2000-01-01

    The comparison of three electricity generating options were considered: coal-fired thermal power plant without flue gas desulfurization (FGD), coal-fired thermal power plant with FGD and combined cycle power plant with 300 MW generating capacity each. The system expansion for both the nuclear and non-nuclear options show an increasing share in the installed capacity of coal fired plants. Considering the high ash content and sulfur emission of coal-fired power plants, this plant type was assessed further in terms of health and environmental effects using the IAEA's computer tools such as QUERI, RUWM and DAM. Since all coal-fired power plants in the country have likewise not been equipped with FGD, another option studied is that of coal plant using FGD, with the objective of eliminating excess pollutants, which was assumed to have 90% removal efficiency. In view of this forecast on coal plants there is an ever-increasing need for the use of new technologies to be used in order to decrease health impacts of the use of coal as fuel. Thus, based on the analysis of three alternatives, the use of coal-fired power plants equipped with flue gas desulfurization (FGD) should be seriously considered. (Author)

  4. Laparoscopic Pelvic Floor Repair Using Polypropylene Mesh

    Shih-Shien Weng

    2008-09-01

    Conclusion: Laparoscopic pelvic floor repair using a single piece of polypropylene mesh combined with uterosacral ligament suspension appears to be a feasible procedure for the treatment of advanced vaginal vault prolapse and enterocele. Fewer mesh erosions and postoperative pain syndromes were seen in patients who had no previous pelvic floor reconstructive surgery.

  5. Robust diamond meshes with unique wettability properties.

    Yang, Yizhou; Li, Hongdong; Cheng, Shaoheng; Zou, Guangtian; Wang, Chuanxi; Lin, Quan

    2014-03-18

    Robust diamond meshes with excellent superhydrophobic and superoleophilic properties have been fabricated. Superhydrophobicity is observed for water with varying pH from 1 to 14 with good recyclability. Reversible superhydrophobicity and hydrophilicity can be easily controlled. The diamond meshes show highly efficient water-oil separation and water pH droplet transference.

  6. Mesh-graft urethroplasty: a case report

    田中, 敏博; 滝川, 浩; 香川, 征; 長江, 浩朗

    1987-01-01

    We used a meshed free-foreskin transplant in a two-stage procedure for reconstruction of the extended stricture of urethra after direct vision urethrotomy. The results were excellent. Mesh-graft urethroplasty is a useful method for patients with extended strictures of the urethra or recurrent strictures after several operations.

  7. 7th International Meshing Roundtable '98

    Eldred, T.J.

    1998-10-01

    The goal of the 7th International Meshing Roundtable is to bring together researchers and developers from industry, academia, and government labs in a stimulating, open environment for the exchange of technical information related to the meshing process. In the past, the Roundtable has enjoyed significant participation from each of these groups from a wide variety of countries.

  8. Postoperative pain outcomes after transvaginal mesh revision.

    Danford, Jill M; Osborn, David J; Reynolds, W Stuart; Biller, Daniel H; Dmochowski, Roger R

    2015-01-01

    Although the current literature discusses mesh complications including pain, as well as suggesting different techniques for removing mesh, there is little literature regarding pain outcomes after surgical removal or revision. The purpose of this study is to determine if surgical removal or revision of vaginal mesh improves patient's subjective complaints of pelvic pain associated with original placement of mesh. After obtaining approval from the Vanderbilt University Medical Center Institutional Review Board, a retrospective review of female patients with pain secondary to previous mesh placement who underwent excision or revision of vaginal mesh from January 2000 to August 2012 was performed. Patient age, relevant medical history including menopause status, previous hysterectomy, smoking status, and presence of diabetes, fibromyalgia, interstitial cystitis, and chronic pelvic pain, was obtained. Patients' postoperative pain complaints were assessed. Of the 481 patients who underwent surgery for mesh revision, removal or urethrolysis, 233 patients met our inclusion criteria. One hundred and sixty-nine patients (73 %) reported that their pain improved, 19 (8 %) reported that their pain worsened, and 45 (19 %) reported that their pain remained unchanged after surgery. Prior history of chronic pelvic pain was associated with increased risk of failure of the procedure to relieve pain (OR 0.28, 95 % CI 0.12-0.64, p = 0.003). Excision or revision of vaginal mesh appears to be effective in improving patients' pain symptoms most of the time. Patients with a history of chronic pelvic pain are at an increased risk of no improvement or of worsening pain.

  9. Converting skeletal structures to quad dominant meshes

    Bærentzen, Jakob Andreas; Misztal, Marek Krzysztof; Welnicka, Katarzyna

    2012-01-01

    We propose the Skeleton to Quad-dominant polygonal Mesh algorithm (SQM), which converts skeletal structures to meshes composed entirely of polar and annular regions. Both types of regions have a regular structure where all faces are quads except for a single ring of triangles at the center of each...

  10. Adaptive mesh refinement in titanium

    Colella, Phillip; Wen, Tong

    2005-01-21

    In this paper, we evaluate Titanium's usability as a high-level parallel programming language through a case study, where we implement a subset of Chombo's functionality in Titanium. Chombo is a software package applying the Adaptive Mesh Refinement methodology to numerical Partial Differential Equations at the production level. In Chombo, the library approach is used to parallel programming (C++ and Fortran, with MPI), whereas Titanium is a Java dialect designed for high-performance scientific computing. The performance of our implementation is studied and compared with that of Chombo in solving Poisson's equation based on two grid configurations from a real application. Also provided are the counts of lines of code from both sides.

  11. MCR2S unstructured mesh capabilities for use in shutdown dose rate analysis

    Eade, T.; Stonell, D.; Turner, A.

    2015-01-01

    Highlights: • Advancements in shutdown dose rate calculations will be needed as fusion moves from experimental reactors to full scale demonstration reactors in order to ensure the safety of personnel. • The MCR2S shutdown dose rate tool has been modified to allow shutdown dose rates calculations using an unstructured mesh. • The unstructured mesh capability of MCR2S was used on three shutdown dose rate models, a simple sphere, the ITER computational benchmark and the DEMO computational benchmark. • The results showed a reasonable agreement between an unstructured mesh approach and the CSG approach and highlighted the need to carefully choose the unstructured mesh resolution. - Abstract: As nuclear fusion progresses towards a sustainable energy source and the power of tokamak devices increases, a greater understanding of the radiation fields will be required. As well as on-load radiation fields, off-load or shutdown radiation field are an important consideration for the safety and economic viability of a commercial fusion reactor. Previously codes such as MCR2S have been written in order to predict the shutdown dose rates within, and in regions surrounding, a fusion reactor. MCR2S utilises a constructive solid geometry (CSG) model and a superimposed structured mesh to calculate 3-D maps of the shutdown dose rate. A new approach to MCR2S calculations is proposed and implemented using a single unstructured mesh to replace both the CSG model and the superimposed structured mesh. This new MCR2S approach has been demonstrated on three models of increasing complexity. These models were: a sphere, the ITER computational shutdown dose rate benchmark and the DEMO computational shutdown dose rate benchmark. In each case the results were compared to MCR2S calculations performed using MCR2S with CSG geometry and a superimposed structured mesh. It was concluded that the results from the unstructured mesh implementation of MCR2S compared well to the CSG structured mesh

  12. Fog water collection effectiveness: Mesh intercomparisons

    Fernandez, Daniel; Torregrosa, Alicia; Weiss-Penzias, Peter; Zhang, Bong June; Sorensen, Deckard; Cohen, Robert; McKinley, Gareth; Kleingartner, Justin; Oliphant, Andrew; Bowman, Matthew

    2018-01-01

    To explore fog water harvesting potential in California, we conducted long-term measurements involving three types of mesh using standard fog collectors (SFC). Volumetric fog water measurements from SFCs and wind data were collected and recorded in 15-minute intervals over three summertime fog seasons (2014–2016) at four California sites. SFCs were deployed with: standard 1.00 m2 double-layer 35% shade coefficient Raschel; stainless steel mesh coated with the MIT-14 hydrophobic formulation; and FogHa-Tin, a German manufactured, 3-dimensional spacer fabric deployed in two orientations. Analysis of 3419 volumetric samples from all sites showed strong relationships between mesh efficiency and wind speed. Raschel mesh collected 160% more fog water than FogHa-Tin at wind speeds less than 1 m s–1 and 45% less for wind speeds greater than 5 m s–1. MIT-14 coated stainless-steel mesh collected more fog water than Raschel mesh at all wind speeds. At low wind speeds of steel mesh collected 3% more and at wind speeds of 4–5 m s–1, it collected 41% more. FogHa-Tin collected 5% more fog water when the warp of the weave was oriented vertically, per manufacturer specification, than when the warp of the weave was oriented horizontally. Time series measurements of three distinct mesh across similar wind regimes revealed inconsistent lags in fog water collection and inconsistent performance. Since such differences occurred under similar wind-speed regimes, we conclude that other factors play important roles in mesh performance, including in-situ fog event and aerosol dynamics that affect droplet-size spectra and droplet-to-mesh surface interactions.

  13. Interoperable mesh components for large-scale, distributed-memory simulations

    Devine, K; Leung, V; Diachin, L; Miller, M

    2009-01-01

    SciDAC applications have a demonstrated need for advanced software tools to manage the complexities associated with sophisticated geometry, mesh, and field manipulation tasks, particularly as computer architectures move toward the petascale. In this paper, we describe a software component - an abstract data model and programming interface - designed to provide support for parallel unstructured mesh operations. We describe key issues that must be addressed to successfully provide high-performance, distributed-memory unstructured mesh services and highlight some recent research accomplishments in developing new load balancing and MPI-based communication libraries appropriate for leadership class computing. Finally, we give examples of the use of parallel adaptive mesh modification in two SciDAC applications.

  14. 3D-2D Deformable Image Registration Using Feature-Based Nonuniform Meshes.

    Zhong, Zichun; Guo, Xiaohu; Cai, Yiqi; Yang, Yin; Wang, Jing; Jia, Xun; Mao, Weihua

    2016-01-01

    By using prior information of planning CT images and feature-based nonuniform meshes, this paper demonstrates that volumetric images can be efficiently registered with a very small portion of 2D projection images of a Cone-Beam Computed Tomography (CBCT) scan. After a density field is computed based on the extracted feature edges from planning CT images, nonuniform tetrahedral meshes will be automatically generated to better characterize the image features according to the density field; that is, finer meshes are generated for features. The displacement vector fields (DVFs) are specified at the mesh vertices to drive the deformation of original CT images. Digitally reconstructed radiographs (DRRs) of the deformed anatomy are generated and compared with corresponding 2D projections. DVFs are optimized to minimize the objective function including differences between DRRs and projections and the regularity. To further accelerate the above 3D-2D registration, a procedure to obtain good initial deformations by deforming the volume surface to match 2D body boundary on projections has been developed. This complete method is evaluated quantitatively by using several digital phantoms and data from head and neck cancer patients. The feature-based nonuniform meshing method leads to better results than either uniform orthogonal grid or uniform tetrahedral meshes.

  15. 3D-2D Deformable Image Registration Using Feature-Based Nonuniform Meshes

    Zichun Zhong

    2016-01-01

    Full Text Available By using prior information of planning CT images and feature-based nonuniform meshes, this paper demonstrates that volumetric images can be efficiently registered with a very small portion of 2D projection images of a Cone-Beam Computed Tomography (CBCT scan. After a density field is computed based on the extracted feature edges from planning CT images, nonuniform tetrahedral meshes will be automatically generated to better characterize the image features according to the density field; that is, finer meshes are generated for features. The displacement vector fields (DVFs are specified at the mesh vertices to drive the deformation of original CT images. Digitally reconstructed radiographs (DRRs of the deformed anatomy are generated and compared with corresponding 2D projections. DVFs are optimized to minimize the objective function including differences between DRRs and projections and the regularity. To further accelerate the above 3D-2D registration, a procedure to obtain good initial deformations by deforming the volume surface to match 2D body boundary on projections has been developed. This complete method is evaluated quantitatively by using several digital phantoms and data from head and neck cancer patients. The feature-based nonuniform meshing method leads to better results than either uniform orthogonal grid or uniform tetrahedral meshes.

  16. Element Partition Trees For H-Refined Meshes to Optimize Direct Solver Performance. Part I: Dynamic Programming

    AbouEisha, Hassan M.

    2017-07-13

    We consider a class of two-and three-dimensional h-refined meshes generated by an adaptive finite element method. We introduce an element partition tree, which controls the execution of the multi-frontal solver algorithm over these refined grids. We propose and study algorithms with polynomial computational cost for the optimization of these element partition trees. The trees provide an ordering for the elimination of unknowns. The algorithms automatically optimize the element partition trees using extensions of dynamic programming. The construction of the trees by the dynamic programming approach is expensive. These generated trees cannot be used in practice, but rather utilized as a learning tool to propose fast heuristic algorithms. In this first part of our paper we focus on the dynamic programming approach, and draw a sketch of the heuristic algorithm. The second part will be devoted to a more detailed analysis of the heuristic algorithm extended for the case of hp-adaptive

  17. Element Partition Trees For H-Refined Meshes to Optimize Direct Solver Performance. Part I: Dynamic Programming

    AbouEisha, Hassan M.; Calo, Victor Manuel; Jopek, Konrad; Moshkov, Mikhail; Paszyńka, Anna; Paszyński, Maciej; Skotniczny, Marcin

    2017-01-01

    We consider a class of two-and three-dimensional h-refined meshes generated by an adaptive finite element method. We introduce an element partition tree, which controls the execution of the multi-frontal solver algorithm over these refined grids. We propose and study algorithms with polynomial computational cost for the optimization of these element partition trees. The trees provide an ordering for the elimination of unknowns. The algorithms automatically optimize the element partition trees using extensions of dynamic programming. The construction of the trees by the dynamic programming approach is expensive. These generated trees cannot be used in practice, but rather utilized as a learning tool to propose fast heuristic algorithms. In this first part of our paper we focus on the dynamic programming approach, and draw a sketch of the heuristic algorithm. The second part will be devoted to a more detailed analysis of the heuristic algorithm extended for the case of hp-adaptive

  18. Artificial Neural Networks as an Architectural Design Tool-Generating New Detail Forms Based On the Roman Corinthian Order Capital

    Radziszewski, Kacper

    2017-10-01

    The following paper presents the results of the research in the field of the machine learning, investigating the scope of application of the artificial neural networks algorithms as a tool in architectural design. The computational experiment was held using the backward propagation of errors method of training the artificial neural network, which was trained based on the geometry of the details of the Roman Corinthian order capital. During the experiment, as an input training data set, five local geometry parameters combined has given the best results: Theta, Pi, Rho in spherical coordinate system based on the capital volume centroid, followed by Z value of the Cartesian coordinate system and a distance from vertical planes created based on the capital symmetry. Additionally during the experiment, artificial neural network hidden layers optimal count and structure was found, giving results of the error below 0.2% for the mentioned before input parameters. Once successfully trained artificial network, was able to mimic the details composition on any other geometry type given. Despite of calculating the transformed geometry locally and separately for each of the thousands of surface points, system could create visually attractive and diverse, complex patterns. Designed tool, based on the supervised learning method of machine learning, gives possibility of generating new architectural forms- free of the designer’s imagination bounds. Implementing the infinitely broad computational methods of machine learning, or Artificial Intelligence in general, not only could accelerate and simplify the design process, but give an opportunity to explore never seen before, unpredictable forms or everyday architectural practice solutions.

  19. Hazard assessment of exhaust emissions - The next generation of fast and reliable tools for in vitro screening

    Rothen-Rutishauser, B.

    2017-12-01

    Hazard assessment of exhaust emissions - The next generation of fast and reliable tools for in vitro screening Barbara Rothen-Rutishauser Adolphe Merkle Institute, University of Fribourg, Switzerland; barbara.rothen@unifr.ch Pollution by vehicles is a major problem for the environment due to the various components in the exhaust gasses that are emitted into the atmosphere. A large number of epidemiological studies demonstrate the profound impact of vehicle emissions upon human health [1-3]. Such studies however, are unable to attribute a given subset of emissions to a certain adverse effect, which renders decision making difficult. Standardized protocols for exhaust toxicity assessment are lacking and it relies in many aspects on epidemiological and in vivo studies (animals), which are very time and cost-intensive and suffer from considerable ethical issues. An overview about the current state of research and clinical aspects in the field, as well as about the development of sophisticated in vitro approaches mimicking the inhalation of airborne particles / exhaust for the toxicological testing of engine emissions will be provided. Data will be presented that show that the combination of an air-liquid exposure system and 3D lung-cell culture model offers an adequate tool for fast and reliable investigations of complete exhaust toxicity as well as the effects of particulate fraction [4,5]. This approach yields important results for novel and improved emission technologies in the early stages of product development. [1] Donaldson et al. Part Fibre Toxicol 2005, 2: 10. [2] Ghio et al. J Toxicol Environ Health B Crit Rev 2012, 15: 1-21. [3] Peters et al. Res Rep Health Eff Inst 2009, 5-77. [4] Bisig et al. Emiss Control Sci Technol 2015, 1: 237-246. [5] Steiner et al. Atmos Environ 2013, 81: 380-388.

  20. Proceedings of the 20th International Meshing Roundtable

    2012-01-01

    This volume contains the articles presented at the 20th International Meshing Roundtable (IMR) organized, in part, by Sandia National Laboratories and was held in Paris, France on Oct 23-26, 2011. This is the first year the IMR was held outside the United States territory. Other sponsors of the 20th IMR are Systematic Paris Region Systems & ICT Cluster, AIAA, NAFEMS, CEA, and NSF. The Sandia National Laboratories started the first IMR in 1992, and the conference has been held annually since. Each year the IMR brings together researchers, developers, and application experts, from a variety of disciplines, to present and discuss ideas on mesh generation and related topics. The topics covered by the IMR have applications in numerical analysis, computational geometry, computer graphics, as well as other areas, and the presentations describe novel work ranging from theory to application.     .

  1. Smart-Home Architecture Based on Bluetooth mesh Technology

    Wan, Qing; Liu, Jianghua

    2018-03-01

    This paper describes the smart home network system based on Nordic nrf52832 device. Nrf52832 is new generation RF SOC device focus on sensor monitor and low power Bluetooth connection applications. In this smart home system, we set up a self-organizing network system which consists of one control node and a lot of monitor nodes. The control node manages the whole network works; the monitor nodes collect the sensor information such as light intensity, temperature, humidity, PM2.5, etc. Then update to the control node by Bluetooth mesh network. The design results show that the Bluetooth mesh wireless network system is flexible and construction cost is low, which is suitable for the communication characteristics of a smart home network. We believe it will be wildly used in the future.

  2. [CLINICAL EVALUATION OF THE NEW ANTISEPTIC MESHES].

    Gogoladze, M; Kiladze, M; Chkhikvadze, T; Jiqia, D

    2016-12-01

    Improving the results of hernia treatment and prevention of complications became a goal of our research which included two parts - experimental and clinical. Histomorphological and bacteriological researches showed that the best result out of the 3 control groups was received in case of covering implant "Coladerm"+ with chlorhexidine. Based on the experiment results working process continued in clinics in order to test and introduce new "coladerm"+ chlorhexidine covered poliprophilene meshes into practice. For clinical illustration there were 60 patients introduced to the research who had hernioplasty procedures by different nets: I group - standard meshes+"coladerm"+chlorhexidine, 35 patients; II group - standard meshes +"coladerm", 15 patients; III group - standard meshes, 10 patients. Assessment of the wound and echo-control was done post-surgery on the 8th, 30th and 90th days. This clinical research based on the experimental results once again showed the best anti-microbe features of new antiseptic polymeric biocomposite meshes (standard meshes+"coladerm"+chlorhexidine); timely termination of regeneration and reparation processes without any post-surgery suppurative complications. We hope that new antiseptic polymeric biocomposite meshes presented by us will be successfully used in surgical practice of hernia treatment based on and supported by expermental-clinical research.

  3. Analysis of the Numerical Diffusion in Anisotropic Mediums: Benchmarks for Magnetic Field Aligned Meshes in Space Propulsion Simulations

    Daniel Pérez-Grande

    2016-11-01

    Full Text Available This manuscript explores numerical errors in highly anisotropic diffusion problems. First, the paper addresses the use of regular structured meshes in numerical solutions versus meshes aligned with the preferential directions of the problem. Numerical diffusion in structured meshes is quantified by solving the classical anisotropic diffusion problem; the analysis is exemplified with the application to a numerical model of conducting fluids under magnetic confinement, where rates of transport in directions parallel and perpendicular to a magnetic field are quite different. Numerical diffusion errors in this problem promote the use of magnetic field aligned meshes (MFAM. The generation of this type of meshes presents some challenges; several meshing strategies are implemented and analyzed in order to provide insight into achieving acceptable mesh regularity. Second, Gradient Reconstruction methods for magnetically aligned meshes are addressed and numerical errors are compared for the structured and magnetically aligned meshes. It is concluded that using the latter provides a more correct and straightforward approach to solving problems where anisotropicity is present, especially, if the anisotropicity level is high or difficult to quantify. The conclusions of the study may be extrapolated to the study of anisotropic flows different from conducting fluids.

  4. Modeling NIF experimental designs with adaptive mesh refinement and Lagrangian hydrodynamics

    Koniges, A. E.; Anderson, R. W.; Wang, P.; Gunney, B. T. N.; Becker, R.; Eder, D. C.; MacGowan, B. J.; Schneider, M. B.

    2006-06-01

    Incorporation of adaptive mesh refinement (AMR) into Lagrangian hydrodynamics algorithms allows for the creation of a highly powerful simulation tool effective for complex target designs with three-dimensional structure. We are developing an advanced modeling tool that includes AMR and traditional arbitrary Lagrangian-Eulerian (ALE) techniques. Our goal is the accurate prediction of vaporization, disintegration and fragmentation in National Ignition Facility (NIF) experimental target elements. Although our focus is on minimizing the generation of shrapnel in target designs and protecting the optics, the general techniques are applicable to modern advanced targets that include three-dimensional effects such as those associated with capsule fill tubes. Several essential computations in ordinary radiation hydrodynamics need to be redesigned in order to allow for AMR to work well with ALE, including algorithms associated with radiation transport. Additionally, for our goal of predicting fragmentation, we include elastic/plastic flow into our computations. We discuss the integration of these effects into a new ALE-AMR simulation code. Applications of this newly developed modeling tool as well as traditional ALE simulations in two and three dimensions are applied to NIF early-light target designs.

  5. Modeling NIF Experimental Designs with Adaptive Mesh Refinement and Lagrangian Hydrodynamics

    Koniges, A E; Anderson, R W; Wang, P; Gunney, B N; Becker, R; Eder, D C; MacGowan, B J

    2005-01-01

    Incorporation of adaptive mesh refinement (AMR) into Lagrangian hydrodynamics algorithms allows for the creation of a highly powerful simulation tool effective for complex target designs with three-dimensional structure. We are developing an advanced modeling tool that includes AMR and traditional arbitrary Lagrangian-Eulerian (ALE) techniques. Our goal is the accurate prediction of vaporization, disintegration and fragmentation in National Ignition Facility (NIF) experimental target elements. Although our focus is on minimizing the generation of shrapnel in target designs and protecting the optics, the general techniques are applicable to modern advanced targets that include three-dimensional effects such as those associated with capsule fill tubes. Several essential computations in ordinary radiation hydrodynamics need to be redesigned in order to allow for AMR to work well with ALE, including algorithms associated with radiation transport. Additionally, for our goal of predicting fragmentation, we include elastic/plastic flow into our computations. We discuss the integration of these effects into a new ALE-AMR simulation code. Applications of this newly developed modeling tool as well as traditional ALE simulations in two and three dimensions are applied to NIF early-light target designs

  6. Modeling Nif experimental designs with adaptive mesh refinement and Lagrangian hydrodynamics

    Koniges, A.E.; Anderson, R.W.; Wang, P.; Gunney, B.T.N.; Becker, R.; Eder, D.C.; MacGowan, B.J.; Schneider, M.B.

    2006-01-01

    Incorporation of adaptive mesh refinement (AMR) into Lagrangian hydrodynamics algorithms allows for the creation of a highly powerful simulation tool effective for complex target designs with three-dimensional structure. We are developing an advanced modeling tool that includes AMR and traditional arbitrary Lagrangian-Eulerian (ALE) techniques. Our goal is the accurate prediction of vaporization, disintegration and fragmentation in National Ignition Facility (NIF) experimental target elements. Although our focus is on minimizing the generation of shrapnel in target designs and protecting the optics, the general techniques are applicable to modern advanced targets that include three-dimensional effects such as those associated with capsule fill tubes. Several essential computations in ordinary radiation hydrodynamics need to be redesigned in order to allow for AMR to work well with ALE, including algorithms associated with radiation transport. Additionally, for our goal of predicting fragmentation, we include elastic/plastic flow into our computations. We discuss the integration of these effects into a new ALE-AMR simulation code. Applications of this newly developed modeling tool as well as traditional ALE simulations in two and three dimensions are applied to NIF early-light target designs. (authors)

  7. Fog water collection effectiveness: Mesh intercomparisons

    Fernandez, Daniel; Torregrosa, Alicia; Weiss-Penzias, Peter; Zhang, Bong June; Sorensen, Deckard; Cohen, Robert; McKinley, Gareth; Kleingartner, Justin; Oliphant, Andrew; Bowman, Matthew

    2018-01-01

    To explore fog water harvesting potential in California, we conducted long-term measurements involving three types of mesh using standard fog collectors (SFC). Volumetric fog water measurements from SFCs and wind data were collected and recorded in 15-minute intervals over three summertime fog seasons (2014–2016) at four California sites. SFCs were deployed with: standard 1.00 m2 double-layer 35% shade coefficient Raschel; stainless steel mesh coated with the MIT-14 hydrophobic formulation; and FogHa-Tin, a German manufactured, 3-dimensional spacer fabric deployed in two orientations. Analysis of 3419 volumetric samples from all sites showed strong relationships between mesh efficiency and wind speed. Raschel mesh collected 160% more fog water than FogHa-Tin at wind speeds less than 1 m s–1 and 45% less for wind speeds greater than 5 m s–1. MIT-14 coated stainless-steel mesh collected more fog water than Raschel mesh at all wind speeds. At low wind speeds of wind speeds of 4–5 m s–1, it collected 41% more. FogHa-Tin collected 5% more fog water when the warp of the weave was oriented vertically, per manufacturer specification, than when the warp of the weave was oriented horizontally. Time series measurements of three distinct mesh across similar wind regimes revealed inconsistent lags in fog water collection and inconsistent performance. Since such differences occurred under similar wind-speed regimes, we conclude that other factors play important roles in mesh performance, including in-situ fog event and aerosol dynamics that affect droplet-size spectra and droplet-to-mesh surface interactions.

  8. Transvaginal mesh procedures for pelvic organ prolapse.

    Walter, Jens-Erik

    2011-02-01

    To provide an update on transvaginal mesh procedures, newly available minimally invasive surgical techniques for pelvic floor repair. The discussion is limited to minimally invasive transvaginal mesh procedures. PubMed and Medline were searched for articles published in English, using the key words "pelvic organ prolapse," transvaginal mesh," and "minimally invasive surgery." Results were restricted to systematic reviews, randomized control trials/controlled clinical trials, and observational studies. Searches were updated on a regular basis, and articles were incorporated in the guideline to May 2010. Grey (unpublished) literature was identified through searching the websites of health technology assessment and health technology assessment-related agencies, clinical practice guideline collections, clinical trial registries, and national and international medical specialty societies. The quality of evidence was rated using the criteria described in the Report of the Canadian Task Force on the Preventive Health Care. Recommendations for practice were ranked according to the method described in that report (Table 1). Counselling for the surgical treatment of pelvic organ prolapse should consider all benefits, harms, and costs of the surgical procedure, with particular emphasis on the use of mesh. 1. Patients should be counselled that transvaginal mesh procedures are considered novel techniques for pelvic floor repair that demonstrate high rates of anatomical cure in uncontrolled short-term case series. (II-2B) 2. Patients should be informed of the range of success rates until stronger evidence of superiority is published. (II-2B) 3. Training specific to transvaginal mesh procedures should be undertaken before procedures are performed. (III-C) 4. Patients should undergo thorough preoperative counselling regarding (a) the potential serious adverse sequelae of transvaginal mesh repairs, including mesh exposure, pain, and dyspareunia; and (b) the limited data available

  9. Mesh optimization for microbial fuel cell cathodes constructed around stainless steel mesh current collectors

    Zhang, Fang

    2011-02-01

    Mesh current collectors made of stainless steel (SS) can be integrated into microbial fuel cell (MFC) cathodes constructed of a reactive carbon black and Pt catalyst mixture and a poly(dimethylsiloxane) (PDMS) diffusion layer. It is shown here that the mesh properties of these cathodes can significantly affect performance. Cathodes made from the coarsest mesh (30-mesh) achieved the highest maximum power of 1616 ± 25 mW m-2 (normalized to cathode projected surface area; 47.1 ± 0.7 W m-3 based on liquid volume), while the finest mesh (120-mesh) had the lowest power density (599 ± 57 mW m-2). Electrochemical impedance spectroscopy showed that charge transfer and diffusion resistances decreased with increasing mesh opening size. In MFC tests, the cathode performance was primarily limited by reaction kinetics, and not mass transfer. Oxygen permeability increased with mesh opening size, accounting for the decreased diffusion resistance. At higher current densities, diffusion became a limiting factor, especially for fine mesh with low oxygen transfer coefficients. These results demonstrate the critical nature of the mesh size used for constructing MFC cathodes. © 2010 Elsevier B.V. All rights reserved.

  10. Automatic mesh adaptivity for CADIS and FW-CADIS neutronics modeling of difficult shielding problems

    Ibrahim, A. M.; Peplow, D. E.; Mosher, S. W.; Wagner, J. C.; Evans, T. M.; Wilson, P. P.; Sawan, M. E.

    2013-01-01

    The CADIS and FW-CADIS hybrid Monte Carlo/deterministic techniques dramatically increase the efficiency of neutronics modeling, but their use in the accurate design analysis of very large and geometrically complex nuclear systems has been limited by the large number of processors and memory requirements for their preliminary deterministic calculations and final Monte Carlo calculation. Three mesh adaptivity algorithms were developed to reduce the memory requirements of CADIS and FW-CADIS without sacrificing their efficiency improvement. First, a macro-material approach enhances the fidelity of the deterministic models without changing the mesh. Second, a deterministic mesh refinement algorithm generates meshes that capture as much geometric detail as possible without exceeding a specified maximum number of mesh elements. Finally, a weight window coarsening algorithm de-couples the weight window mesh and energy bins from the mesh and energy group structure of the deterministic calculations in order to remove the memory constraint of the weight window map from the deterministic mesh resolution. The three algorithms were used to enhance an FW-CADIS calculation of the prompt dose rate throughout the ITER experimental facility. Using these algorithms resulted in a 23.3% increase in the number of mesh tally elements in which the dose rates were calculated in a 10-day Monte Carlo calculation and, additionally, increased the efficiency of the Monte Carlo simulation by a factor of at least 3.4. The three algorithms enabled this difficult calculation to be accurately solved using an FW-CADIS simulation on a regular computer cluster, obviating the need for a world-class super computer. (authors)

  11. Automatic mesh adaptivity for hybrid Monte Carlo/deterministic neutronics modeling of difficult shielding problems

    Ibrahim, Ahmad M.; Wilson, Paul P.H.; Sawan, Mohamed E.; Mosher, Scott W.; Peplow, Douglas E.; Wagner, John C.; Evans, Thomas M.; Grove, Robert E.

    2015-01-01

    The CADIS and FW-CADIS hybrid Monte Carlo/deterministic techniques dramatically increase the efficiency of neutronics modeling, but their use in the accurate design analysis of very large and geometrically complex nuclear systems has been limited by the large number of processors and memory requirements for their preliminary deterministic calculations and final Monte Carlo calculation. Three mesh adaptivity algorithms were developed to reduce the memory requirements of CADIS and FW-CADIS without sacrificing their efficiency improvement. First, a macromaterial approach enhances the fidelity of the deterministic models without changing the mesh. Second, a deterministic mesh refinement algorithm generates meshes that capture as much geometric detail as possible without exceeding a specified maximum number of mesh elements. Finally, a weight window coarsening algorithm decouples the weight window mesh and energy bins from the mesh and energy group structure of the deterministic calculations in order to remove the memory constraint of the weight window map from the deterministic mesh resolution. The three algorithms were used to enhance an FW-CADIS calculation of the prompt dose rate throughout the ITER experimental facility. Using these algorithms resulted in a 23.3% increase in the number of mesh tally elements in which the dose rates were calculated in a 10-day Monte Carlo calculation and, additionally, increased the efficiency of the Monte Carlo simulation by a factor of at least 3.4. The three algorithms enabled this difficult calculation to be accurately solved using an FW-CADIS simulation on a regular computer cluster, eliminating the need for a world-class super computer

  12. Mesh Optimization for Ground Vehicle Aerodynamics

    Adrian Gaylard; Essam F Abo-Serie; Nor Elyana Ahmad

    2010-01-01

    Mesh optimization strategy for estimating accurate drag of a ground vehicle is proposed based on examining the effect of different mesh parameters.  The optimized mesh parameters were selected using design of experiment (DOE) method to be able to work in a...

  13. Engagement of Metal Debris into Gear Mesh

    handschuh, Robert F.; Krantz, Timothy L.

    2010-01-01

    A series of bench-top experiments was conducted to determine the effects of metallic debris being dragged through meshing gear teeth. A test rig that is typically used to conduct contact fatigue experiments was used for these tests. Several sizes of drill material, shim stock and pieces of gear teeth were introduced and then driven through the meshing region. The level of torque required to drive the "chip" through the gear mesh was measured. From the data gathered, chip size sufficient to jam the mechanism can be determined.

  14. Mesh requirements for neutron transport calculations

    Askew, J.R.

    1967-07-01

    Fine-structure calculations are reported for a cylindrical natural uranium-graphite cell using different solution methods (discrete ordinate and collision probability codes) and varying the spatial mesh. It is suggested that of formulations assuming the source constant in a mesh interval the differential approach is generally to be preferred. Due to cancellation between approximations made in the derivation of the finite difference equations and the errors in neglecting source variation, the discrete ordinate code gave a more accurate estimate of fine structure for a given mesh even for unusually coarse representations. (author)

  15. Coarse-mesh discretized low-order quasi-diffusion equations for subregion averaged scalar fluxes

    Anistratov, D. Y.

    2004-01-01

    In this paper we develop homogenization procedure and discretization for the low-order quasi-diffusion equations on coarse grids for core-level reactor calculations. The system of discretized equations of the proposed method is formulated in terms of the subregion averaged group scalar fluxes. The coarse-mesh solution is consistent with a given fine-mesh discretization of the transport equation in the sense that it preserves a set of average values of the fine-mesh transport scalar flux over subregions of coarse-mesh cells as well as the surface currents, and eigenvalue. The developed method generates numerical solution that mimics the large-scale behavior of the transport solution within assemblies. (authors)

  16. Charged particle tracking through electrostatic wire meshes using the finite element method

    Devlin, L. J.; Karamyshev, O.; Welsch, C. P., E-mail: carsten.welsch@cockcroft.ac.uk [The Cockcroft Institute, Daresbury Laboratory, Warrington (United Kingdom); Department of Physics, University of Liverpool, Liverpool (United Kingdom)

    2016-06-15

    Wire meshes are used across many disciplines to accelerate and focus charged particles, however, analytical solutions are non-exact and few codes exist which simulate the exact fields around a mesh with physical sizes. A tracking code based in Matlab-Simulink using field maps generated using finite element software has been developed which tracks electrons or ions through electrostatic wire meshes. The fields around such a geometry are presented as an analytical expression using several basic assumptions, however, it is apparent that computational calculations are required to obtain realistic values of electric potential and fields, particularly when multiple wire meshes are deployed. The tracking code is flexible in that any quantitatively describable particle distribution can be used for both electrons and ions as well as other benefits such as ease of export to other programs for analysis. The code is made freely available and physical examples are highlighted where this code could be beneficial for different applications.

  17. Numerical simulation of deformation of dynamic mesh in the human vocal tract model

    Řidký Václav

    2015-01-01

    Full Text Available Numerical simulation of the acoustic signal generation in the human vocal tract is a very complex problem. The computational mesh is not static; it is deformed due to vibration of vocal folds. Movement of vocal folds is in this case prescribed as function of translation and rotation. A new boundary condition for the 2DOF motion of the vocal folds was implemented in OpenFOAM, an open-source software package based on finite volume method Work is focused on the dynamic mesh and deformation of structured meshes in the computation a package OpenFOAM. These methods are compared with focus onquality of the mesh (non-orthogonality, aspect ratio and skewness.

  18. Stress adapted embroidered meshes with a graded pattern design for abdominal wall hernia repair

    Hahn, J.; Bittrich, L.; Breier, A.; Spickenheuer, A.

    2017-10-01

    Abdominal wall hernias are one of the most relevant injuries of the digestive system with 25 million patients in 2013. Surgery is recommended primarily using allogenic non-absorbable wrap-knitted meshes. These meshes have in common that their stress-strain behaviour is not adapted to the anisotropic behaviour of native abdominal wall tissue. The ideal mesh should possess an adequate mechanical behaviour and a suitable porosity at the same time. An alternative fabrication method to wrap-knitting is the embroidery technology with a high flexibility in pattern design and adaption of mechanical properties. In this study, a pattern generator was created for pattern designs consisting of a base and a reinforcement pattern. The embroidered mesh structures demonstrated different structural and mechanical characteristics. Additionally, the investigation of the mechanical properties exhibited an anisotropic mechanical behaviour for the embroidered meshes. As a result, the investigated pattern generator and the embroidery technology allow the production of stress adapted mesh structures that are a promising approach for hernia reconstruction.

  19. Monte Carlo event generators in atomic collisions: A new tool to tackle the few-body dynamics

    Ciappina, M. F.; Kirchner, T.; Schulz, M.

    2010-04-01

    We present a set of routines to produce theoretical event files, for both single and double ionization of atoms by ion impact, based on a Monte Carlo event generator (MCEG) scheme. Such event files are the theoretical counterpart of the data obtained from a kinematically complete experiment; i.e. they contain the momentum components of all collision fragments for a large number of ionization events. Among the advantages of working with theoretical event files is the possibility to incorporate the conditions present in a real experiment, such as the uncertainties in the measured quantities. Additionally, by manipulating them it is possible to generate any type of cross sections, specially those that are usually too complicated to compute with conventional methods due to a lack of symmetry. Consequently, the numerical effort of such calculations is dramatically reduced. We show examples for both single and double ionization, with special emphasis on a new data analysis tool, called four-body Dalitz plots, developed very recently. Program summaryProgram title: MCEG Catalogue identifier: AEFV_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEFV_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 2695 No. of bytes in distributed program, including test data, etc.: 18 501 Distribution format: tar.gz Programming language: FORTRAN 77 with parallelization directives using scripting Computer: Single machines using Linux and Linux servers/clusters (with cores with any clock speed, cache memory and bits in a word) Operating system: Linux (any version and flavor) and FORTRAN 77 compilers Has the code been vectorised or parallelized?: Yes RAM: 64-128 kBytes (the codes are very cpu intensive) Classification: 2.6 Nature of problem: The code deals with single and double

  20. Multi-scale salient feature extraction on mesh models

    Yang, Yongliang; Shen, ChaoHui

    2012-01-01

    We present a new method of extracting multi-scale salient features on meshes. It is based on robust estimation of curvature on multiple scales. The coincidence between salient feature and the scale of interest can be established straightforwardly, where detailed feature appears on small scale and feature with more global shape information shows up on large scale. We demonstrate this multi-scale description of features accords with human perception and can be further used for several applications as feature classification and viewpoint selection. Experiments exhibit that our method as a multi-scale analysis tool is very helpful for studying 3D shapes. © 2012 Springer-Verlag.

  1. Obtuse triangle suppression in anisotropic meshes

    Sun, Feng; Choi, Yi King; Wang, Wen Ping; Yan, Dongming; Liu, Yang; Lé vy, Bruno L.

    2011-01-01

    Anisotropic triangle meshes are used for efficient approximation of surfaces and flow data in finite element analysis, and in these applications it is desirable to have as few obtuse triangles as possible to reduce the discretization error. We present a variational approach to suppressing obtuse triangles in anisotropic meshes. Specifically, we introduce a hexagonal Minkowski metric, which is sensitive to triangle orientation, to give a new formulation of the centroidal Voronoi tessellation (CVT) method. Furthermore, we prove several relevant properties of the CVT method with the newly introduced metric. Experiments show that our algorithm produces anisotropic meshes with much fewer obtuse triangles than using existing methods while maintaining mesh anisotropy. © 2011 Elsevier B.V. All rights reserved.

  2. Connectivity editing for quad-dominant meshes

    Peng, Chihan; Wonka, Peter

    2013-01-01

    and illustrate the advantages and disadvantages of different strategies for quad-dominant mesh design. © 2013 The Author(s) Computer Graphics Forum © 2013 The Eurographics Association and John Wiley & Sons Ltd.

  3. Grid adaptation using chimera composite overlapping meshes

    Kao, Kai-Hsiung; Liou, Meng-Sing; Chow, Chuen-Yen

    1994-01-01

    The objective of this paper is to perform grid adaptation using composite overlapping meshes in regions of large gradient to accurately capture the salient features during computation. The chimera grid scheme, a multiple overset mesh technique, is used in combination with a Navier-Stokes solver. The numerical solution is first converged to a steady state based on an initial coarse mesh. Solution-adaptive enhancement is then performed by using a secondary fine grid system which oversets on top of the base grid in the high-gradient region, but without requiring the mesh boundaries to join in any special way. Communications through boundary interfaces between those separated grids are carried out using trilinear interpolation. Application to the Euler equations for shock reflections and to shock wave/boundary layer interaction problem are tested. With the present method, the salient features are well-resolved.

  4. Grid adaption using Chimera composite overlapping meshes

    Kao, Kai-Hsiung; Liou, Meng-Sing; Chow, Chuen-Yen

    1993-01-01

    The objective of this paper is to perform grid adaptation using composite over-lapping meshes in regions of large gradient to capture the salient features accurately during computation. The Chimera grid scheme, a multiple overset mesh technique, is used in combination with a Navier-Stokes solver. The numerical solution is first converged to a steady state based on an initial coarse mesh. Solution-adaptive enhancement is then performed by using a secondary fine grid system which oversets on top of the base grid in the high-gradient region, but without requiring the mesh boundaries to join in any special way. Communications through boundary interfaces between those separated grids are carried out using tri-linear interpolation. Applications to the Euler equations for shock reflections and to a shock wave/boundary layer interaction problem are tested. With the present method, the salient features are well resolved.

  5. Shape space exploration of constrained meshes

    Yang, Yongliang

    2011-12-12

    We present a general computational framework to locally characterize any shape space of meshes implicitly prescribed by a collection of non-linear constraints. We computationally access such manifolds, typically of high dimension and co-dimension, through first and second order approximants, namely tangent spaces and quadratically parameterized osculant surfaces. Exploration and navigation of desirable subspaces of the shape space with regard to application specific quality measures are enabled using approximants that are intrinsic to the underlying manifold and directly computable in the parameter space of the osculant surface. We demonstrate our framework on shape spaces of planar quad (PQ) meshes, where each mesh face is constrained to be (nearly) planar, and circular meshes, where each face has a circumcircle. We evaluate our framework for navigation and design exploration on a variety of inputs, while keeping context specific properties such as fairness, proximity to a reference surface, etc. © 2011 ACM.

  6. Shape space exploration of constrained meshes

    Yang, Yongliang; Yang, Yijun; Pottmann, Helmut; Mitra, Niloy J.

    2011-01-01

    We present a general computational framework to locally characterize any shape space of meshes implicitly prescribed by a collection of non-linear constraints. We computationally access such manifolds, typically of high dimension and co-dimension, through first and second order approximants, namely tangent spaces and quadratically parameterized osculant surfaces. Exploration and navigation of desirable subspaces of the shape space with regard to application specific quality measures are enabled using approximants that are intrinsic to the underlying manifold and directly computable in the parameter space of the osculant surface. We demonstrate our framework on shape spaces of planar quad (PQ) meshes, where each mesh face is constrained to be (nearly) planar, and circular meshes, where each face has a circumcircle. We evaluate our framework for navigation and design exploration on a variety of inputs, while keeping context specific properties such as fairness, proximity to a reference surface, etc. © 2011 ACM.

  7. Obtuse triangle suppression in anisotropic meshes

    Sun, Feng

    2011-12-01

    Anisotropic triangle meshes are used for efficient approximation of surfaces and flow data in finite element analysis, and in these applications it is desirable to have as few obtuse triangles as possible to reduce the discretization error. We present a variational approach to suppressing obtuse triangles in anisotropic meshes. Specifically, we introduce a hexagonal Minkowski metric, which is sensitive to triangle orientation, to give a new formulation of the centroidal Voronoi tessellation (CVT) method. Furthermore, we prove several relevant properties of the CVT method with the newly introduced metric. Experiments show that our algorithm produces anisotropic meshes with much fewer obtuse triangles than using existing methods while maintaining mesh anisotropy. © 2011 Elsevier B.V. All rights reserved.

  8. Mesh Processing in Medical Image Analysis

    The following topics are dealt with: mesh processing; medical image analysis; interactive freeform modeling; statistical shape analysis; clinical CT images; statistical surface recovery; automated segmentation; cerebral aneurysms; and real-time particle-based representation....

  9. Energy-efficient wireless mesh networks

    Ntlatlapa, N

    2007-06-01

    Full Text Available This paper outlines the objectives of a recently formed research group at Meraka Institute. The authors consider application of wireless mesh networks in rural infrastructure deficient parts of the African continent where nodes operate on batteries...

  10. LR: Compact connectivity representation for triangle meshes

    Gurung, T; Luffel, M; Lindstrom, P; Rossignac, J

    2011-01-28

    We propose LR (Laced Ring) - a simple data structure for representing the connectivity of manifold triangle meshes. LR provides the option to store on average either 1.08 references per triangle or 26.2 bits per triangle. Its construction, from an input mesh that supports constant-time adjacency queries, has linear space and time complexity, and involves ordering most vertices along a nearly-Hamiltonian cycle. LR is best suited for applications that process meshes with fixed connectivity, as any changes to the connectivity require the data structure to be rebuilt. We provide an implementation of the set of standard random-access, constant-time operators for traversing a mesh, and show that LR often saves both space and traversal time over competing representations.

  11. Seeking new surgical predictors of mesh exposure after transvaginal mesh repair.

    Wu, Pei-Ying; Chang, Chih-Hung; Shen, Meng-Ru; Chou, Cheng-Yang; Yang, Yi-Ching; Huang, Yu-Fang

    2016-10-01

    The purpose of this study was to explore new preventable risk factors for mesh exposure. A retrospective review of 92 consecutive patients treated with transvaginal mesh (TVM) in the urogynecological unit of our university hospital. An analysis of perioperative predictors was conducted in patients after vaginal repairs using a type 1 mesh. Mesh complications were recorded according to International Urogynecological Association (IUGA) definitions. Mesh-exposure-free durations were calculated by using the Kaplan-Meier method and compared between different closure techniques using log-rank test. Hazard ratios (HR) of predictors for mesh exposure were estimated by univariate and multivariate analyses using Cox proportional hazards regression models. The median surveillance interval was 24.1 months. Two late occurrences were found beyond 1 year post operation. No statistically significant correlation was observed between mesh exposure and concomitant hysterectomy. Exposure risks were significantly higher in patients with interrupted whole-layer closure in univariate analysis. In the multivariate analysis, hematoma [HR 5.42, 95 % confidence interval (CI) 1.26-23.35, P = 0.024), Prolift mesh (HR 5.52, 95 % CI 1.15-26.53, P = 0.033), and interrupted whole-layer closure (HR 7.02, 95 % CI 1.62-30.53, P = 0.009) were the strongest predictors of mesh exposure. Findings indicate the risks of mesh exposure and reoperation may be prevented by avoiding hematoma, large amount of mesh, or interrupted whole-layer closure in TVM surgeries. If these risk factors are prevented, hysterectomy may not be a relative contraindication for TVM use. We also provide evidence regarding mesh exposure and the necessity for more than 1 year of follow-up and preoperative counselling.

  12. Evaluation of mesh morphing and mapping techniques in patient specific modeling of the human pelvis.

    Salo, Zoryana; Beek, Maarten; Whyne, Cari Marisa

    2013-01-01

    Robust generation of pelvic finite element models is necessary to understand the variation in mechanical behaviour resulting from differences in gender, aging, disease and injury. The objective of this study was to apply and evaluate mesh morphing and mapping techniques to facilitate the creation and structural analysis of specimen-specific finite element (FE) models of the pelvis. A specimen-specific pelvic FE model (source mesh) was generated following a traditional user-intensive meshing scheme. The source mesh was morphed onto a computed tomography scan generated target surface of a second pelvis using a landmarked-based approach, in which exterior source nodes were shifted to target surface vertices, while constrained along a normal. A second copy of the morphed model was further refined through mesh mapping, in which surface nodes of the initial morphed model were selected in patches and remapped onto the surfaces of the target model. Computed tomography intensity based material properties were assigned to each model. The source, target, morphed and mapped models were analyzed under axial compression using linear static FE analysis and their strain distributions evaluated. Morphing and mapping techniques were effectively applied to generate good quality geometrically complex specimen-specific pelvic FE models. Mapping significantly improved strain concurrence with the target pelvis FE model. Copyright © 2012 John Wiley & Sons, Ltd.

  13. Evaluation of mesh morphing and mapping techniques in patient specific modelling of the human pelvis.

    Salo, Zoryana; Beek, Maarten; Whyne, Cari Marisa

    2012-08-01

    Robust generation of pelvic finite element models is necessary to understand variation in mechanical behaviour resulting from differences in gender, aging, disease and injury. The objective of this study was to apply and evaluate mesh morphing and mapping techniques to facilitate the creation and structural analysis of specimen-specific finite element (FE) models of the pelvis. A specimen-specific pelvic FE model (source mesh) was generated following a traditional user-intensive meshing scheme. The source mesh was morphed onto a computed tomography scan generated target surface of a second pelvis using a landmarked-based approach, in which exterior source nodes were shifted to target surface vertices, while constrained along a normal. A second copy of the morphed model was further refined through mesh mapping, in which surface nodes of the initial morphed model were selected in patches and remapped onto the surfaces of the target model. Computed tomography intensity-based material properties were assigned to each model. The source, target, morphed and mapped models were analyzed under axial compression using linear static FE analysis, and their strain distributions were evaluated. Morphing and mapping techniques were effectively applied to generate good quality and geometrically complex specimen-specific pelvic FE models. Mapping significantly improved strain concurrence with the target pelvis FE model. Copyright © 2012 John Wiley & Sons, Ltd.

  14. Evaluation on Bending Properties of Biomaterial GUM Metal Meshed Plates for Bone Graft Applications

    Suzuki, Hiromichi; He, Jianmei

    2017-11-01

    There are three bone graft methods for bone defects caused by diseases such as cancer and accident injuries: Autogenous bone grafts, Allografts and Artificial bone grafts. In this study, meshed GUM Metal plates with lower elasticity, high strength and high biocompatibility are introduced to solve the over stiffness & weight problems of ready-used metal implants. Basic mesh shapes are designed and applied to GUM Metal plates using 3D CAD modeling tools. Bending properties of prototype meshed GUM Metal plates are evaluated experimentally and analytically. Meshed plate specimens with 180°, 120° and 60° axis-symmetrical types were fabricated for 3-point bending tests. The pseudo bending elastic moduli of meshed plate specimens obtained from 3-point bending test are ranged from 4.22 GPa to 16.07 GPa, within the elasticity range of natural cortical bones from 2.0 GPa to 30.0 GPa. Analytical approach method is validated by comparison with experimental and analytical results for evaluation on bending property of meshed plates.

  15. MHD simulations on an unstructured mesh

    Strauss, H.R.; Park, W.; Belova, E.; Fu, G.Y.; Sugiyama, L.E.

    1998-01-01

    Two reasons for using an unstructured computational mesh are adaptivity, and alignment with arbitrarily shaped boundaries. Two codes which use finite element discretization on an unstructured mesh are described. FEM3D solves 2D and 3D RMHD using an adaptive grid. MH3D++, which incorporates methods of FEM3D into the MH3D generalized MHD code, can be used with shaped boundaries, which might be 3D

  16. Towards Blockchain-enabled Wireless Mesh Networks

    Selimi, Mennan; Kabbinale, Aniruddh Rao; Ali, Anwaar; Navarro, Leandro; Sathiaseelan, Arjuna

    2018-01-01

    Recently, mesh networking and blockchain are two of the hottest technologies in the telecommunications industry. Combining both can reformulate internet access and make connecting to the Internet not only easy, but affordable too. Hyperledger Fabric (HLF) is a blockchain framework implementation and one of the Hyperledger projects hosted by The Linux Foundation. We evaluate HLF in a real production mesh network and in the laboratory, quantify its performance, bottlenecks and limitations of th...

  17. A novel partitioning method for block-structured adaptive meshes

    Fu, Lin; Litvinov, Sergej; Hu, Xiangyu Y.; Adams, Nikolaus A.

    2017-07-01

    We propose a novel partitioning method for block-structured adaptive meshes utilizing the meshless Lagrangian particle concept. With the observation that an optimum partitioning has high analogy to the relaxation of a multi-phase fluid to steady state, physically motivated model equations are developed to characterize the background mesh topology and are solved by multi-phase smoothed-particle hydrodynamics. In contrast to well established partitioning approaches, all optimization objectives are implicitly incorporated and achieved during the particle relaxation to stationary state. Distinct partitioning sub-domains are represented by colored particles and separated by a sharp interface with a surface tension model. In order to obtain the particle relaxation, special viscous and skin friction models, coupled with a tailored time integration algorithm are proposed. Numerical experiments show that the present method has several important properties: generation of approximately equal-sized partitions without dependence on the mesh-element type, optimized interface communication between distinct partitioning sub-domains, continuous domain decomposition which is physically localized and implicitly incremental. Therefore it is particularly suitable for load-balancing of high-performance CFD simulations.

  18. Gear Mesh Loss-of-Lubrication Experiments and Analytical Simulation

    Handschuh, Robert F.; Polly, Joseph; Morales, Wilfredo

    2011-01-01

    An experimental program to determine the loss-of-lubrication (LOL) characteristics of spur gears in an aerospace simulation test facility has been completed. Tests were conducted using two different emergency lubricant types: (1) an oil mist system (two different misted lubricants) and (2) a grease injection system (two different grease types). Tests were conducted using a NASA Glenn test facility normally used for conducting contact fatigue. Tests were run at rotational speeds up to 10000 rpm using two different gear designs and two different gear materials. For the tests conducted using an air-oil misting system, a minimum lubricant injection rate was determined to permit the gear mesh to operate without failure for at least 1 hr. The tests allowed an elevated steady state temperature to be established. A basic 2-D heat transfer simulation has been developed to investigate temperatures of a simulated gear as a function of frictional behavior. The friction (heat generation source) between the meshing surfaces is related to the position in the meshing cycle, the load applied, and the amount of lubricant in the contact. Experimental conditions will be compared to those from the 2-D simulation.

  19. A novel partitioning method for block-structured adaptive meshes

    Fu, Lin, E-mail: lin.fu@tum.de; Litvinov, Sergej, E-mail: sergej.litvinov@aer.mw.tum.de; Hu, Xiangyu Y., E-mail: xiangyu.hu@tum.de; Adams, Nikolaus A., E-mail: nikolaus.adams@tum.de

    2017-07-15

    We propose a novel partitioning method for block-structured adaptive meshes utilizing the meshless Lagrangian particle concept. With the observation that an optimum partitioning has high analogy to the relaxation of a multi-phase fluid to steady state, physically motivated model equations are developed to characterize the background mesh topology and are solved by multi-phase smoothed-particle hydrodynamics. In contrast to well established partitioning approaches, all optimization objectives are implicitly incorporated and achieved during the particle relaxation to stationary state. Distinct partitioning sub-domains are represented by colored particles and separated by a sharp interface with a surface tension model. In order to obtain the particle relaxation, special viscous and skin friction models, coupled with a tailored time integration algorithm are proposed. Numerical experiments show that the present method has several important properties: generation of approximately equal-sized partitions without dependence on the mesh-element type, optimized interface communication between distinct partitioning sub-domains, continuous domain decomposition which is physically localized and implicitly incremental. Therefore it is particularly suitable for load-balancing of high-performance CFD simulations.

  20. Simulating control rod and fuel assembly motion using moving meshes

    Gilbert, D. [Department of Electrical and Computer Engineering, McMaster University, 1280 Main Street West, Hamilton Ontario, L8S 4K1 (Canada)], E-mail: gilbertdw1@gmail.com; Roman, J.E. [Departamento de Sistemas Informaticos y Computacion, Universidad Politecnica de Valencia, Camino de Vera s/n, 46022 Valencia (Spain); Garland, Wm. J. [Department of Engineering Physics, McMaster University, 1280 Main Street West, Hamilton Ontario, L8S 4K1 (Canada); Poehlman, W.F.S. [Department of Computing and Software, McMaster University, 1280 Main Street West, Hamilton Ontario, L8S 4K1 (Canada)

    2008-02-15

    A prerequisite for designing a transient simulation experiment which includes the motion of control and fuel assemblies is the careful verification of a steady state model which computes k{sub eff} versus assembly insertion distance. Previous studies in nuclear engineering have usually approached the problem of the motion of control rods with the use of nonlinear nodal models. Nodal methods employ special approximations for the leading and trailing cells of the moving assemblies to avoid the rod cusping problem which results from the naive volume weighted cell cross-section approximation. A prototype framework called the MOOSE has been developed for modeling moving components in the presence of diffusion phenomena. A linear finite difference model is constructed, solutions for which are computed by SLEPc, a high performance parallel eigenvalue solver. Design techniques for the implementation of a patched non-conformal mesh which links groups of sub-meshes that can move relative to one another are presented. The generation of matrices which represent moving meshes which conserve neutron current at their boundaries, and the performance of the framework when applied to model reactivity insertion experiments is also discussed.

  1. MHD simulations on an unstructured mesh

    Strauss, H.R.; Park, W.

    1996-01-01

    We describe work on a full MHD code using an unstructured mesh. MH3D++ is an extension of the PPPL MH3D resistive full MHD code. MH3D++ replaces the structured mesh and finite difference / fourier discretization of MH3D with an unstructured mesh and finite element / fourier discretization. Low level routines which perform differential operations, solution of PDEs such as Poisson's equation, and graphics, are encapsulated in C++ objects to isolate the finite element operations from the higher level code. The high level code is the same, whether it is run in structured or unstructured mesh versions. This allows the unstructured mesh version to be benchmarked against the structured mesh version. As a preliminary example, disruptions in DIIID reverse shear equilibria are studied numerically with the MH3D++ code. Numerical equilibria were first produced starting with an EQDSK file containing equilibrium data of a DIII-D L-mode negative central shear discharge. Using these equilibria, the linearized equations are time advanced to get the toroidal mode number n = 1 linear growth rate and eigenmode, which is resistively unstable. The equilibrium and linear mode are used to initialize 3D nonlinear runs. An example shows poloidal slices of 3D pressure surfaces: initially, on the left, and at an intermediate time, on the right

  2. [Implants for genital prolapse : Contra mesh surgery].

    Hampel, C

    2017-12-01

    Alloplastic transvaginal meshes have become very popular in the surgery of pelvic organ prolapse (POP) as did alloplastic suburethral slings in female stress incontinence surgery, but without adequate supporting data. The simplicity of the mesh procedure facilitates its propagation with acceptance of higher revision and complication rates. Since attending physicians do more and more prolapse surgeries without practicing or teaching alternative techniques, expertise in these alternatives, which might be very useful in cases of recurrence, persistence or complications, is permanently lost. It is doubtful that proper and detailed information about alternatives, risks, and benefits of transvaginal alloplastic meshes is provided to every single prolapse patient according to the recommendations of the German POP guidelines, since the number of implanted meshes exceeds the number of properly indicated mesh candidates by far. Although there is no dissent internationally about the available mesh data, thousands of lawsuits in the USA, insolvency of companies due to claims for compensation and unambiguous warnings from foreign urological societies leave German urogynecologists still unimpressed. The existing literature in pelvic organ prolapse exclusively focusses on POP stage and improvement of that stage with surgical therapy. Instead, typical prolapse symptoms should trigger therapy and improvement of these symptoms should be the utmost treatment goal. It is strongly recommended for liability reasons to obtain specific written informed consent.

  3. Maintenance Tools applied to Electric Generators to Improve Energy Efficiency and Power Quality of Thermoelectric Power Plants

    Milton Fonseca Junior

    2017-07-01

    Full Text Available This paper presents a specific method to improve the reliability of the equipment and the quality of power supplied to the electrical systems with the frequency and voltage control of a thermoelectric plant, to guarantee a more stable system. The method has the novelty of combining Total Productive Maintenance (TPM using only four pillars, with Electrical Predictive Maintenance based in failure analysis and diagnostic. It prevents voltage drops caused by excessive reactive consumption, thus guaranteeing the company a perfect functioning of its equipment and providing a longer life of them. The Maintenance Management Program (MMP seeks to prevent failures from causing the equipment to be shut down from the electrical system, which means large financial losses, either by reducing billing or by paying fines to the regulatory agency, in addition to prejudice the reliability of the system. Using management tools, but applying only four TPM pillars, it was possible to achieve innovation in power plants with internal combustion engines. This study aims to provide maintenance with a more reliable process, through the implantation of measurement, control and diagnostic devices, thus allowing the management to reduce breakdown of plant equipment. Some results have been achieved after the implementation, such as reduction of annual maintenance cost, reduction of corrective maintenance, increase of MTBF (Mean Time between Failures and reduction of MTTR (Mean Time to Repair in all areas. Probabilistic models able to describe real processes in a more realistic way, and facilitate the optimization at maximum reliability or minimum costs are presented. Such results are reflected in more reliable and continual power generation.

  4. A Deep Penetration Problem Calculation Using AETIUS:An Easy Modeling Discrete Ordinates Transport Code UsIng Unstructured Tetrahedral Mesh, Shared Memory Parallel

    KIM, Jong Woon; LEE, Young-Ouk

    2017-09-01

    As computing power gets better and better, computer codes that use a deterministic method seem to be less useful than those using the Monte Carlo method. In addition, users do not like to think about space, angles, and energy discretization for deterministic codes. However, a deterministic method is still powerful in that we can obtain a solution of the flux throughout the problem, particularly as when particles can barely penetrate, such as in a deep penetration problem with small detection volumes. Recently, a new state-of-the-art discrete-ordinates code, ATTILA, was developed and has been widely used in several applications. ATTILA provides the capabilities to solve geometrically complex 3-D transport problems by using an unstructured tetrahedral mesh. Since 2009, we have been developing our own code by benchmarking ATTILA. AETIUS is a discrete ordinates code that uses an unstructured tetrahedral mesh such as ATTILA. For pre- and post- processing, Gmsh is used to generate an unstructured tetrahedral mesh by importing a CAD file (*.step) and visualizing the calculation results of AETIUS. Using a CAD tool, the geometry can be modeled very easily. In this paper, we describe a brief overview of AETIUS and provide numerical results from both AETIUS and a Monte Carlo code, MCNP5, in a deep penetration problem with small detection volumes. The results demonstrate the effectiveness and efficiency of AETIUS for such calculations.

  5. Parametric Quadrilateral Meshes for the Design and Optimization of Superconducting Magnets

    Aleksa, Martin; Völlinger, Christine

    2002-01-01

    The program package ROXIE has been developed at CERN for the design and optimization of accelerator magnets. The necessity of extremely uniform fields in the superconducting accelerator magnets for LHC requires very accurate methods of field computation. For this purpose the coupled boundary-element / finite-element technique (BEM-FEM) is used. Quadrilateral higher order finite-element meshes are generated for the discretization of the iron domain (yoke) and stainless steel collars. A new mesh generator using geometrically optimized domain decomposition which was developed at the University of Stuttgart, Germany has been implemented into the ROXIE program providing fully automatic and user friendly mesh generation. The structure of the magnet cross-section can be modeled using parametric objects such as holes of different forms, elliptic, parabolic or hyperbolic arcs, notches, slots, .... For sensitivity analysis and parametric studies, point based morphing algorithms are applied to guarantee smooth adaptatio...

  6. Fire performance of basalt FRP mesh reinforced HPC thin plates

    Hulin, Thomas; Hodicky, Kamil; Schmidt, Jacob Wittrup

    2013-01-01

    An experimental program was carried out to investigate the influence of basalt FRP (BFRP) reinforcing mesh on the fire behaviour of thin high performance concrete (HPC) plates applied to sandwich elements. Samples with BFRP mesh were compared to samples with no mesh, samples with steel mesh...

  7. A Graphic Symbol Tool for the Evaluation of Communication, Satisfaction and Priorities of Individuals with Intellectual Disability Who Use a Speech Generating Device

    Valiquette, Christine; Sutton, Ann; Ska, Bernadette

    2010-01-01

    This article reports on the views of individuals with learning disability (LD) on their use of their speech generating devices (SGDs), their satisfaction about their communication, and their priorities. The development of an interview tool made of graphic symbols and entitled Communication, Satisfaction and Priorities of SGD Users (CSPU) is…

  8. Unstructured Mesh Movement and Viscous Mesh Generation for CFD-Based Design Optimization, Phase I

    National Aeronautics and Space Administration — The innovations proposed by ResearchSouth are: 1) a robust method to automatically insert high quality anisotropic prismatic (viscous boundary layer) cells into any...

  9. Prolapse Recurrence after Transvaginal Mesh Removal.

    Rawlings, Tanner; Lavelle, Rebecca S; Coskun, Burhan; Alhalabi, Feras; Zimmern, Philippe E

    2015-11-01

    We determined the rate of pelvic organ prolapse recurrence after transvaginal mesh removal. Following institutional review board approval a longitudinally collected database of women undergoing transvaginal mesh removal for complications after transvaginal mesh placement with at least 1 year minimum followup was queried for pelvic organ prolapse recurrence. Recurrent prolapse was defined as greater than stage 1 on examination or the need for reoperation at the site of transvaginal mesh removal. Outcome measures were based on POP-Q (Pelvic Organ Prolapse Quantification System) at the last visit. Patients were grouped into 3 groups, including group 1--recurrent prolapse in the same compartment as transvaginal mesh removal, 2--persistent prolapse and 3--prolapse in a compartment different than transvaginal mesh removal. Of 73 women 52 met study inclusion criteria from 2007 to 2013, including 73% who presented with multiple indications for transvaginal mesh removal. The mean interval between insertion and removal was 45 months (range 10 to 165). Overall mean followup after transvaginal mesh removal was 30 months (range 12 to 84). In group 1 (recurrent prolapse) the rate was 15% (6 of 40 patients). Four women underwent surgery for recurrent prolapse at a mean 7 of months (range 5 to 10). Two patients elected observation. The rate of persistent prolapse (group 2) was 23% (12 of 52 patients). Three women underwent prolapse reoperation at a mean of 10 months (range 8 to 12). In group 3 (de novo/different compartment prolapse) the rate was 6% (3 of 52 patients). One woman underwent surgical repair at 52 months. At a mean 2.5-year followup 62% of patients (32 of 52) did not have recurrent or persistent prolapse after transvaginal mesh removal and 85% (44 of 52) did not undergo any further procedure for prolapse. Specifically for pelvic organ prolapse in the same compartment as transvaginal mesh removal 12% of patients had recurrence, of whom 8% underwent prolapse repair

  10. SU-F-BRB-16: A Spreadsheet Based Automatic Trajectory GEnerator (SAGE): An Open Source Tool for Automatic Creation of TrueBeam Developer Mode Robotic Trajectories

    Etmektzoglou, A; Mishra, P; Svatos, M [Varian Medical Systems, Palo Alto, CA (United States)

    2015-06-15

    Purpose: To automate creation and delivery of robotic linac trajectories with TrueBeam Developer Mode, an open source spreadsheet-based trajectory generation tool has been developed, tested and made freely available. The computing power inherent in a spreadsheet environment plus additional functions programmed into the tool insulate users from the underlying schema tedium and allow easy calculation, parameterization, graphical visualization, validation and finally automatic generation of Developer Mode XML scripts which are directly loadable on a TrueBeam linac. Methods: The robotic control system platform that allows total coordination of potentially all linac moving axes with beam (continuous, step-and-shoot, or combination thereof) becomes available in TrueBeam Developer Mode. Many complex trajectories are either geometric or can be described in analytical form, making the computational power, graphing and programmability available in a spreadsheet environment an easy and ideal vehicle for automatic trajectory generation. The spreadsheet environment allows also for parameterization of trajectories thus enabling the creation of entire families of trajectories using only a few variables. Standard spreadsheet functionality has been extended for powerful movie-like dynamic graphic visualization of the gantry, table, MLC, room, lasers, 3D observer placement and beam centerline all as a function of MU or time, for analysis of the motions before requiring actual linac time. Results: We used the tool to generate and deliver extended SAD “virtual isocenter” trajectories of various shapes such as parameterized circles and ellipses. We also demonstrated use of the tool in generating linac couch motions that simulate respiratory motion using analytical parameterized functions. Conclusion: The SAGE tool is a valuable resource to experiment with families of complex geometric trajectories for a TrueBeam Linac. It makes Developer Mode more accessible as a vehicle to quickly

  11. SU-F-BRB-16: A Spreadsheet Based Automatic Trajectory GEnerator (SAGE): An Open Source Tool for Automatic Creation of TrueBeam Developer Mode Robotic Trajectories

    Etmektzoglou, A; Mishra, P; Svatos, M

    2015-01-01

    Purpose: To automate creation and delivery of robotic linac trajectories with TrueBeam Developer Mode, an open source spreadsheet-based trajectory generation tool has been developed, tested and made freely available. The computing power inherent in a spreadsheet environment plus additional functions programmed into the tool insulate users from the underlying schema tedium and allow easy calculation, parameterization, graphical visualization, validation and finally automatic generation of Developer Mode XML scripts which are directly loadable on a TrueBeam linac. Methods: The robotic control system platform that allows total coordination of potentially all linac moving axes with beam (continuous, step-and-shoot, or combination thereof) becomes available in TrueBeam Developer Mode. Many complex trajectories are either geometric or can be described in analytical form, making the computational power, graphing and programmability available in a spreadsheet environment an easy and ideal vehicle for automatic trajectory generation. The spreadsheet environment allows also for parameterization of trajectories thus enabling the creation of entire families of trajectories using only a few variables. Standard spreadsheet functionality has been extended for powerful movie-like dynamic graphic visualization of the gantry, table, MLC, room, lasers, 3D observer placement and beam centerline all as a function of MU or time, for analysis of the motions before requiring actual linac time. Results: We used the tool to generate and deliver extended SAD “virtual isocenter” trajectories of various shapes such as parameterized circles and ellipses. We also demonstrated use of the tool in generating linac couch motions that simulate respiratory motion using analytical parameterized functions. Conclusion: The SAGE tool is a valuable resource to experiment with families of complex geometric trajectories for a TrueBeam Linac. It makes Developer Mode more accessible as a vehicle to quickly

  12. A constrained Delaunay discretization method for adaptively meshing highly discontinuous geological media

    Wang, Yang; Ma, Guowei; Ren, Feng; Li, Tuo

    2017-12-01

    A constrained Delaunay discretization method is developed to generate high-quality doubly adaptive meshes of highly discontinuous geological media. Complex features such as three-dimensional discrete fracture networks (DFNs), tunnels, shafts, slopes, boreholes, water curtains, and drainage systems are taken into account in the mesh generation. The constrained Delaunay triangulation method is used to create adaptive triangular elements on planar fractures. Persson's algorithm (Persson, 2005), based on an analogy between triangular elements and spring networks, is enriched to automatically discretize a planar fracture into mesh points with varying density and smooth-quality gradient. The triangulated planar fractures are treated as planar straight-line graphs (PSLGs) to construct piecewise-linear complex (PLC) for constrained Delaunay tetrahedralization. This guarantees the doubly adaptive characteristic of the resulted mesh: the mesh is adaptive not only along fractures but also in space. The quality of elements is compared with the results from an existing method. It is verified that the present method can generate smoother elements and a better distribution of element aspect ratios. Two numerical simulations are implemented to demonstrate that the present method can be applied to various simulations of complex geological media that contain a large number of discontinuities.

  13. Positive Contrast MRI Techniques for Visualization of Iron-Loaded Hernia Mesh Implants in Patients.

    Alexander Ciritsis

    Full Text Available In MRI, implants and devices can be delineated via susceptibility artefacts. To discriminate susceptibility voids from proton-free structures, different positive contrast techniques were implemented. The purpose of this study was to evaluate a pulse sequence-based positive contrast technique (PCSI and a post-processing susceptibility gradient mapping algorithm (SGM for visualization of iron loaded mesh implants in patients.Five patients with iron-loaded MR-visible inguinal hernia mesh implants were examined at 1.5 Tesla. A gradient echo sequence (GRE; parameters: TR: 8.3ms; TE: 4.3ms; NSA:2; FA:20°; FOV:350mm² and a PCSI sequence (parameters: TR: 25ms; TE: 4.6ms; NSA:4; FA:20°; FOV:350mm² with on-resonant proton suppression were performed. SGM maps were calculated using two algorithms. Image quality and mesh delineation were independently evaluated by three radiologists.On GRE, the iron-loaded meshes generated distinct susceptibility-induced signal voids. PCSI exhibited susceptibility differences including the meshes as hyperintense signals. SGM exhibited susceptibility differences with positive contrast. Visually, the different algorithms presented no significant differences. Overall, the diagnostic value was rated best in GRE whereas PCSI and SGM were barely "sufficient".Both "positive contrast" techniques depicted implanted meshes with hyperintense signal. SGM comes without additional acquisition time and can therefore be utilized in every patient.

  14. A 3D coarse-mesh time dependent code for nuclear reactor kinetic calculations

    Montagnini, B.; Raffaelli, P.; Sumini, M.; Zardini, D.M.

    1996-01-01

    A course-mesh code for time-dependent multigroup neutron diffusion calculation based on a direct integration scheme for the time dependence and a low order nodal flux expansion approximation for the space variables has been implemented as a fast tool for transient analysis. (Author)

  15. Laparoscopic mesh explantation and drainage of sacral abscess remote from transvaginal excision of exposed sacral colpopexy mesh.

    Roth, Ted M; Reight, Ian

    2012-07-01

    Sacral colpopexy may be complicated by mesh exposure, and the surgical treatment of mesh exposure typically results in minor postoperative morbidity and few delayed complications. A 75-year-old woman presented 7 years after a laparoscopic sacral colpopexy, with Mersilene mesh, with an apical mesh exposure. She underwent an uncomplicated transvaginal excision and was asymptomatic until 8 months later when she presented with vaginal drainage and a sacral abscess. This was successfully treated with laparoscopic enterolysis, drainage of the abscess, and explantation of the remaining mesh. Incomplete excision of exposed colpopexy mesh can lead to ascending infection and sacral abscess. Laparoscopic drainage and mesh removal may be considered in these patients.

  16. A study on the dependency between turbulent models and mesh configurations of CFD codes

    Bang, Jungjin; Heo, Yujin; Jerng, Dong-Wook

    2015-01-01

    This paper focuses on the analysis of the behavior of hydrogen mixing and hydrogen stratification, using the GOTHIC code and the CFD code. Specifically, we examined the mesh sensitivity and how the turbulence model affects hydrogen stratification or hydrogen mixing, depending on the mesh configuration. In this work, sensitivity analyses for the meshes and the turbulence models were conducted for missing and stratification phenomena. During severe accidents in a nuclear power plants, the generation of hydrogen may occur and this will complicate the atmospheric condition of the containment by causing stratification of air, steam, and hydrogen. This could significantly impact containment integrity analyses, as hydrogen could be accumulated in local region. From this need arises the importance of research about stratification of gases in the containment. Two computation fluid dynamics code, i.e. GOTHIC and STAR-CCM+ were adopted and the computational results were benchmarked against the experimental data from PANDA facility. The main findings observed through the present work can be summarized as follows: 1) In the case of the GOTHIC code, it was observed that the aspect ratio of the mesh was found more important than the mesh size. Also, if the number of the mesh is over 3,000, the effects of the turbulence models were marginal. 2) For STAR-CCM+, the tendency is quite different from the GOTHIC code. That is, the effects of the turbulence models were small for fewer number of the mesh, however, as the number of mesh increases, the effects of the turbulence models becomes significant. Another observation is that away from the injection orifice, the role of the turbulence models tended to be important due to the nature of mixing process and inducted jet stream

  17. A study on the dependency between turbulent models and mesh configurations of CFD codes

    Bang, Jungjin; Heo, Yujin; Jerng, Dong-Wook [CAU, Seoul (Korea, Republic of)

    2015-10-15

    This paper focuses on the analysis of the behavior of hydrogen mixing and hydrogen stratification, using the GOTHIC code and the CFD code. Specifically, we examined the mesh sensitivity and how the turbulence model affects hydrogen stratification or hydrogen mixing, depending on the mesh configuration. In this work, sensitivity analyses for the meshes and the turbulence models were conducted for missing and stratification phenomena. During severe accidents in a nuclear power plants, the generation of hydrogen may occur and this will complicate the atmospheric condition of the containment by causing stratification of air, steam, and hydrogen. This could significantly impact containment integrity analyses, as hydrogen could be accumulated in local region. From this need arises the importance of research about stratification of gases in the containment. Two computation fluid dynamics code, i.e. GOTHIC and STAR-CCM+ were adopted and the computational results were benchmarked against the experimental data from PANDA facility. The main findings observed through the present work can be summarized as follows: 1) In the case of the GOTHIC code, it was observed that the aspect ratio of the mesh was found more important than the mesh size. Also, if the number of the mesh is over 3,000, the effects of the turbulence models were marginal. 2) For STAR-CCM+, the tendency is quite different from the GOTHIC code. That is, the effects of the turbulence models were small for fewer number of the mesh, however, as the number of mesh increases, the effects of the turbulence models becomes significant. Another observation is that away from the injection orifice, the role of the turbulence models tended to be important due to the nature of mixing process and inducted jet stream.

  18. ABC Triblock Copolymer Vesicles with Mesh-like Morphology

    Zhao, Wei; Russell, Thomas; Grason, Gregory

    2010-03-01

    Polymer vesicles can be made from poly(isoprene-b-styrene-b-2-vinylpyridene) (PI-b-PS-b-P2VP) triblock copolymer under the confinement of anodic aluminum oxide (AAO) membrane. It was found that these vesicles have well-defined, nanoscopic size and a microphase-separated hydrophobic core, comprised of PS and PI blocks. Vesicle formation was tracked using both transmission and scanning electron microscopy. A mesh-like morphology formed in the core at a well-defined composition of three blocks. Confinement played an important role in generating these vesicles with such an unusual morphology.

  19. Advanced Mesh-Enabled Monte carlo capability for Multi-Physics Reactor Analysis

    Wilson, Paul; Evans, Thomas; Tautges, Tim

    2012-12-24

    This project will accumulate high-precision fluxes throughout reactor geometry on a non- orthogonal grid of cells to support multi-physics coupling, in order to more accurately calculate parameters such as reactivity coefficients and to generate multi-group cross sections. This work will be based upon recent developments to incorporate advanced geometry and mesh capability in a modular Monte Carlo toolkit with computational science technology that is in use in related reactor simulation software development. Coupling this capability with production-scale Monte Carlo radiation transport codes can provide advanced and extensible test-beds for these developments. Continuous energy Monte Carlo methods are generally considered to be the most accurate computational tool for simulating radiation transport in complex geometries, particularly neutron transport in reactors. Nevertheless, there are several limitations for their use in reactor analysis. Most significantly, there is a trade-off between the fidelity of results in phase space, statistical accuracy, and the amount of computer time required for simulation. Consequently, to achieve an acceptable level of statistical convergence in high-fidelity results required for modern coupled multi-physics analysis, the required computer time makes Monte Carlo methods prohibitive for design iterations and detailed whole-core analysis. More subtly, the statistical uncertainty is typically not uniform throughout the domain, and the simulation quality is limited by the regions with the largest statistical uncertainty. In addition, the formulation of neutron scattering laws in continuous energy Monte Carlo methods makes it difficult to calculate adjoint neutron fluxes required to properly determine important reactivity parameters. Finally, most Monte Carlo codes available for reactor analysis have relied on orthogonal hexahedral grids for tallies that do not conform to the geometric boundaries and are thus generally not well

  20. Early experience with mesh excision for adverse outcomes after transvaginal mesh placement using prolapse kits.

    Ridgeway, Beri; Walters, Mark D; Paraiso, Marie Fidela R; Barber, Matthew D; McAchran, Sarah E; Goldman, Howard B; Jelovsek, J Eric

    2008-12-01

    The purpose of this study was to determine the complications, treatments, and outcomes in patients choosing to undergo removal of mesh previously placed with a mesh procedural kit. This was a retrospective review of all patients who underwent surgical removal of transvaginal mesh for mesh-related complications during a 3-year period at Cleveland Clinic. At last follow-up, patients reported degree of pain, level of improvement, sexual activity, and continued symptoms. Nineteen patients underwent removal of mesh during the study period. Indications for removal included chronic pain (6/19), dyspareunia (6/19), recurrent pelvic organ prolapse (8/19), mesh erosion (12/19), and vesicovaginal fistula (3/19), with most patients (16/19) citing more than 1 reason. There were few complications related to the mesh removal. Most patients reported significant relief of symptoms. Mesh removal can be technically difficult but appears to be safe with few complications and high relief of symptoms, although some symptoms can persist.