WorldWideScience

Sample records for general purpose code

  1. General purpose code for Monte Carlo simulations

    International Nuclear Information System (INIS)

    Wilcke, W.W.

    1983-01-01

    A general-purpose computer called MONTHY has been written to perform Monte Carlo simulations of physical systems. To achieve a high degree of flexibility the code is organized like a general purpose computer, operating on a vector describing the time dependent state of the system under simulation. The instruction set of the computer is defined by the user and is therefore adaptable to the particular problem studied. The organization of MONTHY allows iterative and conditional execution of operations

  2. SRAC95; general purpose neutronics code system

    International Nuclear Information System (INIS)

    Okumura, Keisuke; Tsuchihashi, Keichiro; Kaneko, Kunio.

    1996-03-01

    SRAC is a general purpose neutronics code system applicable to core analyses of various types of reactors. Since the publication of JAERI-1302 for the revised SRAC in 1986, a number of additions and modifications have been made for nuclear data libraries and programs. Thus, the new version SRAC95 has been completed. The system consists of six kinds of nuclear data libraries(ENDF/B-IV, -V, -VI, JENDL-2, -3.1, -3.2), five modular codes integrated into SRAC95; collision probability calculation module (PIJ) for 16 types of lattice geometries, Sn transport calculation modules(ANISN, TWOTRAN), diffusion calculation modules(TUD, CITATION) and two optional codes for fuel assembly and core burn-up calculations(newly developed ASMBURN, revised COREBN). In this version, many new functions and data are implemented to support nuclear design studies of advanced reactors, especially for burn-up calculations. SRAC95 is available not only on conventional IBM-compatible computers but also on scalar or vector computers with the UNIX operating system. This report is the SRAC95 users manual which contains general description, contents of revisions, input data requirements, detail information on usage, sample input data and list of available libraries. (author)

  3. SRAC95; general purpose neutronics code system

    Energy Technology Data Exchange (ETDEWEB)

    Okumura, Keisuke; Tsuchihashi, Keichiro [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment; Kaneko, Kunio

    1996-03-01

    SRAC is a general purpose neutronics code system applicable to core analyses of various types of reactors. Since the publication of JAERI-1302 for the revised SRAC in 1986, a number of additions and modifications have been made for nuclear data libraries and programs. Thus, the new version SRAC95 has been completed. The system consists of six kinds of nuclear data libraries(ENDF/B-IV, -V, -VI, JENDL-2, -3.1, -3.2), five modular codes integrated into SRAC95; collision probability calculation module (PIJ) for 16 types of lattice geometries, Sn transport calculation modules(ANISN, TWOTRAN), diffusion calculation modules(TUD, CITATION) and two optional codes for fuel assembly and core burn-up calculations(newly developed ASMBURN, revised COREBN). In this version, many new functions and data are implemented to support nuclear design studies of advanced reactors, especially for burn-up calculations. SRAC95 is available not only on conventional IBM-compatible computers but also on scalar or vector computers with the UNIX operating system. This report is the SRAC95 users manual which contains general description, contents of revisions, input data requirements, detail information on usage, sample input data and list of available libraries. (author).

  4. A general purpose code for Monte Carlo simulations

    International Nuclear Information System (INIS)

    Wilcke, W.W.; Rochester Univ., NY

    1984-01-01

    A general-purpose computer code MONTHY has been written to perform Monte Carlo simulations of physical systems. To achieve a high degree of flexibility the code is organized like a general purpose computer, operating on a vector describing the time dependent state of the system under simulation. The instruction set of the 'computer' is defined by the user and is therefore adaptable to the particular problem studied. The organization of MONTHY allows iterative and conditional execution of operations. (orig.)

  5. Specialized Monte Carlo codes versus general-purpose Monte Carlo codes

    International Nuclear Information System (INIS)

    Moskvin, Vadim; DesRosiers, Colleen; Papiez, Lech; Lu, Xiaoyi

    2002-01-01

    The possibilities of Monte Carlo modeling for dose calculations and optimization treatment are quite limited in radiation oncology applications. The main reason is that the Monte Carlo technique for dose calculations is time consuming while treatment planning may require hundreds of possible cases of dose simulations to be evaluated for dose optimization. The second reason is that general-purpose codes widely used in practice, require an experienced user to customize them for calculations. This paper discusses the concept of Monte Carlo code design that can avoid the main problems that are preventing wide spread use of this simulation technique in medical physics. (authors)

  6. Applications for a general purpose optical beam propagation code

    International Nuclear Information System (INIS)

    Munroe, J.L.; Wallace, N.W.

    1987-01-01

    Real world beam propagation and diffraction problems can rarely be solved by the analytical expressions commonly found in optics and lasers textbooks. These equations are typically valid only for paraxial geometries, for specific boundary conditions (e.g., infinite apertures), or for special assumptions (e.g., at focus). Numerical techniques must be used to solve the equations for the general case. LOTS, a public domain numerical beam propagation software package developed for this purpose, is a widely used and proven tool. The graphical presentation of results combined with a well designed command language make LOTS particularly user-friendly, and the recent implementation of LOTS on the IBM PC/XT family of desktop computes will make this capability available to a much larger group of users. This paper surveys several applications demonstrating the need for such a capability

  7. Development of general-purpose particle and heavy ion transport monte carlo code

    International Nuclear Information System (INIS)

    Iwase, Hiroshi; Nakamura, Takashi; Niita, Koji

    2002-01-01

    The high-energy particle transport code NMTC/JAM, which has been developed at JAERI, was improved for the high-energy heavy ion transport calculation by incorporating the JQMD code, the SPAR code and the Shen formula. The new NMTC/JAM named PHITS (Particle and Heavy-Ion Transport code System) is the first general-purpose heavy ion transport Monte Carlo code over the incident energies from several MeV/nucleon to several GeV/nucleon. (author)

  8. General-purpose Monte Carlo codes for neutron and photon transport calculations. MVP version 3

    International Nuclear Information System (INIS)

    Nagaya, Yasunobu

    2017-01-01

    JAEA has developed a general-purpose neutron/photon transport Monte Carlo code MVP. This paper describes the recent development of the MVP code and reviews the basic features and capabilities. In addition, capabilities implemented in Version 3 are also described. (author)

  9. BALTORO a general purpose code for coupling discrete ordinates and Monte-Carlo radiation transport calculations

    International Nuclear Information System (INIS)

    Zazula, J.M.

    1983-01-01

    The general purpose code BALTORO was written for coupling the three-dimensional Monte-Carlo /MC/ with the one-dimensional Discrete Ordinates /DO/ radiation transport calculations. The quantity of a radiation-induced /neutrons or gamma-rays/ nuclear effect or the score from a radiation-yielding nuclear effect can be analysed in this way. (author)

  10. Implementation of the dynamic Monte Carlo method for transient analysis in the general purpose code Tripoli

    Energy Technology Data Exchange (ETDEWEB)

    Sjenitzer, Bart L.; Hoogenboom, J. Eduard, E-mail: B.L.Sjenitzer@TUDelft.nl, E-mail: J.E.Hoogenboom@TUDelft.nl [Delft University of Technology (Netherlands)

    2011-07-01

    A new Dynamic Monte Carlo method is implemented in the general purpose Monte Carlo code Tripoli 4.6.1. With this new method incorporated, a general purpose code can be used for safety transient analysis, such as the movement of a control rod or in an accident scenario. To make the Tripoli code ready for calculating on dynamic systems, the Tripoli scheme had to be altered to incorporate time steps, to include the simulation of delayed neutron precursors and to simulate prompt neutron chains. The modified Tripoli code is tested on two sample cases, a steady-state system and a subcritical system and the resulting neutron fluxes behave just as expected. The steady-state calculation has a constant neutron flux over time and this result shows the stability of the calculation. The neutron flux stays constant with acceptable variance. This also shows that the starting conditions are determined correctly. The sub-critical case shows that the code can also handle dynamic systems with a varying neutron flux. (author)

  11. Implementation of the dynamic Monte Carlo method for transient analysis in the general purpose code Tripoli

    International Nuclear Information System (INIS)

    Sjenitzer, Bart L.; Hoogenboom, J. Eduard

    2011-01-01

    A new Dynamic Monte Carlo method is implemented in the general purpose Monte Carlo code Tripoli 4.6.1. With this new method incorporated, a general purpose code can be used for safety transient analysis, such as the movement of a control rod or in an accident scenario. To make the Tripoli code ready for calculating on dynamic systems, the Tripoli scheme had to be altered to incorporate time steps, to include the simulation of delayed neutron precursors and to simulate prompt neutron chains. The modified Tripoli code is tested on two sample cases, a steady-state system and a subcritical system and the resulting neutron fluxes behave just as expected. The steady-state calculation has a constant neutron flux over time and this result shows the stability of the calculation. The neutron flux stays constant with acceptable variance. This also shows that the starting conditions are determined correctly. The sub-critical case shows that the code can also handle dynamic systems with a varying neutron flux. (author)

  12. Development of a large-scale general purpose two-phase flow analysis code

    International Nuclear Information System (INIS)

    Terasaka, Haruo; Shimizu, Sensuke

    2001-01-01

    A general purpose three-dimensional two-phase flow analysis code has been developed for solving large-scale problems in industrial fields. The code uses a two-fluid model to describe the conservation equations for two-phase flow in order to be applicable to various phenomena. Complicated geometrical conditions are modeled by FAVOR method in structured grid systems, and the discretization equations are solved by a modified SIMPLEST scheme. To reduce computing time a matrix solver for the pressure correction equation is parallelized with OpenMP. Results of numerical examples show that the accurate solutions can be obtained efficiently and stably. (author)

  13. INGEN: a general-purpose mesh generator for finite element codes

    International Nuclear Information System (INIS)

    Cook, W.A.

    1979-05-01

    INGEN is a general-purpose mesh generator for two- and three-dimensional finite element codes. The basic parts of the code are surface and three-dimensional region generators that use linear-blending interpolation formulas. These generators are based on an i, j, k index scheme that is used to number nodal points, construct elements, and develop displacement and traction boundary conditions. This code can generate truss elements (2 modal points); plane stress, plane strain, and axisymmetry two-dimensional continuum elements (4 to 8 nodal points); plate elements (4 to 8 nodal points); and three-dimensional continuum elements (8 to 21 nodal points). The traction loads generated are consistent with the element generated. The expansion--contraction option is of special interest. This option makes it possible to change an existing mesh such that some regions are refined and others are made coarser than the original mesh. 9 figures

  14. Design of tallying function for general purpose Monte Carlo particle transport code JMCT

    International Nuclear Information System (INIS)

    Shangguan Danhua; Li Gang; Deng Li; Zhang Baoyin

    2013-01-01

    A new postponed accumulation algorithm was proposed. Based on JCOGIN (J combinatorial geometry Monte Carlo transport infrastructure) framework and the postponed accumulation algorithm, the tallying function of the general purpose Monte Carlo neutron-photon transport code JMCT was improved markedly. JMCT gets a higher tallying efficiency than MCNP 4C by 28% for simple geometry model, and JMCT is faster than MCNP 4C by two orders of magnitude for complicated repeated structure model. The available ability of tallying function for JMCT makes firm foundation for reactor analysis and multi-step burnup calculation. (authors)

  15. Modelling of a general purpose irradiation chamber using a Monte Carlo particle transport code

    International Nuclear Information System (INIS)

    Dhiyauddin Ahmad Fauzi; Sheik, F.O.A.; Nurul Fadzlin Hasbullah

    2013-01-01

    Full-text: The aim of this research is to stimulate the effectiveness use of a general purpose irradiation chamber to contain pure neutron particles obtained from a research reactor. The secondary neutron and gamma particles dose discharge from the chamber layers will be used as a platform to estimate the safe dimension of the chamber. The chamber, made up of layers of lead (Pb), shielding, polyethylene (PE), moderator and commercial grade aluminium (Al) cladding is proposed for the use of interacting samples with pure neutron particles in a nuclear reactor environment. The estimation was accomplished through simulation based on general Monte Carlo N-Particle transport code using Los Alamos MCNPX software. Simulations were performed on the model of the chamber subjected to high neutron flux radiation and its gamma radiation product. The model of neutron particle used is based on the neutron source found in PUSPATI TRIGA MARK II research reactor which holds a maximum flux value of 1 x 10 12 neutron/ cm 2 s. The expected outcomes of this research are zero gamma dose in the core of the chamber and neutron dose rate of less than 10 μSv/ day discharge from the chamber system. (author)

  16. Use of a general-purpose heat-transfer code for casting simulation

    International Nuclear Information System (INIS)

    Erickson, W.C.

    1975-07-01

    The practical use of numerical techniques in simulating casting solidification dictate that a general purpose heat transfer code be used and that results be obtained in an easy-to-analyze format. Color film plotting routines were developed for use with NASA's CINDA-3G heat transfer code; the combination of which meet the above criteria. The subroutine LQSLTR written for SINDA, the successor to CINDA-3G, was verified by comparing calculated results obtained using LQSLTR with those obtained using the specific heat method for handling the heat of fusion. Excellent agreement existed when similar data was used. When the more restrictive requirement of a 1 0 F melting range was used, comparable results were obtained. Uranium and lead rod castings were cast in instrumented graphite molds and the solidification sequence simulated using CINDA-3G. Discrepancies attributed to initial assumptions of instantaneous mold filling, uniform melt temperature, and intimate metal/mold contact were encountered. Further calculations using a model incorporating a gap between the mold and casting showed that the intimate contact assumption could not be used; a three-dimensional model also showed that the thermocouple assemblies used with the platinum--platinum-10 percent rhodium were a significant perturbation to the system. An L-shaped steel casting was simulated and the results compared to those reported in the literature. The experimental data for this casting were reproduced within the accuracy permitted by the thermal conductivity of the sand, thus demonstrating that agreement can be obtained when the mold material does not act as a chill. (U.S.)

  17. ABAQUS/EPGEN - a general purpose finite element code with emphasis on nonlinear applications

    International Nuclear Information System (INIS)

    Hibbitt, H.D.

    1984-01-01

    The article contains a summary description of ABAQUS, a finite element program designed for general use in nonlinear as well as linear structural problems, in the context of its application to nuclear structural integrity analysis. The article begins with a discussion of the design criteria and methods upon which the code development has been based. The engineering modelling capabilities, currently implemented in the program - elements, constitutive models and analysis procedures - are then described. Finally, a few demonstration examples are presented, to illustrate some of the program's features that are of interest in structural integrity analysis associated with nuclear power plants. (orig.)

  18. A parallelization study of the general purpose Monte Carlo code MCNP4 on a distributed memory highly parallel computer

    International Nuclear Information System (INIS)

    Yamazaki, Takao; Fujisaki, Masahide; Okuda, Motoi; Takano, Makoto; Masukawa, Fumihiro; Naito, Yoshitaka

    1993-01-01

    The general purpose Monte Carlo code MCNP4 has been implemented on the Fujitsu AP1000 distributed memory highly parallel computer. Parallelization techniques developed and studied are reported. A shielding analysis function of the MCNP4 code is parallelized in this study. A technique to map a history to each processor dynamically and to map control process to a certain processor was applied. The efficiency of parallelized code is up to 80% for a typical practical problem with 512 processors. These results demonstrate the advantages of a highly parallel computer to the conventional computers in the field of shielding analysis by Monte Carlo method. (orig.)

  19. ABAQUS-EPGEN: a general-purpose finite-element code. Volume 1. User's manual

    International Nuclear Information System (INIS)

    Hibbitt, H.D.; Karlsson, B.I.; Sorensen, E.P.

    1982-10-01

    This document is the User's Manual for ABAQUS/EPGEN, a general purpose finite element computer program, designed specifically to serve advanced structural analysis needs. The program contains very general libraries of elements, materials and analysis procedures, and is highly modular, so that complex combinations of features can be put together to model physical problems. The program is aimed at production analysis needs, and for this purpose aspects such as ease-of-use, reliability, flexibility and efficiency have received maximum attention. The input language is designed to make it straightforward to describe complicated models; the analysis procedures are highly automated with the program choosing time or load increments based on user supplied tolerances and controls; and the program offers a wide range of post-processing options for display of the analysis results

  20. Implementation of Finite Volume based Navier Stokes Algorithm Within General Purpose Flow Network Code

    Science.gov (United States)

    Schallhorn, Paul; Majumdar, Alok

    2012-01-01

    This paper describes a finite volume based numerical algorithm that allows multi-dimensional computation of fluid flow within a system level network flow analysis. There are several thermo-fluid engineering problems where higher fidelity solutions are needed that are not within the capacity of system level codes. The proposed algorithm will allow NASA's Generalized Fluid System Simulation Program (GFSSP) to perform multi-dimensional flow calculation within the framework of GFSSP s typical system level flow network consisting of fluid nodes and branches. The paper presents several classical two-dimensional fluid dynamics problems that have been solved by GFSSP's multi-dimensional flow solver. The numerical solutions are compared with the analytical and benchmark solution of Poiseulle, Couette and flow in a driven cavity.

  1. Application of a general purpose user's version of the EGS4 code system to a photon skyshine benchmarking calculation

    International Nuclear Information System (INIS)

    Nojiri, I.; Fukasaku, Y.; Narita, O.

    1994-01-01

    A general purpose user's version of the EGS4 code system has been developed to make EGS4 easily applicable to the safety analysis of nuclear fuel cycle facilities. One such application involves the determination of skyshine dose for a variety of photon sources. To verify the accuracy of the code, it was benchmarked with Kansas State University (KSU) photon skyshine experiment of 1977. The results of the simulation showed that this version of EGS4 would be appicable to the skyshine calculation. (author)

  2. ABAQUS-EPGEN: a general-purpose finite element code. Volume 3. Example problems manual

    International Nuclear Information System (INIS)

    Hibbitt, H.D.; Karlsson, B.I.; Sorensen, E.P.

    1983-03-01

    This volume is the Example and Verification Problems Manual for ABAQUS/EPGEN. Companion volumes are the User's, Theory and Systems Manuals. This volume contains two major parts. The bulk of the manual (Sections 1-8) contains worked examples that are discussed in detail, while Appendix A documents a large set of basic verification cases that provide the fundamental check of the elements in the code. The examples in Sections 1-8 illustrate and verify significant aspects of the program's capability. Most of these problems provide verification, but they have also been chosen to allow discussion of modeling and analysis techniques. Appendix A contains basic verification cases. Each of these cases verifies one element in the program's library. The verification consists of applying all possible load or flux types (including thermal loading of stress elements), and all possible foundation or film/radiation conditions, and checking the resulting force and stress solutions or flux and temperature results. This manual provides program verification. All of the problems described in the manual are run and the results checked, for each release of the program, and these verification results are made available

  3. Generalized concatenated quantum codes

    International Nuclear Information System (INIS)

    Grassl, Markus; Shor, Peter; Smith, Graeme; Smolin, John; Zeng Bei

    2009-01-01

    We discuss the concept of generalized concatenated quantum codes. This generalized concatenation method provides a systematical way for constructing good quantum codes, both stabilizer codes and nonadditive codes. Using this method, we construct families of single-error-correcting nonadditive quantum codes, in both binary and nonbinary cases, which not only outperform any stabilizer codes for finite block length but also asymptotically meet the quantum Hamming bound for large block length.

  4. SU-E-T-254: Optimization of GATE and PHITS Monte Carlo Code Parameters for Uniform Scanning Proton Beam Based On Simulation with FLUKA General-Purpose Code

    Energy Technology Data Exchange (ETDEWEB)

    Kurosu, K [Department of Radiation Oncology, Osaka University Graduate School of Medicine, Osaka (Japan); Department of Medical Physics ' Engineering, Osaka University Graduate School of Medicine, Osaka (Japan); Takashina, M; Koizumi, M [Department of Medical Physics ' Engineering, Osaka University Graduate School of Medicine, Osaka (Japan); Das, I; Moskvin, V [Department of Radiation Oncology, Indiana University School of Medicine, Indianapolis, IN (United States)

    2014-06-01

    Purpose: Monte Carlo codes are becoming important tools for proton beam dosimetry. However, the relationships between the customizing parameters and percentage depth dose (PDD) of GATE and PHITS codes have not been reported which are studied for PDD and proton range compared to the FLUKA code and the experimental data. Methods: The beam delivery system of the Indiana University Health Proton Therapy Center was modeled for the uniform scanning beam in FLUKA and transferred identically into GATE and PHITS. This computational model was built from the blue print and validated with the commissioning data. Three parameters evaluated are the maximum step size, cut off energy and physical and transport model. The dependence of the PDDs on the customizing parameters was compared with the published results of previous studies. Results: The optimal parameters for the simulation of the whole beam delivery system were defined by referring to the calculation results obtained with each parameter. Although the PDDs from FLUKA and the experimental data show a good agreement, those of GATE and PHITS obtained with our optimal parameters show a minor discrepancy. The measured proton range R90 was 269.37 mm, compared to the calculated range of 269.63 mm, 268.96 mm, and 270.85 mm with FLUKA, GATE and PHITS, respectively. Conclusion: We evaluated the dependence of the results for PDDs obtained with GATE and PHITS Monte Carlo generalpurpose codes on the customizing parameters by using the whole computational model of the treatment nozzle. The optimal parameters for the simulation were then defined by referring to the calculation results. The physical model, particle transport mechanics and the different geometrybased descriptions need accurate customization in three simulation codes to agree with experimental data for artifact-free Monte Carlo simulation. This study was supported by Grants-in Aid for Cancer Research (H22-3rd Term Cancer Control-General-043) from the Ministry of Health

  5. CRSEC: a general purpose Hauser--Feshbach code for the calculation of nuclear cross-sections and thermonuclear reaction rates

    International Nuclear Information System (INIS)

    Woosley, S.; Fowler, W.A.

    1977-09-01

    CRSEC is a FORTRAN IV computer code designed for the efficient calculation of average nuclear cross sections in situations where a statistical theory of nuclear reactions is applicable and where compound nuclear formation is the dominant reaction mechanism. This code generates cross sections of roughly factor of 2 accuracy for incident particle energies in the range of 10 keV to 10 MeV for most target nuclei from magnesium to bismuth. Exceptions usually involve reactions that enter the compound nucleus at such a low energy that fewer than 10 levels are present in the ''energy window of interest.'' The incident particle must be a neutron, proton, or alpha particle, and only binary reactions resulting in the emission of a single n, p, α, or γ (cascade) are calculated. CRSEC is quite fast, a complete calculation of 12 different reactions over a grid of roughly 150 energy points and the generation of Maxwellian averaged rates taking about 30 seconds of CDC7600 time. Also the semi-empirical parameterization of nuclear properties contained in CRSEC is very general. Greater accuracy may be obtained, however, by furnishing specific low-lying excited states, level density parameterization, and nuclear strength functions. A more general version of CRSEC, called CRSECI, is available that conserves isospin properly in all reactions and allows the user to specify a given degree of isospin mixing in the highly excited states of the compound nucleus. Besides the cross section as a function of center-of-mass energy, CRSEC also generates the Maxwell--Boltzmann averaged thermonuclear reaction rate and temperature dependent nuclear partition function for a grid of temperatures from 10 8 to 10 10 0 K. Sections of this report describe in greater detail the physics employed in CRSEC and how to use the code. 2 tables

  6. GARLIC - A general purpose atmospheric radiative transfer line-by-line infrared-microwave code: Implementation and evaluation

    Science.gov (United States)

    Schreier, Franz; Gimeno García, Sebastián; Hedelt, Pascal; Hess, Michael; Mendrok, Jana; Vasquez, Mayte; Xu, Jian

    2014-04-01

    A suite of programs for high resolution infrared-microwave atmospheric radiative transfer modeling has been developed with emphasis on efficient and reliable numerical algorithms and a modular approach appropriate for simulation and/or retrieval in a variety of applications. The Generic Atmospheric Radiation Line-by-line Infrared Code - GARLIC - is suitable for arbitrary observation geometry, instrumental field-of-view, and line shape. The core of GARLIC's subroutines constitutes the basis of forward models used to implement inversion codes to retrieve atmospheric state parameters from limb and nadir sounding instruments. This paper briefly introduces the physical and mathematical basics of GARLIC and its descendants and continues with an in-depth presentation of various implementation aspects: An optimized Voigt function algorithm combined with a two-grid approach is used to accelerate the line-by-line modeling of molecular cross sections; various quadrature methods are implemented to evaluate the Schwarzschild and Beer integrals; and Jacobians, i.e. derivatives with respect to the unknowns of the atmospheric inverse problem, are implemented by means of automatic differentiation. For an assessment of GARLIC's performance, a comparison of the quadrature methods for solution of the path integral is provided. Verification and validation are demonstrated using intercomparisons with other line-by-line codes and comparisons of synthetic spectra with spectra observed on Earth and from Venus.

  7. Optimization of GATE and PHITS Monte Carlo code parameters for uniform scanning proton beam based on simulation with FLUKA general-purpose code

    Energy Technology Data Exchange (ETDEWEB)

    Kurosu, Keita [Department of Medical Physics and Engineering, Osaka University Graduate School of Medicine, Suita, Osaka 565-0871 (Japan); Department of Radiation Oncology, Osaka University Graduate School of Medicine, Suita, Osaka 565-0871 (Japan); Takashina, Masaaki; Koizumi, Masahiko [Department of Medical Physics and Engineering, Osaka University Graduate School of Medicine, Suita, Osaka 565-0871 (Japan); Das, Indra J. [Department of Radiation Oncology, Indiana University School of Medicine, Indianapolis, IN 46202 (United States); Moskvin, Vadim P., E-mail: vadim.p.moskvin@gmail.com [Department of Radiation Oncology, Indiana University School of Medicine, Indianapolis, IN 46202 (United States)

    2014-10-01

    Although three general-purpose Monte Carlo (MC) simulation tools: Geant4, FLUKA and PHITS have been used extensively, differences in calculation results have been reported. The major causes are the implementation of the physical model, preset value of the ionization potential or definition of the maximum step size. In order to achieve artifact free MC simulation, an optimized parameters list for each simulation system is required. Several authors have already proposed the optimized lists, but those studies were performed with a simple system such as only a water phantom. Since particle beams have a transport, interaction and electromagnetic processes during beam delivery, establishment of an optimized parameters-list for whole beam delivery system is therefore of major importance. The purpose of this study was to determine the optimized parameters list for GATE and PHITS using proton treatment nozzle computational model. The simulation was performed with the broad scanning proton beam. The influences of the customizing parameters on the percentage depth dose (PDD) profile and the proton range were investigated by comparison with the result of FLUKA, and then the optimal parameters were determined. The PDD profile and the proton range obtained from our optimized parameters list showed different characteristics from the results obtained with simple system. This led to the conclusion that the physical model, particle transport mechanics and different geometry-based descriptions need accurate customization in planning computational experiments for artifact-free MC simulation.

  8. Optimization of GATE and PHITS Monte Carlo code parameters for uniform scanning proton beam based on simulation with FLUKA general-purpose code

    International Nuclear Information System (INIS)

    Kurosu, Keita; Takashina, Masaaki; Koizumi, Masahiko; Das, Indra J.; Moskvin, Vadim P.

    2014-01-01

    Although three general-purpose Monte Carlo (MC) simulation tools: Geant4, FLUKA and PHITS have been used extensively, differences in calculation results have been reported. The major causes are the implementation of the physical model, preset value of the ionization potential or definition of the maximum step size. In order to achieve artifact free MC simulation, an optimized parameters list for each simulation system is required. Several authors have already proposed the optimized lists, but those studies were performed with a simple system such as only a water phantom. Since particle beams have a transport, interaction and electromagnetic processes during beam delivery, establishment of an optimized parameters-list for whole beam delivery system is therefore of major importance. The purpose of this study was to determine the optimized parameters list for GATE and PHITS using proton treatment nozzle computational model. The simulation was performed with the broad scanning proton beam. The influences of the customizing parameters on the percentage depth dose (PDD) profile and the proton range were investigated by comparison with the result of FLUKA, and then the optimal parameters were determined. The PDD profile and the proton range obtained from our optimized parameters list showed different characteristics from the results obtained with simple system. This led to the conclusion that the physical model, particle transport mechanics and different geometry-based descriptions need accurate customization in planning computational experiments for artifact-free MC simulation

  9. GARLIC — A general purpose atmospheric radiative transfer line-by-line infrared-microwave code: Implementation and evaluation

    International Nuclear Information System (INIS)

    Schreier, Franz; Gimeno García, Sebastián; Hedelt, Pascal; Hess, Michael; Mendrok, Jana; Vasquez, Mayte; Xu, Jian

    2014-01-01

    A suite of programs for high resolution infrared-microwave atmospheric radiative transfer modeling has been developed with emphasis on efficient and reliable numerical algorithms and a modular approach appropriate for simulation and/or retrieval in a variety of applications. The Generic Atmospheric Radiation Line-by-line Infrared Code — GARLIC — is suitable for arbitrary observation geometry, instrumental field-of-view, and line shape. The core of GARLIC's subroutines constitutes the basis of forward models used to implement inversion codes to retrieve atmospheric state parameters from limb and nadir sounding instruments. This paper briefly introduces the physical and mathematical basics of GARLIC and its descendants and continues with an in-depth presentation of various implementation aspects: An optimized Voigt function algorithm combined with a two-grid approach is used to accelerate the line-by-line modeling of molecular cross sections; various quadrature methods are implemented to evaluate the Schwarzschild and Beer integrals; and Jacobians, i.e. derivatives with respect to the unknowns of the atmospheric inverse problem, are implemented by means of automatic differentiation. For an assessment of GARLIC's performance, a comparison of the quadrature methods for solution of the path integral is provided. Verification and validation are demonstrated using intercomparisons with other line-by-line codes and comparisons of synthetic spectra with spectra observed on Earth and from Venus. - Highlights: • High resolution infrared-microwave radiative transfer model. • Discussion of algorithmic and computational aspects. • Jacobians by automatic/algorithmic differentiation. • Performance evaluation by intercomparisons, verification, validation

  10. Optimization of GATE and PHITS Monte Carlo code parameters for spot scanning proton beam based on simulation with FLUKA general-purpose code

    Energy Technology Data Exchange (ETDEWEB)

    Kurosu, Keita [Department of Radiation Oncology, Indiana University School of Medicine, Indianapolis, IN 46202 (United States); Department of Radiation Oncology, Osaka University Graduate School of Medicine, Suita, Osaka 565-0871 (Japan); Department of Radiology, Osaka University Hospital, Suita, Osaka 565-0871 (Japan); Das, Indra J. [Department of Radiation Oncology, Indiana University School of Medicine, Indianapolis, IN 46202 (United States); Moskvin, Vadim P. [Department of Radiation Oncology, Indiana University School of Medicine, Indianapolis, IN 46202 (United States); Department of Radiation Oncology, St. Jude Children’s Research Hospital, Memphis, TN 38105 (United States)

    2016-01-15

    Spot scanning, owing to its superior dose-shaping capability, provides unsurpassed dose conformity, in particular for complex targets. However, the robustness of the delivered dose distribution and prescription has to be verified. Monte Carlo (MC) simulation has the potential to generate significant advantages for high-precise particle therapy, especially for medium containing inhomogeneities. However, the inherent choice of computational parameters in MC simulation codes of GATE, PHITS and FLUKA that is observed for uniform scanning proton beam needs to be evaluated. This means that the relationship between the effect of input parameters and the calculation results should be carefully scrutinized. The objective of this study was, therefore, to determine the optimal parameters for the spot scanning proton beam for both GATE and PHITS codes by using data from FLUKA simulation as a reference. The proton beam scanning system of the Indiana University Health Proton Therapy Center was modeled in FLUKA, and the geometry was subsequently and identically transferred to GATE and PHITS. Although the beam transport is managed by spot scanning system, the spot location is always set at the center of a water phantom of 600 × 600 × 300 mm{sup 3}, which is placed after the treatment nozzle. The percentage depth dose (PDD) is computed along the central axis using 0.5 × 0.5 × 0.5 mm{sup 3} voxels in the water phantom. The PDDs and the proton ranges obtained with several computational parameters are then compared to those of FLUKA, and optimal parameters are determined from the accuracy of the proton range, suppressed dose deviation, and computational time minimization. Our results indicate that the optimized parameters are different from those for uniform scanning, suggesting that the gold standard for setting computational parameters for any proton therapy application cannot be determined consistently since the impact of setting parameters depends on the proton irradiation

  11. Optimization of GATE and PHITS Monte Carlo code parameters for spot scanning proton beam based on simulation with FLUKA general-purpose code

    International Nuclear Information System (INIS)

    Kurosu, Keita; Das, Indra J.; Moskvin, Vadim P.

    2016-01-01

    Spot scanning, owing to its superior dose-shaping capability, provides unsurpassed dose conformity, in particular for complex targets. However, the robustness of the delivered dose distribution and prescription has to be verified. Monte Carlo (MC) simulation has the potential to generate significant advantages for high-precise particle therapy, especially for medium containing inhomogeneities. However, the inherent choice of computational parameters in MC simulation codes of GATE, PHITS and FLUKA that is observed for uniform scanning proton beam needs to be evaluated. This means that the relationship between the effect of input parameters and the calculation results should be carefully scrutinized. The objective of this study was, therefore, to determine the optimal parameters for the spot scanning proton beam for both GATE and PHITS codes by using data from FLUKA simulation as a reference. The proton beam scanning system of the Indiana University Health Proton Therapy Center was modeled in FLUKA, and the geometry was subsequently and identically transferred to GATE and PHITS. Although the beam transport is managed by spot scanning system, the spot location is always set at the center of a water phantom of 600 × 600 × 300 mm 3 , which is placed after the treatment nozzle. The percentage depth dose (PDD) is computed along the central axis using 0.5 × 0.5 × 0.5 mm 3 voxels in the water phantom. The PDDs and the proton ranges obtained with several computational parameters are then compared to those of FLUKA, and optimal parameters are determined from the accuracy of the proton range, suppressed dose deviation, and computational time minimization. Our results indicate that the optimized parameters are different from those for uniform scanning, suggesting that the gold standard for setting computational parameters for any proton therapy application cannot be determined consistently since the impact of setting parameters depends on the proton irradiation technique

  12. GOC: General Orbit Code

    International Nuclear Information System (INIS)

    Maddox, L.B.; McNeilly, G.S.

    1979-08-01

    GOC (General Orbit Code) is a versatile program which will perform a variety of calculations relevant to isochronous cyclotron design studies. In addition to the usual calculations of interest (e.g., equilibrium and accelerated orbits, focusing frequencies, field isochronization, etc.), GOC has a number of options to calculate injections with a charge change. GOC provides both printed and plotted output, and will follow groups of particles to allow determination of finite-beam properties. An interactive PDP-10 program called GIP, which prepares input data for GOC, is available. GIP is a very easy and convenient way to prepare complicated input data for GOC. Enclosed with this report are several microfiche containing source listings of GOC and other related routines and the printed output from a multiple-option GOC run

  13. General Purpose (office) Network reorganisation

    CERN Multimedia

    IT Department

    2016-01-01

    On Saturday 27 August, the IT Department’s Communication Systems group will perform a major reorganisation of CERN’s General Purpose Network.   This reorganisation will cause network interruptions on Saturday 27 August (and possibly Sunday 28 August) and will be followed by a change to the IP addresses of connected systems that will come into effect on Monday 3 October. For further details and information about the actions you may need to take, please see: https://information-technology.web.cern.ch/news/general-purpose-office-network-reorganisation.

  14. General purpose programmable accelerator board

    Science.gov (United States)

    Robertson, Perry J.; Witzke, Edward L.

    2001-01-01

    A general purpose accelerator board and acceleration method comprising use of: one or more programmable logic devices; a plurality of memory blocks; bus interface for communicating data between the memory blocks and devices external to the board; and dynamic programming capabilities for providing logic to the programmable logic device to be executed on data in the memory blocks.

  15. MVP/GMVP Version 3. General purpose Monte Carlo codes for neutron and photon transport calculations based on continuous energy and multigroup methods (Translated document)

    International Nuclear Information System (INIS)

    Nagaya, Yasunobu; Okumura, Keisuke; Sakurai, Takeshi; Mori, Takamasa

    2017-03-01

    In order to realize fast and accurate Monte Carlo simulation of neutron and photon transport problems, two Monte Carlo codes MVP (continuous-energy method) and GMVP (multigroup method) have been developed at Japan Atomic Energy Agency. The codes have adopted a vectorized algorithm and have been developed for vector-type supercomputers. They also support parallel processing with a standard parallelization library MPI and thus a speed-up of Monte Carlo calculations can be achieved on general computing platforms. The first and second versions of the codes were released in 1994 and 2005, respectively. They have been extensively improved and new capabilities have been implemented. The major improvements and new capabilities are as follows: (1) perturbation calculation for effective multiplication factor, (2) exact resonant elastic scattering model, (3) calculation of reactor kinetics parameters, (4) photo-nuclear model, (5) simulation of delayed neutrons, (6) generation of group constants. This report describes the physical model, geometry description method used in the codes, new capabilities and input instructions. (author)

  16. MVP/GMVP version 3. General purpose Monte Carlo codes for neutron and photon transport calculations based on continuous energy and multigroup methods

    International Nuclear Information System (INIS)

    Nagaya, Yasunobu; Okumura, Keisuke; Sakurai, Takeshi; Mori, Takamasa

    2017-03-01

    In order to realize fast and accurate Monte Carlo simulation of neutron and photon transport problems, two Monte Carlo codes MVP (continuous-energy method) and GMVP (multigroup method) have been developed at Japan Atomic Energy Agency. The codes have adopted a vectorized algorithm and have been developed for vector-type supercomputers. They also support parallel processing with a standard parallelization library MPI and thus a speed-up of Monte Carlo calculations can be achieved on general computing platforms. The first and second versions of the codes were released in 1994 and 2005, respectively. They have been extensively improved and new capabilities have been implemented. The major improvements and new capabilities are as follows: (1) perturbation calculation for effective multiplication factor, (2) exact resonant elastic scattering model, (3) calculation of reactor kinetics parameters, (4) photo-nuclear model, (5) simulation of delayed neutrons, (6) generation of group constants. This report describes the physical model, geometry description method used in the codes, new capabilities and input instructions. (author)

  17. General Purpose Heat Source Simulator

    Science.gov (United States)

    Emrich, Bill

    2008-01-01

    The General Purpose Heat Source (GPHS) simulator project is designed to replicate through the use of electrical heaters, the form, fit, and function of actual GPHS modules which generate heat through the radioactive decay of Pu238. The use of electrically heated modules rather than modules containing Pu238 facilitates the testing of spacecraft subsystems and systems without sacrificing the quantity and quality of the test data gathered. Previous GPHS activities are centered around developing robust heater designs with sizes and weights that closely matched those of actual Pu238 fueled GPHS blocks. These efforts were successful, although their maximum temperature capabilities were limited to around 850 C. New designs are being pursued which also replicate the sizes and weights of actual Pu238 fueled GPHS blocks but will allow operation up to 1100 C.

  18. MVP/GMVP 2: general purpose Monte Carlo codes for neutron and photon transport calculations based on continuous energy and multigroup methods

    International Nuclear Information System (INIS)

    Nagaya, Yasunobu; Okumura, Keisuke; Mori, Takamasa; Nakagawa, Masayuki

    2005-06-01

    In order to realize fast and accurate Monte Carlo simulation of neutron and photon transport problems, two vectorized Monte Carlo codes MVP and GMVP have been developed at JAERI. MVP is based on the continuous energy model and GMVP is on the multigroup model. Compared with conventional scalar codes, these codes achieve higher computation speed by a factor of 10 or more on vector super-computers. Both codes have sufficient functions for production use by adopting accurate physics model, geometry description capability and variance reduction techniques. The first version of the codes was released in 1994. They have been extensively improved and new functions have been implemented. The major improvements and new functions are (1) capability to treat the scattering model expressed with File 6 of the ENDF-6 format, (2) time-dependent tallies, (3) reaction rate calculation with the pointwise response function, (4) flexible source specification, (5) continuous-energy calculation at arbitrary temperatures, (6) estimation of real variances in eigenvalue problems, (7) point detector and surface crossing estimators, (8) statistical geometry model, (9) function of reactor noise analysis (simulation of the Feynman-α experiment), (10) arbitrary shaped lattice boundary, (11) periodic boundary condition, (12) parallelization with standard libraries (MPI, PVM), (13) supporting many platforms, etc. This report describes the physical model, geometry description method used in the codes, new functions and how to use them. (author)

  19. General Monte Carlo code MONK

    International Nuclear Information System (INIS)

    Moore, J.G.

    1974-01-01

    The Monte Carlo code MONK is a general program written to provide a high degree of flexibility to the user. MONK is distinguished by its detailed representation of nuclear data in point form i.e., the cross-section is tabulated at specific energies instead of the more usual group representation. The nuclear data are unadjusted in the point form but recently the code has been modified to accept adjusted group data as used in fast and thermal reactor applications. The various geometrical handling capabilities and importance sampling techniques are described. In addition to the nuclear data aspects, the following features are also described; geometrical handling routines, tracking cycles, neutron source and output facilities. 12 references. (U.S.)

  20. 10 CFR 205.350 - General purpose.

    Science.gov (United States)

    2010-01-01

    ... 10 Energy 3 2010-01-01 2010-01-01 false General purpose. 205.350 Section 205.350 Energy DEPARTMENT OF ENERGY OIL ADMINISTRATIVE PROCEDURES AND SANCTIONS Electric Power System Permits and Reports....350 General purpose. The purpose of this rule is to establish a procedure for the Office of...

  1. 12 CFR 1703.31 - General purposes.

    Science.gov (United States)

    2010-01-01

    ... 12 Banks and Banking 7 2010-01-01 2010-01-01 false General purposes. 1703.31 Section 1703.31 Banks and Banking OFFICE OF FEDERAL HOUSING ENTERPRISE OVERSIGHT, DEPARTMENT OF HOUSING AND URBAN... Legal Proceedings in Which OFHEO Is Not a Named Party § 1703.31 General purposes. The purposes of this...

  2. 7 CFR 254.1 - General purpose.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false General purpose. 254.1 Section 254.1 Agriculture... GENERAL REGULATIONS AND POLICIES-FOOD DISTRIBUTION ADMINISTRATION OF THE FOOD DISTRIBUTION PROGRAM FOR INDIAN HOUSEHOLDS IN OKLAHOMA § 254.1 General purpose. This part sets the requirement under which...

  3. 22 CFR 309.1 - General purpose.

    Science.gov (United States)

    2010-04-01

    ... 22 Foreign Relations 2 2010-04-01 2010-04-01 true General purpose. 309.1 Section 309.1 Foreign Relations PEACE CORPS DEBT COLLECTION General Provisions § 309.1 General purpose. This part prescribes the procedures to be used by the United States Peace Corps (Peace Corps) in the collection and/or disposal of non...

  4. Development of a General Purpose Gamification Framework

    OpenAIRE

    Vea, Eivind

    2016-01-01

    This report describes the design and implementation of a general purpose gamification framework developed in JavaScript on the Metor platform. Gamification is described as the use of game elements in none-game contexts. The purpose is to encourage and change user behaviour. Examples of existing gamification use cases and frameworks are described. A demo game shows how a general purpose framework can be used.

  5. Towers of generalized divisible quantum codes

    Science.gov (United States)

    Haah, Jeongwan

    2018-04-01

    A divisible binary classical code is one in which every code word has weight divisible by a fixed integer. If the divisor is 2ν for a positive integer ν , then one can construct a Calderbank-Shor-Steane (CSS) code, where X -stabilizer space is the divisible classical code, that admits a transversal gate in the ν th level of Clifford hierarchy. We consider a generalization of the divisibility by allowing a coefficient vector of odd integers with which every code word has zero dot product modulo the divisor. In this generalized sense, we construct a CSS code with divisor 2ν +1 and code distance d from any CSS code of code distance d and divisor 2ν where the transversal X is a nontrivial logical operator. The encoding rate of the new code is approximately d times smaller than that of the old code. In particular, for large d and ν ≥2 , our construction yields a CSS code of parameters [[O (dν -1) ,Ω (d ) ,d ] ] admitting a transversal gate at the ν th level of Clifford hierarchy. For our construction we introduce a conversion from magic state distillation protocols based on Clifford measurements to those based on codes with transversal T gates. Our tower contains, as a subclass, generalized triply even CSS codes that have appeared in so-called gauge fixing or code switching methods.

  6. General-purpose radiographic and fluoroscopic table

    International Nuclear Information System (INIS)

    Ishizaki, Noritaka

    1982-01-01

    A new series of diagnostic tables, Model DT-KEL, was developed for general-purpose radiographic and fluoroscopic systems. Through several investigations, the table was so constructed that the basic techniques be general radiography and GI examination, and other techniques be optionally added. The diagnostic tables involve the full series of the type for various purposes and are systematized with the surrounding equipment. A retractable mechanism of grids was adopted first for general use. The fine grids with a density of 57 lines per cm, which was adopted in KEL-2, reduced the X-ray doses by 16 percent. (author)

  7. General purpose computers in real time

    International Nuclear Information System (INIS)

    Biel, J.R.

    1989-01-01

    I see three main trends in the use of general purpose computers in real time. The first is more processing power. The second is the use of higher speed interconnects between computers (allowing more data to be delivered to the processors). The third is the use of larger programs running in the computers. Although there is still work that needs to be done, I believe that all indications are that the online need for general purpose computers should be available for the SCC and LHC machines. 2 figs

  8. Summary of JENDL-2 general purpose file

    Energy Technology Data Exchange (ETDEWEB)

    Nakagawa, Tsuneo [ed.

    1984-06-15

    The general purpose file of the second version of Japanese Evaluated Nuclear Data Library (JENDL-2) was released in December 1982. Recently, descriptive data were added to JENDL-2 and at the same time the first revision of numerical data was performed. JENDL-2 (Rev.1) consists of the data for 89 nuclides and about 211,000 records in the ENDF/B-IV format. In this report, full listings of presently added descriptive data are given to summarize the JENDL-2 general purpose file. The 2200-m/sec and 14-MeV cross sections, resonance integrals, Maxwellian and fission spectrum averaged cross sections are given in a table. Average cross sections were also calculated in suitable energy intervals.

  9. Summary of JENDL-2 general purpose file

    International Nuclear Information System (INIS)

    Nakagawa, Tsuneo

    1984-06-01

    The general purpose file of the second version of Japanese Evaluated Nuclear Data Library (JENDL-2) was released in December 1982. Recently, descriptive data were added to JENDL-2 and at the same time the first revision of numerical data was performed. JENDL-2 (Rev1) consists of the data for 89 nuclides and about 211,000 records in the ENDF/B-IV format. In this report, full listings of presently added descriptive data are given to summarize the JENDL-2 general purpose file. The 2200-m/sec and 14-MeV cross sections, resonance integrals, Maxwellian and fission spectrum averaged cross sections are given in a table. Average cross sections were also calculated in suitable energy intervals. (author)

  10. General-purpose RFQ design program

    International Nuclear Information System (INIS)

    Wadlinger, E.A.

    1984-01-01

    We have written a general-purpose, radio-frequency quadrupole (RFQ) design program that allows maximum flexibility in picking design algorithms. This program optimizes the RFQ on any combination of design parameters while simultaneously satisfying mutually compatible, physically required constraint equations. It can be very useful for deriving various scaling laws for RFQs. This program has a friendly user interface in addition to checking the consistency of the user-defined requirements and is written to minimize the effort needed to incorporate additional constraint equations. We describe the program and present some examples

  11. Report of the general purpose detector group

    International Nuclear Information System (INIS)

    Barbaro-Galtieri, A.; Bartel, W.; Bulos, F.; Cool, R.; Hanson, G.; Koetz, U.; Kottahaus, R.; Loken, S.; Luke, D.; Rothenberg, A.

    1975-01-01

    A general purpose detector for PEP is described. The main components of this detector are a l meter radius, 15 kilogauss superconducting solenoidal magnet with drift chambers to detect and measure the momentum of charged particles, a liquid argon neutral detector and hadron calorimeter, and a system of Cherenkov and time-of-flight counters for identification of charged hadrons. A major consideration in the design of this detector was that it be flexible: the magnet coil and drift chambers form a core around which various apparatus for specialized detection can be placed

  12. General Purpose Crate (GPC) for control applications

    International Nuclear Information System (INIS)

    Singh, Kundan; Munda, Deepak K.; Jain, Mamta; Archunan, M.; Barua, P.; Ajith Kumar, B.P.

    2011-01-01

    A General Purpose Crate (GPC) capable of handling digital and analog Inputs/Outputs signals has been developed at Inter University Accelerator Centre (IUAC), New Delhi, for accelerator control system applications. The system includes back-plane bus with on board plugged-in single board computer with PC104 and Ethernet interface, running Linux operating system. The bus control logic is designed on the back-plane pcb itself, making the system more rugged. The various types of digital and analog input/output modules can be plugged into the back plane bus randomly with standard euro connectors, which provides highly reliable and dust free contacts. Maximum eight modules can be inserted into the crate. The total power consumption for various types of modules and back-plane controller is approximately 50 watts. The multi-output DC power supply from COSEL has been used in the crate. The general purpose crate is software compatible with the CAMAC crates used in the accelerator control system. (author)

  13. Sudan-decoding generalized geometric Goppa codes

    DEFF Research Database (Denmark)

    Heydtmann, Agnes Eileen

    2003-01-01

    Generalized geometric Goppa codes are vector spaces of n-tuples with entries from different extension fields of a ground field. They are derived from evaluating functions similar to conventional geometric Goppa codes, but allowing evaluation in places of arbitrary degree. A decoding scheme...... for these codes based on Sudan's improved algorithm is presented and its error-correcting capacity is analyzed. For the implementation of the algorithm it is necessary that the so-called increasing zero bases of certain spaces of functions are available. A method to obtain such bases is developed....

  14. General-Purpose Software For Computer Graphics

    Science.gov (United States)

    Rogers, Joseph E.

    1992-01-01

    NASA Device Independent Graphics Library (NASADIG) is general-purpose computer-graphics package for computer-based engineering and management applications which gives opportunity to translate data into effective graphical displays for presentation. Features include two- and three-dimensional plotting, spline and polynomial interpolation, control of blanking of areas, multiple log and/or linear axes, control of legends and text, control of thicknesses of curves, and multiple text fonts. Included are subroutines for definition of areas and axes of plots; setup and display of text; blanking of areas; setup of style, interpolation, and plotting of lines; control of patterns and of shading of colors; control of legends, blocks of text, and characters; initialization of devices; and setting of mixed alphabets. Written in FORTRAN 77.

  15. Can Universities Profit from General Purpose Inventions?

    DEFF Research Database (Denmark)

    Barirani, Ahmad; Beaudry, Catherine; Agard, Bruno

    2017-01-01

    The lack of control over downstream assets can hinder universities’ ability to extract rents from their inventive activities. We explore this possibility by assessing the relationship between invention generality and renewal decisions for a sample of Canadian nanotechnology patents. Our results s...

  16. 7 CFR 225.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false General purpose and scope. 225.1 Section 225.1... AGRICULTURE CHILD NUTRITION PROGRAMS SUMMER FOOD SERVICE PROGRAM General § 225.1 General purpose and scope... primary purpose of the Program is to provide food service to children from needy areas during periods when...

  17. 7 CFR 2902.48 - General purpose household cleaners.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 15 2010-01-01 2010-01-01 false General purpose household cleaners. 2902.48 Section... PROCUREMENT Designated Items § 2902.48 General purpose household cleaners. (a) Definition. Products designed... procurement preference for qualifying biobased general purpose household cleaners. By that date, Federal...

  18. 7 CFR 2902.37 - General purpose de-icers.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 15 2010-01-01 2010-01-01 false General purpose de-icers. 2902.37 Section 2902.37... Items § 2902.37 General purpose de-icers. (a) Definition. Chemical products (e.g., salt, fluids) that... preference for qualifying biobased general purpose de-icers. By that date, Federal agencies that have the...

  19. 7 CFR 285.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false General purpose and scope. 285.1 Section 285.1... COMMONWEALTH OF PUERTO RICO § 285.1 General purpose and scope. This part describes the general terms and... government of the Commonwealth of Puerto Rico for the purpose of designing and conducting a nutrition...

  20. 47 CFR 32.6124 - General purpose computers expense.

    Science.gov (United States)

    2010-10-01

    ... 47 Telecommunication 2 2010-10-01 2010-10-01 false General purpose computers expense. 32.6124... General purpose computers expense. This account shall include the costs of personnel whose principal job is the physical operation of general purpose computers and the maintenance of operating systems. This...

  1. 7 CFR 246.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false General purpose and scope. 246.1 Section 246.1... General § 246.1 General purpose and scope. This part announces regulations under which the Secretary of... health by reason of inadequate nutrition or health care, or both. The purpose of the Program is to...

  2. The Efficiency of Linda for General Purpose Scientific Programming

    Directory of Open Access Journals (Sweden)

    Timothy G. Mattson

    1994-01-01

    Full Text Available Linda (Linda is a registered trademark of Scientific Computing Associates, Inc. is a programming language for coordinating the execution and interaction of processes. When combined with a language for computation (such as C or Fortran, the resulting hybrid language can be used to write portable programs for parallel and distributed multiple instruction multiple data (MIMD computers. The Linda programming model is based on operations that read, write, and erase a virtual shared memory. It is easy to use, and lets the programmer code in a very expressive, uncoupled programming style. These benefits, however, are of little value unless Linda programs execute efficiently. The goal of this article is to demonstrate that Linda programs are efficient making Linda an effective general purpose tool for programming MIMD parallel computers. Two arguments for Linda's efficiency are given; the first is based on Linda's implementation and the second on a range of case studies spanning a complete set of parallel algorithm classes.

  3. 7 CFR 250.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false General purpose and scope. 250.1 Section 250.1... AGRICULTURE GENERAL REGULATIONS AND POLICIES-FOOD DISTRIBUTION DONATION OF FOODS FOR USE IN THE UNITED STATES, ITS TERRITORIES AND POSSESSIONS AND AREAS UNDER ITS JURISDICTION General § 250.1 General purpose and...

  4. 7 CFR 277.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false General purpose and scope. 277.1 Section 277.1 Agriculture Regulations of the Department of Agriculture (Continued) FOOD AND NUTRITION SERVICE, DEPARTMENT OF... AGENCIES § 277.1 General purpose and scope. (a) Purpose. This part establishes uniform requirements for the...

  5. The general theory of convolutional codes

    Science.gov (United States)

    Mceliece, R. J.; Stanley, R. P.

    1993-01-01

    This article presents a self-contained introduction to the algebraic theory of convolutional codes. This introduction is partly a tutorial, but at the same time contains a number of new results which will prove useful for designers of advanced telecommunication systems. Among the new concepts introduced here are the Hilbert series for a convolutional code and the class of compact codes.

  6. 7 CFR 253.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false General purpose and scope. 253.1 Section 253.1... AGRICULTURE GENERAL REGULATIONS AND POLICIES-FOOD DISTRIBUTION ADMINISTRATION OF THE FOOD DISTRIBUTION PROGRAM FOR HOUSEHOLDS ON INDIAN RESERVATIONS § 253.1 General purpose and scope. This part describes the terms...

  7. 7 CFR 1485.10 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 10 2010-01-01 2010-01-01 false General purpose and scope. 1485.10 Section 1485.10... FOREIGN MARKETS FOR AGRICULTURAL COMMODITIES Market Access Program § 1485.10 General purpose and scope. (a.../Market Access Program (EIP/MAP). It also establishes the general terms and conditions applicable to MAP...

  8. 7 CFR 215.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false General purpose and scope. 215.1 Section 215.1... AGRICULTURE CHILD NUTRITION PROGRAMS SPECIAL MILK PROGRAM FOR CHILDREN § 215.1 General purpose and scope. This part announces the policies and prescribes the general regulations with respect to the Special Milk...

  9. 7 CFR 248.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false General purpose and scope. 248.1 Section 248.1... AGRICULTURE CHILD NUTRITION PROGRAMS WIC FARMERS' MARKET NUTRITION PROGRAM (FMNP) General § 248.1 General purpose and scope. This part announces regulations under which the Secretary of Agriculture shall carry...

  10. 46 CFR 7.1 - General purpose of boundary lines.

    Science.gov (United States)

    2010-10-01

    ... 46 Shipping 1 2010-10-01 2010-10-01 false General purpose of boundary lines. 7.1 Section 7.1 Shipping COAST GUARD, DEPARTMENT OF HOMELAND SECURITY PROCEDURES APPLICABLE TO THE PUBLIC BOUNDARY LINES General § 7.1 General purpose of boundary lines. The lines in this part delineate the application of the...

  11. 7 CFR 251.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false General purpose and scope. 251.1 Section 251.1... AGRICULTURE GENERAL REGULATIONS AND POLICIES-FOOD DISTRIBUTION THE EMERGENCY FOOD ASSISTANCE PROGRAM § 251.1 General purpose and scope. This part announces the policies and prescribes the regulations necessary to...

  12. 7 CFR 226.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false General purpose and scope. 226.1 Section 226.1 Agriculture Regulations of the Department of Agriculture (Continued) FOOD AND NUTRITION SERVICE, DEPARTMENT OF AGRICULTURE CHILD NUTRITION PROGRAMS CHILD AND ADULT CARE FOOD PROGRAM General § 226.1 General purpose and...

  13. Multiple component codes based generalized LDPC codes for high-speed optical transport.

    Science.gov (United States)

    Djordjevic, Ivan B; Wang, Ting

    2014-07-14

    A class of generalized low-density parity-check (GLDPC) codes suitable for optical communications is proposed, which consists of multiple local codes. It is shown that Hamming, BCH, and Reed-Muller codes can be used as local codes, and that the maximum a posteriori probability (MAP) decoding of these local codes by Ashikhmin-Lytsin algorithm is feasible in terms of complexity and performance. We demonstrate that record coding gains can be obtained from properly designed GLDPC codes, derived from multiple component codes. We then show that several recently proposed classes of LDPC codes such as convolutional and spatially-coupled codes can be described using the concept of GLDPC coding, which indicates that the GLDPC coding can be used as a unified platform for advanced FEC enabling ultra-high speed optical transport. The proposed class of GLDPC codes is also suitable for code-rate adaption, to adjust the error correction strength depending on the optical channel conditions.

  14. MONK - a general purpose Monte Carlo neutronics program

    International Nuclear Information System (INIS)

    Sherriffs, V.S.W.

    1978-01-01

    MONK is a Monte Carlo neutronics code written principally for criticality calculations relevant to the transport, storage, and processing of fissile material. The code exploits the ability of the Monte Carlo method to represent complex shapes with very great accuracy. The nuclear data used is derived from the UK Nuclear Data File processed to the required format by a subsidiary program POND. A general description is given of the MONK code together with the subsidiary program SCAN which produces diagrams of the system specified. Details of the data input required by MONK and SCAN are also given. (author)

  15. 7 CFR 245.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false General purpose and scope. 245.1 Section 245.1 Agriculture Regulations of the Department of Agriculture (Continued) FOOD AND NUTRITION SERVICE, DEPARTMENT OF... SCHOOLS § 245.1 General purpose and scope. (a) This part established the responsibilities of State...

  16. 7 CFR 220.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false General purpose and scope. 220.1 Section 220.1 Agriculture Regulations of the Department of Agriculture (Continued) FOOD AND NUTRITION SERVICE, DEPARTMENT OF AGRICULTURE CHILD NUTRITION PROGRAMS SCHOOL BREAKFAST PROGRAM § 220.1 General purpose and scope. This part...

  17. 7 CFR 281.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false General purpose and scope. 281.1 Section 281.1 Agriculture Regulations of the Department of Agriculture (Continued) FOOD AND NUTRITION SERVICE, DEPARTMENT OF... RESERVATIONS § 281.1 General purpose and scope. (a) These regulations govern the operation of the Food Stamp...

  18. 7 CFR 1728.10 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 11 2010-01-01 2010-01-01 false General purpose and scope. 1728.10 Section 1728.10 Agriculture Regulations of the Department of Agriculture (Continued) RURAL UTILITIES SERVICE, DEPARTMENT OF AGRICULTURE ELECTRIC STANDARDS AND SPECIFICATIONS FOR MATERIALS AND CONSTRUCTION § 1728.10 General purpose and...

  19. 7 CFR 235.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false General purpose and scope. 235.1 Section 235.1 Agriculture Regulations of the Department of Agriculture (Continued) FOOD AND NUTRITION SERVICE, DEPARTMENT OF AGRICULTURE CHILD NUTRITION PROGRAMS STATE ADMINISTRATIVE EXPENSE FUNDS § 235.1 General purpose and scope...

  20. 21 CFR 864.4010 - General purpose reagent.

    Science.gov (United States)

    2010-04-01

    ... 21 Food and Drugs 8 2010-04-01 2010-04-01 false General purpose reagent. 864.4010 Section 864.4010 Food and Drugs FOOD AND DRUG ADMINISTRATION, DEPARTMENT OF HEALTH AND HUMAN SERVICES (CONTINUED) MEDICAL DEVICES HEMATOLOGY AND PATHOLOGY DEVICES Specimen Preparation Reagents § 864.4010 General purpose...

  1. 47 CFR 32.2124 - General purpose computers.

    Science.gov (United States)

    2010-10-01

    ... 47 Telecommunication 2 2010-10-01 2010-10-01 false General purpose computers. 32.2124 Section 32... General purpose computers. (a) This account shall include the original cost of computers and peripheral... financial, statistical, or other business analytical reports; preparation of payroll, customer bills, and...

  2. Standalone General Purpose Data Logger Design and Implementation

    African Journals Online (AJOL)

    This paper describes the design of a general purpose data logger that is compatible with a variety of transducers, potentially permitting the measurement and recording of a wide range of phenomena. The recorded data can be retrieved to a PC via an RS-232 serial port. The standalone general purpose data logger ...

  3. WACC and a Generalized Tax Code

    OpenAIRE

    Husmann, Sven; Kruschwitz, Lutz; Löffler, Andreas

    2001-01-01

    We extend the WACC approach to a tax system having a firm income tax and a personal income tax of the investor as well. We use an artificial tax system incorporating most of the G-7 national tax codes as for example the classical or the imputation systems. On our website (www.wacc.de) WACC formulas according to many of the actual G-7 national tax codes can be found.

  4. Efficient probabilistic model checking on general purpose graphic processors

    NARCIS (Netherlands)

    Bosnacki, D.; Edelkamp, S.; Sulewski, D.; Pasareanu, C.S.

    2009-01-01

    We present algorithms for parallel probabilistic model checking on general purpose graphic processing units (GPGPUs). For this purpose we exploit the fact that some of the basic algorithms for probabilistic model checking rely on matrix vector multiplication. Since this kind of linear algebraic

  5. 7 CFR 227.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... management training of school foodservice personnel, and (c) the conduct of nutrition education activities in... Agriculture Regulations of the Department of Agriculture (Continued) FOOD AND NUTRITION SERVICE, DEPARTMENT OF AGRICULTURE CHILD NUTRITION PROGRAMS NUTRITION EDUCATION AND TRAINING PROGRAM General § 227.1 General purpose...

  6. 24 CFR 902.1 - Purpose and general description.

    Science.gov (United States)

    2010-04-01

    ... assessments. The Real Estate Assessment Center (REAC) is responsible for assessing and scoring the performance... uniform and objective protocols for the physical inspection of properties and the financial assessment of... URBAN DEVELOPMENT PUBLIC HOUSING ASSESSMENT SYSTEM General Provisions § 902.1 Purpose and general...

  7. Development of General Purpose Data Acquisition Shell (GPDAS)

    International Nuclear Information System (INIS)

    Chung, Y.; Kim, K.

    1995-01-01

    This note is intended as an abbreviated introduction to the concept and the structure of General Purpose Data Acquisitions Shell (GPDAS) and assumes the reader has a certain level of familiarity with programming in general. The structure of the following sections consists of brief explanations of the concepts and commands of GPDAS, followed by several examples. Some of these are tabulated in the appendices at the end of this note

  8. Syndrome-source-coding and its universal generalization. [error correcting codes for data compression

    Science.gov (United States)

    Ancheta, T. C., Jr.

    1976-01-01

    A method of using error-correcting codes to obtain data compression, called syndrome-source-coding, is described in which the source sequence is treated as an error pattern whose syndrome forms the compressed data. It is shown that syndrome-source-coding can achieve arbitrarily small distortion with the number of compressed digits per source digit arbitrarily close to the entropy of a binary memoryless source. A 'universal' generalization of syndrome-source-coding is formulated which provides robustly effective distortionless coding of source ensembles. Two examples are given, comparing the performance of noiseless universal syndrome-source-coding to (1) run-length coding and (2) Lynch-Davisson-Schalkwijk-Cover universal coding for an ensemble of binary memoryless sources.

  9. On the System and Engineering Design of the General Purpose ...

    Indian Academy of Sciences (India)

    Home; Journals; Resonance – Journal of Science Education; Volume 13; Issue 5. On the System and Engineering Design of the General Purpose Electronic Digital Computer at TIFR. Rangaswamy Narasimhan. Classics Volume 13 Issue 5 May 2008 pp 490-501 ...

  10. Geographical parthenogenesis: General purpose genotypes and frozen niche variation

    DEFF Research Database (Denmark)

    Vrijenhoek, Robert C.; Parker, Dave

    2009-01-01

    hypotheses concerning the evolution of niche breadth in asexual species - the "general-purpose genotype" (GPG) and "frozen niche-variation" (FNV) models. The two models are often portrayed as mutually exclusive, respectively viewing clonal lineages as generalists versus specialists. Nonetheless...

  11. Sense and purpose of bilateral agreements - a general survey

    International Nuclear Information System (INIS)

    Buehler, O.

    1999-01-01

    Switzerland has concluded with its neighbouring States several bilateral agreements on nuclear information and on mutual help in case of catastrophes. The following general survey explains the sense and purpose of these agreements which complement the IAEA-Conventions referring to the same matters. (orig.) [de

  12. Weldability of general purpose heat source new-process iridium

    International Nuclear Information System (INIS)

    Kanne, W.R.

    1987-01-01

    Weldability tests on General Purpose Heat Source (GPHS) iridium capsules showed that a new iridium fabrication process reduced susceptibility to underbead cracking. Seventeen capsules were welded (a total of 255 welds) in four categories and the number of cracks in each weld was measured

  13. General-purpose heat source development. Phase I: design requirements

    International Nuclear Information System (INIS)

    Snow, E.C.; Zocher, R.W.

    1978-09-01

    Studies have been performed to determine the necessary design requirements for a 238 PuO 2 General-Purpose Heat Source (GPHS). Systems and missions applications, as well as accident conditions, were considered. The results of these studies, along with the recommended GPHS design requirements, are given in this report

  14. A Small Acoustic Goniometer for General Purpose Research.

    Science.gov (United States)

    Pook, Michael L; Loo, Sin Ming

    2016-04-29

    Understanding acoustic events and monitoring their occurrence is a useful aspect of many research projects. In particular, acoustic goniometry allows researchers to determine the source of an event based solely on the sound it produces. The vast majority of acoustic goniometry research projects used custom hardware targeted to the specific application under test. Unfortunately, due to the wide range of sensing applications, a flexible general purpose hardware/firmware system does not exist for this purpose. This article focuses on the development of such a system which encourages the continued exploration of general purpose hardware/firmware and lowers barriers to research in projects requiring the use of acoustic goniometry. Simulations have been employed to verify system feasibility, and a complete hardware implementation of the acoustic goniometer has been designed and field tested. The results are reported, and suggested areas for improvement and further exploration are discussed.

  15. General-Purpose Data Containers for Science and Engineering

    International Nuclear Information System (INIS)

    2015-01-01

    In 2012 the SG38 international committee was formed to develop a modern structure to replace the ENDF-6 format for storing evaluated nuclear reaction data on a computer system. This committee divided the project into seven tasks. One of these tasks, the design of General-Purpose Data Containers (GPDCs), is described in this article. What type of data does SG38 need to store and why is the task called General-Purpose Data Containers? The most common types of data in an evaluated nuclear reaction database are representations of physical functions in tabulated forms. There is also a need to store 1-dimensional functions using truncated Legendre or polynomial (or others) expansions. The phrase General-Purpose implies that the containers are to be designed to store generic forms of tabulated data rather than one for each physical function. Also, where possible, it would be beneficial to design containers that can store data forms not currently used in evaluated nuclear database or at least be easily extended. In addition to containers for storing physical functions as tabulated data, other types of containers are needed. There exists a desire within SG38 to support the storage of documentation at various levels within an evaluated file. Containers for storing non-functional data (e.g., a list of numbers) as well as units and labels for axes are also needed. Herein, containers for storing physical functions are called functional containers. One of the goals for the general-purpose data containers task is to design containers that will be useful to other scientific and engineering applications. To meet this goal, task members should think outside of the immediate needs of evaluated nuclear data to ensure that the containers are general- purpose rather than simply repackaged versions of existing containers. While the examples in this article may be specific to nuclear reaction data, it is hoped that the end product will be useful for other applications. To this end, some

  16. Generalized Sudan's List Decoding for Order Domain Codes

    DEFF Research Database (Denmark)

    Geil, Hans Olav; Matsumoto, Ryutaroh

    2007-01-01

    We generalize Sudan's list decoding algorithm without multiplicity to evaluation codes coming from arbitrary order domains. The number of correctable errors by the proposed method is larger than the original list decoding without multiplicity....

  17. On locality of Generalized Reed-Muller codes over the broadcast erasure channel

    KAUST Repository

    Alloum, Amira

    2016-07-28

    One to Many communications are expected to be among the killer applications for the currently discussed 5G standard. The usage of coding mechanisms is impacting broadcasting standard quality, as coding is involved at several levels of the stack, and more specifically at the application layer where Rateless, LDPC, Reed Slomon codes and network coding schemes have been extensively studied, optimized and standardized in the past. Beyond reusing, extending or adapting existing application layer packet coding mechanisms based on previous schemes and designed for the foregoing LTE or other broadcasting standards; our purpose is to investigate the use of Generalized Reed Muller codes and the value of their locality property in their progressive decoding for Broadcast/Multicast communication schemes with real time video delivery. Our results are meant to bring insight into the use of locally decodable codes in Broadcasting. © 2016 IEEE.

  18. How General-Purpose can a GPU be?

    Directory of Open Access Journals (Sweden)

    Philip Machanick

    2015-12-01

    Full Text Available The use of graphics processing units (GPUs in general-purpose computation (GPGPU is a growing field. GPU instruction sets, while implementing a graphics pipeline, draw from a range of single instruction multiple datastream (SIMD architectures characteristic of the heyday of supercomputers. Yet only one of these SIMD instruction sets has been of application on a wide enough range of problems to survive the era when the full range of supercomputer design variants was being explored: vector instructions. This paper proposes a reconceptualization of the GPU as a multicore design with minimal exotic modes of parallelism so as to make GPGPU truly general.

  19. Speed Control of General Purpose Engine with Electronic Governor

    Science.gov (United States)

    Sawut, Umerujan; Tohti, Gheyret; Takigawa, Buso; Tsuji, Teruo

    This paper presents a general purpose engine speed control system with an electronic governor in order to improve the current system with a mechanical governor which shows unstable characteristics by change of mecanical friction or A/F ratio (Air/Fuel ratio). For the control system above, there are problems that the feedback signal is only a crank angle because of cost and the controlled object is a general purpose engine which is strongly nonlinear. In order to overcome these problems, the system model is shown for the dynamic estimation of the amount of air flow and the robust controller is designed. That is, the proposed system includes the robust sliding-mode controller by the feedback signal of only a crank angle where Genetic Algorithm is applied for the controller design. The simulation and the experiments by MATLAB/Simulink are performed to show the effectiveness of our proposal.

  20. The ATLAS Trigger Algorithms for General Purpose Graphics Processor Units

    CERN Document Server

    Tavares Delgado, Ademar; The ATLAS collaboration

    2016-01-01

    The ATLAS Trigger Algorithms for General Purpose Graphics Processor Units Type: Talk Abstract: We present the ATLAS Trigger algorithms developed to exploit General­ Purpose Graphics Processor Units. ATLAS is a particle physics experiment located on the LHC collider at CERN. The ATLAS Trigger system has two levels, hardware-­based Level 1 and the High Level Trigger implemented in software running on a farm of commodity CPU. Performing the trigger event selection within the available farm resources presents a significant challenge that will increase future LHC upgrades. are being evaluated as a potential solution for trigger algorithms acceleration. Key factors determining the potential benefit of this new technology are the relative execution speedup, the number of GPUs required and the relative financial cost of the selected GPU. We have developed a trigger demonstrator which includes algorithms for reconstructing tracks in the Inner Detector and Muon Spectrometer and clusters of energy deposited in the Cal...

  1. Space shuttle general purpose computers (GPCs) (current and future versions)

    Science.gov (United States)

    1988-01-01

    Current and future versions of general purpose computers (GPCs) for space shuttle orbiters are represented in this frame. The two boxes on the left (AP101B) represent the current GPC configuration, with the input-output processor at far left and the central processing unit (CPU) at its side. The upgraded version combines both elements in a single unit (far right, AP101S).

  2. New Generation General Purpose Computer (GPC) compact IBM unit

    Science.gov (United States)

    1991-01-01

    New Generation General Purpose Computer (GPC) compact IBM unit replaces a two-unit earlier generation computer. The new IBM unit is documented in table top views alone (S91-26867, S91-26868), with the onboard equipment it supports including the flight deck CRT screen and keypad (S91-26866), and next to the two earlier versions it replaces (S91-26869).

  3. Survey of advanced general-purpose software for robot manipulators

    International Nuclear Information System (INIS)

    Latombe, J.C.

    1983-01-01

    Computer-controlled sensor-based robots will more and more common in industry. This paper attempts to survey the main trends of the development of advanced general-purpose software for robot manipulators. It is intended to make clear that robots are not only mechanical devices. They are truly programmable machines, and their programming, which occurs in an imperfectly modelled world,is somewhat different from conventional computer programming. (orig.)

  4. General Purpose Multimedia Dataset - GarageBand 2008

    DEFF Research Database (Denmark)

    Meng, Anders

    This document describes a general purpose multimedia data-set to be used in cross-media machine learning problems. In more detail we describe the genre taxonomy applied at http://www.garageband.com, from where the data-set was collected, and how the taxonomy have been fused into a more human...... understandable taxonomy. Finally, a description of various features extracted from both the audio and text are presented....

  5. ENDF/B-4 General Purpose File 1974

    International Nuclear Information System (INIS)

    Schwerer, O.

    1980-04-01

    This document summarizes contents and documentation of the 1974 version of the General Purpose File of the ENDF/B Library maintained by the National Nuclear Data Center (NNDC) at the Brookhaven National Laboratory, USA. The Library contains numerical neutron reaction data for 90 isotopes or elements. The entire Library or selective retrievals from it can be obtained on magnetic tape from the IAEA Nuclear Data Section. (author)

  6. Interfacing a General Purpose Fluid Network Flow Program with the SINDA/G Thermal Analysis Program

    Science.gov (United States)

    Schallhorn, Paul; Popok, Daniel

    1999-01-01

    A general purpose, one dimensional fluid flow code is currently being interfaced with the thermal analysis program Systems Improved Numerical Differencing Analyzer/Gaski (SINDA/G). The flow code, Generalized Fluid System Simulation Program (GFSSP), is capable of analyzing steady state and transient flow in a complex network. The flow code is capable of modeling several physical phenomena including compressibility effects, phase changes, body forces (such as gravity and centrifugal) and mixture thermodynamics for multiple species. The addition of GFSSP to SINDA/G provides a significant improvement in convective heat transfer modeling for SINDA/G. The interface development is conducted in multiple phases. This paper describes the first phase of the interface which allows for steady and quasi-steady (unsteady solid, steady fluid) conjugate heat transfer modeling.

  7. The KFA-Version of the high-energy transport code HETC and the generalized evaluation code SIMPEL

    International Nuclear Information System (INIS)

    Cloth, P.; Filges, D.; Sterzenbach, G.; Armstrong, T.W.; Colborn, B.L.

    1983-03-01

    This document describes the updates that have been made to the high-energy transport code HETC for use in the German spallation-neutron source project SNQ. Performance and purpose of the subsidiary code SIMPEL that has been written for general analysis of the HETC output are also described. In addition means of coupling to low energy transport programs, such as the Monte-Carlo code MORSE is provided. As complete input descriptions for HETC and SIMPEL are given together with a sample problem, this document can serve as a user's manual for these two codes. The document is also an answer to the demand that has been issued by a greater community of HETC users on the ICANS-IV meeting, Oct 20-24 1980, Tsukuba-gun, Japan for a complete description of at least one single version of HETC among the many different versions that exist. (orig.)

  8. Incremental and developmental perspectives for general-purpose learning systems

    Directory of Open Access Journals (Sweden)

    Fernando Martínez-Plumed

    2017-02-01

    Full Text Available The stupefying success of Articial Intelligence (AI for specic problems, from recommender systems to self-driving cars, has not yet been matched with a similar progress in general AI systems, coping with a variety of (dierent problems. This dissertation deals with the long-standing problem of creating more general AI systems, through the analysis of their development and the evaluation of their cognitive abilities. It presents a declarative general-purpose learning system and a developmental and lifelong approach for knowledge acquisition, consolidation and forgetting. It also analyses the use of the use of more ability-oriented evaluation techniques for AI evaluation and provides further insight for the understanding of the concepts of development and incremental learning in AI systems.

  9. An embedded domain specific language for general purpose vectorization

    CERN Document Server

    Karpinski, Przemyslaw

    2017-01-01

    Portable SIMD code generation is an open problem in modern High Performance Computing systems. Performance portability can already be achieved, however it might fail when user-framework interaction is required. Of all portable vectorization techniques, explicit vectorization, using wrapper-class libraries, is proven to achieve the fastest performance, however it does not exploit optimization opportunities outside the simplest algebraic primitives. A more advanced language is therefore required, but the design of a new independent language is not feasible due to its high costs. This work describes an Embedded Domain Specific Language for solving generalized 1-D vectorization problems. The language is implemented using C++ as a host language and published as a lightweight library. By decoupling expression creation from evaluation a wider range of problems can be solved, without sacrificing runtime efficiency. In this paper we discuss design patterns necessary, but not limited, to efficient EDSL implementatio...

  10. The architecture of Newton, a general-purpose dynamics simulator

    Science.gov (United States)

    Cremer, James F.; Stewart, A. James

    1989-01-01

    The architecture for Newton, a general-purpose system for simulating the dynamics of complex physical objects, is described. The system automatically formulates and analyzes equations of motion, and performs automatic modification of this system equations when necessitated by changes in kinematic relationships between objects. Impact and temporary contact are handled, although only using simple models. User-directed influence of simulations is achieved using Newton's module, which can be used to experiment with the control of many-degree-of-freedom articulated objects.

  11. General-purpose heat source development. Phase II: conceptual designs

    International Nuclear Information System (INIS)

    Snow, E.C.; Zocher, R.W.; Grinberg, I.M.; Hulbert, L.E.

    1978-11-01

    Basic geometric module shapes and fuel arrays were studied to determine how well they could be expected to meet the General Purpose Heat Source (GPHS) design requirements. Seven conceptual designs were selected, detailed drawings produced, and these seven concepts analyzed. Three of these design concepts were selected as GPHS Trial Designs to be reanalyzed in more detail and tested. The geometric studies leading to the selection of the seven conceptual designs, the analyses of these designs, and the selection of the three trial designs are discussed

  12. Installation of new Generation General Purpose Computer (GPC) compact unit

    Science.gov (United States)

    1991-01-01

    In the Kennedy Space Center's (KSC's) Orbiter Processing Facility (OPF) high bay 2, Spacecraft Electronics technician Ed Carter (right), wearing clean suit, prepares for (26864) and installs (26865) the new Generation General Purpose Computer (GPC) compact IBM unit in Atlantis', Orbiter Vehicle (OV) 104's, middeck avionics bay as Orbiter Systems Quality Control technician Doug Snider looks on. Both men work for NASA contractor Lockheed Space Operations Company. All three orbiters are being outfitted with the compact IBM unit, which replaces a two-unit earlier generation computer.

  13. Foam: A general purpose Monte Carlo cellular algorithm

    International Nuclear Information System (INIS)

    Jadach, S.

    2003-01-01

    A general-purpose, self-adapting Monte Carlo (MC) algorithm implemented in the program Foam is described. The high efficiency of the MC, that is small maximum weight or variance of the MC weight is achieved by means of dividing the integration domain into small cells. The cells can be n-dimensional simplices, hyperrectangles cells. The next cell to be divided and the position/direction of the division hyperplane is chosen by the algorithm which optimizes the ratio of the maximum weight to the average weight or (optionally) the total variance. The algorithm is able to deal, in principle, with an arbitrary pattern of the singularities in the distribution

  14. Using general-purpose compression algorithms for music analysis

    DEFF Research Database (Denmark)

    Louboutin, Corentin; Meredith, David

    2016-01-01

    General-purpose compression algorithms encode files as dictionaries of substrings with the positions of these strings’ occurrences. We hypothesized that such algorithms could be used for pattern discovery in music. We compared LZ77, LZ78, Burrows–Wheeler and COSIATEC on classifying folk song...... in the input data, COSIATEC outperformed LZ77 with a mean F1 score of 0.123, compared with 0.053 for LZ77. However, when the music was processed a voice at a time, the F1 score for LZ77 more than doubled to 0.124. We also discovered a significant correlation between compression factor and F1 score for all...

  15. Using the general-purpose reactivity indicator: challenging examples.

    Science.gov (United States)

    Anderson, James S M; Melin, Junia; Ayers, Paul W

    2016-03-01

    We elucidate the regioselectivity of nucleophilic attack on substituted benzenesulfonates, quinolines, and pyridines using a general-purpose reactivity indicator (GPRI) for electrophiles. We observe that the GPRI is most accurate when the incoming nucleophile resembles a point charge. We further observe that the GPRI often chooses reactive "dead ends" as the most reactive sites as well as sterically hindered reactive sites. This means that care must be taken to remove sites that are inherently unreactive. Generally, among sites where reactions actually occur, the GPRI identifies the sites in the molecule that lead to the kinetically favored product(s). Furthermore, the GPRI can discern which sites react with hard reagents and which sites react with soft reagents. Because it is currently impossible to use the mathematical framework of conceptual DFT to identify sterically inaccessible sites and reactive dead ends, the GPRI is primarily useful as an interpretative, not a predictive, tool.

  16. Generalized optical code construction for enhanced and Modified Double Weight like codes without mapping for SAC-OCDMA systems

    Science.gov (United States)

    Kumawat, Soma; Ravi Kumar, M.

    2016-07-01

    Double Weight (DW) code family is one of the coding schemes proposed for Spectral Amplitude Coding-Optical Code Division Multiple Access (SAC-OCDMA) systems. Modified Double Weight (MDW) code for even weights and Enhanced Double Weight (EDW) code for odd weights are two algorithms extending the use of DW code for SAC-OCDMA systems. The above mentioned codes use mapping technique to provide codes for higher number of users. A new generalized algorithm to construct EDW and MDW like codes without mapping for any weight greater than 2 is proposed. A single code construction algorithm gives same length increment, Bit Error Rate (BER) calculation and other properties for all weights greater than 2. Algorithm first constructs a generalized basic matrix which is repeated in a different way to produce the codes for all users (different from mapping). The generalized code is analysed for BER using balanced detection and direct detection techniques.

  17. High vacuum general purpose scattering chamber for nuclear reaction study

    International Nuclear Information System (INIS)

    Suresh Kumar; Ojha, S.C.

    2003-01-01

    To study the nuclear reactions induced by beam from medium energy accelerators, one of the most common facility required is a scattering chamber. In the scattering chamber, projectile collides with the target nucleus and the scattered reaction products are detected with various type of nuclear detector at different angles with respect to the beam. The experiments are performed under high vacuum to minimize the background reaction and the energy losses of the charged particles. To make the chamber general purpose various requirement of the experiments are incorporated into it. Changing of targets, changing angle of various detectors while in vacuum are the most desired features. The other features like ascertaining the beam spot size and position on the target, minimizing the background counts by proper beam dump, accurate positioning of the detector as per plan etc. are some of the important requirements

  18. General-purpose event generators for LHC physics

    CERN Document Server

    Buckley, Andy; Gieseke, Stefan; Grellscheid, David; Hoche, Stefan; Hoeth, Hendrik; Krauss, Frank; Lonnblad, Leif; Nurse, Emily; Richardson, Peter; Schumann, Steffen; Seymour, Michael H.; Sjostrand, Torbjorn; Skands, Peter; Webber, Bryan

    2011-01-01

    We review the physics basis, main features and use of general-purpose Monte Carlo event generators for the simulation of proton-proton collisions at the Large Hadron Collider. Topics included are: the generation of hard-scattering matrix elements for processes of interest, at both leading and next-to-leading QCD perturbative order; their matching to approximate treatments of higher orders based on the showering approximation; the parton and dipole shower formulations; parton distribution functions for event generators; non-perturbative aspects such as soft QCD collisions, the underlying event and diffractive processes; the string and cluster models for hadron formation; the treatment of hadron and tau decays; the inclusion of QED radiation and beyond-Standard-Model processes. We describe the principal features of the ARIADNE, Herwig++, PYTHIA 8 and SHERPA generators, together with the Rivet and Professor validation and tuning tools, and discuss the physics philosophy behind the proper use of these generators ...

  19. General purpose heat source task group. Final report

    International Nuclear Information System (INIS)

    1979-01-01

    The results of thermal analyses and impact tests on a modified design of a 238 Pu-fueled general purpose heat source (GPHS) for spacecraft power supplies are presented. This work was performed to establish the safety of a heat source with pyrolytic graphite insulator shells located either inside or outside the graphite impact shell. This safety is dependent on the degree of aerodynamic heating of the heat source during reentry and on the ability of the heat source capsule to withstand impact after reentry. Analysis of wind tunnel and impact test data result in a recommended GPHS design which should meet all temperature and safety requirements. Further wind tunnel tests, drop tests, and impact tests are recommended to verify the safety of this design

  20. General-purpose parallel simulator for quantum computing

    International Nuclear Information System (INIS)

    Niwa, Jumpei; Matsumoto, Keiji; Imai, Hiroshi

    2002-01-01

    With current technologies, it seems to be very difficult to implement quantum computers with many qubits. It is therefore of importance to simulate quantum algorithms and circuits on the existing computers. However, for a large-size problem, the simulation often requires more computational power than is available from sequential processing. Therefore, simulation methods for parallel processors are required. We have developed a general-purpose simulator for quantum algorithms/circuits on the parallel computer (Sun Enterprise4500). It can simulate algorithms/circuits with up to 30 qubits. In order to test efficiency of our proposed methods, we have simulated Shor's factorization algorithm and Grover's database search, and we have analyzed robustness of the corresponding quantum circuits in the presence of both decoherence and operational errors. The corresponding results, statistics, and analyses are presented in this paper

  1. The RHIC general purpose multiplexed analog to digital converter system

    International Nuclear Information System (INIS)

    Michnoff, R.

    1995-01-01

    A general purpose multiplexed analog to digital converter system is currently under development to support acquisition of analog signals for the Relativistic Heavy Ion Collider (RHIC) at Brookhaven National Laboratory. The system consists of a custom intelligent VME based controller module (V113) and a 14-bit 64 channel multiplexed A/D converter module (V114). The design features two independent scan groups, where one scan group is capable of acquiring 64 channels at 60 Hz, concurrently with the second scan group acquiring data at an aggregate rate of up to 80 k samples/second. An interface to the RHIC serially encoded event line is used to synchronize acquisition. Data is stored in a circular static RAM buffer on the controller module, then transferred to a commercial VMEbus CPU board and higher level workstations for plotting, report Generation, analysis and storage

  2. Using a cognitive architecture for general purpose service robot control

    Science.gov (United States)

    Puigbo, Jordi-Ysard; Pumarola, Albert; Angulo, Cecilio; Tellez, Ricardo

    2015-04-01

    A humanoid service robot equipped with a set of simple action skills including navigating, grasping, recognising objects or people, among others, is considered in this paper. By using those skills the robot should complete a voice command expressed in natural language encoding a complex task (defined as the concatenation of a number of those basic skills). As a main feature, no traditional planner has been used to decide skills to be activated, as well as in which sequence. Instead, the SOAR cognitive architecture acts as the reasoner by selecting which action the robot should complete, addressing it towards the goal. Our proposal allows to include new goals for the robot just by adding new skills (without the need to encode new plans). The proposed architecture has been tested on a human-sized humanoid robot, REEM, acting as a general purpose service robot.

  3. Foam: A general purpose Monte Carlo cellular algorithm

    International Nuclear Information System (INIS)

    Jadach, S.

    2002-01-01

    A general-purpose, self-adapting Monte Carlo (MC) algorithm implemented in the program Foam is described. The high efficiency of the MC, that is small maximum weight or variance of the MC weight is achieved by means of dividing the integration domain into small cells. The cells can be n-dimensional simplices, hyperrectangles or a Cartesian product of them. The grid of cells, called 'foam', is produced in the process of the binary split of the cells. The choice of the next cell to be divided and the position/direction of the division hyperplane is driven by the algorithm which optimizes the ratio of the maximum weight to the average weight or (optionally) the total variance. The algorithm is able to deal, in principle, with an arbitrary pattern of the singularities in the distribution. (author)

  4. Foam A General purpose Monte Carlo Cellular Algorithm

    CERN Document Server

    Jadach, Stanislaw

    2002-01-01

    A general-purpose, self-adapting Monte Carlo (MC) algorithm implemented in the program {\\tt Foam} is described. The high efficiency of the MC, that is small maximum weight or variance of the MC weight is achieved by means of dividing the integration domain into small cells. The cells can be $n$-dimensional simplices, hyperrectangles or a Cartesian product of them. The grid of cells, ``foam'', is produced in the process of the binary split of the cells. The next cell to be divided and the position/direction of the division hyperplane is chosen by the algorithm which optimizes the ratio of the maximum weight to the average weight or (optionally) the total variance. The algorithm is able to deal, in principle, with an arbitrary pattern of the singularities in the distribution.

  5. General-purpose software for science technology calculation

    International Nuclear Information System (INIS)

    Aikawa, Hiroshi

    1999-01-01

    We have developed many general-purpose softwares for parallel processing of science technology calculation. This paper reported six softwares such as STA (Seamless Thinking Aid) basic soft, parallel numerical computation library, grid formation software for parallel computer, real-time visualizing system, parallel benchmark test system and object-oriented parallel programing method. STA is a user interface software to perform a total environment for parallel programing, a network computing environment for various parallel computers and a desktop computing environment via Web. Some examples using the above softwares are explained. One of them is a simultaneous parallel calculation of both analysis of flow and structure of supersonic transport to design of them. The other is various kinds of computer parallel calculations for nuclear fusion reaction such as a molecular dynamic calculation and a calculation of reactor structure and fluid. These softs are opened to the public by the home page {http://guide.tokai.jaeri.go.jp/ccse/}. (S.Y.)

  6. Selecting a general-purpose data compression algorithm

    Science.gov (United States)

    Mathews, Gary Jason

    1995-01-01

    The National Space Science Data Center's Common Data Formate (CDF) is capable of storing many types of data such as scalar data items, vectors, and multidimensional arrays of bytes, integers, or floating point values. However, regardless of the dimensionality and data type, the data break down into a sequence of bytes that can be fed into a data compression function to reduce the amount of data without losing data integrity and thus remaining fully reconstructible. Because of the diversity of data types and high performance speed requirements, a general-purpose, fast, simple data compression algorithm is required to incorporate data compression into CDF. The questions to ask are how to evaluate and compare compression algorithms, and what compression algorithm meets all requirements. The object of this paper is to address these questions and determine the most appropriate compression algorithm to use within the CDF data management package that would be applicable to other software packages with similar data compression needs.

  7. General Purpose Technologies and their Implications for International Trade

    Directory of Open Access Journals (Sweden)

    Petsas Iordanis

    2015-09-01

    Full Text Available This paper develops a simple model of trade and “quality-ladders” growth without scale effects to study the implications of general purpose technologies (GPTs for international trade. GPTs refer to a certain type of drastic innovations, such as electrification, the transistor, and the Internet, that are characterized by the pervasiveness in use, innovational complementarities, and technological dynamism. The model presents a two-country (Home and Foreign dynamic general equilibrium framework and incorporates GPT diffusion within Home that exhibits endogenous Schumpeterian growth. The model analyzes the long-run and transitional dynamic effects of a new GPT on the pattern of trade and relative wages. The main findings of the paper are: 1 when the GPT diffusion across industries is governed by S-curve dynamics, there are two steady-state equilibria: the initial steadystate arises before the adoption of the new GPT and the final one is reached after the GPT diffusion process has been completed, 2 when all industries at Home have adopted the new GPT, Home enjoys comparative advantage in a greater range of industries compared to Foreign, 3 during the transitional dynamics, Foreign gains back its competitiveness in some of the industries that lost its comparative advantage to Home.

  8. Generalized instantly decodable network coding for relay-assisted networks

    KAUST Repository

    Elmahdy, Adel M.

    2013-09-01

    In this paper, we investigate the problem of minimizing the frame completion delay for Instantly Decodable Network Coding (IDNC) in relay-assisted wireless multicast networks. We first propose a packet recovery algorithm in the single relay topology which employs generalized IDNC instead of strict IDNC previously proposed in the literature for the same relay-assisted topology. This use of generalized IDNC is supported by showing that it is a super-set of the strict IDNC scheme, and thus can generate coding combinations that are at least as efficient as strict IDNC in reducing the average completion delay. We then extend our study to the multiple relay topology and propose a joint generalized IDNC and relay selection algorithm. This proposed algorithm benefits from the reception diversity of the multiple relays to further reduce the average completion delay in the network. Simulation results show that our proposed solutions achieve much better performance compared to previous solutions in the literature. © 2013 IEEE.

  9. standalone general purpose data logger design and implementation

    African Journals Online (AJOL)

    eobe

    volatile EEPROM data memory, four AT24C256 2-wire serial EEPROM chips were used for data storage. wire serial ... analog electrical signal that relays information about .... in-code, ADC, Serial Communication and short term memory ...

  10. Use of general purpose graphics processing units with MODFLOW

    Science.gov (United States)

    Hughes, Joseph D.; White, Jeremy T.

    2013-01-01

    To evaluate the use of general-purpose graphics processing units (GPGPUs) to improve the performance of MODFLOW, an unstructured preconditioned conjugate gradient (UPCG) solver has been developed. The UPCG solver uses a compressed sparse row storage scheme and includes Jacobi, zero fill-in incomplete, and modified-incomplete lower-upper (LU) factorization, and generalized least-squares polynomial preconditioners. The UPCG solver also includes options for sequential and parallel solution on the central processing unit (CPU) using OpenMP. For simulations utilizing the GPGPU, all basic linear algebra operations are performed on the GPGPU; memory copies between the central processing unit CPU and GPCPU occur prior to the first iteration of the UPCG solver and after satisfying head and flow criteria or exceeding a maximum number of iterations. The efficiency of the UPCG solver for GPGPU and CPU solutions is benchmarked using simulations of a synthetic, heterogeneous unconfined aquifer with tens of thousands to millions of active grid cells. Testing indicates GPGPU speedups on the order of 2 to 8, relative to the standard MODFLOW preconditioned conjugate gradient (PCG) solver, can be achieved when (1) memory copies between the CPU and GPGPU are optimized, (2) the percentage of time performing memory copies between the CPU and GPGPU is small relative to the calculation time, (3) high-performance GPGPU cards are utilized, and (4) CPU-GPGPU combinations are used to execute sequential operations that are difficult to parallelize. Furthermore, UPCG solver testing indicates GPGPU speedups exceed parallel CPU speedups achieved using OpenMP on multicore CPUs for preconditioners that can be easily parallelized.

  11. The purpose of the general practice consultation from the patients perspective - theoretical aspects

    DEFF Research Database (Denmark)

    Thorsen, Hanne; Witt, Klaus; Malterud, Kirsti

    2001-01-01

    Consultation purposes, general practice, patients´expectations, patients satosfaction, patientcenteredness......Consultation purposes, general practice, patients´expectations, patients satosfaction, patientcenteredness...

  12. General-purpose event generators for LHC physics

    Energy Technology Data Exchange (ETDEWEB)

    Buckley, Andy [PPE Group, School of Physics and Astronomy, University of Edinburgh, EH25 9PN (United Kingdom); Butterworth, Jonathan [Department of Physics and Astronomy, University College London, WC1E 6BT (United Kingdom); Gieseke, Stefan [Institute for Theoretical Physics, Karlsruhe Institute of Technology, D-76128 Karlsruhe (Germany); Grellscheid, David [Institute for Particle Physics Phenomenology, Durham University, DH1 3LE (United Kingdom); Hoeche, Stefan [SLAC National Accelerator Laboratory, Menlo Park, CA 94025 (United States); Hoeth, Hendrik; Krauss, Frank [Institute for Particle Physics Phenomenology, Durham University, DH1 3LE (United Kingdom); Loennblad, Leif [Department of Astronomy and Theoretical Physics, Lund University (Sweden); PH Department, TH Unit, CERN, CH-1211 Geneva 23 (Switzerland); Nurse, Emily [Department of Physics and Astronomy, University College London, WC1E 6BT (United Kingdom); Richardson, Peter [Institute for Particle Physics Phenomenology, Durham University, DH1 3LE (United Kingdom); Schumann, Steffen [Institute for Theoretical Physics, University of Heidelberg, 69120 Heidelberg (Germany); Seymour, Michael H. [School of Physics and Astronomy, University of Manchester, M13 9PL (United Kingdom); Sjoestrand, Torbjoern [Department of Astronomy and Theoretical Physics, Lund University (Sweden); Skands, Peter [PH Department, TH Unit, CERN, CH-1211 Geneva 23 (Switzerland); Webber, Bryan, E-mail: webber@hep.phy.cam.ac.uk [Cavendish Laboratory, J.J. Thomson Avenue, Cambridge CB3 0HE (United Kingdom)

    2011-07-15

    We review the physics basis, main features and use of general-purpose Monte Carlo event generators for the simulation of proton-proton collisions at the Large Hadron Collider. Topics included are: the generation of hard scattering matrix elements for processes of interest, at both leading and next-to-leading QCD perturbative order; their matching to approximate treatments of higher orders based on the showering approximation; the parton and dipole shower formulations; parton distribution functions for event generators; non-perturbative aspects such as soft QCD collisions, the underlying event and diffractive processes; the string and cluster models for hadron formation; the treatment of hadron and tau decays; the inclusion of QED radiation and beyond Standard Model processes. We describe the principal features of the ARIADNE, Herwig++, PYTHIA 8 and SHERPA generators, together with the Rivet and Professor validation and tuning tools, and discuss the physics philosophy behind the proper use of these generators and tools. This review is aimed at phenomenologists wishing to understand better how parton-level predictions are translated into hadron-level events as well as experimentalists seeking a deeper insight into the tools available for signal and background simulation at the LHC.

  13. General-purpose event generators for LHC physics

    International Nuclear Information System (INIS)

    Buckley, Andy; Butterworth, Jonathan; Gieseke, Stefan; Grellscheid, David; Hoeche, Stefan; Hoeth, Hendrik; Krauss, Frank; Loennblad, Leif; Nurse, Emily; Richardson, Peter; Schumann, Steffen; Seymour, Michael H.; Sjoestrand, Torbjoern; Skands, Peter; Webber, Bryan

    2011-01-01

    We review the physics basis, main features and use of general-purpose Monte Carlo event generators for the simulation of proton-proton collisions at the Large Hadron Collider. Topics included are: the generation of hard scattering matrix elements for processes of interest, at both leading and next-to-leading QCD perturbative order; their matching to approximate treatments of higher orders based on the showering approximation; the parton and dipole shower formulations; parton distribution functions for event generators; non-perturbative aspects such as soft QCD collisions, the underlying event and diffractive processes; the string and cluster models for hadron formation; the treatment of hadron and tau decays; the inclusion of QED radiation and beyond Standard Model processes. We describe the principal features of the ARIADNE, Herwig++, PYTHIA 8 and SHERPA generators, together with the Rivet and Professor validation and tuning tools, and discuss the physics philosophy behind the proper use of these generators and tools. This review is aimed at phenomenologists wishing to understand better how parton-level predictions are translated into hadron-level events as well as experimentalists seeking a deeper insight into the tools available for signal and background simulation at the LHC.

  14. A VMEbus general-purpose data acquisition system

    International Nuclear Information System (INIS)

    Ninane, A.; Nemry, M.; Martou, J.L.; Somers, F.

    1992-01-01

    We present a general-purpose, VMEbus based, multiprocessor data acquisition and monitoring system. Events, handled by a master CPU, are kept at the disposal of data storage and monitoring processes which can run on distinct processors. They access either the complete set of data or a fraction of them, minimizing the acquisition dead-time. The system is built with the VxWorks 5.0 real time kernel to which we have added device drivers for data acquisition and monitoring. The acquisition is controlled and the data are displayed on a workstation. The user interface is written in C ++ and re-uses the classes of the Interviews and the NIH libraries. The communication between the control workstation and the VMEbus processors is made through SUN RPCs on an Ethernet link. The system will be used for, CAMAC based, data acquisition for nuclear physics experiments as well as for the VXI data taking with the 4π configuration (100 neutron detectors) of the Brussels-Caen-Louvian-Strasbourg DEMON collaboration. (author)

  15. High-Speed General Purpose Genetic Algorithm Processor.

    Science.gov (United States)

    Hoseini Alinodehi, Seyed Pourya; Moshfe, Sajjad; Saber Zaeimian, Masoumeh; Khoei, Abdollah; Hadidi, Khairollah

    2016-07-01

    In this paper, an ultrafast steady-state genetic algorithm processor (GAP) is presented. Due to the heavy computational load of genetic algorithms (GAs), they usually take a long time to find optimum solutions. Hardware implementation is a significant approach to overcome the problem by speeding up the GAs procedure. Hence, we designed a digital CMOS implementation of GA in [Formula: see text] process. The proposed processor is not bounded to a specific application. Indeed, it is a general-purpose processor, which is capable of performing optimization in any possible application. Utilizing speed-boosting techniques, such as pipeline scheme, parallel coarse-grained processing, parallel fitness computation, parallel selection of parents, dual-population scheme, and support for pipelined fitness computation, the proposed processor significantly reduces the processing time. Furthermore, by relying on a built-in discard operator the proposed hardware may be used in constrained problems that are very common in control applications. In the proposed design, a large search space is achievable through the bit string length extension of individuals in the genetic population by connecting the 32-bit GAPs. In addition, the proposed processor supports parallel processing, in which the GAs procedure can be run on several connected processors simultaneously.

  16. Interfacial Properties of EXXPRO(TM) and General Purpose Elastomers

    Science.gov (United States)

    Zhang, Y.; Rafailovich, M.; Sokolov, Jon; Qu, S.; Ge, S.; Ngyuen, D.; Li, Z.; Peiffer, D.; Song, L.; Dias, J. A.; McElrath, K. O.

    1998-03-01

    EXXPRO(Trademark) elastomers are used for tires and many other applications. This elastomer (denoted as BIMS) is a random copolymer of p-methylstyrene (MS) and polyisobutylene (I) with varying degrees of PMS content and bromination (B) on the p-methyl group. BIMS is impermeable to gases, and has good heat, ozone and flex resistance. Very often general purpose elastomers are blended with BIMS. The interfacial width between polybutadiene and BIMS is a sensitive function of the Br level and PMS content. By neutron reflectivity (NR), we studied the dynamics of interface formation as a function of time and temperature for BIMS with varying degrees of PMS and Br. We found that in addition to the bulk parameters, the total film thickness and the proximity of an interactive surface can affect the interfacial interaction rates. The interfacial properties can also be modified by inclusion of particles, such as carbon black (a filler component in tire rubbers). Results will be presented on the relation between the interfacial width as measured by NR and compatibilization studies via AFM and LFM.

  17. General-purpose event generators for LHC physics

    Energy Technology Data Exchange (ETDEWEB)

    Buckley, Andy; /Edinburgh U.; Butterworth, Jonathan; /University Coll. London; Gieseke, Stefan; /Karlsruhe U., ITP; Grellscheid, David; /Durham U., IPPP; Hoche, Stefan; /SLAC; Hoeth, Hendrik; Krauss, Frank; /Durham U., IPPP; Lonnblad, Leif; /Lund U., Dept. Theor. Phys. /CERN; Nurse, Emily; /University Coll. London; Richardson, Peter; /Durham U., IPPP; Schumann, Steffen; /Heidelberg U.; Seymour, Michael H.; /Manchester U.; Sjostrand, Torbjorn; /Lund U., Dept. Theor. Phys.; Skands, Peter; /CERN; Webber, Bryan; /Cambridge U.

    2011-03-03

    We review the physics basis, main features and use of general-purpose Monte Carlo event generators for the simulation of proton-proton collisions at the Large Hadron Collider. Topics included are: the generation of hard-scattering matrix elements for processes of interest, at both leading and next-to-leading QCD perturbative order; their matching to approximate treatments of higher orders based on the showering approximation; the parton and dipole shower formulations; parton distribution functions for event generators; non-perturbative aspects such as soft QCD collisions, the underlying event and diffractive processes; the string and cluster models for hadron formation; the treatment of hadron and tau decays; the inclusion of QED radiation and beyond-Standard-Model processes. We describe the principal features of the Ariadne, Herwig++, Pythia 8 and Sherpa generators, together with the Rivet and Professor validation and tuning tools, and discuss the physics philosophy behind the proper use of these generators and tools. This review is aimed at phenomenologists wishing to understand better how parton-level predictions are translated into hadron-level events as well as experimentalists wanting a deeper insight into the tools available for signal and background simulation at the LHC.

  18. Foam A General Purpose Cellular Monte Carlo Event Generator

    CERN Document Server

    Jadach, Stanislaw

    2003-01-01

    A general purpose, self-adapting, Monte Carlo (MC) event generator (simulator) is described. The high efficiency of the MC, that is small maximum weight or variance of the MC weight is achieved by means of dividing the integration domain into small cells. The cells can be $n$-dimensional simplices, hyperrectangles or Cartesian product of them. The grid of cells, called ``foam'', is produced in the process of the binary split of the cells. The choice of the next cell to be divided and the position/direction of the division hyper-plane is driven by the algorithm which optimizes the ratio of the maximum weight to the average weight or (optionally) the total variance. The algorithm is able to deal, in principle, with an arbitrary pattern of the singularities in the distribution. As any MC generator, it can also be used for the MC integration. With the typical personal computer CPU, the program is able to perform adaptive integration/simulation at relatively small number of dimensions ($\\leq 16$). With the continu...

  19. Generalized Fluid System Simulation Program (GFSSP) Version 6 - General Purpose Thermo-Fluid Network Analysis Software

    Science.gov (United States)

    Majumdar, Alok; Leclair, Andre; Moore, Ric; Schallhorn, Paul

    2011-01-01

    GFSSP stands for Generalized Fluid System Simulation Program. It is a general-purpose computer program to compute pressure, temperature and flow distribution in a flow network. GFSSP calculates pressure, temperature, and concentrations at nodes and calculates flow rates through branches. It was primarily developed to analyze Internal Flow Analysis of a Turbopump Transient Flow Analysis of a Propulsion System. GFSSP development started in 1994 with an objective to provide a generalized and easy to use flow analysis tool for thermo-fluid systems.

  20. General purpose graphics-processing-unit implementation of cosmological domain wall network evolution.

    Science.gov (United States)

    Correia, J R C C C; Martins, C J A P

    2017-10-01

    Topological defects unavoidably form at symmetry breaking phase transitions in the early universe. To probe the parameter space of theoretical models and set tighter experimental constraints (exploiting the recent advances in astrophysical observations), one requires more and more demanding simulations, and therefore more hardware resources and computation time. Improving the speed and efficiency of existing codes is essential. Here we present a general purpose graphics-processing-unit implementation of the canonical Press-Ryden-Spergel algorithm for the evolution of cosmological domain wall networks. This is ported to the Open Computing Language standard, and as a consequence significant speedups are achieved both in two-dimensional (2D) and 3D simulations.

  1. General purpose graphics-processing-unit implementation of cosmological domain wall network evolution

    Science.gov (United States)

    Correia, J. R. C. C. C.; Martins, C. J. A. P.

    2017-10-01

    Topological defects unavoidably form at symmetry breaking phase transitions in the early universe. To probe the parameter space of theoretical models and set tighter experimental constraints (exploiting the recent advances in astrophysical observations), one requires more and more demanding simulations, and therefore more hardware resources and computation time. Improving the speed and efficiency of existing codes is essential. Here we present a general purpose graphics-processing-unit implementation of the canonical Press-Ryden-Spergel algorithm for the evolution of cosmological domain wall networks. This is ported to the Open Computing Language standard, and as a consequence significant speedups are achieved both in two-dimensional (2D) and 3D simulations.

  2. SNAP: A General Purpose Network Analysis and Graph Mining Library.

    Science.gov (United States)

    Leskovec, Jure; Sosič, Rok

    2016-10-01

    Large networks are becoming a widely used abstraction for studying complex systems in a broad set of disciplines, ranging from social network analysis to molecular biology and neuroscience. Despite an increasing need to analyze and manipulate large networks, only a limited number of tools are available for this task. Here, we describe Stanford Network Analysis Platform (SNAP), a general-purpose, high-performance system that provides easy to use, high-level operations for analysis and manipulation of large networks. We present SNAP functionality, describe its implementational details, and give performance benchmarks. SNAP has been developed for single big-memory machines and it balances the trade-off between maximum performance, compact in-memory graph representation, and the ability to handle dynamic graphs where nodes and edges are being added or removed over time. SNAP can process massive networks with hundreds of millions of nodes and billions of edges. SNAP offers over 140 different graph algorithms that can efficiently manipulate large graphs, calculate structural properties, generate regular and random graphs, and handle attributes and meta-data on nodes and edges. Besides being able to handle large graphs, an additional strength of SNAP is that networks and their attributes are fully dynamic, they can be modified during the computation at low cost. SNAP is provided as an open source library in C++ as well as a module in Python. We also describe the Stanford Large Network Dataset, a set of social and information real-world networks and datasets, which we make publicly available. The collection is a complementary resource to our SNAP software and is widely used for development and benchmarking of graph analytics algorithms.

  3. MCNP: a general Monte Carlo code for neutron and photon transport. Version 3A. Revision 2

    International Nuclear Information System (INIS)

    Briesmeister, J.F.

    1986-09-01

    This manual is a practical guide for the use of our general-purpose Monte Carlo code MCNP. The first chapter is a primer for the novice user. The second chapter describes the mathematics, data, physics, and Monte Carlo simulation found in MCNP. This discussion is not meant to be exhaustive - details of the particular techniques and of the Monte Carlo method itself will have to be found elsewhere. The third chapter shows the user how to prepare input for the code. The fourth chapter contains several examples, and the fifth chapter explains the output. The appendices show how to use MCNP on particular computer systems at the Los Alamos National Laboratory and also give details about some of the code internals that those who wish to modify the code may find useful. 57 refs

  4. 21 CFR 880.6890 - General purpose disinfectants.

    Science.gov (United States)

    2010-04-01

    ... (CONTINUED) MEDICAL DEVICES GENERAL HOSPITAL AND PERSONAL USE DEVICES General Hospital and Personal Use... disinfectant is a germicide intended to process noncritical medical devices and equipment surfaces. A general... prior to terminal sterilization or high level disinfection. Noncritical medical devices make only...

  5. SPIDR, a general-purpose readout system for pixel ASICs

    International Nuclear Information System (INIS)

    Heijden, B. van der; Visser, J.; Beuzekom, M. van; Boterenbrood, H.; Munneke, B.; Schreuder, F.; Kulis, S.

    2017-01-01

    The SPIDR (Speedy PIxel Detector Readout) system is a flexible general-purpose readout platform that can be easily adapted to test and characterize new and existing detector readout ASICs. It is originally designed for the readout of pixel ASICs from the Medipix/Timepix family, but other types of ASICs or front-end circuits can be read out as well. The SPIDR system consists of an FPGA board with memory and various communication interfaces, FPGA firmware, CPU subsystem and an API library on the PC . The FPGA firmware can be adapted to read out other ASICs by re-using IP blocks. The available IP blocks include a UDP packet builder, 1 and 10 Gigabit Ethernet MAC's and a 'soft core' CPU . Currently the firmware is targeted at the Xilinx VC707 development board and at a custom board called Compact-SPIDR . The firmware can easily be ported to other Xilinx 7 series and ultra scale FPGAs. The gap between an ASIC and the data acquisition back-end is bridged by the SPIDR system. Using the high pin count VITA 57 FPGA Mezzanine Card (FMC) connector only a simple chip carrier PCB is required. A 1 and a 10 Gigabit Ethernet interface handle the connection to the back-end. These can be used simultaneously for high-speed data and configuration over separate channels. In addition to the FMC connector, configurable inputs and outputs are available for synchronization with other detectors. A high resolution (≈ 27 ps bin size) Time to Digital converter is provided for time stamping events in the detector. The SPIDR system is frequently used as readout for the Medipix3 and Timepix3 ASICs. Using the 10 Gigabit Ethernet interface it is possible to read out a single chip at full bandwidth or up to 12 chips at a reduced rate. Another recent application is the test-bed for the VeloPix ASIC, which is developed for the Vertex Detector of the LHCb experiment. In this case the SPIDR system processes the 20 Gbps scrambled data stream from the VeloPix and distributes it over four

  6. A damage mechanics based general purpose interface/contact element

    Science.gov (United States)

    Yan, Chengyong

    laboratory test data presented in the literature. The results demonstrate that the proposed element and the damage law perform very well. The most important scientific contribution of this dissertation is the proposed damage criterion based on second law of thermodynamic and entropy of the system. The proposed general purpose interface/contact element is another contribution of this research. Compared to the previous adhoc interface elements proposed in the literature, the new one is, much more powerful and includes creep, plastic deformations, sliding, temperature, damage, cyclic behavior and fatigue life in a unified formulation.

  7. CLOUDCLOUD : general-purpose instrument monitoring and data managing software

    Science.gov (United States)

    Dias, António; Amorim, António; Tomé, António

    2016-04-01

    An effective experiment is dependent on the ability to store and deliver data and information to all participant parties regardless of their degree of involvement in the specific parts that make the experiment a whole. Having fast, efficient and ubiquitous access to data will increase visibility and discussion, such that the outcome will have already been reviewed several times, strengthening the conclusions. The CLOUD project aims at providing users with a general purpose data acquisition, management and instrument monitoring platform that is fast, easy to use, lightweight and accessible to all participants of an experiment. This work is now implemented in the CLOUD experiment at CERN and will be fully integrated with the experiment as of 2016. Despite being used in an experiment of the scale of CLOUD, this software can also be used in any size of experiment or monitoring station, from single computers to large networks of computers to monitor any sort of instrument output without influencing the individual instrument's DAQ. Instrument data and meta data is stored and accessed via a specially designed database architecture and any type of instrument output is accepted using our continuously growing parsing application. Multiple databases can be used to separate different data taking periods or a single database can be used if for instance an experiment is continuous. A simple web-based application gives the user total control over the monitored instruments and their data, allowing data visualization and download, upload of processed data and the ability to edit existing instruments or add new instruments to the experiment. When in a network, new computers are immediately recognized and added to the system and are able to monitor instruments connected to them. Automatic computer integration is achieved by a locally running python-based parsing agent that communicates with a main server application guaranteeing that all instruments assigned to that computer are

  8. WORM: A general-purpose input deck specification language

    International Nuclear Information System (INIS)

    Jones, T.

    1999-01-01

    Using computer codes to perform criticality safety calculations has become common practice in the industry. The vast majority of these codes use simple text-based input decks to represent the geometry, materials, and other parameters that describe the problem. However, the data specified in input files are usually processed results themselves. For example, input decks tend to require the geometry specification in linear dimensions and materials in atom or weight fractions, while the parameter of interest might be mass or concentration. The calculations needed to convert from the item of interest to the required parameter in the input deck are usually performed separately and then incorporated into the input deck. This process of calculating, editing, and renaming files to perform a simple parameter study is tedious at best. In addition, most computer codes require dimensions to be specified in centimeters, while drawings or other materials used to create the input decks might be in other units. This also requires additional calculation or conversion prior to composition of the input deck. These additional calculations, while extremely simple, introduce a source for error in both the calculations and transcriptions. To overcome these difficulties, WORM (Write One, Run Many) was created. It is an easy-to-use programming language to describe input decks and can be used with any computer code that uses standard text files for input. WORM is available, via the Internet, at worm.lanl.gov. A user's guide, tutorials, example models, and other WORM-related materials are also available at this Web site. Questions regarding WORM should be directed to wormatlanl.gov

  9. Linear-time general decoding algorithm for the surface code

    Science.gov (United States)

    Darmawan, Andrew S.; Poulin, David

    2018-05-01

    A quantum error correcting protocol can be substantially improved by taking into account features of the physical noise process. We present an efficient decoder for the surface code which can account for general noise features, including coherences and correlations. We demonstrate that the decoder significantly outperforms the conventional matching algorithm on a variety of noise models, including non-Pauli noise and spatially correlated noise. The algorithm is based on an approximate calculation of the logical channel using a tensor-network description of the noisy state.

  10. 7 CFR 249.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... Agriculture Regulations of the Department of Agriculture (Continued) FOOD AND NUTRITION SERVICE, DEPARTMENT OF AGRICULTURE CHILD NUTRITION PROGRAMS SENIOR FARMERS' MARKET NUTRITION PROGRAM (SFMNP) General § 249.1 General.... 2011, et seq.), and to any other Federal or State food or nutrition assistance program under which...

  11. On A Nonlinear Generalization of Sparse Coding and Dictionary Learning.

    Science.gov (United States)

    Xie, Yuchen; Ho, Jeffrey; Vemuri, Baba

    2013-01-01

    Existing dictionary learning algorithms are based on the assumption that the data are vectors in an Euclidean vector space ℝ d , and the dictionary is learned from the training data using the vector space structure of ℝ d and its Euclidean L 2 -metric. However, in many applications, features and data often originated from a Riemannian manifold that does not support a global linear (vector space) structure. Furthermore, the extrinsic viewpoint of existing dictionary learning algorithms becomes inappropriate for modeling and incorporating the intrinsic geometry of the manifold that is potentially important and critical to the application. This paper proposes a novel framework for sparse coding and dictionary learning for data on a Riemannian manifold, and it shows that the existing sparse coding and dictionary learning methods can be considered as special (Euclidean) cases of the more general framework proposed here. We show that both the dictionary and sparse coding can be effectively computed for several important classes of Riemannian manifolds, and we validate the proposed method using two well-known classification problems in computer vision and medical imaging analysis.

  12. 7 CFR 210.1 - General purpose and scope.

    Science.gov (United States)

    2010-01-01

    ... general and special cash assistance and donations of foods acquired by the Department to be used to assist..., preparation and service of nutritious lunches, payment of funds, use of program funds, program monitoring, and...

  13. RoboCon: A general purpose telerobotic control center

    International Nuclear Information System (INIS)

    Draper, J.V.; Noakes, M.W.; Blair, L.M.

    1997-01-01

    This report describes human factors issues involved in the design of RoboCon, a multi-purpose control center for use in US Department of Energy remote handling applications. RoboCon is intended to be a flexible, modular control center capable of supporting a wide variety of robotic devices

  14. RoboCon: A general purpose telerobotic control center

    Energy Technology Data Exchange (ETDEWEB)

    Draper, J.V.; Noakes, M.W. [Oak Ridge National Lab., TN (United States). Robotics and Process Systems Div.; Schempf, H. [Carnegie Mellon Univ., Pittsburgh, PA (United States); Blair, L.M. [Human Machine Interfaces, Inc., Knoxville, TN (United States)

    1997-02-01

    This report describes human factors issues involved in the design of RoboCon, a multi-purpose control center for use in US Department of Energy remote handling applications. RoboCon is intended to be a flexible, modular control center capable of supporting a wide variety of robotic devices.

  15. DYNSYL: a general-purpose dynamic simulator for chemical processes

    International Nuclear Information System (INIS)

    Patterson, G.K.; Rozsa, R.B.

    1978-01-01

    Lawrence Livermore Laboratory is conducting a safeguards program for the Nuclear Regulatory Commission. The goal of the Material Control Project of this program is to evaluate material control and accounting (MCA) methods in plants that handle special nuclear material (SNM). To this end we designed and implemented the dynamic chemical plant simulation program DYNSYL. This program can be used to generate process data or to provide estimates of process performance; it simulates both steady-state and dynamic behavior. The MCA methods that may have to be evaluated range from sophisticated on-line material trackers such as Kalman filter estimators, to relatively simple material balance procedures. This report describes the overall structure of DYNSYL and includes some example problems. The code is still in the experimental stage and revision is continuing

  16. Test and intercomparisons of data fitting with general least squares code GMA versus Bayesian code GLUCS

    International Nuclear Information System (INIS)

    Pronyaev, V.G.

    2003-01-01

    Data fitting with GMA and GLUCS gives consistent results. Difference in the evaluated central values obtained with different formalisms can be related to the general accuracy with which fits could be done in different formalisms. It has stochastic nature and should be accounted in the final results of the data evaluation as small SERC uncertainty. Some shift in central values of data evaluated with GLUCS and GMA relative the central values evaluated with the R-matrix model code RAC is observed for cases of fitting strongly varying data and is related to the PPP. The procedure of evaluation, free from PPP, should be elaborated. (author)

  17. General Purpose Segmentation for Microorganisms in Microscopy Images

    DEFF Research Database (Denmark)

    Jensen, Sebastian H. Nesgaard; Moeslund, Thomas B.; Rankl, Christian

    2014-01-01

    In this paper, we propose an approach for achieving generalized segmentation of microorganisms in mi- croscopy images. It employs a pixel-wise classification strategy based on local features. Multilayer percep- trons are utilized for classification of the local features and is trained for each sp...

  18. A general purpose tomographic program with combined inversions

    International Nuclear Information System (INIS)

    Xu Wenbin; Dong Jiafu; Li Fanzhu

    1996-01-01

    A general tomographic program has been developed by combining the Bessel expansion with the Zernicke expansion. It is useful for studying of the magnetic island structure of the tearing mode and in reconstructing the density profiles of impurities in tokamak plasmas. This combined method have the advantages of both expansions, i.e. there will be no spurious images in the edge and it will be of high inverse precision in the center of plasma

  19. GOC: General Orbit Code. [In FORTRAN for IBM 360

    Energy Technology Data Exchange (ETDEWEB)

    Maddox, L.B.; McNeilly, G.S.

    1979-08-01

    GOC (General Orbit Code) is a versatile program which will perform a variety of calculations relevant to isochronous cyclotron design studies. In addition to the usual calculations of interest (e.g., equilibrium and accelerated orbits, focusing frequencies, field isochronization, etc.), GOC has a number of options to calculate injections with a charge change. GOC provides both printed and plotted output, and will follow groups of particles to allow determination of finite-beam properties. An interactive PDP-10 program called GIP, which prepares input data for GOC, is available. GIP is a very easy and convenient way to prepare complicated input data for GOC. Enclosed with this report are several microfiche containing source listings of GOC and other related routines and the printed output from a multiple-option GOC run.

  20. General purpose dynamic Monte Carlo with continuous energy for transient analysis

    Energy Technology Data Exchange (ETDEWEB)

    Sjenitzer, B. L.; Hoogenboom, J. E. [Delft Univ. of Technology, Dept. of Radiation, Radionuclide and Reactors, Mekelweg 15, 2629JB Delft (Netherlands)

    2012-07-01

    For safety assessments transient analysis is an important tool. It can predict maximum temperatures during regular reactor operation or during an accident scenario. Despite the fact that this kind of analysis is very important, the state of the art still uses rather crude methods, like diffusion theory and point-kinetics. For reference calculations it is preferable to use the Monte Carlo method. In this paper the dynamic Monte Carlo method is implemented in the general purpose Monte Carlo code Tripoli4. Also, the method is extended for use with continuous energy. The first results of Dynamic Tripoli demonstrate that this kind of calculation is indeed accurate and the results are achieved in a reasonable amount of time. With the method implemented in Tripoli it is now possible to do an exact transient calculation in arbitrary geometry. (authors)

  1. General Purpose Graphics Processing Unit Based High-Rate Rice Decompression and Reed-Solomon Decoding

    Energy Technology Data Exchange (ETDEWEB)

    Loughry, Thomas A. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2015-02-01

    As the volume of data acquired by space-based sensors increases, mission data compression/decompression and forward error correction code processing performance must likewise scale. This competency development effort was explored using the General Purpose Graphics Processing Unit (GPGPU) to accomplish high-rate Rice Decompression and high-rate Reed-Solomon (RS) decoding at the satellite mission ground station. Each algorithm was implemented and benchmarked on a single GPGPU. Distributed processing across one to four GPGPUs was also investigated. The results show that the GPGPU has considerable potential for performing satellite communication Data Signal Processing, with three times or better performance improvements and up to ten times reduction in cost over custom hardware, at least in the case of Rice Decompression and Reed-Solomon Decoding.

  2. GENERAL ASPECTS AND PURPOSE OF THE CONSUMMATION CONTRACT

    Directory of Open Access Journals (Sweden)

    Alexandru MATEESCU

    2015-07-01

    Full Text Available The current paper mainly tackles the consummation contract and the ways in which it forms, adjusts and helps the citizen, in the general frame of the consumer’s rights. The consummation contract is a legal instrument for institutionalizing the commercial operations between a professional and the consumer. This type of instrument has been especially created to help the citizen in being adequately treated both correctly and adequately in all consummation matters. In the 20th century, the need to bring the commercial relations between professionals and citizens under regulation became imperative. Taking into account the socio-humanistic development rhythm, the consummation contact became a way of guaranteeing for the citizens’ rights in consummation matters and, at the same time, it confers a legal frame to guide and direct the merchants. Since 2007, the year when Romania adhered the European Union, the consumers’ rights became even more important, as the fact that the European Union focuses on the rights of all its citizens and the way in which they are treated by the merchants from a commercial point of view, was widely recognized. Once Romania adhered to a commercial frame which embeds over 500 million inhabitants, the dimension of the rights and obligations regarding consummation became an ever stronger argument in bringing the legislation under regulation. Eventually, clarifying the principles that founded the consummation contract will determine the legal frame in which it revolves and will bring to light the utility it has for the citizens, its subsequent employments and all things that should be further added to it in order to ensure an even better protection and settlement, in legal terms, of the commercial relations between a merchant and one or more citizens.

  3. Correlated sampling added to the specific purpose Monte Carlo code McPNL for neutron lifetime log responses

    International Nuclear Information System (INIS)

    Mickael, M.; Verghese, K.; Gardner, R.P.

    1989-01-01

    The specific purpose neutron lifetime oil well logging simulation code, McPNL, has been rewritten for greater user-friendliness and faster execution. Correlated sampling has been added to the code to enable studies of relative changes in the tool response caused by environmental changes. The absolute responses calculated by the code have been benchmarked against laboratory test pit data. The relative responses from correlated sampling are not directly benchmarked, but they are validated using experimental and theoretical results

  4. Control rod computer code IAMCOS: general theory and numerical methods

    International Nuclear Information System (INIS)

    West, G.

    1982-11-01

    IAMCOS is a computer code for the description of mechanical and thermal behavior of cylindrical control rods for fast breeders. This code version was applied, tested and modified from 1979 to 1981. In this report are described the basic model (02 version), theoretical definitions and computation methods [fr

  5. Geometry system used in the General Monte Carlo transport code SPARTAN

    International Nuclear Information System (INIS)

    Bending, R.C.; Easter, P.G.

    1974-01-01

    The geometry routines used in the general-purpose, three-dimensional particle transport code SPARTAN are described. The code is designed to deal with the very complex geometries encountered in lattice cell and fuel handling calculations, health physics, and shielding problems. Regions of the system being studied may be represented by simple shapes (spheres, cylinders, and so on) or by multinomial surfaces of any order, and many simple shapes may be combined to make up a complex layout. The geometry routines are designed to allow the program to carry out a number of tasks (such as sampling for a random point or tracking a path through several regions) in any order, so that the use of the routines is not restricted to a particular tracking or scoring method. Routines for reading, checking, and printing the data are included. (U.S.)

  6. Differentially Encoded LDPC Codes—Part II: General Case and Code Optimization

    Directory of Open Access Journals (Sweden)

    Jing Li (Tiffany

    2008-04-01

    Full Text Available This two-part series of papers studies the theory and practice of differentially encoded low-density parity-check (DE-LDPC codes, especially in the context of noncoherent detection. Part I showed that a special class of DE-LDPC codes, product accumulate codes, perform very well with both coherent and noncoherent detections. The analysis here reveals that a conventional LDPC code, however, is not fitful for differential coding and does not, in general, deliver a desirable performance when detected noncoherently. Through extrinsic information transfer (EXIT analysis and a modified “convergence-constraint” density evolution (DE method developed here, we provide a characterization of the type of LDPC degree profiles that work in harmony with differential detection (or a recursive inner code in general, and demonstrate how to optimize these LDPC codes. The convergence-constraint method provides a useful extension to the conventional “threshold-constraint” method, and can match an outer LDPC code to any given inner code with the imperfectness of the inner decoder taken into consideration.

  7. General features of the neutronics design code EQUICYCLE

    International Nuclear Information System (INIS)

    Jirlow, K.

    1978-10-01

    The neutronics code EQUICYCLE has been developed and improved over a long period of time. It is expecially adapted to survey type design calculations of large fast power reactors with particular emphasis on the nuclear parameters for a realistic equilibrium fuel cycle. Thus the code is used to evaluate the breeding performance, the power distributions and the uranium and plutonium mass balance for realistic refuelling schemes. In addition reactivity coefficients can be calculated and the influence of burnup could be assessed. The code is two-dimensional and treats the reactor core in R-Z geometry. The basic ideas of the calculating scheme are successive iterative improvement of cross-section sets and flux spectra and use of the mid-cycle flux for burning the fuel according to a specified refuelling scheme. Normally given peak burn-ups and maximum power densities are used as boundary conditions. The code is capable of handling the unconventional, so called heterogeneous cores. (author)

  8. A Fast Optimization Method for General Binary Code Learning.

    Science.gov (United States)

    Shen, Fumin; Zhou, Xiang; Yang, Yang; Song, Jingkuan; Shen, Heng; Tao, Dacheng

    2016-09-22

    Hashing or binary code learning has been recognized to accomplish efficient near neighbor search, and has thus attracted broad interests in recent retrieval, vision and learning studies. One main challenge of learning to hash arises from the involvement of discrete variables in binary code optimization. While the widely-used continuous relaxation may achieve high learning efficiency, the pursued codes are typically less effective due to accumulated quantization error. In this work, we propose a novel binary code optimization method, dubbed Discrete Proximal Linearized Minimization (DPLM), which directly handles the discrete constraints during the learning process. Specifically, the discrete (thus nonsmooth nonconvex) problem is reformulated as minimizing the sum of a smooth loss term with a nonsmooth indicator function. The obtained problem is then efficiently solved by an iterative procedure with each iteration admitting an analytical discrete solution, which is thus shown to converge very fast. In addition, the proposed method supports a large family of empirical loss functions, which is particularly instantiated in this work by both a supervised and an unsupervised hashing losses, together with the bits uncorrelation and balance constraints. In particular, the proposed DPLM with a supervised `2 loss encodes the whole NUS-WIDE database into 64-bit binary codes within 10 seconds on a standard desktop computer. The proposed approach is extensively evaluated on several large-scale datasets and the generated binary codes are shown to achieve very promising results on both retrieval and classification tasks.

  9. On the Representation of the Porosity-Pressure Relationship in General Subsurface Flow Codes

    Science.gov (United States)

    Birdsell, Daniel T.; Karra, Satish; Rajaram, Harihar

    2018-02-01

    The governing equations for subsurface flow codes in a deformable porous media are derived from the balance of fluid mass and Darcy's equation. One class of these codes, which we call general subsurface flow codes (GSFs), allow for more general constitutive relations for material properties such as porosity, permeability and density. Examples of GSFs include PFLOTRAN, FEHM, TOUGH2, STOMP, and some reservoir simulators such as BOAST. Depending on the constitutive relations used in GSFs, an inconsistency arises between the standard groundwater flow equation and the governing equation of GSFs, and we clarify that the reason for this inconsistency is because the Darcy's equation used in the GSFs should account for the velocity of fluid with respect to solid. Due to lack of awareness of this inconsistency, users of the GSFs tend to use a porosity-pressure relationship that comes from the standard groundwater flow equation and assumes that the relative velocity is already accounted for. For the Theis problem, we show that using this traditional relationship in the GSFs leads to significantly large errors. We propose an alternate porosity-pressure relationship that is consistent with the derivation of the governing equations in the GSFs where the solid velocity is not tracked, and show that, with this relationship, the results are more accurate for the Theis problem. The purpose of this note is to make the users and developers of these GSFs aware of this inconsistency and to advocate that the alternate porosity model derived here should be incorporated in GSFs.

  10. General data analysis code for TDCR liquid scintillation counting

    Energy Technology Data Exchange (ETDEWEB)

    Rodrigues, D. [Laboratorio de Metrologia de Radioisotopos, Comision Nacional de Energia Atomica, Buenos Aires (Argentina)], E-mail: drodrigu@cae.cnea.gov.ar; Arenillas, P.; Capoulat, M.E.; Balpardo, C. [Laboratorio de Metrologia de Radioisotopos, Comision Nacional de Energia Atomica, Buenos Aires (Argentina)

    2008-06-15

    A non-radionuclide-specific computer code to analyze data, calculate detection efficiency and activity in a TDCR system is presented. The program was developed prioritizing flexibility in measuring conditions, parameters and calculation models. It is also intended to be well structured in order to easily replace subroutines which could eventually be improved by the user. It is written in standard FORTRAN language but a graphical interface is also available. Several tests were performed to check the ability of the code to deal with different decay schemes such as H-3, C-14, Fe-55, Mn-54 and Co-60.

  11. General data analysis code for TDCR liquid scintillation counting

    International Nuclear Information System (INIS)

    Rodrigues, D.; Arenillas, P.; Capoulat, M.E.; Balpardo, C.

    2008-01-01

    A non-radionuclide-specific computer code to analyze data, calculate detection efficiency and activity in a TDCR system is presented. The program was developed prioritizing flexibility in measuring conditions, parameters and calculation models. It is also intended to be well structured in order to easily replace subroutines which could eventually be improved by the user. It is written in standard FORTRAN language but a graphical interface is also available. Several tests were performed to check the ability of the code to deal with different decay schemes such as H-3, C-14, Fe-55, Mn-54 and Co-60

  12. Lipseys Quest for the Micro-foundations of GPT-the General Purpose Engine

    NARCIS (Netherlands)

    Van der Kooij, B.J.G.

    2016-01-01

    The construct of the General Purpose Technology misses its micro-foundation (as observed by Richard Lipsey). We present a possible solution in the General Purpose Engines. These are the basic innovations and the clusters of contributing and derived innovation, that appear in a Schumpeterian 'cluster

  13. 78 FR 65300 - Notice of Availability (NOA) for General Purpose Warehouse and Information Technology Center...

    Science.gov (United States)

    2013-10-31

    ... (NOA) for General Purpose Warehouse and Information Technology Center Construction (GPW/IT)--Tracy Site... proposed action to construct a General Purpose Warehouse and Information Technology Center at Defense..., Suite 02G09, Alexandria, VA 22350- 3100. FOR FURTHER INFORMATION CONTACT: Ann Engelberger at (703) 767...

  14. Evaluation of the General Atomic codes TAP and RECA for HTGR accident analyses

    International Nuclear Information System (INIS)

    Ball, S.J.; Cleveland, J.C.; Sanders, J.P.

    1978-01-01

    The General Atomic codes TAP (Transient Analysis Program) and RECA (Reactor Emergency Cooling Analysis) are evaluated with respect to their capability for predicting the dynamic behavior of high-temperature gas-cooled reactors (HTGRs) for postulated accident conditions. Several apparent modeling problems are noted, and the susceptibility of the codes to misuse and input errors is discussed. A critique of code verification plans is also included. The several cases where direct comparisons could be made between TAP/RECA calculations and those based on other independently developed codes indicated generally good agreement, thus contributing to the credibility of the codes

  15. Some questions of using the algebraic coding theory for construction of special-purpose processors in high energy physics spectrometers

    International Nuclear Information System (INIS)

    Nikityuk, N.M.

    1989-01-01

    The results of investigations of using the algebraic coding theory for the creation of parallel encoders, majority coincidence schemes and coordinate processors for the first and second trigger levels are described. Concrete examples of calculation and structure of special-purpose processor using the table arithmetic method are given for multiplicity t ≤ 5. The question of using parallel and sequential syndrome coding methods for the registration of events with clusters is discussed. 30 refs.; 10 figs

  16. CoCoNuT: General relativistic hydrodynamics code with dynamical space-time evolution

    Science.gov (United States)

    Dimmelmeier, Harald; Novak, Jérôme; Cerdá-Durán, Pablo

    2012-02-01

    CoCoNuT is a general relativistic hydrodynamics code with dynamical space-time evolution. The main aim of this numerical code is the study of several astrophysical scenarios in which general relativity can play an important role, namely the collapse of rapidly rotating stellar cores and the evolution of isolated neutron stars. The code has two flavors: CoCoA, the axisymmetric (2D) magnetized version, and CoCoNuT, the 3D non-magnetized version.

  17. Generalized instantly decodable network coding for relay-assisted networks

    KAUST Repository

    Elmahdy, Adel M.; Sorour, Sameh; Seddik, Karim G.

    2013-01-01

    which employs generalized IDNC instead of strict IDNC previously proposed in the literature for the same relay-assisted topology. This use of generalized IDNC is supported by showing that it is a super-set of the strict IDNC scheme, and thus can generate

  18. A generalized interface module for the coupling of spatial kinetics and thermal-hydraulics codes

    Energy Technology Data Exchange (ETDEWEB)

    Barber, D.A.; Miller, R.M.; Joo, H.G.; Downar, T.J. [Purdue Univ., West Lafayette, IN (United States). Dept. of Nuclear Engineering; Wang, W. [SCIENTECH, Inc., Rockville, MD (United States); Mousseau, V.A.; Ebert, D.D. [Nuclear Regulatory Commission, Washington, DC (United States). Office of Nuclear Regulatory Research

    1999-03-01

    A generalized interface module has been developed for the coupling of any thermal-hydraulics code to any spatial kinetics code. The coupling scheme was designed and implemented with emphasis placed on maximizing flexibility while minimizing modifications to the respective codes. In this design, the thermal-hydraulics, general interface, and spatial kinetics codes function independently and utilize the Parallel Virtual Machine software to manage cross-process communication. Using this interface, the USNRC version of the 3D neutron kinetics code, PARCX, has been coupled to the USNRC system analysis codes RELAP5 and TRAC-M. RELAP5/PARCS assessment results are presented for two NEACRP rod ejection benchmark problems and an NEA/OECD main steam line break benchmark problem. The assessment of TRAC-M/PARCS has only recently been initiated, nonetheless, the capabilities of the coupled code are presented for a typical PWR system/core model.

  19. A generalized interface module for the coupling of spatial kinetics and thermal-hydraulics codes

    International Nuclear Information System (INIS)

    Barber, D.A.; Miller, R.M.; Joo, H.G.; Downar, T.J.; Mousseau, V.A.; Ebert, D.D.

    1999-01-01

    A generalized interface module has been developed for the coupling of any thermal-hydraulics code to any spatial kinetics code. The coupling scheme was designed and implemented with emphasis placed on maximizing flexibility while minimizing modifications to the respective codes. In this design, the thermal-hydraulics, general interface, and spatial kinetics codes function independently and utilize the Parallel Virtual Machine software to manage cross-process communication. Using this interface, the USNRC version of the 3D neutron kinetics code, PARCX, has been coupled to the USNRC system analysis codes RELAP5 and TRAC-M. RELAP5/PARCS assessment results are presented for two NEACRP rod ejection benchmark problems and an NEA/OECD main steam line break benchmark problem. The assessment of TRAC-M/PARCS has only recently been initiated, nonetheless, the capabilities of the coupled code are presented for a typical PWR system/core model

  20. An Aerodynamic Database for the Mk 82 General Purpose Low Drag Bomb

    National Research Council Canada - National Science Library

    Krishnamoorthy, L

    1997-01-01

    The drag database of the Mk 82 General Purpose Low Drag bomb, the primary gravity weapon in the RAAF inventory, has some shortcomings in the quality and traceability of data, and in the variations due...

  1. Implementation elements for conversion of general-purpose freeway lane into high-occupancy-vehicle lane

    Science.gov (United States)

    1997-01-01

    Conversion of a general-purpose freeway into a high-occupancy-vehicle (HOV) lane is an alternative to infrastructure addition for HOV system implementation. Research indicates that lane conversion is feasible technically if sufficient HOV usage and m...

  2. Low Overhead Real-Time Computing With General Purpose Operating Systems

    National Research Council Canada - National Science Library

    Raymond, Michael

    2004-01-01

    .... In larger systems and more recently, general-purpose operating systems such as SGI IRIX and Linux are used for new projects because they already have multiprocessor and device driver support as well a large user base...

  3. Implementation of generalized quantum measurements: Superadditive quantum coding, accessible information extraction, and classical capacity limit

    International Nuclear Information System (INIS)

    Takeoka, Masahiro; Fujiwara, Mikio; Mizuno, Jun; Sasaki, Masahide

    2004-01-01

    Quantum-information theory predicts that when the transmission resource is doubled in quantum channels, the amount of information transmitted can be increased more than twice by quantum-channel coding technique, whereas the increase is at most twice in classical information theory. This remarkable feature, the superadditive quantum-coding gain, can be implemented by appropriate choices of code words and corresponding quantum decoding which requires a collective quantum measurement. Recently, an experimental demonstration was reported [M. Fujiwara et al., Phys. Rev. Lett. 90, 167906 (2003)]. The purpose of this paper is to describe our experiment in detail. Particularly, a design strategy of quantum-collective decoding in physical quantum circuits is emphasized. We also address the practical implication of the gain on communication performance by introducing the quantum-classical hybrid coding scheme. We show how the superadditive quantum-coding gain, even in a small code length, can boost the communication performance of conventional coding techniques

  4. Variation in clinical coding lists in UK general practice: a barrier to consistent data entry?

    Science.gov (United States)

    Tai, Tracy Waize; Anandarajah, Sobanna; Dhoul, Neil; de Lusignan, Simon

    2007-01-01

    Routinely collected general practice computer data are used for quality improvement; poor data quality including inconsistent coding can reduce their usefulness. To document the diversity of data entry systems currently in use in UK general practice and highlight possible implications for data quality. General practice volunteers provided screen shots of the clinical coding screen they would use to code a diagnosis or problem title in the clinical consultation. The six clinical conditions examined were: depression, cystitis, type 2 diabetes mellitus, sore throat, tired all the time, and myocardial infarction. We looked at the picking lists generated for these problem titles in EMIS, IPS, GPASS and iSOFT general practice clinical computer systems, using the Triset browser as a gold standard for comparison. A mean of 19.3 codes is offered in the picking list after entering a diagnosis or problem title. EMIS produced the longest picking lists and GPASS the shortest, with a mean number of choices of 35.2 and 12.7, respectively. Approximately three-quarters (73.5%) of codes are diagnoses, one-eighth (12.5%) symptom codes, and the remainder come from a range of Read chapters. There was no readily detectable consistent order in which codes were displayed. Velocity coding, whereby commonly-used codes are placed higher in the picking list, results in variation between practices even where they have the same brand of computer system. Current systems for clinical coding promote diversity rather than consistency of clinical coding. As the UK moves towards an integrated health IT system consistency of coding will become more important. A standardised, limited list of codes for primary care might help address this need.

  5. Utilizing General Purpose Graphics Processing Units to Improve Performance of Computer Modelling and Visualization

    Science.gov (United States)

    Monk, J.; Zhu, Y.; Koons, P. O.; Segee, B. E.

    2009-12-01

    With the introduction of the G8X series of cards by nVidia an architecture called CUDA was released, virtually all subsequent video cards have had CUDA support. With this new architecture nVidia provided extensions for C/C++ that create an Application Programming Interface (API) allowing code to be executed on the GPU. Since then the concept of GPGPU (general purpose graphics processing unit) has been growing, this is the concept that the GPU is very good a algebra and running things in parallel so we should take use of that power for other applications. This is highly appealing in the area of geodynamic modeling, as multiple parallel solutions of the same differential equations at different points in space leads to a large speedup in simulation speed. Another benefit of CUDA is a programmatic method of transferring large amounts of data between the computer's main memory and the dedicated GPU memory located on the video card. In addition to being able to compute and render on the video card, the CUDA framework allows for a large speedup in the situation, such as with a tiled display wall, where the rendered pixels are to be displayed in a different location than where they are rendered. A CUDA extension for VirtualGL was developed allowing for faster read back at high resolutions. This paper examines several aspects of rendering OpenGL graphics on large displays using VirtualGL and VNC. It demonstrates how performance can be significantly improved in rendering on a tiled monitor wall. We present a CUDA enhanced version of VirtualGL as well as the advantages to having multiple VNC servers. It will discuss restrictions caused by read back and blitting rates and how they are affected by different sizes of virtual displays being rendered.

  6. A fast and compact Fuel Rod Performance Simulator code for predictive, interpretive and educational purpose

    International Nuclear Information System (INIS)

    Lorenzen, J.

    1990-01-01

    A new Fuel rod Performance Simulator code FRPS has been developed, tested and benchmarked and is now available in different versions. The user may choose between the batch version INTERPIN producing results in form of listings or beforehand defined plots, or the interactive simulator code SIMSIM which is stepping through a power history under the control of user. Both versions are presently running on minicomputers and PC:s using EGA-Graphics. A third version is the implementation in a Studsvik Compact Simulator with FRPS being one of its various modules receiving the dynamic inputs from the simulator

  7. 24 CFR 990.310 - Purpose-General policy on financial management, monitoring and reporting.

    Science.gov (United States)

    2010-04-01

    ... Management Systems, Monitoring, and Reporting § 990.310 Purpose—General policy on financial management, monitoring and reporting. All PHA financial management systems, reporting, and monitoring of program... 24 Housing and Urban Development 4 2010-04-01 2010-04-01 false Purpose-General policy on financial...

  8. Computer code for general analysis of radon risks (GARR)

    International Nuclear Information System (INIS)

    Ginevan, M.

    1984-09-01

    This document presents a computer model for general analysis of radon risks that allow the user to specify a large number of possible models with a small number of simple commands. The model is written in a version of BASIC which conforms closely to the American National Standards Institute (ANSI) definition for minimal BASIC and thus is readily modified for use on a wide variety of computers and, in particular, microcomputers. Model capabilities include generation of single-year life tables from 5-year abridged data, calculation of multiple-decrement life tables for lung cancer for the general population, smokers, and nonsmokers, and a cohort lung cancer risk calculation that allows specification of level and duration of radon exposure, the form of the risk model, and the specific population assumed at risk. 36 references, 8 figures, 7 tables

  9. Code of practice for the design of laboratories using radioactive substances for medical purposes

    International Nuclear Information System (INIS)

    1981-01-01

    This Code has been prepared to supplement the radioactive substances acts and regulations implemented in Australia. It is intended as a guide to safe practices but is not legislation. Areas covered include siting, layout, surface finishes, laboratory furniture and fittings, ventilation, containment and release of airborne effluent and storage of radioactive substances

  10. Generalized rank weights of reducible codes, optimal cases and related properties

    DEFF Research Database (Denmark)

    Martinez Peñas, Umberto

    2018-01-01

    in network coding. In this paper, we study their security behavior against information leakage on networks when applied as coset coding schemes, giving the following main results: 1) we give lower and upper bounds on their generalized rank weights (GRWs), which measure worst case information leakage...... to the wire tapper; 2) we find new parameters for which these codes are MRD (meaning that their first GRW is optimal) and use the previous bounds to estimate their higher GRWs; 3) we show that all linear (over the extension field) codes, whose GRWs are all optimal for fixed packet and code sizes but varying...... length are reducible codes up to rank equivalence; and 4) we show that the information leaked to a wire tapper when using reducible codes is often much less than the worst case given by their (optimal in some cases) GRWs. We conclude with some secondary related properties: conditions to be rank...

  11. TRIO-EF a general thermal hydraulics computer code applied to the Avlis process

    International Nuclear Information System (INIS)

    Magnaud, J.P.; Claveau, M.; Coulon, N.; Yala, P.; Guilbaud, D.; Mejane, A.

    1993-01-01

    TRIO(EF is a general purpose Fluid Mechanics 3D Finite Element Code. The system capabilities cover areas such as steady state or transient, laminar or turbulent, isothermal or temperature dependent fluid flows; it is applicable to the study of coupled thermo-fluid problems involving heat conduction and possibly radiative heat transfer. It has been used to study the thermal behaviour of the AVLIS process separation module. In this process, a linear electron beam impinges the free surface of a uranium ingot, generating a two dimensional curtain emission of vapour from a water-cooled crucible. The energy transferred to the metal causes its partial melting, forming a pool where strong convective motion increases heat transfer towards the crucible. In the upper part of the Separation Module, the internal structures are devoted to two main functions: vapor containment and reflux, irradiation and physical separation. They are subjected to very high temperature levels and heat transfer occurs mainly by radiation. Moreover, special attention has to be paid to electron backscattering. These two major points have been simulated numerically with TRIO-EF and the paper presents and comments the results of such a computation, for each of them. After a brief overview of the computer code, two examples of the TRIO-EF capabilities are given: a crucible thermal hydraulics model, a thermal analysis of the internal structures

  12. [Violence for educational purpose: Representations of general practitioners in the Paris area, France. A qualitative study].

    Science.gov (United States)

    de Brie, Claire; Piet, Emmanuelle; Chariot, Patrick

    2018-03-01

    Violence for educational purpose refers to a modality of education that includes threats, verbal abuse, physical abuse and humiliations. Twenty European countries, not including France, have abolished corporal punishment through explicit laws and regulations. The position of general practitioners in the screening and care of violence for educational purpose in France is unknown. In this study, we aimed to assess the representations of this form of violence among general practitioners. We have performed semi-directed interviews of general practitioners in the Paris, France region (Île-de-France). Interviews were conducted until data saturation was achieved. Interviews were recorded, transcribed and analysed by two investigators. Interviews were conducted with 20 physicians (November 2015-January 2016). General practitioners considered that physical, verbal or psychological abuse had possible negative consequences on children. Uncertainty regarding the consequences of violence was a cause of tolerance towards violence for educational purpose, depending on the act committed and the context, as perceived by nearly all practitioners. General practitioners expressed interest in the field. They cited their own education and experience as the main obstacles to action. Most of them expressed a feeling of failure when they screened or took care of violence for educational purpose. This study suggests that doctors can participate in supporting the parents in the prevention of violence for educational purpose. Support to parents would need specific medical training as well as a societal change. Copyright © 2018 Elsevier Masson SAS. All rights reserved.

  13. Determining Market Categorization of United States Zip Codes for Purposes of Army Recruiting

    Science.gov (United States)

    2016-06-01

    Army uses commercial market segmentation data to analyze markets and past accessions to assign recruiters and quotas to maximize production. We use...Army Recruiting Command to rely on proprietary data with 66 market segments per ZIP code for market analysis and predicting recruiting potential...have different densities of potential recruits; the Army uses commercial market segmentation data to analyze markets and past accessions to assign

  14. MCNP: a general Monte Carlo code for neutron and photon transport

    International Nuclear Information System (INIS)

    1978-07-01

    The general-purpose Monte Carlo code MCNP can be used for neutron, photon, or coupled neutron--photon transport. The code treats an arbitrary three-dimensional configuration of materials in geometric cells bounded by first- and second-degree surfaces and some special fourth-degree surfaces (elliptical tori). Pointwise cross-section data are used. For neutrons, all reactions given in a particular cross-section evaluation (such as ENDF/B-IV) are accounted for. For photons, the code takes account of incoherent and coherent scattering, the possibility of fluorescent emission following photoelectric absorption, and absorption in pair production with local emission of annihilation radiation. Standard optional variance reduction schemes include geometry splitting and Russian roulette, the exponential transformation, energy splitting, forced collisions in designated cells, flux estimates at point detectors, track-length estimators, and source biasing. The standard output of MCNP includes two-way current as a function of energy, time, and angle with the normal, across any subset of bounding surfaces in the problem. Fluxes across any set of bounding surfaces are available as a function of time and energy. Similarly, the flux at designated points and the average flux in a cell (track length per unit volume) are standard tallies. Reactions such as fissions or absorptions may be obtained in a subset of geometric cells. The heating tallies give the energy deposition per starting particle. In addition, particles may be flagged when they cross specified surfaces or enter designated cells, and the contributions of these flagged particles to certain of the tallies are listed separately. All quantities printed out have their relative errors listed also. 11 figures, 27 tables

  15. MCNP: a general Monte Carlo code for neutron and photon transport

    International Nuclear Information System (INIS)

    1979-11-01

    The general-purpose Monte Carlo code MCNP ca be used for neutron, photon, or coupled neutron-photon transport, including the capability to calculate eigenvalues for critical systems. The code treats an arbitrary three-dimensional configuration of materials in geometric cells bounded by first- and second-degree surfaces and some special fourth-degree surfaces (elliptical tori). Pointwise cross-section data are used. For neutrons, all reactions given in a particular cross-section evaluation are accounted for. Thermal neutrons are described by both the free-gas and S(α,β) models. For photons, the code takes account of incoherent and coherent scattering, the possibility of fluorescent emission following photoelectric absorption, and absorption in pair production with local emission of annihilation radiation. MCNP includes an elaborate, interactive plotting capability that allows the user to view his input geometry to help check for setup errors. Standard features which are available to improve computational efficiency include geometry splitting and Russian roulette, weight cutoff with Russian roulette, correlated sampling, analog capture or capture by weight reduction, the exponential transformation, energy splitting, forced collisions in designated cells, flux estimates at point or ring detectors, deterministically transporting pseudo-particles to designated regions, track-length estimators, source biasing, and several parameter cutoffs. Extensive summary information is provided to help the user better understand the physics and Monte Carlo simulation of his problem. The standard, user-defined output of MCNP includes two-way current as a function of direction across any set of surfaces or surface segments in the problem. Flux across any set of surfaces or surface segments is available. 58 figures, 28 tables

  16. Development of a general coupling interface for the fuel performance code transuranus tested with the reactor dynamic code DYN3D

    International Nuclear Information System (INIS)

    Holt, L.; Rohde, U.; Seidl, M.; Schubert, A.; Van Uffelen, P.

    2013-01-01

    Several institutions plan to couple the fuel performance code TRANSURANUS developed by the European Institute for Transuranium Elements with their own codes. One of these codes is the reactor dynamic code DYN3D maintained by the Helmholtz-Zentrum Dresden - Rossendorf. DYN3D was developed originally for VVER type reactors and was extended later to western type reactors. Usually, the fuel rod behavior is modeled in thermal hydraulics and neutronic codes in a simplified manner. The main idea of this coupling is to describe the fuel rod behavior in the frame of core safety analysis in a more detailed way, e.g. including the influence of the high burn-up structure, geometry changes and fission gas release. It allows to take benefit from the improved computational power and software achieved over the last two decades. The coupling interface was developed in a general way from the beginning. Thence it can be easily used also by other codes for a coupling with TRANSURANUS. The user can choose between a one-way as well as a two-way online coupling option. For a one-way online coupling, DYN3D provides only the time-dependent rod power and thermal hydraulics conditions to TRANSURANUS, but the fuel performance code doesn’t transfer any variable back to DYN3D. In a two-way online coupling, TRANSURANUS in addition transfers parameters like fuel temperature and cladding temperature back to DYN3D. This list of variables can be extended easily by geometric and further variables of interest. First results of the code system DYN3D-TRANSURANUS will be presented for a control rod ejection transient in a modern western type reactor. Pre-analyses show already that a detailed fuel rod behavior modeling will influence the thermal hydraulics and thence also the neutronics due to the Doppler reactivity effect of the fuel temperature. The coupled code system has therefore a potential to improve the assessment of safety criteria. The developed code system DYN3D-TRANSURANUS can be used also

  17. Evaluating Coding Accuracy in General Surgery Residents' Accreditation Council for Graduate Medical Education Procedural Case Logs.

    Science.gov (United States)

    Balla, Fadi; Garwe, Tabitha; Motghare, Prasenjeet; Stamile, Tessa; Kim, Jennifer; Mahnken, Heidi; Lees, Jason

    The Accreditation Council for Graduate Medical Education (ACGME) case log captures resident operative experience based on Current Procedural Terminology (CPT) codes and is used to track operative experience during residency. With increasing emphasis on resident operative experiences, coding is more important than ever. It has been shown in other surgical specialties at similar institutions that the residents' ACGME case log may not accurately reflect their operative experience. What barriers may influence this remains unclear. As the only objective measure of resident operative experience, an accurate case log is paramount in representing one's operative experience. This study aims to determine the accuracy of procedural coding by general surgical residents at a single institution. Data were collected from 2 consecutive graduating classes of surgical residents' ACGME case logs from 2008 to 2014. A total of 5799 entries from 7 residents were collected. The CPT codes entered by residents were compared to departmental billing records submitted by the attending surgeon for each procedure. Assigned CPT codes by institutional American Academy of Professional Coders certified abstract coders were considered the "gold standard." A total of 4356 (75.12%) of 5799 entries were identified in billing records. Excel 2010 and SAS 9.3 were used for analysis. In the event of multiple codes for the same patient, any match between resident codes and billing record codes was considered a "correct" entry. A 4-question survey was distributed to all current general surgical residents at our institution for feedback on coding habits, limitations to accurate coding, and opinions on ACGME case log representation of their operative experience. All 7 residents had a low percentage of correctly entered CPT codes. The overall accuracy proportion for all residents was 52.82% (range: 43.32%-60.07%). Only 1 resident showed significant improvement in accuracy during his/her training (p = 0

  18. Design of a general purpose (RS-232C) analog-to-digital data converter

    International Nuclear Information System (INIS)

    Ali, Q.

    1995-01-01

    The purpose of this project is to design a general purpose hardware that interfaces analog devices with any desirable computer supporting the RS-232 interface. The hardware incorporates bidirectional data transmission at 1,200 bps, 2,400 bps, 4800 bps, 9,600 bps, 19,200 pbs and 38400 bps. The communication / processing software has been written in C language that incorporates the idea of the potability of the software from one environment to the other. (author)

  19. Design of low-cost general purpose microcontroller based neuromuscular stimulator.

    Science.gov (United States)

    Koçer, S; Rahmi Canal, M; Güler, I

    2000-04-01

    In this study, a general purpose, low-cost, programmable, portable and high performance stimulator is designed and implemented. For this purpose, a microcontroller is used in the design of the stimulator. The duty cycle and amplitude of the designed system can be controlled using a keyboard. The performance test of the system has shown that the results are reliable. The overall system can be used as the neuromuscular stimulator under safe conditions.

  20. Differentially Encoded LDPC Codes—Part II: General Case and Code Optimization

    Directory of Open Access Journals (Sweden)

    Li (Tiffany Jing

    2008-01-01

    Full Text Available This two-part series of papers studies the theory and practice of differentially encoded low-density parity-check (DE-LDPC codes, especially in the context of noncoherent detection. Part I showed that a special class of DE-LDPC codes, product accumulate codes, perform very well with both coherent and noncoherent detections. The analysis here reveals that a conventional LDPC code, however, is not fitful for differential coding and does not, in general, deliver a desirable performance when detected noncoherently. Through extrinsic information transfer (EXIT analysis and a modified "convergence-constraint" density evolution (DE method developed here, we provide a characterization of the type of LDPC degree profiles that work in harmony with differential detection (or a recursive inner code in general, and demonstrate how to optimize these LDPC codes. The convergence-constraint method provides a useful extension to the conventional "threshold-constraint" method, and can match an outer LDPC code to any given inner code with the imperfectness of the inner decoder taken into consideration.

  1. Catalog of physical protection equipment. Book 3: Volume VII. General purpose display components

    International Nuclear Information System (INIS)

    1977-06-01

    A catalog of commercially available physical protection equipment has been prepared under MITRE contract AT(49-24)-0376 for use by the U. S. Nuclear Regulatory Commission (NRC). Included is information on barrier structures and equipment, interior and exterior intrusion detection sensors, entry (access) control devices, surveillance and alarm assessment equipment, contraband detection sensors, automated response equipment, general purpose displays and general purpose communications, with one volume devoted to each of these eight areas. For each item of equipment the information included consists of performance, physical, cost and supply/logistics data. The entire catalog is contained in three notebooks for ease in its use by licensing and inspection staff at NRC

  2. Typical performance of regular low-density parity-check codes over general symmetric channels

    International Nuclear Information System (INIS)

    Tanaka, Toshiyuki; Saad, David

    2003-01-01

    Typical performance of low-density parity-check (LDPC) codes over a general binary-input output-symmetric memoryless channel is investigated using methods of statistical mechanics. Relationship between the free energy in statistical-mechanics approach and the mutual information used in the information-theory literature is established within a general framework; Gallager and MacKay-Neal codes are studied as specific examples of LDPC codes. It is shown that basic properties of these codes known for particular channels, including their potential to saturate Shannon's bound, hold for general symmetric channels. The binary-input additive-white-Gaussian-noise channel and the binary-input Laplace channel are considered as specific channel models

  3. Typical performance of regular low-density parity-check codes over general symmetric channels

    Energy Technology Data Exchange (ETDEWEB)

    Tanaka, Toshiyuki [Department of Electronics and Information Engineering, Tokyo Metropolitan University, 1-1 Minami-Osawa, Hachioji-shi, Tokyo 192-0397 (Japan); Saad, David [Neural Computing Research Group, Aston University, Aston Triangle, Birmingham B4 7ET (United Kingdom)

    2003-10-31

    Typical performance of low-density parity-check (LDPC) codes over a general binary-input output-symmetric memoryless channel is investigated using methods of statistical mechanics. Relationship between the free energy in statistical-mechanics approach and the mutual information used in the information-theory literature is established within a general framework; Gallager and MacKay-Neal codes are studied as specific examples of LDPC codes. It is shown that basic properties of these codes known for particular channels, including their potential to saturate Shannon's bound, hold for general symmetric channels. The binary-input additive-white-Gaussian-noise channel and the binary-input Laplace channel are considered as specific channel models.

  4. Report on the operation and utilization of general purpose use computer system 2001

    Energy Technology Data Exchange (ETDEWEB)

    Watanabe, Kunihiko; Watanabe, Reiko; Tsugawa, Kazuko; Tsuda, Kenzo; Yamamoto, Takashi; Nakamura, Osamu; Kamimura, Tetsuo [National Inst. for Fusion Science, Toki, Gifu (Japan)

    2001-09-01

    The General Purpose Use Computer System of National Institute for Fusion Science was replaced in January, 2001. The System is almost fully used after the first three months operation. Reported here is the process of the introduction of the new system and the state of the operation and utilization of the System between January and March, 2001, especially the detailed utilization of March. (author)

  5. Some thermo-electromagnetic applications to fusion technology of a general purpose CAD package

    International Nuclear Information System (INIS)

    Girdinio, P.; Molfino, P.; Molinari, G.; Raia, G.; Rosatelli, F.; Viviani, A.

    1985-01-01

    A general purpose CAD package is applied to the solution of problems related to fusion technology. The problems solved are the interacting electromagnetic and thermal fields in a resistive toroidal coil and the design of the poloidal field coils in Tokamak machines. In both cases, the procedure used is reported and the results obtained are displayed and discussed

  6. Some thermo-electromagnetic applications to fusion technology of a general purpose CAD package

    International Nuclear Information System (INIS)

    Girdinio, P.; Molfino, P.; Molinari, G.; Viviani, A.; Raia, G.; Rosatelli, F.

    1984-01-01

    A general purpose CAD package is applied to the solution of problems related to fusion technology. The problems solved are the interacting electromagnetic and thermal fields in a resistive toroidal coil and the design of the poloidal field coils in Tokamak machines. In both cases, the procedure used is reported and the results obtained are displayed and discussed. (author)

  7. An Evaluation of Classroom Activities and Exercises in ELT Classroom for General Purposes Course

    Science.gov (United States)

    Zohrabi, Mohammad

    2011-01-01

    It is through effective implementation of activities and exercises which students can be motivated and consequently lead to language learning. However, as an insider, the experience of teaching English for General Purposes (EGP) course indicates that it has some problems which need to be modified. In order to evaluate the EGP course,…

  8. Experience of application of the general-purpose pressure and pressure drop transformers on nitrogen tetroxide

    International Nuclear Information System (INIS)

    Grishchuk, M.Kh.

    1979-01-01

    An experience of application of the general-purpose pressure and pressure drop transformers at the Nuclear Power Engineering Institute of the BSSR Academy of Sciences for measurements on nitrogen tetroxide has been described. The concrete recommendations on the types of transformers and the volume of preparational work before putting them into operation have been given

  9. A general-purpose trigger processor system and its application to fast vertex trigger

    International Nuclear Information System (INIS)

    Hazumi, M.; Banas, E.; Natkaniec, Z.; Ostrowicz, W.

    1997-12-01

    A general-purpose hardware trigger system has been developed. The system comprises programmable trigger processors and pattern generator/samplers. The hardware design of the system is described. An application as a prototype of the very fast vertex trigger in an asymmetric B-factory at KEK is also explained. (author)

  10. A general purpose program system for high energy physics experiment data acquisition and analysis

    International Nuclear Information System (INIS)

    Li Shuren; Xing Yuguo; Jin Bingnian

    1985-01-01

    This paper introduced the functions, structure and system generation of a general purpose program system (Fermilab MULTI) for high energy physics experiment data acquisition and analysis. Works concerning the reconstruction of MULTI system level 0.5 which can be run on the computer PDP-11/23 are also introduced briefly

  11. Perbandingan Kemampuan Embedded Computer dengan General Purpose Computer untuk Pengolahan Citra

    Directory of Open Access Journals (Sweden)

    Herryawan Pujiharsono

    2017-08-01

    Full Text Available Perkembangan teknologi komputer membuat pengolahan citra saat ini banyak dikembangkan untuk dapat membantu manusia di berbagai bidang pekerjaan. Namun, tidak semua bidang pekerjaan dapat dikembangkan dengan pengolahan citra karena tidak mendukung penggunaan komputer sehingga mendorong pengembangan pengolahan citra dengan mikrokontroler atau mikroprosesor khusus. Perkembangan mikrokontroler dan mikroprosesor memungkinkan pengolahan citra saat ini dapat dikembangkan dengan embedded computer atau single board computer (SBC. Penelitian ini bertujuan untuk menguji kemampuan embedded computer dalam mengolah citra dan membandingkan hasilnya dengan komputer pada umumnya (general purpose computer. Pengujian dilakukan dengan mengukur waktu eksekusi dari empat operasi pengolahan citra yang diberikan pada sepuluh ukuran citra. Hasil yang diperoleh pada penelitian ini menunjukkan bahwa optimasi waktu eksekusi embedded computer lebih baik jika dibandingkan dengan general purpose computer dengan waktu eksekusi rata-rata embedded computer adalah 4-5 kali waktu eksekusi general purpose computer dan ukuran citra maksimal yang tidak membebani CPU terlalu besar untuk embedded computer adalah 256x256 piksel dan untuk general purpose computer adalah 400x300 piksel.

  12. Application of a general purpose finite element program system in pressure vessel technology

    International Nuclear Information System (INIS)

    Aamodt, B.; Sandsmark, N.; Medonos, S.

    1977-01-01

    Main advantages of using general purpose finite element program systems in structural analysis are summarized. Several illustrative applications of the program system SESAM-69 to pressure vessel problems are described. The first example is a dynamic analysis of the motor housing of the internal main circulation pump of a BWR nuclear reactor. The next example is a transient heat conduction and stress analysis of deflector of feeding nozzle of PWR nuclear reactor. Then, numerical calculations of stress intensity factors and fatigue crack growth of semi-elliptical surface cracks are discussed. And finally, an elasto-plastic analysis of a thick plate with edge-cracks is considered. It is concluded that due to the fact that general purpose finite element program systems are general and user-orientated, they will gain increasingly higher popularity in the years ahead

  13. Thresholds of surface codes on the general lattice structures suffering biased error and loss

    International Nuclear Information System (INIS)

    Tokunaga, Yuuki; Fujii, Keisuke

    2014-01-01

    A family of surface codes with general lattice structures is proposed. We can control the error tolerances against bit and phase errors asymmetrically by changing the underlying lattice geometries. The surface codes on various lattices are found to be efficient in the sense that their threshold values universally approach the quantum Gilbert-Varshamov bound. We find that the error tolerance of the surface codes depends on the connectivity of the underlying lattices; the error chains on a lattice of lower connectivity are easier to correct. On the other hand, the loss tolerance of the surface codes exhibits an opposite behavior; the logical information on a lattice of higher connectivity has more robustness against qubit loss. As a result, we come upon a fundamental trade-off between error and loss tolerances in the family of surface codes with different lattice geometries

  14. Thresholds of surface codes on the general lattice structures suffering biased error and loss

    Energy Technology Data Exchange (ETDEWEB)

    Tokunaga, Yuuki [NTT Secure Platform Laboratories, NTT Corporation, 3-9-11 Midori-cho, Musashino, Tokyo 180-8585, Japan and Japan Science and Technology Agency, CREST, 5 Sanban-cho, Chiyoda-ku, Tokyo 102-0075 (Japan); Fujii, Keisuke [Graduate School of Engineering Science, Osaka University, Toyonaka, Osaka 560-8531 (Japan)

    2014-12-04

    A family of surface codes with general lattice structures is proposed. We can control the error tolerances against bit and phase errors asymmetrically by changing the underlying lattice geometries. The surface codes on various lattices are found to be efficient in the sense that their threshold values universally approach the quantum Gilbert-Varshamov bound. We find that the error tolerance of the surface codes depends on the connectivity of the underlying lattices; the error chains on a lattice of lower connectivity are easier to correct. On the other hand, the loss tolerance of the surface codes exhibits an opposite behavior; the logical information on a lattice of higher connectivity has more robustness against qubit loss. As a result, we come upon a fundamental trade-off between error and loss tolerances in the family of surface codes with different lattice geometries.

  15. Benchmark of physics design of a proposed 30 MW Multi Purpose Research Reactor using a Monte Carlo code MCNP

    International Nuclear Information System (INIS)

    Singh, Tej; Kumar, Jainendra; Sharma, Archana; Singh, Kanchhi; Raina, V.K.; Srinivasan, P.

    2009-01-01

    At present Dhruva and Cirus reactors provide majority of research reactor based experimental/irradiation facilities to cater to various needs of the vast pool of researchers in the field of sciences research and development work for nuclear power plants and production of radioisotopes. With a view to further consolidate and expand the scope of research and development in nuclear and allied sciences, a new 30 MWt Multi Purpose Research Reactor is proposed to be constructed. This paper describes some of the physics design features of this reactor using MCNP code to validate the deterministic methods. The criticality calculations for 100 material testing reactor (JHR) of France and 610 MW SAVANNAH thermal reactor were performed using MCNP computer codes to boost the confidence level in designing the physics design of reactor core. (author)

  16. 21 CFR 862.2050 - General purpose laboratory equipment labeled or promoted for a specific medical use.

    Science.gov (United States)

    2010-04-01

    ... 21 Food and Drugs 8 2010-04-01 2010-04-01 false General purpose laboratory equipment labeled or... TOXICOLOGY DEVICES Clinical Laboratory Instruments § 862.2050 General purpose laboratory equipment labeled or promoted for a specific medical use. (a) Identification. General purpose laboratory equipment labeled or...

  17. RUMD: A general purpose molecular dynamics package optimized to utilize GPU hardware down to a few thousand particles

    Directory of Open Access Journals (Sweden)

    Nicholas P. Bailey, Trond S. Ingebrigtsen, Jesper Schmidt Hansen, Arno A. Veldhorst, Lasse Bøhling, Claire A. Lemarchand, Andreas E. Olsen, Andreas K. Bacher, Lorenzo Costigliola, Ulf R. Pedersen, Heine Larsen, Jeppe C. Dyre, Thomas B. Schrøder

    2017-12-01

    Full Text Available RUMD is a general purpose, high-performance molecular dynamics (MD simulation package running on graphical processing units (GPU's. RUMD addresses the challenge of utilizing the many-core nature of modern GPU hardware when simulating small to medium system sizes (roughly from a few thousand up to hundred thousand particles. It has a performance that is comparable to other GPU-MD codes at large system sizes and substantially better at smaller sizes.RUMD is open-source and consists of a library written in C++ and the CUDA extension to C, an easy-to-use Python interface, and a set of tools for set-up and post-simulation data analysis. The paper describes RUMD's main features, optimizations and performance benchmarks.

  18. VoxelMages: a general-purpose graphical interface for designing geometries and processing DICOM images for PENELOPE.

    Science.gov (United States)

    Giménez-Alventosa, V; Ballester, F; Vijande, J

    2016-12-01

    The design and construction of geometries for Monte Carlo calculations is an error-prone, time-consuming, and complex step in simulations describing particle interactions and transport in the field of medical physics. The software VoxelMages has been developed to help the user in this task. It allows to design complex geometries and to process DICOM image files for simulations with the general-purpose Monte Carlo code PENELOPE in an easy and straightforward way. VoxelMages also allows to import DICOM-RT structure contour information as delivered by a treatment planning system. Its main characteristics, usage and performance benchmarking are described in detail. Copyright © 2016 Elsevier Ltd. All rights reserved.

  19. Human factors in equipment development for the Space Shuttle - A study of the general purpose work station

    Science.gov (United States)

    Junge, M. K.; Giacomi, M. J.

    1981-01-01

    The results of a human factors test to assay the suitability of a prototype general purpose work station (GPWS) for biosciences experiments on the fourth Spacelab mission are reported. The evaluation was performed to verify that users of the GPWS would optimally interact with the GPWS configuration and instrumentation. Six male subjects sat on stools positioned to allow assimilation of the zero-g body posture. Trials were run concerning the operator viewing angles facing the console, the console color, procedures for injecting rates with dye, a rat blood cell count, mouse dissection, squirrel monkey transfer, and plant fixation. The trials were run for several days in order to gage improvement or poor performance conditions. Better access to the work surface was found necessary, together with more distinct and better located LEDs, better access window latches, clearer sequences on control buttons, color-coded sequential buttons, and provisions made for an intercom system when operators of the GPWS work in tandem.

  20. Real-time radar signal processing using GPGPU (general-purpose graphic processing unit)

    Science.gov (United States)

    Kong, Fanxing; Zhang, Yan Rockee; Cai, Jingxiao; Palmer, Robert D.

    2016-05-01

    This study introduces a practical approach to develop real-time signal processing chain for general phased array radar on NVIDIA GPUs(Graphical Processing Units) using CUDA (Compute Unified Device Architecture) libraries such as cuBlas and cuFFT, which are adopted from open source libraries and optimized for the NVIDIA GPUs. The processed results are rigorously verified against those from the CPUs. Performance benchmarked in computation time with various input data cube sizes are compared across GPUs and CPUs. Through the analysis, it will be demonstrated that GPGPUs (General Purpose GPU) real-time processing of the array radar data is possible with relatively low-cost commercial GPUs.

  1. Discrete-Event Execution Alternatives on General Purpose Graphical Processing Units

    International Nuclear Information System (INIS)

    Perumalla, Kalyan S.

    2006-01-01

    Graphics cards, traditionally designed as accelerators for computer graphics, have evolved to support more general-purpose computation. General Purpose Graphical Processing Units (GPGPUs) are now being used as highly efficient, cost-effective platforms for executing certain simulation applications. While most of these applications belong to the category of time-stepped simulations, little is known about the applicability of GPGPUs to discrete event simulation (DES). Here, we identify some of the issues and challenges that the GPGPU stream-based interface raises for DES, and present some possible approaches to moving DES to GPGPUs. Initial performance results on simulation of a diffusion process show that DES-style execution on GPGPU runs faster than DES on CPU and also significantly faster than time-stepped simulations on either CPU or GPGPU.

  2. Design method of general-purpose driving circuit for CCD based on CPLD

    International Nuclear Information System (INIS)

    Zhang Yong; Tang Benqi; Xiao Zhigang; Wang Zujun; Huang Shaoyan

    2005-01-01

    It is very important for studying the radiation damage effects and mechanism systematically about CCD to develop a general-purpose test platform. The paper discusses the design method of general-purpose driving circuit for CCD based on CPLD and the realization approach. A main controller has being designed to read the data file from the outer memory, setup the correlative parameter registers and produce the driving pulses according with parameter request strictly, which is based on MAX7000S by using MAX-PLUS II software. The basic driving circuit module has being finished based on this method. The output waveform of the module is the same figure as the simulation waveform. The result indicates that the design method is feasible. (authors)

  3. General-purpose heat source project and space nuclear safety and fuels program. Progress report

    International Nuclear Information System (INIS)

    Maraman, W.J.

    1979-12-01

    This formal monthly report covers the studies related to the use of 238 PuO 2 in radioisotopic power systems carried out for the Advanced Nuclear Systems and Projects Division of the Los Alamos Scientific Laboratory. The two programs involved are general-purpose heat source development and space nuclear safety and fuels. Most of the studies discussed hear are of a continuing nature. Results and conclusions described may change as the work continues

  4. Comparison of progressive addition lenses for general purpose and for computer vision: an office field study.

    Science.gov (United States)

    Jaschinski, Wolfgang; König, Mirjam; Mekontso, Tiofil M; Ohlendorf, Arne; Welscher, Monique

    2015-05-01

    Two types of progressive addition lenses (PALs) were compared in an office field study: 1. General purpose PALs with continuous clear vision between infinity and near reading distances and 2. Computer vision PALs with a wider zone of clear vision at the monitor and in near vision but no clear distance vision. Twenty-three presbyopic participants wore each type of lens for two weeks in a double-masked four-week quasi-experimental procedure that included an adaptation phase (Weeks 1 and 2) and a test phase (Weeks 3 and 4). Questionnaires on visual and musculoskeletal conditions as well as preferences regarding the type of lenses were administered. After eight more weeks of free use of the spectacles, the preferences were assessed again. The ergonomic conditions were analysed from photographs. Head inclination when looking at the monitor was significantly lower by 2.3 degrees with the computer vision PALs than with the general purpose PALs. Vision at the monitor was judged significantly better with computer PALs, while distance vision was judged better with general purpose PALs; however, the reported advantage of computer vision PALs differed in extent between participants. Accordingly, 61 per cent of the participants preferred the computer vision PALs, when asked without information about lens design. After full information about lens characteristics and additional eight weeks of free spectacle use, 44 per cent preferred the computer vision PALs. On average, computer vision PALs were rated significantly better with respect to vision at the monitor during the experimental part of the study. In the final forced-choice ratings, approximately half of the participants preferred either the computer vision PAL or the general purpose PAL. Individual factors seem to play a role in this preference and in the rated advantage of computer vision PALs. © 2015 The Authors. Clinical and Experimental Optometry © 2015 Optometry Australia.

  5. A new general purpose event horizon finder for 3D numerical spacetimes

    International Nuclear Information System (INIS)

    Diener, Peter

    2003-01-01

    I present a new general purpose event horizon finder for full 3D numerical spacetimes. It works by evolving a complete null surface backwards in time. The null surface is described as the zero-level set of a scalar function, which in principle is defined everywhere. This description of the surface allows the surface, trivially, to change topology, making this event horizon finder able to handle numerical spacetimes where two (or more) black holes merge into a single final black hole

  6. A prepaid case study: Ready Credit’s general-purpose & transit-fare programs

    OpenAIRE

    Philip Keitel

    2012-01-01

    Today, prepaid cards are used in dozens of payment applications. To examine the most recent developments, the Payment Cards Center of the Federal Reserve Bank of Philadelphia hosted a workshop on August 22, 2011. Leading the workshop was Tim Walsh, president and chief executive officer of Ready Credit Corporation, a firm that developed network-branded prepaid cards for use in transit-fare systems and also markets general-purpose, reloadable prepaid cards to consumers. Walsh discussed the uniq...

  7. Design of General-purpose Industrial signal acquisition system in a large scientific device

    Science.gov (United States)

    Ren, Bin; Yang, Lei

    2018-02-01

    In order to measure the industrial signal of a large scientific device experiment, a set of industrial data general-purpose acquisition system has been designed. It can collect 4~20mA current signal and 0~10V voltage signal. Through the practical experiments, it shows that the system is flexible, reliable, convenient and economical, and the system has characters of high definition and strong anti-interference ability. Thus, the system fully meets the design requirements..

  8. Child first language and adult second language are both tied to general-purpose learning systems.

    Science.gov (United States)

    Hamrick, Phillip; Lum, Jarrad A G; Ullman, Michael T

    2018-02-13

    Do the mechanisms underlying language in fact serve general-purpose functions that preexist this uniquely human capacity? To address this contentious and empirically challenging issue, we systematically tested the predictions of a well-studied neurocognitive theory of language motivated by evolutionary principles. Multiple metaanalyses were performed to examine predicted links between language and two general-purpose learning systems, declarative and procedural memory. The results tied lexical abilities to learning only in declarative memory, while grammar was linked to learning in both systems in both child first language and adult second language, in specific ways. In second language learners, grammar was associated with only declarative memory at lower language experience, but with only procedural memory at higher experience. The findings yielded large effect sizes and held consistently across languages, language families, linguistic structures, and tasks, underscoring their reliability and validity. The results, which met the predicted pattern, provide comprehensive evidence that language is tied to general-purpose systems both in children acquiring their native language and adults learning an additional language. Crucially, if language learning relies on these systems, then our extensive knowledge of the systems from animal and human studies may also apply to this domain, leading to predictions that might be unwarranted in the more circumscribed study of language. Thus, by demonstrating a role for these systems in language, the findings simultaneously lay a foundation for potentially important advances in the study of this critical domain.

  9. Computer-assisted analyses of (/sup 14/C)2-DG autoradiographs employing a general purpose image processing system

    Energy Technology Data Exchange (ETDEWEB)

    Porro, C; Biral, G P [Modena Univ. (Italy). Ist. di Fisiologia Umana; Fonda, S; Baraldi, P [Modena Univ. (Italy). Lab. di Bioingegneria della Clinica Oculistica; Cavazzuti, M [Modena Univ. (Italy). Clinica Neurologica

    1984-09-01

    A general purpose image processing system is described including B/W TV camera, high resolution image processor and display system (TESAK VDC 501), computer (DEC PDP 11/23) and monochrome and color monitors. Images may be acquired from a microscope equipped with a TV camera or using the TV in direct viewing; the A/D converter and the image processor provides fast (40 ms) and precise (512x512 data points) digitization of TV signal with a 256 gray levels maximum resolution. Computer programs have been developed in order to perform qualitative and quantitative analyses of autoradiographs obtained with the 2-DG method, which are written in FORTRAN and MACRO 11 Assembly Language. They include: (1) procedures designed to recognize errors in acquisition due to possible image shading and correct them via software; (2) routines suitable for qualitative analyses of the whole image or selected regions of it, providing the opportunity for pseudocolor coding, statistics, graphic overlays; (3) programs permitting the conversion of gray levels into metabolic rates of glucose utilization and the display of gray- or color-coded metabolic maps.

  10. Minus 3: a general purpose data acquisition system at LBL's 88''-cyclotron and superhilac

    International Nuclear Information System (INIS)

    Maples, C.; Sivak, J.

    1979-05-01

    MINUS 3 is a general, multi-tasked data acquisition package operating on the ModComp IV/25 computers at both the 88''-Cyclotron and SuperHILAC. It currently can acquire data via three different channels: interrupt; serial DMA link; and remote slave units for histogram type data. Two additional acquisition paths, CAMAC (with programmable differential branch drivers) and MODACS (for multiple CPU linkages and control) are scheduled to be added in the near future. The package operates in a prioritized, time-available mode which permits it to dynamically adapt to microscopic data rate structures due to beam characteristics at different accelerators. Special hardware has been added to the graphics system to provide enhanced high-speed interactive capability. The program framework is also designed as a parasitic environment in which users may, in parallel, attach their own specialized and independent code

  11. General productivity code: productivity optimization of gaseous diffusion cascades. The programmer's guide

    International Nuclear Information System (INIS)

    Tunstall, J.N.

    1975-05-01

    The General Productivity Code is a FORTRAN IV computer program for the IBM System 360. With its model of the productivity of gaseous diffusion cascades, the program is used to determine optimum cascade performance based on specified operating conditions and to aid in the calculation of optimum operating conditions for a complex of diffusion cascades. This documentation of the program is directed primarily to programmers who will be responsible for updating the code as requested by the users. It is also intended to be an aid in training new Productivity Code users and to serve as a general reference manual. Elements of the mathematical model, the input data requirements, the definitions of the various tasks (Instructions) that can be performed, and a detailed description of most FORTRAN variables and program subroutines are presented. A sample problem is also included. (auth)

  12. Conversion of a general quantum stabilizer code to an entanglement distillation protocol

    Energy Technology Data Exchange (ETDEWEB)

    Matsumoto, Ryutaroh [Department of Communications and Integrated Systems, Tokyo Institute of Technology, Tokyo 152-8552 (Japan)

    2003-07-25

    We show how to convert a quantum stabilizer code to a one- or two-way entanglement distillation protocol. The proposed conversion method is a generalization of those of Shor-Preskill and Nielsen-Chuang. The recurrence protocol and the quantum privacy amplification protocol are equivalent to the protocols converted from [[2, 1

  13. Conversion of a general quantum stabilizer code to an entanglement distillation protocol

    International Nuclear Information System (INIS)

    Matsumoto, Ryutaroh

    2003-01-01

    We show how to convert a quantum stabilizer code to a one- or two-way entanglement distillation protocol. The proposed conversion method is a generalization of those of Shor-Preskill and Nielsen-Chuang. The recurrence protocol and the quantum privacy amplification protocol are equivalent to the protocols converted from [[2, 1

  14. SACRD: a data base for fast reactor safety computer codes, general description

    International Nuclear Information System (INIS)

    Greene, N.M.; Forsberg, V.M.; Raiford, G.B.; Arwood, J.W.; Simpson, D.B.; Flanagan, G.F.

    1979-01-01

    SACRD is a data base of material properties and other handbook data needed in computer codes used for fast reactor safety studies. Data are available in the thermodynamics, heat transfer, fluid mechanics, structural mechanics, aerosol transport, meteorology, neutronics, and dosimetry areas. Tabular, graphical and parameterized data are provided in many cases. A general description of the SACRD system is presented in the report

  15. General-purpose readout electronics for white neutron source at China Spallation Neutron Source.

    Science.gov (United States)

    Wang, Q; Cao, P; Qi, X; Yu, T; Ji, X; Xie, L; An, Q

    2018-01-01

    The under-construction White Neutron Source (WNS) at China Spallation Neutron Source is a facility for accurate measurements of neutron-induced cross section. Seven spectrometers are planned at WNS. As the physical objectives of each spectrometer are different, the requirements for readout electronics are not the same. In order to simplify the development of the readout electronics, this paper presents a general method for detector signal readout. This method has advantages of expansibility and flexibility, which makes it adaptable to most detectors at WNS. In the WNS general-purpose readout electronics, signals from any kinds of detectors are conditioned by a dedicated signal conditioning module corresponding to this detector, and then digitized by a common waveform digitizer with high speed and high precision (1 GSPS at 12-bit) to obtain the full waveform data. The waveform digitizer uses a field programmable gate array chip to process the data stream and trigger information in real time. PXI Express platform is used to support the functionalities of data readout, clock distribution, and trigger information exchange between digitizers and trigger modules. Test results show that the performance of the WNS general-purpose readout electronics can meet the requirements of the WNS spectrometers.

  16. Literature Review: Weldability of Iridium DOP-26 Alloy for General Purpose Heat Source

    Energy Technology Data Exchange (ETDEWEB)

    Burgardt, Paul [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Pierce, Stanley W. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2016-10-19

    The basic purpose of this paper is to provide a literature review relative to fabrication of the General Purpose Heat Source (GPHS) that is used to provide electrical power for deep space missions of NASA. The particular fabrication operation to be addressed here is arc welding of the GPHS encapsulation. A considerable effort was made to optimize the fabrication of the fuel pellets and of other elements of the encapsulation; that work will not be directly addressed in this paper. This report consists of three basic sections: 1) a brief description of the GPHS will be provided as background information for the reader; 2) mechanical properties and the optimization thereof as relevant to welding will be discussed; 3) a review of the arc welding process development and optimization will be presented. Since the welding equipment must be upgraded for future production, some discussion of the historical establishment of relevant welding variables and possible changes thereto will also be discussed.

  17. Proposal to change General Consideration 5 and Principle 2 of the International Code of Nomenclature of Prokaryotes.

    Science.gov (United States)

    Oren, Aharon; Garrity, George M

    2014-01-01

    A proposal is submitted to the ICSP to change the wording of General Consideration 5 of the International Code of Nomenclature of Prokaryotes (ICNP), deleting the words Schizophycetes, Cyanophyceae and Cyanobacteria from the groups of organisms whose nomenclature is covered by the Code. It is further proposed to change the terms Zoological Code and International Code of Botanical Nomenclature in General Consideration 5 and in Principle 2 to International Code of Zoological Nomenclature and International Code of Nomenclature for algae, fungi and plants, respectively.

  18. General-purpose chemical analyzer for on-line analyses of radioactive solutions

    International Nuclear Information System (INIS)

    Spencer, W.A.; Kronberg, J.W.

    1983-01-01

    An automated analyzer is being developed to perform analytical measurements on radioactive solutions on-line in a hostile environment. This General Purpose Chemical Analyzer (GPCA) samples a process stream, adds reagents, measures solution absorbances or electrode potentials, and automatically calculates the results. The use of modular components, under microprocessor control, permits a single analyzer design to carry out many types of analyses. This paper discusses the more important design criteria for the GPCA, and describes the equipment being tested in a prototype unit

  19. Software design of a general purpose data acquisition and control executive

    International Nuclear Information System (INIS)

    Labiak, W.G.; Minor, E.G.

    1981-01-01

    The software design of an executive which performs general purpose data acquisition, monitoring, and control is presented. The executive runs on a memory-based mini or micro-computer and communicates with a disk-based computer where data analysis and display are done. The executive design stresses reliability and versatility, and has yielded software which can provide control and monitoring for widely different hardware systems. Applications of this software on two major fusion energy experiments at Lawrence Livermore National Laboratory will be described

  20. A Real-Time Programmer's Tour of General-Purpose L4 Microkernels

    OpenAIRE

    Ruocco Sergio

    2008-01-01

    Abstract L4-embedded is a microkernel successfully deployed in mobile devices with soft real-time requirements. It now faces the challenges of tightly integrated systems, in which user interface, multimedia, OS, wireless protocols, and even software-defined radios must run on a single CPU. In this paper we discuss the pros and cons of L4-embedded for real-time systems design, focusing on the issues caused by the extreme speed optimisations it inherited from its general-purpose ancestors. Sinc...

  1. A Real-Time Programmer's Tour of General-Purpose L4 Microkernels

    OpenAIRE

    Sergio Ruocco

    2008-01-01

    L4-embedded is a microkernel successfully deployed in mobile devices with soft real-time requirements. It now faces the challenges of tightly integrated systems, in which user interface, multimedia, OS, wireless protocols, and even software-defined radios must run on a single CPU. In this paper we discuss the pros and cons of L4-embedded for real-time systems design, focusing on the issues caused by the extreme speed optimisations it inherited from its general-purpose ancestors. Since these i...

  2. General-purpose heat source project and space nuclear safety fuels program. Progress report, February 1980

    International Nuclear Information System (INIS)

    Maraman, W.J.

    1980-05-01

    This formal monthly report covers the studies related to the use of 238 PuO 2 in radioisotopic power systems carried out for the Advanced Nuclear Systems and Projects Division of the Los Alamos Scientific Laboratory. The two programs involved are: General-Purpose Heat Source Development and Space Nuclear Safety and Fuels. Most of the studies discussed here are of a continuing nature. Results and conclusions described may change as the work continues. Published reference to the results cited in this report should not be made without the explicit permission of the person in charge of the work

  3. Minimal-memory realization of pearl-necklace encoders of general quantum convolutional codes

    International Nuclear Information System (INIS)

    Houshmand, Monireh; Hosseini-Khayat, Saied

    2011-01-01

    Quantum convolutional codes, like their classical counterparts, promise to offer higher error correction performance than block codes of equivalent encoding complexity, and are expected to find important applications in reliable quantum communication where a continuous stream of qubits is transmitted. Grassl and Roetteler devised an algorithm to encode a quantum convolutional code with a ''pearl-necklace'' encoder. Despite their algorithm's theoretical significance as a neat way of representing quantum convolutional codes, it is not well suited to practical realization. In fact, there is no straightforward way to implement any given pearl-necklace structure. This paper closes the gap between theoretical representation and practical implementation. In our previous work, we presented an efficient algorithm to find a minimal-memory realization of a pearl-necklace encoder for Calderbank-Shor-Steane (CSS) convolutional codes. This work is an extension of our previous work and presents an algorithm for turning a pearl-necklace encoder for a general (non-CSS) quantum convolutional code into a realizable quantum convolutional encoder. We show that a minimal-memory realization depends on the commutativity relations between the gate strings in the pearl-necklace encoder. We find a realization by means of a weighted graph which details the noncommutative paths through the pearl necklace. The weight of the longest path in this graph is equal to the minimal amount of memory needed to implement the encoder. The algorithm has a polynomial-time complexity in the number of gate strings in the pearl-necklace encoder.

  4. Status of the general description of fission observables by the GEF code

    CERN Document Server

    Jurado, B

    2014-01-01

    The GEneral Fission (GEF) model treats spontaneous fission and fission up to an excitation energy of about 100 MeV of a wide range of heavy nuclei. GEF makes use of general laws of statistical and quantum mechanics, assuring a high predictive power. It is unique in providing a general description of essentially all fission observables in a consistent way while preserving the correlations between all of them. In this contribution we present some of the physical aspects on which the model is based, give an overview on the results that can be obtained with the code and show an example that illustrates how the GEF code can serve as a framework for revealing the sensitivity of the fission observables to some basic nuclear properties.

  5. Conceptual analyses of neutronic and equilibrium refueling parameters to develop a cost-effective multi-purpose pool-type research reactor using WIMSD and CITVAP codes

    Energy Technology Data Exchange (ETDEWEB)

    Hedayat, Afshin, E-mail: ahedayat@aeoi.org.ir

    2016-12-01

    Highlights: • Introducing a high-beneficent and low-cost multipurpose research reactor. • High technical documents and standard safety issues are introduced coherently. • High effective conceptual neutronic analyses and fuel management strategy. • Gaining high score design criteria and safety margins via 3-D core modeling. • Capacity and capability to produce all medical and industrial radioisotopes. - Abstract: In this paper, neutronic and equilibrium refueling parameters of a multi-purpose cost-effective research reactor have been studied and analyzed. It has been tried to provide periodic and long-term requirements of the irradiating applications coherently. The WIMSD5B and CITVAP codes are used to calculate neutronic parameters and simulate fuel management strategy. The used nuclear data, codes, and calculating methods have been severally benchmarked and verified, successfully. Fundamental concepts, design criteria, and safety issues are introduced and discussed, coherently. Design criteria are selected to gain the most economic benefits per capital costs via minimum required reactor power. Accurate, fast and simplified models have been tried for an integrated decision making and analyses using deterministic codes. Core management, power effects, fuel consumption and burn up effects, and also a complete simulation of the fuel management strategy are presented and analyzed. Results show that the supposed reactor core design can be promisingly suitable in accordance with the commercial multi-purpose irradiating applications. It also retains Operating Limits and Conditions (OLCs) due to standard safety issues, conservatively where safety parameters are calculated using best estimate tools. Such reactor core configuration and integrated refueling task can effectively enhance the Quality Assurance (QA) of the general irradiating applications of the current MTR within their power limits and corresponding OLCs.

  6. Power performance of the general-purpose heat source radioisotope thermoelectric generator

    International Nuclear Information System (INIS)

    Bennett, G.L.; Lombardo, J.J.; Rock, B.J.

    1986-01-01

    The General-Purpose Heat Source Radioisotope Thermoelectric Generator (GRHS-RTG) has been developed under the sponsorship of the Department of Energy (DOE) to provide electrical power for the National Aeronautics and Space Administration (NASA) Galileo mission to Jupiter and the joint NASA/European Space Agency (ESA) Ulysses mission to study the polar regions of the sun. A total of five nuclear-heated generators and one electrically heated generator have been built and tested, proving out the design concept and meeting the specification requirements. The GPHS-RTG design is built upon the successful-technology used in the RTGs flown on the two NASA Voyager spacecraft and two US Air Force communications satellites. THe GPHS-RTG converts about 4400 W(t) from the nuclear heat source into at least 285 W(e) at beginning of mission (BOM). The GPHS-RTG consists of two major components: the General-Purpose Heat Source (GPHS) and the Converter. A conceptual drawing of the GPHs-RTG is presented and its design and performance are described

  7. Economic selection index development for Beefmaster cattle II: General-purpose breeding objective.

    Science.gov (United States)

    Ochsner, K P; MacNeil, M D; Lewis, R M; Spangler, M L

    2017-05-01

    An economic selection index was developed for Beefmaster cattle in a general-purpose production system in which bulls are mated to a combination of heifers and mature cows, with resulting progeny retained as replacements or sold at weaning. National average prices from 2010 to 2014 were used to establish income and expenses for the system. Genetic parameters were obtained from the literature. Economic values were estimated by simulating 100,000 animals and approximating the partial derivatives of the profit function by perturbing traits 1 at a time, by 1 unit, while holding the other traits constant at their respective means. Relative economic values for the objective traits calving difficultly direct (CDd), calving difficulty maternal (CDm), weaning weight direct (WWd), weaning weight maternal (WWm), mature cow weight (MW), and heifer pregnancy (HP) were -2.11, -1.53, 18.49, 11.28, -33.46, and 1.19, respectively. Consequently, under the scenario assumed herein, the greatest improvements in profitability could be made by decreasing maintenance energy costs associated with MW followed by improvements in weaning weight. The accuracy of the index lies between 0.218 (phenotypic-based index selection) and 0.428 (breeding values known without error). Implementation of this index would facilitate genetic improvement and increase profitability of Beefmaster cattle operations with a general-purpose breeding objective when replacement females are retained and with weaned calves as the sale end point.

  8. General Purpose Data-Driven Online System Health Monitoring with Applications to Space Operations

    Science.gov (United States)

    Iverson, David L.; Spirkovska, Lilly; Schwabacher, Mark

    2010-01-01

    Modern space transportation and ground support system designs are becoming increasingly sophisticated and complex. Determining the health state of these systems using traditional parameter limit checking, or model-based or rule-based methods is becoming more difficult as the number of sensors and component interactions grows. Data-driven monitoring techniques have been developed to address these issues by analyzing system operations data to automatically characterize normal system behavior. System health can be monitored by comparing real-time operating data with these nominal characterizations, providing detection of anomalous data signatures indicative of system faults, failures, or precursors of significant failures. The Inductive Monitoring System (IMS) is a general purpose, data-driven system health monitoring software tool that has been successfully applied to several aerospace applications and is under evaluation for anomaly detection in vehicle and ground equipment for next generation launch systems. After an introduction to IMS application development, we discuss these NASA online monitoring applications, including the integration of IMS with complementary model-based and rule-based methods. Although the examples presented in this paper are from space operations applications, IMS is a general-purpose health-monitoring tool that is also applicable to power generation and transmission system monitoring.

  9. Operation of general purpose stepping motor controllers at the National Synchrotron Light Source

    International Nuclear Information System (INIS)

    Stubblefield, F.W.

    1986-10-01

    A prototype and four copies of a general purpose subsystem for mechanical positioning of detectors, samples, and beam line optical elements which constitute experiments at the National Synchrotron Light Source facility of Brookhaven National Laboratory have been constructed and placed into operation. Construction of a sixth subsystem is nearing completion. The subsystems effect mechanical positioning by controlling a set of stepping motors and their associated position encoders. The units are general purpose in the sense that they receive commands over a standard 9600 baud asynchronous serial line compatible with the RS-232-C electrical signal standard, generate TTL-compatible streams of stepping pulses which can be used with a wide variety of stepping motors, and read back position values from a number of different types and models of position encoder. The basic structure of the motor controller subsystem will be briefly reviewed. Short descriptions of the positioning apparatus actuated at each of the test and experiment stations employing a motor control unit are given. Additions and enhancements to the subsystem made in response to problems indicated by actual operation of the four installed units are described in more detail

  10. Operation of general purpose stepping motor controllers at the National Synchrotron Light Source

    International Nuclear Information System (INIS)

    Stubblefield, F.W.

    1987-01-01

    A prototype and four copies of a general purpose subsystem for mechanical positioning of detectors, samples, and beam line optical elements which constitute experiments at the National Synchrotron Light Source facility of Brookhaven National Laboratory have been constructed and placed into operation. Construction of a sixth subsystem is nearing completion. The subsystems effect mechanical positioning by controlling a set of stepping motors and their associated position encoders. The units are general purpose in the sense that they receive commands over a standard 9600 baud asynchronous serial line compatible with the RS-232-C electrical signal standard, generate TTL-compatible streams of stepping pulses which can be used with a wide variety of stepping motors, and read back position values from a number of different types and models of position encoder. The basic structure of the motor controller subsystem is briefly reviewed. Short descriptions of the positioning apparatus actuated at each of the test and experiment stations employing a motor control unit are given. Additions and enhancements to the sub-system made in response to problems indicated by actual operation of the four installed units are described in more detail

  11. Knowledge Management Systems as an Interdisciplinary Communication and Personalized General-Purpose Technology

    Directory of Open Access Journals (Sweden)

    Ulrich Schmitt

    2015-10-01

    Full Text Available As drivers of human civilization, Knowledge Management (KM processes have co-evolved in line with General-Purpose-Technologies (GPT, such as writing, printing, and information and communication systems. As evidenced by the recent shift from information scarcity to abundance, GPTs are capable of drastically altering societies due to their game-changing impact on our spheres of work and personal development. This paper looks at the prospect of whether a novel Personal Knowledge Management (PKM concept supported by a prototype system has got what it takes to grow into a transformative General-Purpose-Technology. Following up on a series of papers, the KM scenario of a decentralizing revolution where individuals and self-organized groups yield more power and autonomy is examined according to a GPT's essential characteristics, including a wide scope for improvement and elaboration (in people's private, professional and societal life, applicability across a broad range of uses in a wide variety of products and processes (in multi-disciplinary educational and work contexts, and strong complementarities with existing or potential new technologies (like organizational KM Systems and a proposed World Heritage of Memes Repository. The result portrays the PKM concept as a strong candidate due to its personal, autonomous, bottom-up, collaborative, interdisciplinary, and creativity-supporting approach destined to advance the availability, quantity, and quality of the world extelligence and to allow for a wider sharing and faster diffusion of ideas across current disciplinary and opportunity divides.

  12. Comparison of the General Electric BWR/6 standard plant design to the IAEA NUSS codes and guides

    International Nuclear Information System (INIS)

    D'Ardenne, W.H.; Sherwood, G.G.

    1985-01-01

    The General Electric BWR/6 Mark III standard plant design meets or exceeds current requirements of published International Atomic Energy Agency (IAEA) Nuclear Safety Standards (NUSS) codes and guides. This conclusion is based on a review of the NUSS codes and guides by General Electric and by the co-ordinated US review of the NUSS codes and guides during their development. General Electric compared the published IAEA NUSS codes and guides with the General Electric design. The applicability of each code and guide to the BWR/6 Mark III standard plant design was determined. Each code or guide was reviewed by a General Electric engineer knowledgeable about the structures, systems and components addressed and the technical area covered by that code or guide. The results of this review show that the BWR/6 Mark III standard plant design meets or exceeds the applicable requirements of the published IAEA NUSS codes and guides. The co-ordinated US review of the IAEA NUSS codes and guides corroborates the General Electric review. In the co-ordinated US review, the USNRC and US industry organizations (including General Electric) review the NUSS codes and guides during their development. This review ensures that the NUSS codes and guides are consistent with the current US government regulations, guidance and regulatory practices, US voluntary industry codes and standards, and accepted US industry design, construction and operational practices. If any inconsistencies are identified, comments are submitted to the IAEA by the USNRC. All US concerns submitted to the IAEA have been resolved. General Electric design reviews and the Final Design Approval (FDA) issued by the USNRC have verified that the General Electric BWR/6 Mark III standard plant design meets or exceeds the current US requirements, guidance and practices. Since these requirements, guidance and practices meet or exceed those of the NUSS codes and guides, so does the General Electric design. (author)

  13. Development of a general coupling interface for the fuel performance code TRANSURANUS – Tested with the reactor dynamics code DYN3D

    International Nuclear Information System (INIS)

    Holt, L.; Rohde, U.; Seidl, M.; Schubert, A.; Van Uffelen, P.; Macián-Juan, R.

    2015-01-01

    Highlights: • A general coupling interface was developed for couplings of the TRANSURANUS code. • With this new tool simplified fuel behavior models in codes can be replaced. • Applicable e.g. for several reactor types and from normal operation up to DBA. • The general coupling interface was applied to the reactor dynamics code DYN3D. • The new coupled code system DYN3D–TRANSURANUS was successfully tested for RIA. - Abstract: A general interface is presented for coupling the TRANSURANUS fuel performance code with thermal hydraulics system, sub-channel thermal hydraulics, computational fluid dynamics (CFD) or reactor dynamics codes. As first application the reactor dynamics code DYN3D was coupled at assembly level in order to describe the fuel behavior in more detail. In the coupling, DYN3D provides process time, time-dependent rod power and thermal hydraulics conditions to TRANSURANUS, which in case of the two-way coupling approach transfers parameters like fuel temperature and cladding temperature back to DYN3D. Results of the coupled code system are presented for the reactivity transient scenario, initiated by control rod ejection. More precisely, the two-way coupling approach systematically calculates higher maximum values for the node fuel enthalpy. These differences can be explained thanks to the greater detail in fuel behavior modeling. The numerical performance for DYN3D–TRANSURANUS was proved to be fast and stable. The coupled code system can therefore improve the assessment of safety criteria, at a reasonable computational cost

  14. A randomised comparison between an inexpensive, general-purpose headlight and a purpose-built surgical headlight on users' visual acuity and colour vision.

    Science.gov (United States)

    Street, I; Sayles, M; Nistor, M; McRae, A R

    2014-02-01

    To determine if there are any differences in near visual acuity and colour vision between an inexpensive general-purpose light emitting diode (LED) headlight and a purpose-built surgical LED headlight. A prospective study was conducted sequentially comparing near visual acuity and colour vision, the headlights being tested in random order, in a testing room with a constant minimal amount of background light. The participants were NHS employee volunteers, with self-declared normal (or corrected) vision, working in occupations requiring full literacy. For visual acuity, outcome was measured by recording the smallest font legible when using each headlight when the subject read a near visual acuity test card. For colour vision, the outcome was passing or failing the Ishihara test. There was no statistically significant difference between the general-purpose and the purpose-built headlights in users' near visual acuity or colour vision.

  15. A general concurrent algorithm for plasma particle-in-cell simulation codes

    International Nuclear Information System (INIS)

    Liewer, P.C.; Decyk, V.K.

    1989-01-01

    We have developed a new algorithm for implementing plasma particle-in-cell (PIC) simulation codes on concurrent processors with distributed memory. This algorithm, named the general concurrent PIC algorithm (GCPIC), has been used to implement an electrostatic PIC code on the 33-node JPL Mark III Hypercube parallel computer. To decompose at PIC code using the GCPIC algorithm, the physical domain of the particle simulation is divided into sub-domains, equal in number to the number of processors, such that all sub-domains have roughly equal numbers of particles. For problems with non-uniform particle densities, these sub-domains will be of unequal physical size. Each processor is assigned a sub-domain and is responsible for updating the particles in its sub-domain. This algorithm has led to a a very efficient parallel implementation of a well-benchmarked 1-dimensional PIC code. The dominant portion of the code, updating the particle positions and velocities, is nearly 100% efficient when the number of particles is increased linearly with the number of hypercube processors used so that the number of particles per processor is constant. For example, the increase in time spent updating particles in going from a problem with 11,264 particles run on 1 processor to 360,448 particles on 32 processors was only 3% (parallel efficiency of 97%). Although implemented on a hypercube concurrent computer, this algorithm should also be efficient for PIC codes on other parallel architectures and for large PIC codes on sequential computers where part of the data must reside on external disks. copyright 1989 Academic Press, Inc

  16. Design of a general-purpose European compound screening library for EU-OPENSCREEN.

    Science.gov (United States)

    Horvath, Dragos; Lisurek, Michael; Rupp, Bernd; Kühne, Ronald; Specker, Edgar; von Kries, Jens; Rognan, Didier; Andersson, C David; Almqvist, Fredrik; Elofsson, Mikael; Enqvist, Per-Anders; Gustavsson, Anna-Lena; Remez, Nikita; Mestres, Jordi; Marcou, Gilles; Varnek, Alexander; Hibert, Marcel; Quintana, Jordi; Frank, Ronald

    2014-10-01

    This work describes a collaborative effort to define and apply a protocol for the rational selection of a general-purpose screening library, to be used by the screening platforms affiliated with the EU-OPENSCREEN initiative. It is designed as a standard source of compounds for primary screening against novel biological targets, at the request of research partners. Given the general nature of the potential applications of this compound collection, the focus of the selection strategy lies on ensuring chemical stability, absence of reactive compounds, screening-compliant physicochemical properties, loose compliance to drug-likeness criteria (as drug design is a major, but not exclusive application), and maximal diversity/coverage of chemical space, aimed at providing hits for a wide spectrum of drugable targets. Finally, practical availability/cost issues cannot be avoided. The main goal of this publication is to inform potential future users of this library about its conception, sources, and characteristics. The outline of the selection procedure, notably of the filtering rules designed by a large committee of European medicinal chemists and chemoinformaticians, may be of general methodological interest for the screening/medicinal chemistry community. The selection task of 200K molecules out of a pre-filtered set of 1.4M candidates was shared by five independent European research groups, each picking a subset of 40K compounds according to their own in-house methodology and expertise. An in-depth analysis of chemical space coverage of the library serves not only to characterize the collection, but also to compare the various chemoinformatics-driven selection procedures of maximal diversity sets. Compound selections contributed by various participating groups were mapped onto general-purpose self-organizing maps (SOMs) built on the basis of marketed drugs and bioactive reference molecules. In this way, the occupancy of chemical space by the EU-OPENSCREEN library could

  17. A Real-Time Programmer's Tour of General-Purpose L4 Microkernels

    Directory of Open Access Journals (Sweden)

    Ruocco Sergio

    2008-01-01

    Full Text Available Abstract L4-embedded is a microkernel successfully deployed in mobile devices with soft real-time requirements. It now faces the challenges of tightly integrated systems, in which user interface, multimedia, OS, wireless protocols, and even software-defined radios must run on a single CPU. In this paper we discuss the pros and cons of L4-embedded for real-time systems design, focusing on the issues caused by the extreme speed optimisations it inherited from its general-purpose ancestors. Since these issues can be addressed with a minimal performance loss, we conclude that, overall, the design of real-time systems based on L4-embedded is possible, and facilitated by a number of design features unique to microkernels and the L4 family.

  18. Generic functional requirements for a NASA general-purpose data base management system

    Science.gov (United States)

    Lohman, G. M.

    1981-01-01

    Generic functional requirements for a general-purpose, multi-mission data base management system (DBMS) for application to remotely sensed scientific data bases are detailed. The motivation for utilizing DBMS technology in this environment is explained. The major requirements include: (1) a DBMS for scientific observational data; (2) a multi-mission capability; (3) user-friendly; (4) extensive and integrated information about data; (5) robust languages for defining data structures and formats; (6) scientific data types and structures; (7) flexible physical access mechanisms; (8) ways of representing spatial relationships; (9) a high level nonprocedural interactive query and data manipulation language; (10) data base maintenance utilities; (11) high rate input/output and large data volume storage; and adaptability to a distributed data base and/or data base machine configuration. Detailed functions are specified in a top-down hierarchic fashion. Implementation, performance, and support requirements are also given.

  19. A General Purpose Feature Extractor for Light Detection and Ranging Data

    Directory of Open Access Journals (Sweden)

    Edwin B. Olson

    2010-11-01

    Full Text Available Feature extraction is a central step of processing Light Detection and Ranging (LIDAR data. Existing detectors tend to exploit characteristics of specific environments: corners and lines from indoor (rectilinear environments, and trees from outdoor environments. While these detectors work well in their intended environments, their performance in different environments can be poor. We describe a general purpose feature detector for both 2D and 3D LIDAR data that is applicable to virtually any environment. Our method adapts classic feature detection methods from the image processing literature, specifically the multi-scale Kanade-Tomasi corner detector. The resulting method is capable of identifying highly stable and repeatable features at a variety of spatial scales without knowledge of environment, and produces principled uncertainty estimates and corner descriptors at same time. We present results on both software simulation and standard datasets, including the 2D Victoria Park and Intel Research Center datasets, and the 3D MIT DARPA Urban Challenge dataset.

  20. A general purpose feature extractor for light detection and ranging data.

    Science.gov (United States)

    Li, Yangming; Olson, Edwin B

    2010-01-01

    Feature extraction is a central step of processing Light Detection and Ranging (LIDAR) data. Existing detectors tend to exploit characteristics of specific environments: corners and lines from indoor (rectilinear) environments, and trees from outdoor environments. While these detectors work well in their intended environments, their performance in different environments can be poor. We describe a general purpose feature detector for both 2D and 3D LIDAR data that is applicable to virtually any environment. Our method adapts classic feature detection methods from the image processing literature, specifically the multi-scale Kanade-Tomasi corner detector. The resulting method is capable of identifying highly stable and repeatable features at a variety of spatial scales without knowledge of environment, and produces principled uncertainty estimates and corner descriptors at same time. We present results on both software simulation and standard datasets, including the 2D Victoria Park and Intel Research Center datasets, and the 3D MIT DARPA Urban Challenge dataset.

  1. Edge corrections to electromagnetic Casimir energies from general-purpose Mathieu-function routines

    Science.gov (United States)

    Blose, Elizabeth Noelle; Ghimire, Biswash; Graham, Noah; Stratton-Smith, Jeremy

    2015-01-01

    Scattering theory methods make it possible to calculate the Casimir energy of a perfectly conducting elliptic cylinder opposite a perfectly conducting plane in terms of Mathieu functions. In the limit of zero radius, the elliptic cylinder becomes a finite-width strip, which allows for the study of edge effects. However, existing packages for computing Mathieu functions are insufficient for this calculation because none can compute Mathieu functions of both the first and second kind for complex arguments. To address this shortcoming, we have written a general-purpose Mathieu-function package, based on algorithms developed by Alhargan. We use these routines to find edge corrections to the proximity force approximation for the Casimir energy of a perfectly conducting strip opposite a perfectly conducting plane.

  2. A General Purpose Connections type CTI Server Based on SIP Protocol and Its Implementation

    Science.gov (United States)

    Watanabe, Toru; Koizumi, Hisao

    In this paper, we propose a general purpose connections type CTI (Computer Telephony Integration) server that provides various CTI services such as voice logging where the CTI server communicates with IP-PBX using the SIP (Session Initiation Protocol), and accumulates voice packets of external line telephone call flowing between an IP telephone for extension and a VoIP gateway connected to outside line networks. The CTI server realizes CTI services such as voice logging, telephone conference, or IVR (interactive voice response) with accumulating and processing voice packets sampled. Furthermore, the CTI server incorporates a web server function which can provide various CTI services such as a Web telephone directory via a Web browser to PCs, cellular telephones or smart-phones in mobile environments.

  3. The General-Purpose Heat Source Radioisotope Thermoelectric Generator: Power for the Galileo and Ulysses missions

    International Nuclear Information System (INIS)

    Bennett, G.L.; Lombardo, J.J.; Hemler, R.J.; Peterson, J.R.

    1986-01-01

    Electrical power for NASA's Galileo mission to Jupiter and ESA's Ulysses mission to explore the polar regions of the Sun will be provided by General-Purpose Heat Source Radioisotope Thermo-electric Generators (GPHS-RTGs). Building upon the successful RTG technology used in the Voyager program, each GPHS-RTG will provide at least 285 W(e) at beginning-of-mission. The design concept has been proven through extensive tests of an electrically heated Engineering Unit and a nuclear-heated Qualification Unit. Four flight generators have been successfully assembled and tested for use on the Galileo and Ulysses spacecraft. All indications are that the GPHS-RTGs will meet or exceed the power requirement of the missions

  4. Development of general-purpose software to analyze the static thermal characteristic of nuclear power plant

    International Nuclear Information System (INIS)

    Nakao, Yoshinobu; Koda, Eiichi; Takahashi, Toru

    2009-01-01

    We have developed the general-purpose software by which static thermal characteristic of the power generation system is analyzed easily. This software has the notable features as follows. It has the new algorithm to solve non-linear simultaneous equations to analyze the static thermal characteristics such as heat and mass balance, efficiencies, etc. of various power generation systems. It has the flexibility for setting calculation conditions. It is able to be executed on the personal computer easily and quickly. We ensured that it is able to construct heat and mass balance diagrams of main steam system of nuclear power plant and calculate the power output and efficiencies of the system. Furthermore, we evaluated various heat recovery measures of steam generator blowdown water and found that this software could be a useful operation aid for planning effective changes in support of power stretch. (author)

  5. The ICVSIE: A General Purpose Integral Equation Method for Bio-Electromagnetic Analysis.

    Science.gov (United States)

    Gomez, Luis J; Yucel, Abdulkadir C; Michielssen, Eric

    2018-03-01

    An internally combined volume surface integral equation (ICVSIE) for analyzing electromagnetic (EM) interactions with biological tissue and wide ranging diagnostic, therapeutic, and research applications, is proposed. The ICVSIE is a system of integral equations in terms of volume and surface equivalent currents in biological tissue subject to fields produced by externally or internally positioned devices. The system is created by using equivalence principles and solved numerically; the resulting current values are used to evaluate scattered and total electric fields, specific absorption rates, and related quantities. The validity, applicability, and efficiency of the ICVSIE are demonstrated by EM analysis of transcranial magnetic stimulation, magnetic resonance imaging, and neuromuscular electrical stimulation. Unlike previous integral equations, the ICVSIE is stable regardless of the electric permittivities of the tissue or frequency of operation, providing an application-agnostic computational framework for EM-biomedical analysis. Use of the general purpose and robust ICVSIE permits streamlining the development, deployment, and safety analysis of EM-biomedical technologies.

  6. Real-time traffic sign recognition based on a general purpose GPU and deep-learning.

    Science.gov (United States)

    Lim, Kwangyong; Hong, Yongwon; Choi, Yeongwoo; Byun, Hyeran

    2017-01-01

    We present a General Purpose Graphics Processing Unit (GPGPU) based real-time traffic sign detection and recognition method that is robust against illumination changes. There have been many approaches to traffic sign recognition in various research fields; however, previous approaches faced several limitations when under low illumination or wide variance of light conditions. To overcome these drawbacks and improve processing speeds, we propose a method that 1) is robust against illumination changes, 2) uses GPGPU-based real-time traffic sign detection, and 3) performs region detecting and recognition using a hierarchical model. This method produces stable results in low illumination environments. Both detection and hierarchical recognition are performed in real-time, and the proposed method achieves 0.97 F1-score on our collective dataset, which uses the Vienna convention traffic rules (Germany and South Korea).

  7. A low cost general purpose portable programmable master/slave manipulative appliance

    International Nuclear Information System (INIS)

    Cameron, W.

    1984-01-01

    The TRIUMF 100 μA 500 MeV cyclotron, located at the University of British Columbia, required a low cost, portable master/slave manipulative capability for experimental beam line servicing. A programmable capability was also required for the hot cell manipulators. A general purpose unit was developed that might also have applications in light manufacturing and medical rehabilitation. The project now in prototype testing represents a modular portable robot costing less than $5000 that is lead-through-teach programmable by either a master controller or hands-on lead-through. Task programs are stored and retrieved on any 32 k personal computer. An on-board proportional integral derivative controller (Motorola 6809 based) gives discrete positioning of the six degrees of freedom 2 kg capacity end effector

  8. Simrank: Rapid and sensitive general-purpose k-mer search tool

    Energy Technology Data Exchange (ETDEWEB)

    DeSantis, T.Z.; Keller, K.; Karaoz, U.; Alekseyenko, A.V; Singh, N.N.S.; Brodie, E.L; Pei, Z.; Andersen, G.L; Larsen, N.

    2011-04-01

    Terabyte-scale collections of string-encoded data are expected from consortia efforts such as the Human Microbiome Project (http://nihroadmap.nih.gov/hmp). Intra- and inter-project data similarity searches are enabled by rapid k-mer matching strategies. Software applications for sequence database partitioning, guide tree estimation, molecular classification and alignment acceleration have benefited from embedded k-mer searches as sub-routines. However, a rapid, general-purpose, open-source, flexible, stand-alone k-mer tool has not been available. Here we present a stand-alone utility, Simrank, which allows users to rapidly identify database strings the most similar to query strings. Performance testing of Simrank and related tools against DNA, RNA, protein and human-languages found Simrank 10X to 928X faster depending on the dataset. Simrank provides molecular ecologists with a high-throughput, open source choice for comparing large sequence sets to find similarity.

  9. A Real-Time Programmer's Tour of General-Purpose L4 Microkernels

    Directory of Open Access Journals (Sweden)

    Sergio Ruocco

    2008-02-01

    Full Text Available L4-embedded is a microkernel successfully deployed in mobile devices with soft real-time requirements. It now faces the challenges of tightly integrated systems, in which user interface, multimedia, OS, wireless protocols, and even software-defined radios must run on a single CPU. In this paper we discuss the pros and cons of L4-embedded for real-time systems design, focusing on the issues caused by the extreme speed optimisations it inherited from its general-purpose ancestors. Since these issues can be addressed with a minimal performance loss, we conclude that, overall, the design of real-time systems based on L4-embedded is possible, and facilitated by a number of design features unique to microkernels and the L4 family.

  10. General-purpose stepping motor-encoder positioning subsystem with standard asynchronous serial-line interface

    International Nuclear Information System (INIS)

    Stubblefield, F.W.; Alberi, J.L.

    1982-01-01

    A general-purpose mechanical positioning subsystem for open-loop control of experiment devices which have their positions established and read out by stepping motor-encoder combinations has been developed. The subsystem is to be used mainly for experiments to be conducted at the National Synchrotron Light Source at Brookhaven National Laboratory. The subsystem unit has been designed to be compatible with a wide variety of stepping motor and encoder types. The unit may be operated by any device capable of driving a standard RS-232-C asynchronous serial communication line. An informal survey has shown that several experiments at the Light Source will use one particular type of computer, operating system, and programming language. Accordingly, a library of subroutines compatible with this combination of computer system elements has been written to facilitate driving the positioning subsystem unit

  11. Development and application of General Purpose Data Acquisition Shell (GPDAS) at advanced photon source

    International Nuclear Information System (INIS)

    Chung, Youngjoo; Kim, Keeman.

    1991-01-01

    An operating system shell GPDAS (General Purpose Data Acquisition Shell) on MS-DOS-based microcomputers has been developed to provide flexibility in data acquisition and device control for magnet measurements at the Advanced Photon Source. GPDAS is both a command interpreter and an integrated script-based programming environment. It also incorporates the MS-DOS shell to make use of the existing utility programs for file manipulation and data analysis. Features include: alias definition, virtual memory, windows, graphics, data and procedure backup, background operation, script programming language, and script level debugging. Data acquisition system devices can be controlled through IEEE488 board, multifunction I/O board, digital I/O board and Gespac crate via Euro G-64 bus. GPDAS is now being used for diagnostics R ampersand D and accelerator physics studies as well as for magnet measurements. Their hardware configurations will also be discussed. 3 refs., 3 figs

  12. Developing wearable bio-feedback systems: a general-purpose platform.

    Science.gov (United States)

    Bianchi, Luigi; Babiloni, Fabio; Cincotti, Febo; Arrivas, Marco; Bollero, Patrizio; Marciani, Maria Grazia

    2003-06-01

    Microprocessors, even those in PocketPCs, have adequate power for many real-time biofeedback applications for disabled people. This power allows design of portable or wearable devices that are smaller and lighter, and that have longer battery life compared to notebook-based systems. In this paper, we discuss a general-purpose hardware/software solution based on industrial or consumer devices and a C++ framework. Its flexibility and modularity make it adaptable to a wide range of situations. Moreover, its design minimizes system requirements and programming effort, thus allowing efficient systems to be built quickly and easily. Our design has been used to build two brain computer interface systems that were easily ported from the Win32 platform.

  13. General-purpose heat source safety verification test series: SVT-11 through SVT-13

    International Nuclear Information System (INIS)

    George, T.G.; Pavone, D.

    1986-05-01

    The General-Purpose Heat Source (GPHS) is a modular component of the radioisotope thermoelectric generator that will provide power for the Galileo and Ulysses (formerly ISPM) space missions. The GPHS provides power by transmitting the heat of 238 Pu α-decay to an array of thermoelectric elements. Because the possibility of an orbital abort always exists, the heat source was designed and constructed to minimize plutonia release in any accident environment. The Safety Verification Test (SVT) series was formulated to evaluate the effectiveness of GPHS plutonia containment after atmospheric reentry and Earth impact. The first two reports (covering SVT-1 through SVT-10) described the results of flat, side-on, and angular module impacts against steel targets at 54 m/s. This report describes flat-on module impacts against concrete and granite targets, at velocities equivalent to or higher than previous SVTs

  14. ''Sheiva'' : a general purpose multi-parameter data acquisition and processing system at VECC

    International Nuclear Information System (INIS)

    Viyogi, Y.P.; Ganguly, N.K.

    1982-01-01

    A general purpose interactive software to be used with the PDP-15/76 on-line computer at VEC Centre for the acquisition and processing of data in nuclear physics experiments is described. The program can accommodate a maximum of thirty two inputs although the present hardware limits the number of inputs to eight. Particular emphasis is given to the problems of flexibility and ease of operation, memory optimisation and techniques dealing with experimenter-computer interaction. Various graphical methods for one- and two-dimensional data presentation are discussed. Specific problems of particle identification using detector telescopes have been dealt with carefully to handle experiments using several detector telescopes and those involving light particle-heavy particle coincidence studies. Steps needed to tailor this program towards utilisation for special experiments are also described. (author)

  15. A general purpose subroutine for fast fourier transform on a distributed memory parallel machine

    Science.gov (United States)

    Dubey, A.; Zubair, M.; Grosch, C. E.

    1992-01-01

    One issue which is central in developing a general purpose Fast Fourier Transform (FFT) subroutine on a distributed memory parallel machine is the data distribution. It is possible that different users would like to use the FFT routine with different data distributions. Thus, there is a need to design FFT schemes on distributed memory parallel machines which can support a variety of data distributions. An FFT implementation on a distributed memory parallel machine which works for a number of data distributions commonly encountered in scientific applications is presented. The problem of rearranging the data after computing the FFT is also addressed. The performance of the implementation on a distributed memory parallel machine Intel iPSC/860 is evaluated.

  16. General purpose graphic processing unit implementation of adaptive pulse compression algorithms

    Science.gov (United States)

    Cai, Jingxiao; Zhang, Yan

    2017-07-01

    This study introduces a practical approach to implement real-time signal processing algorithms for general surveillance radar based on NVIDIA graphical processing units (GPUs). The pulse compression algorithms are implemented using compute unified device architecture (CUDA) libraries such as CUDA basic linear algebra subroutines and CUDA fast Fourier transform library, which are adopted from open source libraries and optimized for the NVIDIA GPUs. For more advanced, adaptive processing algorithms such as adaptive pulse compression, customized kernel optimization is needed and investigated. A statistical optimization approach is developed for this purpose without needing much knowledge of the physical configurations of the kernels. It was found that the kernel optimization approach can significantly improve the performance. Benchmark performance is compared with the CPU performance in terms of processing accelerations. The proposed implementation framework can be used in various radar systems including ground-based phased array radar, airborne sense and avoid radar, and aerospace surveillance radar.

  17. A general-purpose process modelling framework for marine energy systems

    International Nuclear Information System (INIS)

    Dimopoulos, George G.; Georgopoulou, Chariklia A.; Stefanatos, Iason C.; Zymaris, Alexandros S.; Kakalis, Nikolaos M.P.

    2014-01-01

    Highlights: • Process modelling techniques applied in marine engineering. • Systems engineering approaches to manage the complexity of modern ship machinery. • General purpose modelling framework called COSSMOS. • Mathematical modelling of conservation equations and related chemical – transport phenomena. • Generic library of ship machinery component models. - Abstract: High fuel prices, environmental regulations and current shipping market conditions impose ships to operate in a more efficient and greener way. These drivers lead to the introduction of new technologies, fuels, and operations, increasing the complexity of modern ship energy systems. As a means to manage this complexity, in this paper we present the introduction of systems engineering methodologies in marine engineering via the development of a general-purpose process modelling framework for ships named as DNV COSSMOS. Shifting the focus from components – the standard approach in shipping- to systems, widens the space for optimal design and operation solutions. The associated computer implementation of COSSMOS is a platform that models, simulates and optimises integrated marine energy systems with respect to energy efficiency, emissions, safety/reliability and costs, under both steady-state and dynamic conditions. DNV COSSMOS can be used in assessment and optimisation of design and operation problems in existing vessels, new builds as well as new technologies. The main features and our modelling approach are presented and key capabilities are illustrated via two studies on the thermo-economic design and operation optimisation of a combined cycle system for large bulk carriers, and the transient operation simulation of an electric marine propulsion system

  18. Design of the SLAC RCE Platform: A General Purpose ATCA Based Data Acquisition System

    International Nuclear Information System (INIS)

    Herbst, R.; Claus, R.; Freytag, M.; Haller, G.; Huffer, M.; Maldonado, S.; Nishimura, K.; O'Grady, C.; Panetta, J.; Perazzo, A.; Reese, B.; Ruckman, L.; Thayer, J.G.; Weaver, M.

    2015-01-01

    The SLAC RCE platform is a general purpose clustered data acquisition system implemented on a custom ATCA compliant blade, called the Cluster On Board (COB). The core of the system is the Reconfigurable Cluster Element (RCE), which is a system-on-chip design based upon the Xilinx Zynq family of FPGAs, mounted on custom COB daughter-boards. The Zynq architecture couples a dual core ARM Cortex A9 based processor with a high performance 28nm FPGA. The RCE has 12 external general purpose bi-directional high speed links, each supporting serial rates of up to 12Gbps. 8 RCE nodes are included on a COB, each with a 10Gbps connection to an on-board 24-port Ethernet switch integrated circuit. The COB is designed to be used with a standard full-mesh ATCA backplane allowing multiple RCE nodes to be tightly interconnected with minimal interconnect latency. Multiple shelves can be clustered using the front panel 10-gbps connections. The COB also supports local and inter-blade timing and trigger distribution. An experiment specific Rear Transition Module adapts the 96 high speed serial links to specific experiments and allows an experiment-specific timing and busy feedback connection. This coupling of processors with a high performance FPGA fabric in a low latency, multiple node cluster allows high speed data processing that can be easily adapted to any physics experiment. RTEMS and Linux are both ported to the module. The RCE has been used or is the baseline for several current and proposed experiments (LCLS, HPS, LSST, ATLAS-CSC, LBNE, DarkSide, ILC-SiD, etc).

  19. Uniform physical theory of diffraction equivalent edge currents for implementation in general computer codes

    DEFF Research Database (Denmark)

    Johansen, Peter Meincke

    1996-01-01

    New uniform closed-form expressions for physical theory of diffraction equivalent edge currents are derived for truncated incremental wedge strips. In contrast to previously reported expressions, the new expressions are well-behaved for all directions of incidence and observation and take a finite...... value for zero strip length. Consequently, the new equivalent edge currents are, to the knowledge of the author, the first that are well-suited for implementation in general computer codes...

  20. Transforming the ASDEX Upgrade discharge control system to a general-purpose plasma control platform

    International Nuclear Information System (INIS)

    Treutterer, Wolfgang; Cole, Richard; Gräter, Alexander; Lüddecke, Klaus; Neu, Gregor; Rapson, Christopher; Raupp, Gerhard; Zasche, Dieter; Zehetbauer, Thomas

    2015-01-01

    Highlights: • Control framework split in core and custom part. • Core framework deployable in other fusion device environments. • Adaptible through customizable modules, plug-in support and generic interfaces. - Abstract: The ASDEX Upgrade Discharge Control System DCS is a modern and mature product, originally designed to regulate and supervise ASDEX Upgrade Tokamak plasma operation. In its core DCS is based on a generic, versatile real-time software framework with a plugin architecture that allows to easily combine, modify and extend control function modules in order to tailor the system to required features and let it continuously evolve with the progress of an experimental fusion device. Due to these properties other fusion experiments like the WEST project have expressed interest in adopting DCS. For this purpose, essential parts of DCS must be unpinned from the ASDEX Upgrade environment by exposure or introduction of generalised interfaces. Re-organisation of DCS modules allows distinguishing between intrinsic framework core functions and device-specific applications. In particular, DCS must be prepared for deployment in different system environments with their own realisations for user interface, pulse schedule preparation, parameter server, time and event distribution, diagnostic and actuator systems, network communication and data archiving. The article explains the principles of the revised DCS structure, derives the necessary interface definitions and describes major steps to achieve the separation between general-purpose framework and fusion device specific components.

  1. Transforming the ASDEX Upgrade discharge control system to a general-purpose plasma control platform

    Energy Technology Data Exchange (ETDEWEB)

    Treutterer, Wolfgang, E-mail: Wolfgang.Treutterer@ipp.mpg.de [Max-Planck-Institut für Plasmaphysik, Boltzmannstr. 2, 85748 Garching (Germany); Cole, Richard [Unlimited Computer Systems, Seeshaupter Str. 15, 82393 Iffeldorf (Germany); Gräter, Alexander [Max-Planck-Institut für Plasmaphysik, Boltzmannstr. 2, 85748 Garching (Germany); Lüddecke, Klaus [Unlimited Computer Systems, Seeshaupter Str. 15, 82393 Iffeldorf (Germany); Neu, Gregor; Rapson, Christopher; Raupp, Gerhard; Zasche, Dieter; Zehetbauer, Thomas [Max-Planck-Institut für Plasmaphysik, Boltzmannstr. 2, 85748 Garching (Germany)

    2015-10-15

    Highlights: • Control framework split in core and custom part. • Core framework deployable in other fusion device environments. • Adaptible through customizable modules, plug-in support and generic interfaces. - Abstract: The ASDEX Upgrade Discharge Control System DCS is a modern and mature product, originally designed to regulate and supervise ASDEX Upgrade Tokamak plasma operation. In its core DCS is based on a generic, versatile real-time software framework with a plugin architecture that allows to easily combine, modify and extend control function modules in order to tailor the system to required features and let it continuously evolve with the progress of an experimental fusion device. Due to these properties other fusion experiments like the WEST project have expressed interest in adopting DCS. For this purpose, essential parts of DCS must be unpinned from the ASDEX Upgrade environment by exposure or introduction of generalised interfaces. Re-organisation of DCS modules allows distinguishing between intrinsic framework core functions and device-specific applications. In particular, DCS must be prepared for deployment in different system environments with their own realisations for user interface, pulse schedule preparation, parameter server, time and event distribution, diagnostic and actuator systems, network communication and data archiving. The article explains the principles of the revised DCS structure, derives the necessary interface definitions and describes major steps to achieve the separation between general-purpose framework and fusion device specific components.

  2. Optimization of a general-purpose, actively scanned proton beamline for ocular treatments: Geant4 simulations.

    Science.gov (United States)

    Piersimoni, Pierluigi; Rimoldi, Adele; Riccardi, Cristina; Pirola, Michele; Molinelli, Silvia; Ciocca, Mario

    2015-03-08

    The Italian National Center for Hadrontherapy (CNAO, Centro Nazionale di Adroterapia Oncologica), a synchrotron-based hospital facility, started the treatment of patients within selected clinical trials in late 2011 and 2012 with actively scanned proton and carbon ion beams, respectively. The activation of a new clinical protocol for the irradiation of uveal melanoma using the existing general-purpose proton beamline is foreseen for late 2014. Beam characteristics and patient treatment setup need to be tuned to meet the specific requirements for such a type of treatment technique. The aim of this study is to optimize the CNAO transport beamline by adding passive components and minimizing air gap to achieve the optimal conditions for ocular tumor irradiation. The CNAO setup with the active and passive components along the transport beamline, as well as a human eye-modeled detector also including a realistic target volume, were simulated using the Monte Carlo Geant4 toolkit. The strong reduction of the air gap between the nozzle and patient skin, as well as the insertion of a range shifter plus a patient-specific brass collimator at a short distance from the eye, were found to be effective tools to be implemented. In perspective, this simulation toolkit could also be used as a benchmark for future developments and testing purposes on commercial treatment planning systems.

  3. General relativistic radiative transfer code in rotating black hole space-time: ARTIST

    Science.gov (United States)

    Takahashi, Rohta; Umemura, Masayuki

    2017-02-01

    We present a general relativistic radiative transfer code, ARTIST (Authentic Radiative Transfer In Space-Time), that is a perfectly causal scheme to pursue the propagation of radiation with absorption and scattering around a Kerr black hole. The code explicitly solves the invariant radiation intensity along null geodesics in the Kerr-Schild coordinates, and therefore properly includes light bending, Doppler boosting, frame dragging, and gravitational redshifts. The notable aspect of ARTIST is that it conserves the radiative energy with high accuracy, and is not subject to the numerical diffusion, since the transfer is solved on long characteristics along null geodesics. We first solve the wavefront propagation around a Kerr black hole that was originally explored by Hanni. This demonstrates repeated wavefront collisions, light bending, and causal propagation of radiation with the speed of light. We show that the decay rate of the total energy of wavefronts near a black hole is determined solely by the black hole spin in late phases, in agreement with analytic expectations. As a result, the ARTIST turns out to correctly solve the general relativistic radiation fields until late phases as t ˜ 90 M. We also explore the effects of absorption and scattering, and apply this code for a photon wall problem and an orbiting hotspot problem. All the simulations in this study are performed in the equatorial plane around a Kerr black hole. The ARTIST is the first step to realize the general relativistic radiation hydrodynamics.

  4. TACO: a general-purpose tool for predicting cell-type-specific transcription factor dimers.

    Science.gov (United States)

    Jankowski, Aleksander; Prabhakar, Shyam; Tiuryn, Jerzy

    2014-03-19

    Cooperative binding of transcription factor (TF) dimers to DNA is increasingly recognized as a major contributor to binding specificity. However, it is likely that the set of known TF dimers is highly incomplete, given that they were discovered using ad hoc approaches, or through computational analyses of limited datasets. Here, we present TACO (Transcription factor Association from Complex Overrepresentation), a general-purpose standalone software tool that takes as input any genome-wide set of regulatory elements and predicts cell-type-specific TF dimers based on enrichment of motif complexes. TACO is the first tool that can accommodate motif complexes composed of overlapping motifs, a characteristic feature of many known TF dimers. Our method comprehensively outperforms existing tools when benchmarked on a reference set of 29 known dimers. We demonstrate the utility and consistency of TACO by applying it to 152 DNase-seq datasets and 94 ChIP-seq datasets. Based on these results, we uncover a general principle governing the structure of TF-TF-DNA ternary complexes, namely that the flexibility of the complex is correlated with, and most likely a consequence of, inter-motif spacing.

  5. DNA Processing and Reassembly on General Purpose FPGA-based Development Boards

    Directory of Open Access Journals (Sweden)

    SZÁSZ Csaba

    2017-05-01

    Full Text Available The great majority of researchers involved in microelectronics generally agree that many scientific challenges in life sciences have associated with them a powerful computational requirement that must be solved before scientific progress can be made. The current trend in Deoxyribonucleic Acid (DNA computing technologies is to develop special hardware platforms capable to provide the needed processing performance at lower cost. In this endeavor the FPGA-based (Field Programmable Gate Arrays configurations aimed to accelerate genome sequencing and reassembly plays a leading role. This paper emphasizes benefits and advantages using general purpose FPGA-based development boards in DNA reassembly applications beside the special hardware architecture solutions. An original approach is unfolded which outlines the versatility of high performance ready-to-use manufacturer development platforms endowed with powerful hardware resources fully optimized for high speed processing applications. The theoretical arguments are supported via an intuitive implementation example where the designer it is discharged from any hardware development effort and completely assisted in exclusive concentration only on software design issues providing greatly reduced application development cycles. The experiments prove that such boards available on the market are suitable to fulfill in all a wide range of DNA sequencing and reassembly applications.

  6. Litrani: a general purpose Monte-Carlo program simulating light propagation in isotropic or anisotropic media

    International Nuclear Information System (INIS)

    Gentit, F.-X.

    2002-01-01

    Litrani is a general purpose Monte-Carlo program simulating light propagation in any type of setup describable by the shapes provided by ROOT. Each shape may be made of a different material. Dielectric constant, absorption length and diffusion length of materials may depend upon wavelength. Dielectric constant and absorption length may be anisotropic. Each face of a volume is either partially or totally in contact with a face of another volume, or covered with some wrapping having defined characteristics of absorption, reflection and diffusion. When in contact with another face of another volume, the possibility exists to have a thin slice of width d and index n between the two faces. The program has various sources of light: spontaneous photons, photons coming from an optical fibre, photons generated by the crossing of particles or photons generated by an electromagnetic shower. The time and wavelength spectra of emitted photons may reproduce any scintillation spectrum. As detectors, phototubes, APD, or any general type of surface or volume detectors may be specified. The aim is to follow each photon until it is absorbed or detected. Quantities to be delivered by the program are the proportion of photons detected, and the time distribution for the arrival of these, or the various ways photons may be lost

  7. Efficient coding explains the universal law of generalization in human perception.

    Science.gov (United States)

    Sims, Chris R

    2018-05-11

    Perceptual generalization and discrimination are fundamental cognitive abilities. For example, if a bird eats a poisonous butterfly, it will learn to avoid preying on that species again by generalizing its past experience to new perceptual stimuli. In cognitive science, the "universal law of generalization" seeks to explain this ability and states that generalization between stimuli will follow an exponential function of their distance in "psychological space." Here, I challenge existing theoretical explanations for the universal law and offer an alternative account based on the principle of efficient coding. I show that the universal law emerges inevitably from any information processing system (whether biological or artificial) that minimizes the cost of perceptual error subject to constraints on the ability to process or transmit information. Copyright © 2018 The Authors, some rights reserved; exclusive licensee American Association for the Advancement of Science. No claim to original U.S. Government Works.

  8. TOUGH2: A general-purpose numerical simulator for multiphase nonisothermal flows

    Energy Technology Data Exchange (ETDEWEB)

    Pruess, K. [Lawrence Berkeley Lab., CA (United States)

    1991-06-01

    Numerical simulators for multiphase fluid and heat flows in permeable media have been under development at Lawrence Berkeley Laboratory for more than 10 yr. Real geofluids contain noncondensible gases and dissolved solids in addition to water, and the desire to model such `compositional` systems led to the development of a flexible multicomponent, multiphase simulation architecture known as MULKOM. The design of MULKOM was based on the recognition that the mass-and energy-balance equations for multiphase fluid and heat flows in multicomponent systems have the same mathematical form, regardless of the number and nature of fluid components and phases present. Application of MULKOM to different fluid mixtures, such as water and air, or water, oil, and gas, is possible by means of appropriate `equation-of-state` (EOS) modules, which provide all thermophysical and transport parameters of the fluid mixture and the permeable medium as a function of a suitable set of primary thermodynamic variables. Investigations of thermal and hydrologic effects from emplacement of heat-generating nuclear wastes into partially water-saturated formations prompted the development and release of a specialized version of MULKOM for nonisothermal flow of water and air, named TOUGH. TOUGH is an acronym for `transport of unsaturated groundwater and heat` and is also an allusion to the tuff formations at Yucca Mountain, Nevada. The TOUGH2 code is intended to supersede TOUGH. It offers all the capabilities of TOUGH and includes a considerably more general subset of MULKOM modules with added capabilities. The paper briefly describes the simulation methodology and user features.

  9. 15 CFR 744.17 - Restrictions on certain exports and reexports of general purpose microprocessors for “military...

    Science.gov (United States)

    2010-01-01

    ... reexports of general purpose microprocessors for âmilitary end-usesâ and to âmilitary end-users.â 744.17...: END-USER AND END-USE BASED § 744.17 Restrictions on certain exports and reexports of general purpose microprocessors for “military end-uses” and to “military end-users.” (a) General prohibition. In addition to the...

  10. GRHydro: a new open-source general-relativistic magnetohydrodynamics code for the Einstein toolkit

    International Nuclear Information System (INIS)

    Mösta, Philipp; Haas, Roland; Ott, Christian D; Reisswig, Christian; Mundim, Bruno C; Faber, Joshua A; Noble, Scott C; Bode, Tanja; Löffler, Frank; Schnetter, Erik

    2014-01-01

    We present the new general-relativistic magnetohydrodynamics (GRMHD) capabilities of the Einstein toolkit, an open-source community-driven numerical relativity and computational relativistic astrophysics code. The GRMHD extension of the toolkit builds upon previous releases and implements the evolution of relativistic magnetized fluids in the ideal MHD limit in fully dynamical spacetimes using the same shock-capturing techniques previously applied to hydrodynamical evolution. In order to maintain the divergence-free character of the magnetic field, the code implements both constrained transport and hyperbolic divergence cleaning schemes. We present test results for a number of MHD tests in Minkowski and curved spacetimes. Minkowski tests include aligned and oblique planar shocks, cylindrical explosions, magnetic rotors, Alfvén waves and advected loops, as well as a set of tests designed to study the response of the divergence cleaning scheme to numerically generated monopoles. We study the code’s performance in curved spacetimes with spherical accretion onto a black hole on a fixed background spacetime and in fully dynamical spacetimes by evolutions of a magnetized polytropic neutron star and of the collapse of a magnetized stellar core. Our results agree well with exact solutions where these are available and we demonstrate convergence. All code and input files used to generate the results are available on http://einsteintoolkit.org. This makes our work fully reproducible and provides new users with an introduction to applications of the code. (paper)

  11. Delay reduction in lossy intermittent feedback for generalized instantly decodable network coding

    KAUST Repository

    Douik, Ahmed S.; Sorour, Sameh; Alouini, Mohamed-Slim; Ai-Naffouri, Tareq Y.

    2013-01-01

    In this paper, we study the effect of lossy intermittent feedback loss events on the multicast decoding delay performance of generalized instantly decodable network coding. These feedback loss events create uncertainty at the sender about the reception statues of different receivers and thus uncertainty to accurately determine subsequent instantly decodable coded packets. To solve this problem, we first identify the different possibilities of uncertain packets at the sender and their probabilities. We then derive the expression of the mean decoding delay. We formulate the Generalized Instantly Decodable Network Coding (G-IDNC) minimum decoding delay problem as a maximum weight clique problem. Since finding the optimal solution is NP-hard, we design a variant of the algorithm employed in [1]. Our algorithm is compared to the two blind graph update proposed in [2] through extensive simulations. Results show that our algorithm outperforms the blind approaches in all the situations and achieves a tolerable degradation, against the perfect feedback, for large feedback loss period. © 2013 IEEE.

  12. Delay reduction in lossy intermittent feedback for generalized instantly decodable network coding

    KAUST Repository

    Douik, Ahmed S.

    2013-10-01

    In this paper, we study the effect of lossy intermittent feedback loss events on the multicast decoding delay performance of generalized instantly decodable network coding. These feedback loss events create uncertainty at the sender about the reception statues of different receivers and thus uncertainty to accurately determine subsequent instantly decodable coded packets. To solve this problem, we first identify the different possibilities of uncertain packets at the sender and their probabilities. We then derive the expression of the mean decoding delay. We formulate the Generalized Instantly Decodable Network Coding (G-IDNC) minimum decoding delay problem as a maximum weight clique problem. Since finding the optimal solution is NP-hard, we design a variant of the algorithm employed in [1]. Our algorithm is compared to the two blind graph update proposed in [2] through extensive simulations. Results show that our algorithm outperforms the blind approaches in all the situations and achieves a tolerable degradation, against the perfect feedback, for large feedback loss period. © 2013 IEEE.

  13. The PennBMBI: Design of a General Purpose Wireless Brain-Machine-Brain Interface System.

    Science.gov (United States)

    Liu, Xilin; Zhang, Milin; Subei, Basheer; Richardson, Andrew G; Lucas, Timothy H; Van der Spiegel, Jan

    2015-04-01

    In this paper, a general purpose wireless Brain-Machine-Brain Interface (BMBI) system is presented. The system integrates four battery-powered wireless devices for the implementation of a closed-loop sensorimotor neural interface, including a neural signal analyzer, a neural stimulator, a body-area sensor node and a graphic user interface implemented on the PC end. The neural signal analyzer features a four channel analog front-end with configurable bandpass filter, gain stage, digitization resolution, and sampling rate. The target frequency band is configurable from EEG to single unit activity. A noise floor of 4.69 μVrms is achieved over a bandwidth from 0.05 Hz to 6 kHz. Digital filtering, neural feature extraction, spike detection, sensing-stimulating modulation, and compressed sensing measurement are realized in a central processing unit integrated in the analyzer. A flash memory card is also integrated in the analyzer. A 2-channel neural stimulator with a compliance voltage up to ± 12 V is included. The stimulator is capable of delivering unipolar or bipolar, charge-balanced current pulses with programmable pulse shape, amplitude, width, pulse train frequency and latency. A multi-functional sensor node, including an accelerometer, a temperature sensor, a flexiforce sensor and a general sensor extension port has been designed. A computer interface is designed to monitor, control and configure all aforementioned devices via a wireless link, according to a custom designed communication protocol. Wireless closed-loop operation between the sensory devices, neural stimulator, and neural signal analyzer can be configured. The proposed system was designed to link two sites in the brain, bridging the brain and external hardware, as well as creating new sensory and motor pathways for clinical practice. Bench test and in vivo experiments are performed to verify the functions and performances of the system.

  14. General-Purpose Heat Source development: Safety Verification Test Program. Bullet/fragment test series

    Energy Technology Data Exchange (ETDEWEB)

    George, T.G.; Tate, R.E.; Axler, K.M.

    1985-05-01

    The radioisotope thermoelectric generator (RTG) that will provide power for space missions contains 18 General-Purpose Heat Source (GPHS) modules. Each module contains four /sup 238/PuO/sub 2/-fueled clads and generates 250 W/sub (t)/. Because a launch-pad or post-launch explosion is always possible, we need to determine the ability of GPHS fueled clads within a module to survive fragment impact. The bullet/fragment test series, part of the Safety Verification Test Plan, was designed to provide information on clad response to impact by a compact, high-energy, aluminum-alloy fragment and to establish a threshold value of fragment energy required to breach the iridium cladding. Test results show that a velocity of 555 m/s (1820 ft/s) with an 18-g bullet is at or near the threshold value of fragment velocity that will cause a clad breach. Results also show that an exothermic Ir/Al reaction occurs if aluminum and hot iridium are in contact, a contact that is possible and most damaging to the clad within a narrow velocity range. The observed reactions between the iridium and the aluminum were studied in the laboratory and are reported in the Appendix.

  15. General-Purpose Heat Source Safety Verification Test program: Edge-on flyer plate tests

    International Nuclear Information System (INIS)

    George, T.G.

    1987-03-01

    The radioisotope thermoelectric generator (RTG) that will supply power for the Galileo and Ulysses space missions contains 18 General-Purpose Heat Source (GPHS) modules. The GPHS modules provide power by transmitting the heat of 238 Pu α-decay to an array of thermoelectric elements. Each module contains four 238 PuO 2 -fueled clads and generates 250 W(t). Because the possibility of a launch vehicle explosion always exists, and because such an explosion could generate a field of high-energy fragments, the fueled clads within each GPHS module must survive fragment impact. The edge-on flyer plate tests were included in the Safety Verification Test series to provide information on the module/clad response to the impact of high-energy plate fragments. The test results indicate that the edge-on impact of a 3.2-mm-thick, aluminum-alloy (2219-T87) plate traveling at 915 m/s causes the complete release of fuel from capsules contained within a bare GPHS module, and that the threshold velocity sufficient to cause the breach of a bare, simulant-fueled clad impacted by a 3.5-mm-thick, aluminum-alloy (5052-T0) plate is approximately 140 m/s

  16. Environmental assessment of general-purpose heat source safety verification testing

    International Nuclear Information System (INIS)

    1995-02-01

    This Environmental Assessment (EA) was prepared to identify and evaluate potential environmental, safety, and health impacts associated with the Proposed Action to test General-Purpose Heat Source (GPHS) Radioisotope Thermoelectric Generator (RTG) assemblies at the Sandia National Laboratories (SNL) 10,000-Foot Sled Track Facility, Albuquerque, New Mexico. RTGs are used to provide a reliable source of electrical power on board some spacecraft when solar power is inadequate during long duration space missions. These units are designed to convert heat from the natural decay of radioisotope fuel into electrical power. Impact test data are required to support DOE's mission to provide radioisotope power systems to NASA and other user agencies. The proposed tests will expand the available safety database regarding RTG performance under postulated accident conditions. Direct observations and measurements of GPHS/RTG performance upon impact with hard, unyielding surfaces are required to verify model predictions and to ensure the continual evolution of the RTG designs that perform safely under varied accident environments. The Proposed Action is to conduct impact testing of RTG sections containing GPHS modules with simulated fuel. End-On and Side-On impact test series are planned

  17. Evaluation and characterization of General Purpose Heat Source girth welds for the Cassini mission

    International Nuclear Information System (INIS)

    Lynch, C.M.; Moniz, P.F.; Reimus, M.A.H.

    1998-01-01

    General Purpose Heat Sources (GPHSs) are components of Radioisotopic thermoelectric Generators (RTGs) which provide electric power for deep space missions. Each GPHS consists of a 238 Pu oxide ceramic pellet encapsulated in a welded iridium alloy shell which forms a protective barrier against the release of plutonia in the unlikely event of a launch-pad failure or reentry incident. GPHS fueled clad girth weld flaw detection was paramount to ensuring this safety function, and was accomplished using both destructive and non-destructive evaluation techniques. The first girth weld produced from each welding campaign was metallographically examined for flaws such as incomplete weld penetration, cracks, or porosity which would render a GPHS unacceptable for flight applications. After an acceptable example weld was produced, the subsequently welded heat sources were evaluated non-destructively for flaws using ultrasonic immersion testing. Selected heat sources which failed ultrasonic testing would be radiographed, and/or, destructively evaluated to further characterize and document anomalous indications. Metallography was also performed on impacted heat sources to determine the condition of the welds

  18. General purpose nonlinear analysis program FINAS for elevated temperature design of FBR components

    International Nuclear Information System (INIS)

    Iwata, K.; Atsumo, H.; Kano, T.; Takeda, H.

    1982-01-01

    This paper presents currently available capabilities of a general purpose finite element nonlinear analysis program FINAS (FBR Inelastic Structural Analysis System) which has been developed at Power Reactor and Nuclear Fuel Development Corporation (PNC) since 1976 to support structural design of fast breeder reactor (FBR) components in Japan. This program is capable of treating inelastic responses of arbitrary complex structures subjected to static and dynamic load histories. Various types of finite element covering rods, beams, pipes, axisymmetric, two and three dimensional solids, plates and shells, are implemented in the program. The thermal elastic-plastic creep analysis is possible for each element type, with primary emphasis on the application to FBR components subjected to sustained or cyclic loads at elevated temperature. The program permits large deformation, buckling, fracture mechanics, and dynamic analyses for some of the element types and provides a number of options for automatic mesh generation and computer graphics. Some examples including elevated temperature effects are shown to demonstrate the accuracy and the efficiency of the program

  19. Nondestructive inspection of General Purpose Heat Source (GPHS) fueled clad girth welds

    International Nuclear Information System (INIS)

    Reimus, M. A. H.; George, T. G.; Lynch, C.; Padilla, M.; Moniz, P.; Guerrero, A.; Moyer, M. W.; Placr, A.

    1998-01-01

    The General-Purpose Heat Source (GPHS) provides power for space missions by transmitting the heat of 238 Pu decay to an array of thermoelectric elements. The GPHS is fabricated using an iridium-alloy to contain the 238 PuO 2 fuel pellet. GPHS capsules will be utilized in the upcoming Cassini mission to explore Saturn and its moons. The physical integrity of the girth weld is important to mission safety and performance. Because past experience had revealed a potential for initiation of small cracks in the girth weld overlap zone, a nondestructive inspection of each capsule weld is required. An ultrasonic method was used to inspect the welds of capsules fabricated for the Galileo mission. The instrument, transducer, and method used were state of the art at the time (early 1980s). The ultrasonic instrumentation and methods used to inspect the Cassini GPHSs was significantly upgraded from those used for the Galileo mission. GPHSs that had ultrasonic reflectors in excess of the reject specification level were subsequently inspected with radiography to provide additional engineering data used to accept/reject the heat source. This paper describes the Galileo-era ultrasonic instrumentation and methods and the subsequent upgrades made to support testing of Cassini GPHSs. Also discussed is the data obtained from radiographic examination and correlation to ultrasonic examination results

  20. Nondestructive inspection of General Purpose Heat Source (GPHS) fueled clad girth welds

    International Nuclear Information System (INIS)

    Reimus, M.A.; George, T.G.; Lynch, C.; Padilla, M.; Moniz, P.; Guerrero, A.; Moyer, M.W.; Placr, A.

    1998-01-01

    The General-Purpose Heat Source (GPHS) provides power for space missions by transmitting the heat of 238 Pu decay to an array of thermoelectric elements. The GPHS is fabricated using an iridium-alloy to contain the 238 PuO 2 fuel pellet. GPHS capsules will be utilized in the upcoming Cassini mission to explore Saturn and its moons. The physical integrity of the girth weld is important to mission safety and performance. Because past experience had revealed a potential for initiation of small cracks in the girth weld overlap zone, a nondestructive inspection of each capsule weld is required. An ultrasonic method was used to inspect the welds of capsules fabricated for the Galileo mission. The instrument, transducer, and method used were state of the art at the time (early 1980s). The ultrasonic instrumentation and methods used to inspect the Cassini GPHSs was significantly upgraded from those used for the Galileo mission. GPHSs that had ultrasonic reflectors in excess of the reject specification level were subsequently inspected with radiography to provide additional engineering data used to accept/reject the heat source. This paper describes the Galileo-era ultrasonic instrumentation and methods and the subsequent upgrades made to support testing of Cassini GPHSs. Also discussed is the data obtained from radiographic examination and correlation to ultrasonic examination results. copyright 1998 American Institute of Physics

  1. Applications of artificial intelligence to space station: General purpose intelligent sensor interface

    Science.gov (United States)

    Mckee, James W.

    1988-01-01

    This final report describes the accomplishments of the General Purpose Intelligent Sensor Interface task of the Applications of Artificial Intelligence to Space Station grant for the period from October 1, 1987 through September 30, 1988. Portions of the First Biannual Report not revised will not be included but only referenced. The goal is to develop an intelligent sensor system that will simplify the design and development of expert systems using sensors of the physical phenomena as a source of data. This research will concentrate on the integration of image processing sensors and voice processing sensors with a computer designed for expert system development. The result of this research will be the design and documentation of a system in which the user will not need to be an expert in such areas as image processing algorithms, local area networks, image processor hardware selection or interfacing, television camera selection, voice recognition hardware selection, or analog signal processing. The user will be able to access data from video or voice sensors through standard LISP statements without any need to know about the sensor hardware or software.

  2. Low-cost general purpose spectral display unit using an IBM PC

    International Nuclear Information System (INIS)

    Robinson, S.L.

    1985-10-01

    Many physics experiments require acquisition and analysis of spectral data. commercial minicomputer-based multichannel analyzers collect detected counts at various energies, create a histogram of the counts in memory, and display the resultant spectra. They acquire data and provide the user-to-display interface. The system discussed separates functions into the three modular components of data acquisition, storage, and display. This decoupling of functions allows the experimenter to use any number of detectors for data collection before forwarding up to 64 spectra to the display unit, thereby increasing data throughput over that available with commercial systems. An IBM PC was chosen for the low-cost, general purpose display unit. Up to four spectra may be displayed simultaneously in different colors. The histogram saves 1024 channels per detector, 640 of which may be distinctly displayed per spectra. The IEEE-488 standard provides the data path between the IBM PC and the data collection unit. Data is sent to the PC under interrupt control, using direct memory access. Display manipulations available via keyboard are also discussed

  3. General-Purpose Heat Source Development: Safety Test Program. Postimpact evaluation, Design Iteration Test 3

    International Nuclear Information System (INIS)

    Schonfeld, F.W.; George, T.G.

    1984-07-01

    The General-Purpose Heat Source(GPHS) provides power for space missions by transmitting the heat of 238 PuO 2 decay to thermoelectric elements. Because of the inevitable return of certain aborted missions, the heat source must be designed and constructed to survive both re-entry and Earth impact. The Design Iteration Test (DIT) series is part of an ongoing test program. In the third test (DIT-3), a full GPHS module was impacted at 58 m/s and 930 0 C. The module impacted the target at an angle of 30 0 to the pole of the large faces. The four capsules used in DIT-3 survived impact with minimal deformation; no internal cracks other than in the regions indicated by Savannah River Plant (SRP) preimpact nondestructive testing were observed in any of the capsules. The 30 0 impact orientation used in DIT-3 was considerably less severe than the flat-on impact utilized in DIT-1 and DIT-2. The four capsules used in DIT-1 survived, while two of the capsules used in DIT-2 breached; a small quantity (approx. = 50 μg) of 238 PuO 2 was released from the capsules breached in the DIT-2 impact. All of the capsules used in DIT-1 and DIT-2 were severely deformed and contained large internal cracks. Postimpact analyses of the DIT-3 test components are described, with emphasis on weld structure and the behavior of defects identified by SRP nondestructive testing

  4. ICECAP: an integrated, general-purpose, automation-assisted IC50/EC50 assay platform.

    Science.gov (United States)

    Li, Ming; Chou, Judy; King, Kristopher W; Jing, Jing; Wei, Dong; Yang, Liyu

    2015-02-01

    IC50 and EC50 values are commonly used to evaluate drug potency. Mass spectrometry (MS)-centric bioanalytical and biomarker labs are now conducting IC50/EC50 assays, which, if done manually, are tedious and error-prone. Existing bioanalytical sample preparation automation systems cannot meet IC50/EC50 assay throughput demand. A general-purpose, automation-assisted IC50/EC50 assay platform was developed to automate the calculations of spiking solutions and the matrix solutions preparation scheme, the actual spiking and matrix solutions preparations, as well as the flexible sample extraction procedures after incubation. In addition, the platform also automates the data extraction, nonlinear regression curve fitting, computation of IC50/EC50 values, graphing, and reporting. The automation-assisted IC50/EC50 assay platform can process the whole class of assays of varying assay conditions. In each run, the system can handle up to 32 compounds and up to 10 concentration levels per compound, and it greatly improves IC50/EC50 assay experimental productivity and data processing efficiency. © 2014 Society for Laboratory Automation and Screening.

  5. A Fast General-Purpose Clustering Algorithm Based on FPGAs for High-Throughput Data Processing

    CERN Document Server

    Annovi, A; The ATLAS collaboration; Castegnaro, A; Gatta, M

    2012-01-01

    We present a fast general-purpose algorithm for high-throughput clustering of data ”with a two dimensional organization”. The algorithm is designed to be implemented with FPGAs or custom electronics. The key feature is a processing time that scales linearly with the amount of data to be processed. This means that clustering can be performed in pipeline with the readout, without suffering from combinatorial delays due to looping multiple times through all the data. This feature makes this algorithm especially well suited for problems where the data has high density, e.g. in the case of tracking devices working under high-luminosity condition such as those of LHC or Super-LHC. The algorithm is organized in two steps: the first step (core) clusters the data; the second step analyzes each cluster of data to extract the desired information. The current algorithm is developed as a clustering device for modern high-energy physics pixel detectors. However, the algorithm has much broader field of applications. In ...

  6. Explosion overpressure test series: General-Purpose Heat Source development: Safety Verification Test program

    International Nuclear Information System (INIS)

    Cull, T.A.; George, T.G.; Pavone, D.

    1986-09-01

    The General-Purpose Heat Source (GPHS) is a modular, radioisotope heat source that will be used in radioisotope thermoelectric generators (RTGs) to supply electric power for space missions. The first two uses will be the NASA Galileo and the ESA Ulysses missions. The RTG for these missions will contain 18 GPHS modules, each of which contains four 238 PuO 2 -fueled clads and generates 250 W/sub (t)/. A series of Safety Verification Tests (SVTs) was conducted to assess the ability of the GPHS modules to contain the plutonia in accident environments. Because a launch pad or postlaunch explosion of the Space Transportation System vehicle (space shuttle) is a conceivable accident, the SVT plan included a series of tests that simulated the overpressure exposure the RTG and GPHS modules could experience in such an event. Results of these tests, in which we used depleted UO 2 as a fuel simulant, suggest that exposure to overpressures as high as 15.2 MPa (2200 psi), without subsequent impact, does not result in a release of fuel

  7. A low-cost general purpose spectral display unit using an IBM PC

    International Nuclear Information System (INIS)

    Robinson, S.L.

    1986-01-01

    Many physics experiments require acquisition and analysis of spectral data. Commercial minicomputer-based multichannel analyzers collect detected counts at various energies, create a histogram of the counts in memory, and display the resultant spectra. They acquire data and provide the user-to-display interface. The system discussed separates functions into the three modular components of data acquisition, storage, and display. This decoupling of functions allows the experimenter to use any number of detectors for data collection before forwarding up to 64 spectra to the display unit, thereby increasing data throughput over that available with commercial systems. An IBM PC was chosen for the low-cost, general purpose display unit. Up to four spectra may be displayed simultaneously in different colors. The histogram saves 1024 channels per detector, 640 of which may be distinctly displayed per spectra. The IEEE-488 standard provides the data path between the IBM PC and the data collection unit. Data is sent to the PC under interrupt control, using direct memory access. Display manipulations available via keyboard are also discussed

  8. General-purpose heat source project and space nuclear safety and fuels program. Progress report

    International Nuclear Information System (INIS)

    Maraman, W.J.

    1980-02-01

    Studies related to the use of 238 PuO 2 in radioisotopic power systems carried out for the Advanced Nuclear Systems and Projects Division of LASL are presented. The three programs involved are: general-purpose heat source development; space nuclear safety; and fuels program. Three impact tests were conducted to evaluate the effects of a high temperature reentry pulse and the use of CBCF on impact performance. Additionally, two 238 PuO 2 pellets were encapsulated in Ir-0.3% W for impact testing. Results of the clad development test and vent testing are noted. Results of the environmental tests are summarized. Progress on the Stirling isotope power systems test and the status of the improved MHW tests are indicated. The examination of the impact failure of the iridium shell of MHFT-65 at a fuel pass-through continued. A test plan was written for vibration testing of the assembled light-weight radioisotopic heater unit. Progress on fuel processing is reported

  9. Computing OpenSURF on OpenCL and General Purpose GPU

    Directory of Open Access Journals (Sweden)

    Wanglong Yan

    2013-10-01

    Full Text Available Speeded-Up Robust Feature (SURF algorithm is widely used for image feature detecting and matching in computer vision area. Open Computing Language (OpenCL is a framework for writing programs that execute across heterogeneous platforms consisting of CPUs, GPUs, and other processors. This paper introduces how to implement an open-sourced SURF program, namely OpenSURF, on general purpose GPU by OpenCL, and discusses the optimizations in terms of the thread architectures and memory models in detail. Our final OpenCL implementation of OpenSURF is on average 37% and 64% faster than the OpenCV SURF v2.4.5 CUDA implementation on NVidia's GTX660 and GTX460SE GPUs, repectively. Our OpenCL program achieved real-time performance (>25 Frames Per Second for almost all the input images with different sizes from 320*240 to 1024*768 on NVidia's GTX660 GPU, NVidia's GTX460SE GPU and AMD's Radeon HD 6850 GPU. Our OpenCL approach on NVidia's GTX660 GPU is more than 22.8 times faster than its original CPU version on Intel's Dual-Core E5400 2.7G on average.

  10. Geometric correction of radiographic images using general purpose image processing program

    International Nuclear Information System (INIS)

    Kim, Eun Kyung; Cheong, Ji Seong; Lee, Sang Hoon

    1994-01-01

    The present study was undertaken to compare geometric corrected image by general-purpose image processing program for the Apple Macintosh II computer (NIH Image, Adobe Photoshop) with standardized image by individualized custom fabricated alignment instrument. Two non-standardized periapical films with XCP film holder only were taken at the lower molar portion of 19 volunteers. Two standardized periapical films with customized XCP film holder with impression material on the bite-block were taken for each person. Geometric correction was performed with Adobe Photoshop and NIH Image program. Specially, arbitrary image rotation function of 'Adobe Photoshop' and subtraction with transparency function of 'NIH Image' were utilized. The standard deviations of grey values of subtracted images were used to measure image similarity. Average standard deviation of grey values of subtracted images if standardized group was slightly lower than that of corrected group. However, the difference was found to be statistically insignificant (p>0.05). It is considered that we can use 'NIH Image' and 'Adobe Photoshop' program for correction of nonstandardized film, taken with XCP film holder at lower molar portion.

  11. Design evolution and verification of the general-purpose heat source

    International Nuclear Information System (INIS)

    Schock, A.

    The General-Purpose Heat Source (GPHS) is a radioisotope heat source for use in space power systems. It employs a modular design, to make it adaptable to a wide range of energy conversion systems and power levels. Each 250 W module is completely autonomous, with its own passive safety provisions to prevent fuel release under all abort modes, including atmospheric reentry and earth impact. Prior development tests had demonstrated good impact survival as long as the iridium fuel capsules retained their ductility. This requires high impact temperatures, typically above 900 0 C and reasonably fine grain size, which in turn requires avoidance of excessive operating temperatures and reentry temperatures. These three requirements - on operating, reentry, and impact temperatures - are in mutual conflict, since thermal design changes to improve any one of these temperatures tend to worsen one or both of the others. This conflict creates a difficult design problem, which for a time threatened the success of the program. The present paper describes how this problem was overcome by successive design revisions, supplemented by thermal analyses and confirmatory vibration and impact tests; and how this may be achieved while raising the specific power of the GPHS to 83 W/lb, a 50% improvement over previously flown radioisotope heat sources

  12. 21 CFR 1240.80 - General requirements for water for drinking and culinary purposes.

    Science.gov (United States)

    2010-04-01

    ... culinary purposes. 1240.80 Section 1240.80 Food and Drugs FOOD AND DRUG ADMINISTRATION, DEPARTMENT OF... requirements for water for drinking and culinary purposes. Only potable water shall be provided for drinking and culinary purposes by any operator of a conveyance engaged in interstate traffic, except as...

  13. Analysis of random point images with the use of symbolic computation codes and generalized Catalan numbers

    Science.gov (United States)

    Reznik, A. L.; Tuzikov, A. V.; Solov'ev, A. A.; Torgov, A. V.

    2016-11-01

    Original codes and combinatorial-geometrical computational schemes are presented, which are developed and applied for finding exact analytical formulas that describe the probability of errorless readout of random point images recorded by a scanning aperture with a limited number of threshold levels. Combinatorial problems encountered in the course of the study and associated with the new generalization of Catalan numbers are formulated and solved. An attempt is made to find the explicit analytical form of these numbers, which is, on the one hand, a necessary stage of solving the basic research problem and, on the other hand, an independent self-consistent problem.

  14. Upscaling from research watersheds: an essential stage of trustworthy general-purpose hydrologic model building

    Science.gov (United States)

    McNamara, J. P.; Semenova, O.; Restrepo, P. J.

    2011-12-01

    Highly instrumented research watersheds provide excellent opportunities for investigating hydrologic processes. A danger, however, is that the processes observed at a particular research watershed are too specific to the watershed and not representative even of the larger scale watershed that contains that particular research watershed. Thus, models developed based on those partial observations may not be suitable for general hydrologic use. Therefore demonstrating the upscaling of hydrologic process from research watersheds to larger watersheds is essential to validate concepts and test model structure. The Hydrograph model has been developed as a general-purpose process-based hydrologic distributed system. In its applications and further development we evaluate the scaling of model concepts and parameters in a wide range of hydrologic landscapes. All models, either lumped or distributed, are based on a discretization concept. It is common practice that watersheds are discretized into so called hydrologic units or hydrologic landscapes possessing assumed homogeneous hydrologic functioning. If a model structure is fixed, the difference in hydrologic functioning (difference in hydrologic landscapes) should be reflected by a specific set of model parameters. Research watersheds provide the possibility for reasonable detailed combining of processes into some typical hydrologic concept such as hydrologic units, hydrologic forms, and runoff formation complexes in the Hydrograph model. And here by upscaling we imply not the upscaling of a single process but upscaling of such unified hydrologic functioning. The simulation of runoff processes for the Dry Creek research watershed, Idaho, USA (27 km2) was undertaken using the Hydrograph model. The information on the watershed was provided by Boise State University and included a GIS database of watershed characteristics and a detailed hydrometeorological observational dataset. The model provided good simulation results in

  15. Motivation for Using Generalized Geometry in the Time Dependent Transport Code TDKENO

    Energy Technology Data Exchange (ETDEWEB)

    Dustin Popp; Zander Mausolff; Sedat Goluoglu

    2016-04-01

    We are proposing to use the code, TDKENO, to model TREAT. TDKENO solves the time dependent, three dimensional Boltzmann transport equation with explicit representation of delayed neutrons. Instead of directly integrating this equation, the neutron flux is factored into two components – a rapidly varying amplitude equation and a slowly varying shape equation and each is solved separately on different time scales. The shape equation is solved using the 3D Monte Carlo transport code KENO, from Oak Ridge National Laboratory’s SCALE code package. Using the Monte Carlo method to solve the shape equation is still computationally intensive, but the operation is only performed when needed. The amplitude equation is solved deterministically and frequently, so the solution gives an accurate time-dependent solution without having to repeatedly We have modified TDKENO to incorporate KENO-VI so that we may accurately represent the geometries within TREAT. This paper explains the motivation behind using generalized geometry, and provides the results of our modifications. TDKENO uses the Improved Quasi-Static method to accomplish this. In this method, the neutron flux is factored into two components. One component is a purely time-dependent and rapidly varying amplitude function, which is solved deterministically and very frequently (small time steps). The other is a slowly varying flux shape function that weakly depends on time and is only solved when needed (significantly larger time steps).

  16. Generalized hypercube graph $\\Q_n(S$, graph products and self-orthogonal codes

    Directory of Open Access Journals (Sweden)

    Pani Seneviratne

    2016-01-01

    Full Text Available A generalized hypercube graph $\\Q_n(S$ has $\\F_{2}^{n}=\\{0,1\\}^n$ as the vertex set and two vertices being adjacent whenever their mutual Hamming distance belongs to $S$, where $n \\ge 1$ and $S\\subseteq \\{1,2,\\ldots, n\\}$. The graph $\\Q_n(\\{1\\}$ is the $n$-cube, usually denoted by $\\Q_n$.We study graph boolean products $G_1 = \\Q_n(S\\times \\Q_1, G_2 = \\Q_{n}(S\\wedge \\Q_1$, $G_3 = \\Q_{n}(S[\\Q_1]$ and show that binary codes from neighborhood designs of $G_1, G_2$ and $G_3$ are self-orthogonal for all choices of $n$ and $S$. More over, we show that the class of codes $C_1$ are self-dual. Further we find subgroups of the automorphism group of these graphs and use these subgroups to obtain PD-sets for permutation decoding. As an example we find a full error-correcting PD set for the binary $[32, 16, 8]$ extremal self-dual code.

  17. A general-purpose development environment for intelligent computer-aided training systems

    Science.gov (United States)

    Savely, Robert T.

    1990-01-01

    Space station training will be a major task, requiring the creation of large numbers of simulation-based training systems for crew, flight controllers, and ground-based support personnel. Given the long duration of space station missions and the large number of activities supported by the space station, the extension of space shuttle training methods to space station training may prove to be impractical. The application of artificial intelligence technology to simulation training can provide the ability to deliver individualized training to large numbers of personnel in a distributed workstation environment. The principal objective of this project is the creation of a software development environment which can be used to build intelligent training systems for procedural tasks associated with the operation of the space station. Current NASA Johnson Space Center projects and joint projects with other NASA operational centers will result in specific training systems for existing space shuttle crew, ground support personnel, and flight controller tasks. Concurrently with the creation of these systems, a general-purpose development environment for intelligent computer-aided training systems will be built. Such an environment would permit the rapid production, delivery, and evolution of training systems for space station crew, flight controllers, and other support personnel. The widespread use of such systems will serve to preserve task and training expertise, support the training of many personnel in a distributed manner, and ensure the uniformity and verifiability of training experiences. As a result, significant reductions in training costs can be realized while safety and the probability of mission success can be enhanced.

  18. An auxiliary frequency tracking system for general purpose lock-in amplifiers

    Science.gov (United States)

    Xie, Kai; Chen, Liuhao; Huang, Anfeng; Zhao, Kai; Zhang, Hanlu

    2018-04-01

    Lock-in amplifiers (LIAs) are designed to measure weak signals submerged by noise. This is achieved with a signal modulator to avoid low-frequency noise and a narrow-band filter to suppress out-of-band noise. In asynchronous measurement, even a slight frequency deviation between the modulator and the reference may lead to measurement error because the filter’s passband is not flat. Because many commercial LIAs are unable to track frequency deviations, in this paper we propose an auxiliary frequency tracking system. We analyze the measurement error caused by the frequency deviation and propose both a tracking method and an auto-tracking system. This approach requires only three basic parameters, which can be obtained from any general purpose LIA via its communications interface, to calculate the frequency deviation from the phase difference. The proposed auxiliary tracking system is designed as a peripheral connected to the LIA’s serial port, removing the need for an additional power supply. The test results verified the effectiveness of the proposed system; the modified commercial LIA (model SR-850) was able to track the frequency deviation and continuous drift. For step frequency deviations, a steady tracking error of less than 0.001% was achieved within three adjustments, and the worst tracking accuracy was still better than 0.1% for a continuous frequency drift. The tracking system can be used to expand the application scope of commercial LIAs, especially for remote measurements in which the modulation clock and the local reference are separated.

  19. VHS-tape system for general purpose computer. For next generation mass storage system

    International Nuclear Information System (INIS)

    Ukai, K.; Takano, M.; Shinohara, M.; Niki, K.; Suzuki, Y.; Hamada, T.; Ogawa, M.

    1994-07-01

    Mass storage is one of the key technology of next generation computer system. A huge amount of data is produced on a field of particle and nuclear physics. These data are raw data of experiments, analysis data, Monte Carlo simulations data, etc. We search a storage device for these data at the point of view of capacity, price, size, transfer speed, etc. We have selected a VHS-tape (12.7 mm-tape, helical scan) from many storage devices. Characteristics of the VHS-tape are as follows; capacity of 14.5 GB, size of 460 cm 3 , price of 1,000 yen (S-VHS tape for video use), and 1.996 MB/sec transfer speed at a sustained mode. Last year, we succeeded to operate the VHS-tape system on a workstation as a I/O device with read/write speed of 1.5 MB/sec. We have tested a VHS-tape system by connecting to the channel of the general purpose computer (Fujitsu M-780/10S) in our institute. We obtained a read and write speeds of 1.07 MB/sec and 1.72 MB/sec by FORTRAN test programs, respectively. Read speeds of an open reel tape and a 3480 type cassete tape by the same test programs are 1.13 MB/sec and 2.54 MB/sec, respectively. Speeds of write operation are 1.09 MB/sec and 2.54 MB/sec for the open reel and 3480 cassete tape, respectively. A start motion of the VHS-tape for read/write operations needs about 60 seconds. (author)

  20. Code conforming determination of cumulative usage factors for general elastic-plastic finite element analyses

    International Nuclear Information System (INIS)

    Rudolph, Juergen; Goetz, Andreas; Hilpert, Roland

    2012-01-01

    The procedures of fatigue analyses of several relevant nuclear and conventional design codes (ASME, KTA, EN, AD) for power plant components differentiate between an elastic, simplified elastic-plastic and elastic-plastic fatigue check. As a rule, operational load levels will exclude the purely elastic fatigue check. The application of the code procedure of the simplified elastic-plastic fatigue check is common practice. Nevertheless, resulting cumulative usage factors may be overly conservative mainly due to high code based plastification penalty factors Ke. As a consequence, the more complex and still code conforming general elastic-plastic fatigue analysis methodology based on non-linear finite element analysis (FEA) is applied for fatigue design as an alternative. The requirements of the FEA and the material law to be applied have to be clarified in a first step. Current design codes only give rough guidelines on these relevant items. While the procedure for the simplified elastic-plastic fatigue analysis and the associated code passages are based on stress related cycle counting and the determination of pseudo elastic equivalent stress ranges, an adaptation to elastic-plastic strains and strain ranges is required for the elastic-plastic fatigue check. The associated requirements are explained in detail in the paper. If the established and implemented evaluation mechanism (cycle counting according to the peak and valley respectively the rainflow method, calculation of stress ranges from arbitrary load-time histories and determination of cumulative usage factors based on all load events) is to be retained, a conversion of elastic-plastic strains and strain ranges into pseudo elastic stress ranges is required. The algorithm to be applied is described in the paper. It has to be implemented in the sense of an extended post processing operation of FEA e.g. by APDL scripts in ANSYS registered . Variations of principal stress (strain) directions during the loading

  1. The EB factory project. I. A fast, neural-net-based, general purpose light curve classifier optimized for eclipsing binaries

    International Nuclear Information System (INIS)

    Paegert, Martin; Stassun, Keivan G.; Burger, Dan M.

    2014-01-01

    We describe a new neural-net-based light curve classifier and provide it with documentation as a ready-to-use tool for the community. While optimized for identification and classification of eclipsing binary stars, the classifier is general purpose, and has been developed for speed in the context of upcoming massive surveys such as the Large Synoptic Survey Telescope. A challenge for classifiers in the context of neural-net training and massive data sets is to minimize the number of parameters required to describe each light curve. We show that a simple and fast geometric representation that encodes the overall light curve shape, together with a chi-square parameter to capture higher-order morphology information results in efficient yet robust light curve classification, especially for eclipsing binaries. Testing the classifier on the ASAS light curve database, we achieve a retrieval rate of 98% and a false-positive rate of 2% for eclipsing binaries. We achieve similarly high retrieval rates for most other periodic variable-star classes, including RR Lyrae, Mira, and delta Scuti. However, the classifier currently has difficulty discriminating between different sub-classes of eclipsing binaries, and suffers a relatively low (∼60%) retrieval rate for multi-mode delta Cepheid stars. We find that it is imperative to train the classifier's neural network with exemplars that include the full range of light curve quality to which the classifier will be expected to perform; the classifier performs well on noisy light curves only when trained with noisy exemplars. The classifier source code, ancillary programs, a trained neural net, and a guide for use, are provided.

  2. The revised APTA code of ethics for the physical therapist and standards of ethical conduct for the physical therapist assistant: theory, purpose, process, and significance.

    Science.gov (United States)

    Swisher, Laura Lee; Hiller, Peggy

    2010-05-01

    In June 2009, the House of Delegates (HOD) of the American Physical Therapy Association (APTA) passed a major revision of the APTA Code of Ethics for physical therapists and the Standards of Ethical Conduct for the Physical Therapist Assistant. The revised documents will be effective July 1, 2010. The purposes of this article are: (1) to provide a historical, professional, and theoretical context for this important revision; (2) to describe the 4-year revision process; (3) to examine major features of the documents; and (4) to discuss the significance of the revisions from the perspective of the maturation of physical therapy as a doctoring profession. PROCESS OF REVISION: The process for revision is delineated within the context of history and the Bylaws of APTA. FORMAT, STRUCTURE, AND CONTENT OF REVISED CORE ETHICS DOCUMENTS: The revised documents represent a significant change in format, level of detail, and scope of application. Previous APTA Codes of Ethics and Standards of Ethical Conduct for the Physical Therapist Assistant have delineated very broad general principles, with specific obligations spelled out in the Ethics and Judicial Committee's Guide for Professional Conduct and Guide for Conduct of the Physical Therapist Assistant. In contrast to the current documents, the revised documents address all 5 roles of the physical therapist, delineate ethical obligations in organizational and business contexts, and align with the tenets of Vision 2020. The significance of this revision is discussed within historical parameters, the implications for physical therapists and physical therapist assistants, the maturation of the profession, societal accountability and moral community, potential regulatory implications, and the inclusive and deliberative process of moral dialogue by which changes were developed, revised, and approved.

  3. Apple-CORE: Microgrids of SVP cores: flexible, general-purpose, fine-grained hardware concurrency management

    NARCIS (Netherlands)

    Poss, R.; Lankamp, M.; Yang, Q.; Fu, J.; van Tol, M.W.; Jesshope, C.; Nair, S.

    2012-01-01

    To harness the potential of CMPs for scalable, energy-efficient performance in general-purpose computers, the Apple-CORE project has co-designed a general machine model and concurrency control interface with dedicated hardware support for concurrency control across multiple cores. Its SVP interface

  4. Variable Conductance Heat Pipe Cooling of Stirling Convertor and General Purpose Heat Source

    Science.gov (United States)

    Tarau, Calin; Schwendeman, Carl; Anderson, William G.; Cornell, Peggy A.; Schifer, Nicholas A.

    2013-01-01

    In a Stirling Radioisotope Power System (RPS), heat must be continuously removed from the General Purpose Heat Source (GPHS) modules to maintain the modules and surrounding insulation at acceptable temperatures. The Stirling convertor normally provides this cooling. If the Stirling convertor stops in the current system, the insulation is designed to spoil, preventing damage to the GPHS at the cost of an early termination of the mission. An alkali-metal Variable Conductance Heat Pipe (VCHP) can be used to passively allow multiple stops and restarts of the Stirling convertor. In a previous NASA SBIR Program, Advanced Cooling Technologies, Inc. (ACT) developed a series of sodium VCHPs as backup cooling systems for Stirling RPS. The operation of these VCHPs was demonstrated using Stirling heater head simulators and GPHS simulators. In the most recent effort, a sodium VCHP with a stainless steel envelope was designed, fabricated and tested at NASA Glenn Research Center (GRC) with a Stirling convertor for two concepts; one for the Advanced Stirling Radioisotope Generator (ASRG) back up cooling system and one for the Long-lived Venus Lander thermal management system. The VCHP is designed to activate and remove heat from the stopped convertor at a 19 degC temperature increase from the nominal vapor temperature. The 19 degC temperature increase from nominal is low enough to avoid risking standard ASRG operation and spoiling of the Multi-Layer Insulation (MLI). In addition, the same backup cooling system can be applied to the Stirling convertor used for the refrigeration system of the Long-lived Venus Lander. The VCHP will allow the refrigeration system to: 1) rest during transit at a lower temperature than nominal; 2) pre-cool the modules to an even lower temperature before the entry in Venus atmosphere; 3) work at nominal temperature on Venus surface; 4) briefly stop multiple times on the Venus surface to allow scientific measurements. This paper presents the experimental

  5. The Chronic Kidney Disease Model: A General Purpose Model of Disease Progression and Treatment

    Directory of Open Access Journals (Sweden)

    Patel Uptal D

    2011-06-01

    Full Text Available Abstract Background Chronic kidney disease (CKD is the focus of recent national policy efforts; however, decision makers must account for multiple therapeutic options, comorbidities and complications. The objective of the Chronic Kidney Disease model is to provide guidance to decision makers. We describe this model and give an example of how it can inform clinical and policy decisions. Methods Monte Carlo simulation of CKD natural history and treatment. Health states include myocardial infarction, stroke with and without disability, congestive heart failure, CKD stages 1-5, bone disease, dialysis, transplant and death. Each cycle is 1 month. Projections account for race, age, gender, diabetes, proteinuria, hypertension, cardiac disease, and CKD stage. Treatment strategies include hypertension control, diabetes control, use of HMG-CoA reductase inhibitors, use of angiotensin converting enzyme inhibitors, nephrology specialty care, CKD screening, and a combination of these. The model architecture is flexible permitting updates as new data become available. The primary outcome is quality adjusted life years (QALYs. Secondary outcomes include health state events and CKD progression rate. Results The model was validated for GFR change/year -3.0 ± 1.9 vs. -1.7 ± 3.4 (in the AASK trial, and annual myocardial infarction and mortality rates 3.6 ± 0.9% and 1.6 ± 0.5% vs. 4.4% and 1.6% in the Go study. To illustrate the model's utility we estimated lifetime impact of a hypothetical treatment for primary prevention of vascular disease. As vascular risk declined, QALY improved but risk of dialysis increased. At baseline, 20% and 60% reduction: QALYs = 17.6, 18.2, and 19.0 and dialysis = 7.7%, 8.1%, and 10.4%, respectively. Conclusions The CKD Model is a valid, general purpose model intended as a resource to inform clinical and policy decisions improving CKD care. Its value as a tool is illustrated in our example which projects a relationship between

  6. Plasduino: An inexpensive, general-purpose data acquisition framework for educational experiments

    International Nuclear Information System (INIS)

    Baldini, L.

    2014-01-01

    Based on the Arduino development platform, plasduino is an open source data acquisition framework specifically designed for educational physics experiments. The source code, schematics and documentation are in the public domain under a GPL license and the system, streamlined for low cost and ease of use, can be replicated on the scale of a typical didactic lab with minimal effort. We describe the basic architecture of the system and illustrate its potential with some real-life examples.

  7. Plasduino: An inexpensive, general-purpose data acquisition framework for educational experiments

    Energy Technology Data Exchange (ETDEWEB)

    Baldini, L. [Universita' di Pisa and INFN Sez. di Pisa, Pisa (Italy)

    2014-07-15

    Based on the Arduino development platform, plasduino is an open source data acquisition framework specifically designed for educational physics experiments. The source code, schematics and documentation are in the public domain under a GPL license and the system, streamlined for low cost and ease of use, can be replicated on the scale of a typical didactic lab with minimal effort. We describe the basic architecture of the system and illustrate its potential with some real-life examples.

  8. A MacWilliams Identity for Convolutional Codes: The General Case

    OpenAIRE

    Gluesing-Luerssen, Heide; Schneider, Gert

    2008-01-01

    A MacWilliams Identity for convolutional codes will be established. It makes use of the weight adjacency matrices of the code and its dual, based on state space realizations (the controller canonical form) of the codes in question. The MacWilliams Identity applies to various notions of duality appearing in the literature on convolutional coding theory.

  9. International common-cause failure data exchange. ICDE general coding guidelines - Technical note

    International Nuclear Information System (INIS)

    Johanson, Gunnar; Werner, Wolfgang; Concepcion Capote, Marina; Kreuser, Albert; Rasmuson, Dale; Jonsson, Esther; Pereira Pagan, Begona; Tirira, Jorge; Morris, Ian; Morales, Rosa; Oxberry, Anna; Kreuser, Albert

    2004-01-01

    Several Member countries of the Nuclear Energy Agency of the Organisation for Economic Co-operation and Development (OECD/NEA) have established the International Common-Cause Failure Data Exchange Project (ICDE Project) to encourage multilateral co-operation in the collection and analysis of data relating to Common-Cause Failure (CCF) events. The objectives of the ICDE Project are to: a) Collect and analyse CCF events over the long term so as to better understand such events, their causes, and their prevention; b) Generate qualitative insights into the root causes of CCF events which can then be used to derive approaches or mechanisms for their prevention or for mitigating their consequences; c) Establish a mechanism for the efficient feedback of experience gained in connection with CCF phenomena, including the development of defences against their occurrence, such as indicators for risk based inspections; and d) Record event attributes to facilitate quantification of CCF frequencies when so decided by the Project Working Group. The ICDE Project is envisaged to comprise all possible events of interest, including both complete and partial ICDE events. The ICDE Project will cover the key components of the main safety systems. Presently, the components listed below are included in the ICDE Project. Data have been collected for the six first components in the list: Centrifugal pumps, Diesel generators, Motor operated valves, Safety relief valves/power operated relief valves, Check valves, Batteries, Level measurement, Breakers, Control rod drive assemblies. Others will be added to this list later on. In this component coding guidelines, explanations on the ICDE General coding format are given. The guide reflects present experience with the data format and with the collected data. Further interpretations and clarifications will be added, should they become necessary. For each component analysed in the ICDE project, separate coding guidance is provided in the appendices

  10. International Common Cause Failure Data Exchange (ICDE). General Coding Guidelines - Updated Version, October 2011

    International Nuclear Information System (INIS)

    Johanson, Gunnar; Werner, Wolfgang; Capote, Marina Concepcion; Kreuser, Albert

    2012-01-01

    Several OECD/NEA member countries have established the International Common-Cause Failure Data Exchange Project ('ICDE Project') to encourage multilateral cooperation in the collection and analysis of data relating to Common-Cause Failure (CCF) events. The objectives of the ICDE Project are to: a) Collect and analyse CCF events over the long term so as to better understand such events, their causes, and their prevention. b) Generate qualitative insights into the root causes of CCF events which can then be used to derive approaches or mechanisms for their prevention or for mitigating their consequences. c) Establish a mechanism for the efficient feedback of experience gained in connection with CCF phenomena, including the development of defenses against their occurrence, such as indicators for risk based inspections. d) Record event attributes to facilitate quantification of CCF frequencies when so decided by the Project Working Group. The ICDE Project is envisaged to comprise all possible events of interest, including both complete and partial ICDE events. The ICDE Project will cover the key components of the main safety systems. Presently, the components listed below are included in the ICDE Project. Data have been collected for the six first components in the list: Centrifugal pumps, Diesel generators, Motor operated valves, Safety relief valves/power operated relief valves, Check valves, Batteries, Level measurement, Breakers, Control rod drive assemblies. Others will be added to this list later on. In this component coding guidelines, explanations on the ICDE general coding format are given. The guide reflects present experience with the data format and with the collected data. Further interpretations and clarifications will be added, should they become necessary. For each component analysed in the ICDE project, separate coding guidance is provided in the appendices ICDECG 01-06, specifying details relevant to the respective components

  11. A neutron spectrum unfolding code based on generalized regression artificial neural networks

    International Nuclear Information System (INIS)

    Rosario Martinez-Blanco, Ma. del

    2016-01-01

    The most delicate part of neutron spectrometry, is the unfolding process. The derivation of the spectral information is not simple because the unknown is not given directly as a result of the measurements. Novel methods based on Artificial Neural Networks have been widely investigated. In prior works, back propagation neural networks (BPNN) have been used to solve the neutron spectrometry problem, however, some drawbacks still exist using this kind of neural nets, i.e. the optimum selection of the network topology and the long training time. Compared to BPNN, it's usually much faster to train a generalized regression neural network (GRNN). That's mainly because spread constant is the only parameter used in GRNN. Another feature is that the network will converge to a global minimum, provided that the optimal values of spread has been determined and that the dataset adequately represents the problem space. In addition, GRNN are often more accurate than BPNN in the prediction. These characteristics make GRNNs to be of great interest in the neutron spectrometry domain. This work presents a computational tool based on GRNN capable to solve the neutron spectrometry problem. This computational code, automates the pre-processing, training and testing stages using a k-fold cross validation of 3 folds, the statistical analysis and the post-processing of the information, using 7 Bonner spheres rate counts as only entrance data. The code was designed for a Bonner Spheres System based on a "6LiI(Eu) neutron detector and a response matrix expressed in 60 energy bins taken from an International Atomic Energy Agency compilation. - Highlights: • Main drawback of neutron spectrometry with BPNN is network topology optimization. • Compared to BPNN, it’s usually much faster to train a (GRNN). • GRNN are often more accurate than BPNN in the prediction. These characteristics make GRNNs to be of great interest. • This computational code, automates the pre-processing, training

  12. Teacher's and Students' Beliefs on English for General Academic Purposes: The Case of Iranian University Students

    Science.gov (United States)

    Kojour, Masoud Kermani; Heirati, Javad Kia

    2015-01-01

    This study was framed in the sociocultural theory to look into the evolution of L2 learners' beliefs about the general English course during a term. One hundred ninety-eight male and female university students and their general English course teacher were randomly selected as the participants of the study. Data were gathered through the…

  13. Parallelized computation for computer simulation of electrocardiograms using personal computers with multi-core CPU and general-purpose GPU.

    Science.gov (United States)

    Shen, Wenfeng; Wei, Daming; Xu, Weimin; Zhu, Xin; Yuan, Shizhong

    2010-10-01

    Biological computations like electrocardiological modelling and simulation usually require high-performance computing environments. This paper introduces an implementation of parallel computation for computer simulation of electrocardiograms (ECGs) in a personal computer environment with an Intel CPU of Core (TM) 2 Quad Q6600 and a GPU of Geforce 8800GT, with software support by OpenMP and CUDA. It was tested in three parallelization device setups: (a) a four-core CPU without a general-purpose GPU, (b) a general-purpose GPU plus 1 core of CPU, and (c) a four-core CPU plus a general-purpose GPU. To effectively take advantage of a multi-core CPU and a general-purpose GPU, an algorithm based on load-prediction dynamic scheduling was developed and applied to setting (c). In the simulation with 1600 time steps, the speedup of the parallel computation as compared to the serial computation was 3.9 in setting (a), 16.8 in setting (b), and 20.0 in setting (c). This study demonstrates that a current PC with a multi-core CPU and a general-purpose GPU provides a good environment for parallel computations in biological modelling and simulation studies. Copyright 2010 Elsevier Ireland Ltd. All rights reserved.

  14. A neutron spectrum unfolding code based on generalized regression artificial neural networks

    International Nuclear Information System (INIS)

    Ortiz R, J. M.; Martinez B, M. R.; Castaneda M, R.; Solis S, L. O.; Vega C, H. R.

    2015-10-01

    The most delicate part of neutron spectrometry, is the unfolding process. Then derivation of the spectral information is not simple because the unknown is not given directly as result of the measurements. Novel methods based on Artificial Neural Networks have been widely investigated. In prior works, back propagation neural networks (BPNN) have been used to solve the neutron spectrometry problem, however, some drawbacks still exist using this kind of neural nets, as the optimum selection of the network topology and the long training time. Compared to BPNN, is usually much faster to train a generalized regression neural network (GRNN). That is mainly because spread constant is the only parameter used in GRNN. Another feature is that the network will converge to a global minimum. In addition, often are more accurate than BPNN in prediction. These characteristics make GRNN be of great interest in the neutron spectrometry domain. In this work is presented a computational tool based on GRNN, capable to solve the neutron spectrometry problem. This computational code, automates the pre-processing, training and testing stages, the statistical analysis and the post-processing of the information, using 7 Bonner spheres rate counts as only entrance data. The code was designed for a Bonner Spheres System based on a 6 LiI(Eu) neutron detector and a response matrix expressed in 60 energy bins taken from an International Atomic Energy Agency compilation. (Author)

  15. A neutron spectrum unfolding code based on generalized regression artificial neural networks

    Energy Technology Data Exchange (ETDEWEB)

    Ortiz R, J. M.; Martinez B, M. R.; Castaneda M, R.; Solis S, L. O. [Universidad Autonoma de Zacatecas, Unidad Academica de Ingenieria Electrica, Av. Ramon Lopez Velarde 801, Col. Centro, 98000 Zacatecas, Zac. (Mexico); Vega C, H. R., E-mail: morvymm@yahoo.com.mx [Universidad Autonoma de Zacatecas, Unidad Academica de Estudios Nucleares, Cipres No. 10, Fracc. La Penuela, 98068 Zacatecas, Zac. (Mexico)

    2015-10-15

    The most delicate part of neutron spectrometry, is the unfolding process. Then derivation of the spectral information is not simple because the unknown is not given directly as result of the measurements. Novel methods based on Artificial Neural Networks have been widely investigated. In prior works, back propagation neural networks (BPNN) have been used to solve the neutron spectrometry problem, however, some drawbacks still exist using this kind of neural nets, as the optimum selection of the network topology and the long training time. Compared to BPNN, is usually much faster to train a generalized regression neural network (GRNN). That is mainly because spread constant is the only parameter used in GRNN. Another feature is that the network will converge to a global minimum. In addition, often are more accurate than BPNN in prediction. These characteristics make GRNN be of great interest in the neutron spectrometry domain. In this work is presented a computational tool based on GRNN, capable to solve the neutron spectrometry problem. This computational code, automates the pre-processing, training and testing stages, the statistical analysis and the post-processing of the information, using 7 Bonner spheres rate counts as only entrance data. The code was designed for a Bonner Spheres System based on a {sup 6}LiI(Eu) neutron detector and a response matrix expressed in 60 energy bins taken from an International Atomic Energy Agency compilation. (Author)

  16. Simulation of the preliminary General Electric SP-100 space reactor concept using the ATHENA computer code

    International Nuclear Information System (INIS)

    Fletcher, C.D.

    1986-01-01

    The capability to perform thermal-hydraulic analyses of a space reactor using the ATHENA computer code is demonstrated. The fast reactor, liquid-lithium coolant loops, and lithium-filled heat pipes of the preliminary General electric SP-100 design were modeled with ATHENA. Two demonstration transient calculations were performed simulating accident conditions. Calculated results are available for display using the Nuclear Plant Analyzer color graphics analysis tool in addition to traditional plots. ATHENA-calculated results appear reasonable, both for steady state full power conditions, and for the two transients. This analysis represents the first known transient thermal-hydraulic simulation using an integral space reactor system model incorporating heat pipes. 6 refs., 17 figs., 1 tab

  17. Beyond the "Ivory Tower": Restoring the Balance of Private and Public Purposes of General Education

    Science.gov (United States)

    Cantor, Nancy; Englot, Peter

    2013-01-01

    We in higher education have allowed the balance of the private and public purposes of liberal education to become skewed in our angst over fully embracing a central role in restoring prosperity and preparing our increasingly diverse population for the evolving responsibility of mending our fraying democracy. As we consider how best to reshape…

  18. Factors Affecting Preservice Teachers' Computer Use for General Purposes: Implications for Computer Training Courses

    Science.gov (United States)

    Zogheib, Salah

    2014-01-01

    As the majority of educational research has focused on preservice teachers' computer use for "educational purposes," the question remains: Do preservice teachers use computer technology for daily life activities and encounters? And do preservice teachers' personality traits and motivational beliefs related to computer training provided…

  19. APL/JHU free flight tests of the General Purpose Heat Source module. Testing: 5-7 March 1984

    International Nuclear Information System (INIS)

    Baker, W.M. II.

    1984-01-01

    Purpose of the test was to obtain statistical information on the dynamics of the General Purpose Heat Source (GPHS) module at terminal speeds. Models were designed to aerodynamically and dynamically represent the GPHS module. Normal and high speed photographic coverage documented the motion of the models. This report documents test parameters and techniques for the free-spin tests. It does not include data analysis

  20. Detectability of T1a lung cancer on digital chest radiographs: an observer-performance comparison among 2-megapixel general-purpose, 2-megapixel medical-purpose, and 3-megapixel medical-purpose liquid-crystal display (LCD) monitors.

    Science.gov (United States)

    Yabuuchi, Hidetake; Matsuo, Yoshio; Kamitani, Takeshi; Jinnnouchi, Mikako; Yonezawa, Masato; Yamasaki, Yuzo; Nagao, Michinobu; Kawanami, Satoshi; Okamoto, Tatsuro; Sasaki, Masayuki; Honda, Hiroshi

    2015-08-01

    There has been no comparison of detectability of small lung cancer between general and medical LCD monitors or no comparison of detectability of small lung cancer between solid and part-solid nodules. To compare the detectabilities of T1a lung cancer on chest radiographs on three LCD monitor types: 2-megapixel (MP) for general purpose (General), 2-MP for medical purpose (Medical), and 3-MP-Medical. Radiographs from forty patients with T1aN0M0 primary lung cancer (27 solid nodules, 13 part-solid nodules) and 60 patients with no abnormalities on both chest X-ray and computed tomography (CT) were consecutively collected. Five readers assessed 100 cases for each monitor. The observations were analyzed using receiver operating characteristic (ROC) analysis. A jackknife method was used for statistical analysis. A P value of General, 2-MP-Medical, and 3-MP-Medical LCD monitors were 0.86, 0.89, and 0.89, respectively; there were no significant differences among them. The average AUC for part-solid nodule detection using a 2-MP-General, 2-MP-Medical, and 3-MP-Medical LCD monitors were 0.77, 0.86, and 0.89, respectively. There were significant differences between the 2-MP-General and 2-MP-Medical LCD monitors (P = 0.043) and between the 2-MP-General and 3-MP-Medical LCD monitors (P = 0.027). There was no significant difference between the 2-MP-Medical and 3-MP-Medical LCD monitors. The average AUC for solid nodule detection using a 2-MP-General, 2-MP-Medical, and 3-MP-Medical LCD monitors were 0.90, 0.90, and 0.88, respectively; there were no significant differences among them. The mean AUC values for all and part-solid nodules of the low-experienced readers were significantly lower than those of the high-experienced readers with the 2 M-GP color LCD monitor (P general-purpose LCD monitor was significantly lower than those using medical-purpose LCD monitors. © The Foundation Acta Radiologica 2014.

  1. KDAS: General-Purpose Data Acquisition System Developed for KAIST-Tokamak

    International Nuclear Information System (INIS)

    Seo, Seong-Heon; Choe, Wonho; Chang, Hong-Young; Jeong, Seung-Ho

    2000-01-01

    The Korea Advanced Institute of Science and Technology (KAIST)-Tokamak Data Acquisition System (KDAS) was originally developed for KAIST-Tokamak (R/a = 0.53 m/0.14 m). It operates on a distributed system based on personal computers and has a driver-based hierarchical structure. Since KDAS can be dynamically composed of any number of available computers, and the hardware-dependent codes can be thoroughly separated into external drivers, it exhibits excellent system performance flexibility and extensibility and can optimize various user needs. It collectively controls the VXI, CAMAC, GPIB, and RS232 instrument hybrids. With these useful and convenient features, it can be applied to any computerized experiment, especially to fusion-related research. The system design and features are discussed in detail

  2. General purpose - expert system for the analysis and design of base plates

    International Nuclear Information System (INIS)

    Al-Shawaf, T.D.; Hahn, W.F.; Ho, A.D.

    1987-01-01

    As an expert system, the IMPLATE program uses plant specific information to make decisions in modeling and analysis of baseplates. The user supplies a minimum of information which is checked for validity and reasonableness. Once this data is supplied, the program automatically generates a compatible mesh and finite element model from its data base accounting for the attachments, stiffeners, anchor bolts and plate/concrete interface. Based on the loading direction, the program deletes certain degrees of freedom and performs a linear or a nonlinear solution, whichever is appropriate. Load step sizes and equilibrium iteration are automatically selected by the program to ensure a convergent solution. Once the analysis is completed, a code check is then performed and a summary of results is produced. Plots of the plate deformation pattern and stress contours are also generated. (orig.)

  3. The ESPAT tool: a general-purpose DSS shell for solving stochastic optimization problems in complex river-aquifer systems

    Science.gov (United States)

    Macian-Sorribes, Hector; Pulido-Velazquez, Manuel; Tilmant, Amaury

    2015-04-01

    Stochastic programming methods are better suited to deal with the inherent uncertainty of inflow time series in water resource management. However, one of the most important hurdles in their use in practical implementations is the lack of generalized Decision Support System (DSS) shells, usually based on a deterministic approach. The purpose of this contribution is to present a general-purpose DSS shell, named Explicit Stochastic Programming Advanced Tool (ESPAT), able to build and solve stochastic programming problems for most water resource systems. It implements a hydro-economic approach, optimizing the total system benefits as the sum of the benefits obtained by each user. It has been coded using GAMS, and implements a Microsoft Excel interface with a GAMS-Excel link that allows the user to introduce the required data and recover the results. Therefore, no GAMS skills are required to run the program. The tool is divided into four modules according to its capabilities: 1) the ESPATR module, which performs stochastic optimization procedures in surface water systems using a Stochastic Dual Dynamic Programming (SDDP) approach; 2) the ESPAT_RA module, which optimizes coupled surface-groundwater systems using a modified SDDP approach; 3) the ESPAT_SDP module, capable of performing stochastic optimization procedures in small-size surface systems using a standard SDP approach; and 4) the ESPAT_DET module, which implements a deterministic programming procedure using non-linear programming, able to solve deterministic optimization problems in complex surface-groundwater river basins. The case study of the Mijares river basin (Spain) is used to illustrate the method. It consists in two reservoirs in series, one aquifer and four agricultural demand sites currently managed using historical (XIV century) rights, which give priority to the most traditional irrigation district over the XX century agricultural developments. Its size makes it possible to use either the SDP or

  4. On locality of Generalized Reed-Muller codes over the broadcast erasure channel

    KAUST Repository

    Alloum, Amira; Lin, Sian Jheng; Al-Naffouri, Tareq Y.

    2016-01-01

    , and more specifically at the application layer where Rateless, LDPC, Reed Slomon codes and network coding schemes have been extensively studied, optimized and standardized in the past. Beyond reusing, extending or adapting existing application layer packet

  5. The Dynamics of a General Purpose Technology in a Research and Assimilation Model

    NARCIS (Netherlands)

    Nahuis, R.

    1998-01-01

    Where is the productivity growth from the IT revolution? Why did the skill premium rise sharply in the early eighties? Were these phenomena related? This paper examines these questions in a general equilibrium model of growth. Technological progress in firms is driven by research aimed at improving

  6. Accuracy of Surface Plate Measurements - General Purpose Software for Flatness Measurement

    NARCIS (Netherlands)

    Meijer, J.; Heuvelman, C.J.

    1990-01-01

    Flatness departures of surface plates are generally obtained from straightness measurements of lines on the surface. A computer program has been developed for on-line measurement and evaluation, based on the simultaneous coupling of measurements in all grid points. Statistical methods are used to

  7. 41 CFR 60-2.10 - General purpose and contents of affirmative action programs.

    Science.gov (United States)

    2010-07-01

    ... central premise underlying affirmative action is that, absent discrimination, over time a contractor's workforce, generally, will reflect the gender, racial and ethnic profile of the labor pools from which the... include action-oriented programs. If women and minorities are not being employed at a rate to be expected...

  8. A Framework for a General Purpose Intelligent Control System for Particle Accelerators. Phase II Final Report

    International Nuclear Information System (INIS)

    Westervelt, Robert; Klein, William; Kroupa, Michael; Olsson, Eric; Rothrock, Rick

    1999-01-01

    Vista Control Systems, Inc. has developed a portable system for intelligent accelerator control. The design is general in scope and is thus configurable to a wide range of accelerator facilities and control problems. The control system employs a multi-layer organization in which knowledge-based decision making is used to dynamically configure lower level optimization and control algorithms

  9. A MacWilliams Identity for Convolutional Codes : The General Case

    NARCIS (Netherlands)

    Gluesing-Luerssen, Heide; Schneider, Gert

    A MacWilliams Identity for convolutional codes will be established. It makes use of the weight adjacency matrices of the code and its dual, based on state space realizations (the controller canonical form) of the codes in question. The MacWilliams Identity applies to various notions of duality

  10. Code of practice for the control and safe handling of radioactive sources used for therapeutic purposes (1988)

    International Nuclear Information System (INIS)

    1988-01-01

    This Code is intended as a guide to safe practices in the use of sealed and unsealed radioactive sources and in the management of patients being treated with them. It covers the procedures for the handling, preparation and use of radioactive sources, precautions to be taken for patients undergoing treatment, storage and transport of radioactive sources within a hospital or clinic, and routine testing of sealed sources [fr

  11. SMILEI: A collaborative, open-source, multi-purpose PIC code for the next generation of super-computers

    Science.gov (United States)

    Grech, Mickael; Derouillat, J.; Beck, A.; Chiaramello, M.; Grassi, A.; Niel, F.; Perez, F.; Vinci, T.; Fle, M.; Aunai, N.; Dargent, J.; Plotnikov, I.; Bouchard, G.; Savoini, P.; Riconda, C.

    2016-10-01

    Over the last decades, Particle-In-Cell (PIC) codes have been central tools for plasma simulations. Today, new trends in High-Performance Computing (HPC) are emerging, dramatically changing HPC-relevant software design and putting some - if not most - legacy codes far beyond the level of performance expected on the new and future massively-parallel super computers. SMILEI is a new open-source PIC code co-developed by both plasma physicists and HPC specialists, and applied to a wide range of physics-related studies: from laser-plasma interaction to astrophysical plasmas. It benefits from an innovative parallelization strategy that relies on a super-domain-decomposition allowing for enhanced cache-use and efficient dynamic load balancing. Beyond these HPC-related developments, SMILEI also benefits from additional physics modules allowing to deal with binary collisions, field and collisional ionization and radiation back-reaction. This poster presents the SMILEI project, its HPC capabilities and illustrates some of the physics problems tackled with SMILEI.

  12. Analysis of impact of general-purpose graphics processor units in supersonic flow modeling

    Science.gov (United States)

    Emelyanov, V. N.; Karpenko, A. G.; Kozelkov, A. S.; Teterina, I. V.; Volkov, K. N.; Yalozo, A. V.

    2017-06-01

    Computational methods are widely used in prediction of complex flowfields associated with off-normal situations in aerospace engineering. Modern graphics processing units (GPU) provide architectures and new programming models that enable to harness their large processing power and to design computational fluid dynamics (CFD) simulations at both high performance and low cost. Possibilities of the use of GPUs for the simulation of external and internal flows on unstructured meshes are discussed. The finite volume method is applied to solve three-dimensional unsteady compressible Euler and Navier-Stokes equations on unstructured meshes with high resolution numerical schemes. CUDA technology is used for programming implementation of parallel computational algorithms. Solutions of some benchmark test cases on GPUs are reported, and the results computed are compared with experimental and computational data. Approaches to optimization of the CFD code related to the use of different types of memory are considered. Speedup of solution on GPUs with respect to the solution on central processor unit (CPU) is compared. Performance measurements show that numerical schemes developed achieve 20-50 speedup on GPU hardware compared to CPU reference implementation. The results obtained provide promising perspective for designing a GPU-based software framework for applications in CFD.

  13. Applications for General Purpose Command Buffers: The Emergency Conjunction Avoidance Maneuver

    Science.gov (United States)

    Scheid, Robert J; England, Martin

    2016-01-01

    A case study is presented for the use of Relative Operation Sequence (ROS) command buffers to quickly execute a propulsive maneuver to avoid a collision with space debris. In this process, a ROS is custom-built with a burn time and magnitude, uplinked to the spacecraft, and executed in 15 percent of the time of the previous method. This new process provides three primary benefits. First, the planning cycle can be delayed until it is certain a burn must be performed, reducing team workload. Second, changes can be made to the burn parameters almost up to the point of execution while still allowing the normal uplink product review process, reducing the risk of leaving the operational orbit because of outdated burn parameters, and minimizing the chance of accidents from human error, such as missed commands, in a high-stress situation. Third, the science impacts can be customized and minimized around the burn, and in the event of an abort can be eliminated entirely in some circumstances. The result is a compact burn process that can be executed in as few as four hours and can be aborted seconds before execution. Operational, engineering, planning, and flight dynamics perspectives are presented, as well as a functional overview of the code and workflow required to implement the process. Future expansions and capabilities are also discussed.

  14. 369 TFlop/s molecular dynamics simulations on the Roadrunner general-purpose heterogeneous supercomputer

    Energy Technology Data Exchange (ETDEWEB)

    Swaminarayan, Sriram [Los Alamos National Laboratory; Germann, Timothy C [Los Alamos National Laboratory; Kadau, Kai [Los Alamos National Laboratory; Fossum, Gordon C [IBM CORPORATION

    2008-01-01

    The authors present timing and performance numbers for a short-range parallel molecular dynamics (MD) code, SPaSM, that has been rewritten for the heterogeneous Roadrunner supercomputer. Each Roadrunner compute node consists of two AMD Opteron dual-core microprocessors and four PowerXCell 8i enhanced Cell microprocessors, so that there are four MPI ranks per node, each with one Opteron and one Cell. The interatomic forces are computed on the Cells (each with one PPU and eight SPU cores), while the Opterons are used to direct inter-rank communication and perform I/O-heavy periodic analysis, visualization, and checkpointing tasks. The performance measured for our initial implementation of a standard Lennard-Jones pair potential benchmark reached a peak of 369 Tflop/s double-precision floating-point performance on the full Roadrunner system (27.7% of peak), corresponding to 124 MFlop/Watt/s at a price of approximately 3.69 MFlops/dollar. They demonstrate an initial target application, the jetting and ejection of material from a shocked surface.

  15. General-Purpose Computation with Neural Networks: A Survey of Complexity Theoretic Results

    Czech Academy of Sciences Publication Activity Database

    Šíma, Jiří; Orponen, P.

    2003-01-01

    Roč. 15, č. 12 (2003), s. 2727-2778 ISSN 0899-7667 R&D Projects: GA AV ČR IAB2030007; GA ČR GA201/02/1456 Institutional research plan: AV0Z1030915 Keywords : computational power * computational complexity * perceptrons * radial basis functions * spiking neurons * feedforward networks * reccurent networks * probabilistic computation * analog computation Subject RIV: BA - General Mathematics Impact factor: 2.747, year: 2003

  16. A Comprehensive Toolset for General-Purpose Private Computing and Outsourcing

    Science.gov (United States)

    2016-12-08

    contexts businesses are also hesitant to make their proprietary available to the cloud [1]. While in general sensitive data can be protected by the...data sources, gathering and maintaining the data needed , and completing and reviewing the collection of information. Send comments regarding this...project and scientific advances made towards each of the research thrusts throughout the project duration. 1 Project Objectives Cloud computing enables

  17. Preparing General Purpose Forces in the United States and British Armies for Counterinsurgent Operations

    Science.gov (United States)

    2010-12-10

    Operations In Iraq: Planning, Combat, And Occupation,” Thomas Ricks’ Fiasco, and reports by Army historian Major Isaiah Wilson, and former CENCTOM J-4...established Multi-National Forces-Iraq, and Lieutenant General Thomas Metz, commander of the Army’s III Corps, assumed the mantle of Multi-National Corps...Donald P. Wright and Colonel Timothy R. Reese or Thomas Ricks’ Fiasco among other books. 18Dr. Carter Malkasian, “Counterinsurgency in Iraq: May 2003

  18. How did the General Purpose Technology ’Electricity’ contribute to the Second Industrial Revolution (I): The Power Engines.

    NARCIS (Netherlands)

    van der Kooij, B.J.G.

    2016-01-01

    The concept of the General Purpose Technology (GPT) of the late 1990s is a culmination of many evolutionairy views in innovation-thinking. By definition the GPT considers the technical, social, and economic effects of meta-technologies like steam-technology and electric technology. This paper uses

  19. [Application of the grayscale standard display function to general purpose liquid-crystal display monitors for clinical use].

    Science.gov (United States)

    Tanaka, Nobukazu; Naka, Kentaro; Sueoka, Masaki; Higashida, Yoshiharu; Morishita, Junji

    2010-01-20

    Interpretations of medical images have been shifting to soft-copy readings with liquid-crystal display (LCD) monitors. The display function of the medical-grade LCD monitor for soft-copy readings is recommended to calibrate the grayscale standard display function (GSDF) in accordance with the guidelines of Japan and other countries. In this study, the luminance and display function of five models of eight general purpose LCD monitors were measured to gain an understanding of their characteristics. Moreover, the display function (gamma 2.2 or gamma 1.8) of general purpose LCD monitors was converted to GSDF through the use of a look-up table, and the detectability of a simulated lung nodule in the chest x-ray image was examined. As a result, the maximum luminance, contrast ratio, and luminance uniformity of general purpose LCD monitors, except for one model of two LCD monitors, met the management grade 1 standard in the guideline JESRA X-0093-2005. In addition, the detectability of simulated lung nodule in the mediastinal space was obviously improved by converting the display function of a general purpose LCD monitor into GSDF.

  20. RUMD: A general purpose molecular dynamics package optimized to utilize GPU hardware down to a few thousand particles

    DEFF Research Database (Denmark)

    Bailey, Nicholas; Ingebrigtsen, Trond; Hansen, Jesper Schmidt

    2017-01-01

    RUMD is a general purpose, high-performance molecular dynamics (MD) simulation package running on graphical processing units (GPU’s). RUMD addresses the challenge of utilizing the many-core nature of modern GPU hardware when simulating small to medium system sizes (roughly from a few thousand up...

  1. How did the General Purpose Technology Electricity contribute to the Second Industrial Revolution (II): The Communication Engines

    NARCIS (Netherlands)

    van der Kooij, B.J.G.

    2017-01-01

    The concept of the General Purpose Technology (GPT) of the late 1990s is a culmination of many evolutionairy views in innovation-thinking. By definition the GPT considers the technical, social, and economic effects of meta-technologies like steam-technology and electric technology. This paper uses

  2. Limits to high-speed simulations of spiking neural networks using general-purpose computers.

    Science.gov (United States)

    Zenke, Friedemann; Gerstner, Wulfram

    2014-01-01

    To understand how the central nervous system performs computations using recurrent neuronal circuitry, simulations have become an indispensable tool for theoretical neuroscience. To study neuronal circuits and their ability to self-organize, increasing attention has been directed toward synaptic plasticity. In particular spike-timing-dependent plasticity (STDP) creates specific demands for simulations of spiking neural networks. On the one hand a high temporal resolution is required to capture the millisecond timescale of typical STDP windows. On the other hand network simulations have to evolve over hours up to days, to capture the timescale of long-term plasticity. To do this efficiently, fast simulation speed is the crucial ingredient rather than large neuron numbers. Using different medium-sized network models consisting of several thousands of neurons and off-the-shelf hardware, we compare the simulation speed of the simulators: Brian, NEST and Neuron as well as our own simulator Auryn. Our results show that real-time simulations of different plastic network models are possible in parallel simulations in which numerical precision is not a primary concern. Even so, the speed-up margin of parallelism is limited and boosting simulation speeds beyond one tenth of real-time is difficult. By profiling simulation code we show that the run times of typical plastic network simulations encounter a hard boundary. This limit is partly due to latencies in the inter-process communications and thus cannot be overcome by increased parallelism. Overall, these results show that to study plasticity in medium-sized spiking neural networks, adequate simulation tools are readily available which run efficiently on small clusters. However, to run simulations substantially faster than real-time, special hardware is a prerequisite.

  3. rFerns: An Implementation of the Random Ferns Method for General-Purpose Machine Learning

    Directory of Open Access Journals (Sweden)

    Miron B. Kursa

    2014-11-01

    Full Text Available Random ferns is a very simple yet powerful classification method originally introduced for specific computer vision tasks. In this paper, I show that this algorithm may be considered as a constrained decision tree ensemble and use this interpretation to introduce a series of modifications which enable the use of random ferns in general machine learning problems. Moreover, I extend the method with an internal error approximation and an attribute importance measure based on corresponding features of the random forest algorithm. I also present the R package rFerns containing an efficient implementation of this modified version of random ferns.

  4. Working memory training mostly engages general-purpose large-scale networks for learning.

    Science.gov (United States)

    Salmi, Juha; Nyberg, Lars; Laine, Matti

    2018-03-21

    The present meta-analytic study examined brain activation changes following working memory (WM) training, a form of cognitive training that has attracted considerable interest. Comparisons with perceptual-motor (PM) learning revealed that WM training engages domain-general large-scale networks for learning encompassing the dorsal attention and salience networks, sensory areas, and striatum. Also the dynamics of the training-induced brain activation changes within these networks showed a high overlap between WM and PM training. The distinguishing feature for WM training was the consistent modulation of the dorso- and ventrolateral prefrontal cortex (DLPFC/VLPFC) activity. The strongest candidate for mediating transfer to similar untrained WM tasks was the frontostriatal system, showing higher striatal and VLPFC activations, and lower DLPFC activations after training. Modulation of transfer-related areas occurred mostly with longer training periods. Overall, our findings place WM training effects into a general perception-action cycle, where some modulations may depend on the specific cognitive demands of a training task. Copyright © 2018 Elsevier Ltd. All rights reserved.

  5. Provisional safety analyses for SGT stage 2 -- Models, codes and general modelling approach

    International Nuclear Information System (INIS)

    2014-12-01

    In the framework of the provisional safety analyses for Stage 2 of the Sectoral Plan for Deep Geological Repositories (SGT), deterministic modelling of radionuclide release from the barrier system along the groundwater pathway during the post-closure period of a deep geological repository is carried out. The calculated radionuclide release rates are interpreted as annual effective dose for an individual and assessed against the regulatory protection criterion 1 of 0.1 mSv per year. These steps are referred to as dose calculations. Furthermore, from the results of the dose calculations so-called characteristic dose intervals are determined, which provide input to the safety-related comparison of the geological siting regions in SGT Stage 2. Finally, the results of the dose calculations are also used to illustrate and to evaluate the post-closure performance of the barrier systems under consideration. The principal objective of this report is to describe comprehensively the technical aspects of the dose calculations. These aspects comprise: · the generic conceptual models of radionuclide release from the solid waste forms, of radionuclide transport through the system of engineered and geological barriers, of radionuclide transfer in the biosphere, as well as of the potential radiation exposure of the population, · the mathematical models for the explicitly considered release and transport processes, as well as for the radiation exposure pathways that are included, · the implementation of the mathematical models in numerical codes, including an overview of these codes and the most relevant verification steps, · the general modelling approach when using the codes, in particular the generic assumptions needed to model the near field and the geosphere, along with some numerical details, · a description of the work flow related to the execution of the calculations and of the software tools that are used to facilitate the modelling process, and · an overview of the

  6. Provisional safety analyses for SGT stage 2 -- Models, codes and general modelling approach

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    2014-12-15

    In the framework of the provisional safety analyses for Stage 2 of the Sectoral Plan for Deep Geological Repositories (SGT), deterministic modelling of radionuclide release from the barrier system along the groundwater pathway during the post-closure period of a deep geological repository is carried out. The calculated radionuclide release rates are interpreted as annual effective dose for an individual and assessed against the regulatory protection criterion 1 of 0.1 mSv per year. These steps are referred to as dose calculations. Furthermore, from the results of the dose calculations so-called characteristic dose intervals are determined, which provide input to the safety-related comparison of the geological siting regions in SGT Stage 2. Finally, the results of the dose calculations are also used to illustrate and to evaluate the post-closure performance of the barrier systems under consideration. The principal objective of this report is to describe comprehensively the technical aspects of the dose calculations. These aspects comprise: · the generic conceptual models of radionuclide release from the solid waste forms, of radionuclide transport through the system of engineered and geological barriers, of radionuclide transfer in the biosphere, as well as of the potential radiation exposure of the population, · the mathematical models for the explicitly considered release and transport processes, as well as for the radiation exposure pathways that are included, · the implementation of the mathematical models in numerical codes, including an overview of these codes and the most relevant verification steps, · the general modelling approach when using the codes, in particular the generic assumptions needed to model the near field and the geosphere, along with some numerical details, · a description of the work flow related to the execution of the calculations and of the software tools that are used to facilitate the modelling process, and · an overview of the

  7. Reassessment of Resuspension Factor Following Radionuclide Dispersal: Toward a General-purpose Rate Constant

    Energy Technology Data Exchange (ETDEWEB)

    Marshall, Shaun [Worcester Polytechnic Inst., Worcester, MA (United States). Dept. of Physics; Potter, Charles [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Medich, David [Worcester Polytechnic Inst., Worcester, MA (United States). Dept. of Physics

    2018-05-01

    A recent analysis of historical radionuclide resuspension datasets con rmed the general applicability of the Anspaugh and modified Anspaugh models of resuspension factors following both controlled and disastrous releases. The observations appear to increase in variance earlier in time, however all points were equally weighted in statistical fit calculations, inducing a positive skewing of resuspension coeffcients. Such data are extracted from the available deposition experiments spanning 2900 days. Measurements within a 3-day window are grouped into singular sample sets to construct standard deviations. A refitting is performed using a relative instrumental weighting of the observations. The resulting best-fit equations produces tamer exponentials which give decreased integrated resuspension factor values relative to those reported by Anspaugh. As expected, the fits attenuate greater error amongst the data at earlier time. The reevaluation provides a sharper contrast between the empirical models, and reafirms their deficiencies in the short-lived timeframe wherein the dynamics of particulate dispersion dominate the resuspension process.

  8. Reassessment of Resuspension Factor Following Radionuclide Dispersal: Toward a General-purpose Rate Constant.

    Science.gov (United States)

    Marshall, Shaun; Potter, Charles; Medich, David

    2018-05-01

    A recent analysis of historical radionuclide resuspension datasets confirmed the general applicability of the Anspaugh and modified Anspaugh models of resuspension factors following both controlled and disastrous releases. While observations appear to have larger variance earlier in time, previous studies equally weighted the data for statistical fit calculations; this could induce a positive skewing of resuspension coefficients in the early time-period. A refitting is performed using a relative instrumental weighting of the observations. Measurements within a 3-d window are grouped into singular sample sets to construct standard deviations. The resulting best-fit equations produce tamer exponentials, which give decreased integrated resuspension factor values relative to those reported by Anspaugh. As expected, the fits attenuate greater error among the data at earlier time. The reevaluation provides a sharper contrast between the empirical models and reaffirms their deficiencies in the short-lived timeframe wherein the dynamics of particulate dispersion dominate the resuspension process.

  9. The Spiral Discovery Network as an Automated General-Purpose Optimization Tool

    Directory of Open Access Journals (Sweden)

    Adam B. Csapo

    2018-01-01

    Full Text Available The Spiral Discovery Method (SDM was originally proposed as a cognitive artifact for dealing with black-box models that are dependent on multiple inputs with nonlinear and/or multiplicative interaction effects. Besides directly helping to identify functional patterns in such systems, SDM also simplifies their control through its characteristic spiral structure. In this paper, a neural network-based formulation of SDM is proposed together with a set of automatic update rules that makes it suitable for both semiautomated and automated forms of optimization. The behavior of the generalized SDM model, referred to as the Spiral Discovery Network (SDN, and its applicability to nondifferentiable nonconvex optimization problems are elucidated through simulation. Based on the simulation, the case is made that its applicability would be worth investigating in all areas where the default approach of gradient-based backpropagation is used today.

  10. Signal processing and general purpose data acquisition system for on-line tomographic measurements

    Science.gov (United States)

    Murari, A.; Martin, P.; Hemming, O.; Manduchi, G.; Marrelli, L.; Taliercio, C.; Hoffmann, A.

    1997-01-01

    New analog signal conditioning electronics and data acquisition systems have been developed for the soft x-ray and bolometric tomography diagnostic in the reverse field pinch experiment (RFX). For the soft x-ray detectors the analog signal processing includes a fully differential current to voltage conversion, with up to a 200 kHz bandwidth. For the bolometers, a 50 kHz carrier frequency amplifier allows a maximum bandwidth of 10 kHz. In both cases the analog signals are digitized with a 1 MHz sampling rate close to the diagnostic and are transmitted via a transparent asynchronous xmitter/receiver interface (TAXI) link to purpose built Versa Module Europa (VME) modules which perform data acquisition. A software library has been developed for data preprocessing and tomographic reconstruction. It has been written in C language and is self-contained, i.e., no additional mathematical library is required. The package is therefore platform-free: in particular it can perform online analysis in a real-time application, such as continuous display and feedback, and is portable for long duration fusion or other physical experiments. Due to the modular organization of the library, new preprocessing and analysis modules can be easily integrated in the environment. This software is implemented in RFX over three different platforms: open VMS, digital Unix, and VME 68040 CPU.

  11. A lossy graph model for delay reduction in generalized instantly decodable network coding

    KAUST Repository

    Douik, Ahmed S.

    2014-06-01

    The problem of minimizing the decoding delay in Generalized instantly decodable network coding (G-IDNC) for both perfect and lossy feedback scenarios is formulated as a maximum weight clique problem over the G-IDNC graph in. In this letter, we introduce a new lossy G-IDNC graph (LG-IDNC) model to further minimize the decoding delay in lossy feedback scenarios. Whereas the G-IDNC graph represents only doubtless combinable packets, the LG-IDNC graph represents also uncertain packet combinations, arising from lossy feedback events, when the expected decoding delay of XORing them among themselves or with other certain packets is lower than that expected when sending these packets separately. We compare the decoding delay performance of LG-IDNC and G-IDNC graphs through extensive simulations. Numerical results show that our new LG-IDNC graph formulation outperforms the G-IDNC graph formulation in all lossy feedback situations and achieves significant improvement in the decoding delay especially when the feedback erasure probability is higher than the packet erasure probability. © 2012 IEEE.

  12. Delay reduction in persistent erasure channels for generalized instantly decodable network coding

    KAUST Repository

    Sorour, Sameh

    2013-06-01

    In this paper, we consider the problem of minimizing the decoding delay of generalized instantly decodable network coding (G-IDNC) in persistent erasure channels (PECs). By persistent erasure channels, we mean erasure channels with memory, which are modeled as a Gilbert-Elliott two-state Markov model with good and bad channel states. In this scenario, the channel erasure dependence, represented by the transition probabilities of this channel model, is an important factor that could be exploited to reduce the decoding delay. We first formulate the G-IDNC minimum decoding delay problem in PECs as a maximum weight clique problem over the G-IDNC graph. Since finding the optimal solution of this formulation is NP-hard, we propose two heuristic algorithms to solve it and compare them using extensive simulations. Simulation results show that each of these heuristics outperforms the other in certain ranges of channel memory levels. They also show that the proposed heuristics significantly outperform both the optimal strict IDNC in the literature and the channel-unaware G-IDNC algorithms. © 2013 IEEE.

  13. Delay reduction in persistent erasure channels for generalized instantly decodable network coding

    KAUST Repository

    Sorour, Sameh; Aboutorab, Neda; Sadeghi, Parastoo; Karim, Mohammad Shahriar; Al-Naffouri, Tareq Y.; Alouini, Mohamed-Slim

    2013-01-01

    In this paper, we consider the problem of minimizing the decoding delay of generalized instantly decodable network coding (G-IDNC) in persistent erasure channels (PECs). By persistent erasure channels, we mean erasure channels with memory, which are modeled as a Gilbert-Elliott two-state Markov model with good and bad channel states. In this scenario, the channel erasure dependence, represented by the transition probabilities of this channel model, is an important factor that could be exploited to reduce the decoding delay. We first formulate the G-IDNC minimum decoding delay problem in PECs as a maximum weight clique problem over the G-IDNC graph. Since finding the optimal solution of this formulation is NP-hard, we propose two heuristic algorithms to solve it and compare them using extensive simulations. Simulation results show that each of these heuristics outperforms the other in certain ranges of channel memory levels. They also show that the proposed heuristics significantly outperform both the optimal strict IDNC in the literature and the channel-unaware G-IDNC algorithms. © 2013 IEEE.

  14. Generalized rate-code model for neuron ensembles with finite populations

    International Nuclear Information System (INIS)

    Hasegawa, Hideo

    2007-01-01

    We have proposed a generalized Langevin-type rate-code model subjected to multiplicative noise, in order to study stationary and dynamical properties of an ensemble containing a finite number N of neurons. Calculations using the Fokker-Planck equation have shown that, owing to the multiplicative noise, our rate model yields various kinds of stationary non-Gaussian distributions such as Γ, inverse-Gaussian-like, and log-normal-like distributions, which have been experimentally observed. The dynamical properties of the rate model have been studied with the use of the augmented moment method (AMM), which was previously proposed by the author from a macroscopic point of view for finite-unit stochastic systems. In the AMM, the original N-dimensional stochastic differential equations (DEs) are transformed into three-dimensional deterministic DEs for the means and fluctuations of local and global variables. The dynamical responses of the neuron ensemble to pulse and sinusoidal inputs calculated by the AMM are in good agreement with those obtained by direct simulation. The synchronization in the neuronal ensemble is discussed. The variabilities of the firing rate and of the interspike interval are shown to increase with increasing magnitude of multiplicative noise, which may be a conceivable origin of the observed large variability in cortical neurons

  15. SUMMARY OF GENERAL WORKING GROUP A+B+D: CODES BENCHMARKING.

    Energy Technology Data Exchange (ETDEWEB)

    WEI, J.; SHAPOSHNIKOVA, E.; ZIMMERMANN, F.; HOFMANN, I.

    2006-05-29

    Computer simulation is an indispensable tool in assisting the design, construction, and operation of accelerators. In particular, computer simulation complements analytical theories and experimental observations in understanding beam dynamics in accelerators. The ultimate function of computer simulation is to study mechanisms that limit the performance of frontier accelerators. There are four goals for the benchmarking of computer simulation codes, namely debugging, validation, comparison and verification: (1) Debugging--codes should calculate what they are supposed to calculate; (2) Validation--results generated by the codes should agree with established analytical results for specific cases; (3) Comparison--results from two sets of codes should agree with each other if the models used are the same; and (4) Verification--results from the codes should agree with experimental measurements. This is the summary of the joint session among working groups A, B, and D of the HI32006 Workshop on computer codes benchmarking.

  16. [Coding in general practice-Will the ICD-11 be a step forward?

    Science.gov (United States)

    Kühlein, Thomas; Virtanen, Martti; Claus, Christoph; Popert, Uwe; van Boven, Kees

    2018-07-01

    Primary care physicians in Germany don't benefit from coding diagnoses-they are coding for the needs of others. For coding, they mostly are using either the thesaurus of the German Institute of Medical Documentation and Information (DIMDI) or self-made cheat-sheets. Coding quality is low but seems to be sufficient for the main use case of the resulting data, which is the morbidity adjusted risk compensation scheme that distributes financial resources between the many German health insurance companies.Neither the International Classification of Diseases and Health Related Problems (ICD-10) nor the German thesaurus as an interface terminology are adequate for coding in primary care. The ICD-11 itself will not recognizably be a step forward from the perspective of primary care. At least the browser database format will be advantageous. An implementation into the 182 different electronic health records (EHR) on the German market would probably standardize the coding process and make code finding easier. This method of coding would still be more cumbersome than the current coding with self-made cheat-sheets.The first steps towards a useful official cheat-sheet for primary care have been taken, awaiting implementation and evaluation. The International Classification of Primary Care (ICPC-2) already provides an adequate classification standard for primary care that can also be used in combination with ICD-10. A new version of ICPC (ICPC-3) is under development. As the ICPC-2 has already been integrated into the foundation layer of ICD-11 it might easily become the future standard for coding in primary care. Improving communication between the different EHR would make taking over codes from other healthcare providers possible. Another opportunity to improve the coding quality might be creating use cases for the resulting data for the primary care physicians themselves.

  17. Interactive general-purpose function minimization for the analysis of neutron scattering data

    International Nuclear Information System (INIS)

    Abel, W.

    1981-12-01

    An on-line graphic display facility has been employed mainly for the peak analysis of time-of-flight spectra measured by inelastic scattering of thermal neutrons. But it is useful also for the analysis of spectra measured with triple axis spectrometers and of diffraction patterns. The spectral lines may be fitted by the following analytical shape functions: (i) a Gaussian, (ii) a Lorentzian, or (iii) a convolution of a Lorentzian with a Gaussian, plus a background continuum. Data reduction or correction may be invoked optionally. For more general applications in analysing of numerical data there is also the possibility to define the analytical shape functions by the user. Three different minimization methods are available which may be used alone or in combination. The parameters of the shape functions may be kept fixed or variable during the minimization steps. The width of variation may be restricted. Global correlation coefficients, parameter errors and the chi 2 are displayed to inform the user about the quality of the fit. A detailed description of the program operations is given. The programs are written in FORTRAN IV and use an IBM/2250-1 graphic display unit. (orig.) [de

  18. Design and validation of a general purpose robotic testing system for musculoskeletal applications.

    Science.gov (United States)

    Noble, Lawrence D; Colbrunn, Robb W; Lee, Dong-Gil; van den Bogert, Antonie J; Davis, Brian L

    2010-02-01

    Orthopaedic research on in vitro forces applied to bones, tendons, and ligaments during joint loading has been difficult to perform because of limitations with existing robotic simulators in applying full-physiological loading to the joint under investigation in real time. The objectives of the current work are as follows: (1) describe the design of a musculoskeletal simulator developed to support in vitro testing of cadaveric joint systems, (2) provide component and system-level validation results, and (3) demonstrate the simulator's usefulness for specific applications of the foot-ankle complex and knee. The musculoskeletal simulator allows researchers to simulate a variety of loading conditions on cadaver joints via motorized actuators that simulate muscle forces while simultaneously contacting the joint with an external load applied by a specialized robot. Multiple foot and knee studies have been completed at the Cleveland Clinic to demonstrate the simulator's capabilities. Using a variety of general-use components, experiments can be designed to test other musculoskeletal joints as well (e.g., hip, shoulder, facet joints of the spine). The accuracy of the tendon actuators to generate a target force profile during simulated walking was found to be highly variable and dependent on stance position. Repeatability (the ability of the system to generate the same tendon forces when the same experimental conditions are repeated) results showed that repeat forces were within the measurement accuracy of the system. It was determined that synchronization system accuracy was 6.7+/-2.0 ms and was based on timing measurements from the robot and tendon actuators. The positioning error of the robot ranged from 10 microm to 359 microm, depending on measurement condition (e.g., loaded or unloaded, quasistatic or dynamic motion, centralized movements or extremes of travel, maximum value, or root-mean-square, and x-, y- or z-axis motion). Algorithms and methods for controlling

  19. Operation of a general purpose stepping motor-encoder positioning subsystem at the National Synchrotron Light Source

    International Nuclear Information System (INIS)

    Stubblefield, F.W.

    1985-11-01

    Four copies of a general purpose subsystem for mechanical positioning of detectors, samples, and beam line optical elements which constitute experiments at the National Synchrotron Light Source facility of Brookhaven National Laboratory have been constructed and placed into operation. Construction of a fifth subsystem unit is nearing completion. The subsystems affect mechanical positioning by controlling a set of stepping motor-encoder pairs. The units are general purpose in the sense that they receive commands over a 9600 baud asynchronous serial line compatible with the RS-232-C electrical signal standard, generate TTL-compatible streams of stepping pulses which can be used with a wide variety of stepping motors, and read back position values from a number of different types and models of position encoder. The basic structure of the motor controller subsystem is briefly reviewed. Additions to the subsystem made in response to problems indicated by actual operation of the four installed units are described in more detail

  20. Recent advances toward a general purpose linear-scaling quantum force field.

    Science.gov (United States)

    Giese, Timothy J; Huang, Ming; Chen, Haoyuan; York, Darrin M

    2014-09-16

    Conspectus There is need in the molecular simulation community to develop new quantum mechanical (QM) methods that can be routinely applied to the simulation of large molecular systems in complex, heterogeneous condensed phase environments. Although conventional methods, such as the hybrid quantum mechanical/molecular mechanical (QM/MM) method, are adequate for many problems, there remain other applications that demand a fully quantum mechanical approach. QM methods are generally required in applications that involve changes in electronic structure, such as when chemical bond formation or cleavage occurs, when molecules respond to one another through polarization or charge transfer, or when matter interacts with electromagnetic fields. A full QM treatment, rather than QM/MM, is necessary when these features present themselves over a wide spatial range that, in some cases, may span the entire system. Specific examples include the study of catalytic events that involve delocalized changes in chemical bonds, charge transfer, or extensive polarization of the macromolecular environment; drug discovery applications, where the wide range of nonstandard residues and protonation states are challenging to model with purely empirical MM force fields; and the interpretation of spectroscopic observables. Unfortunately, the enormous computational cost of conventional QM methods limit their practical application to small systems. Linear-scaling electronic structure methods (LSQMs) make possible the calculation of large systems but are still too computationally intensive to be applied with the degree of configurational sampling often required to make meaningful comparison with experiment. In this work, we present advances in the development of a quantum mechanical force field (QMFF) suitable for application to biological macromolecules and condensed phase simulations. QMFFs leverage the benefits provided by the LSQM and QM/MM approaches to produce a fully QM method that is able to

  1. A new code for automatic detection and analysis of the lineament patterns for geophysical and geological purposes (ADALGEO)

    Science.gov (United States)

    Soto-Pinto, C.; Arellano-Baeza, A.; Sánchez, G.

    2013-08-01

    We present a new numerical method for automatic detection and analysis of changes in lineament patterns caused by seismic and volcanic activities. The method is implemented as a series of modules: (i) normalization of the image contrast, (ii) extraction of small linear features (stripes) through convolution of the part of the image in the vicinity of each pixel with a circular mask or through Canny algorithm, and (iii) posterior detection of main lineaments using the Hough transform. We demonstrate that our code reliably detects changes in the lineament patterns related to the stress evolution in the Earth's crust: specifically, a significant number of new lineaments appear approximately one month before an earthquake, while one month after the earthquake the lineament configuration returns to its initial state. Application of our software to the deformations caused by volcanic activity yields the opposite results: the number of lineaments decreases with the onset of microseismicity. This discrepancy can be explained assuming that the plate tectonic earthquakes are caused by the compression and accumulation of stress in the Earth's crust due to subduction of tectonic plates, whereas in the case of volcanic activity we deal with the inflation of a volcano edifice due to elevation of pressure and magma intrusion and the resulting stretching of the surface.

  2. General Labour Code (Act No. 2 of 1986), 5 April 1986. [Selected provisions].

    Science.gov (United States)

    1989-01-01

    Guinea-Bissau's General Labor Code, Act No. 2 of 1986, requires that employers adopt non-discriminatory procedures in dealing with their employees, disallowing discrimination based on gender, race, national origin, religion, political ideology or inclination, or on whether or not an employee is affiliated to a worker's union. Women are guaranteed equality with men in work opportunity and employment treatment; they are assured access to any type of job, profession, or position which does not pose any potential or actual threat to their genetic functioning. Additional legislation shall establish the conditions and prohibitions for contracting woman for heavy labor, for jobs performed under insalubrious conditions, for underground work, and for other types of jobs which might endanger a woman's genetic functioning. Job offers must not discriminate, unless the restrictions or specifications are essential to the nature of the job, rendering the job qualitatively differentiable when performed by a man or a woman. The employer must ensure that female workers have the same job opportunity and treatment for professional training and career development. Different specific job categories may not be created for men and women if a woman's salary is inferior to that of a man performing equivalent work in a corresponding professional category. Protection during pregnancy and delivery guarantees a woman the right to decline medically inadvisable tasks without loss of wages; to decline overtime work; to be absent from work when necessary for maternal medical care without loss of wages; to interrupt daily work to nurse her children for 1 hour or 2 for half-hour periods for up to 1 year without the loss of wages. Every female employee has the right to 60 days pregnancy and maternity leave, without loss of wages, for whatever clinical type of delivery, including for still births and for births where the infant died shortly thereafter.

  3. Accuracy of clinical coding from 1210 appendicectomies in a British district general hospital.

    Science.gov (United States)

    Bhangu, Aneel; Nepogodiev, Dmitri; Taylor, Caroline; Durkin, Natalie; Patel, Rajan

    2012-01-01

    The primary aim of this study was to assess the accuracy of clinical coding in identifying negative appendicectomies. The secondary aim was to analyse trends over time in rates of simple, complex (gangrenous or perforated) and negative appendicectomies. Retrospective review of 1210 patients undergoing emergency appendicectomy during a five year period (2006-2010). Histopathology reports were taken as gold standard for diagnosis and compared to clinical coding lists. Clinical coding is the process by which non-medical administrators apply standardised diagnostic codes to patients, based upon clinical notes at discharge. These codes then contribute to national databases. Statistical analysis included correlation studies and regression analyses. Clinical coding had only moderate correlation with histopathology, with an overall kappa of 0.421. Annual kappa values varied between 0.378 and 0.500. Overall 14% of patients were incorrectly coded as having had appendicitis when in fact they had a histopathologically normal appendix (153/1107), whereas 4% were falsely coded as having received a negative appendicectomy when they had appendicitis (48/1107). There was an overall significant fall and then rise in the rate of simple appendicitis (B coefficient -0.239 (95% confidence interval -0.426, -0.051), p = 0.014) but no change in the rate of complex appendicitis (B coefficient 0.008 (-0.015, 0.031), p = 0.476). Clinical coding for negative appendicectomy was unreliable. Negative rates may be higher than suspected. This has implications for the validity of national database analyses. Using this form of data as a quality indictor for appendicitis should be reconsidered until its quality is improved. Copyright © 2012 Surgical Associates Ltd. Published by Elsevier Ltd. All rights reserved.

  4. TRIO a general computer code for reactor 3-D flows analysis. Application to a LMFBR hot plenum

    International Nuclear Information System (INIS)

    Magnaud, J.P.; Rouzaud, P.

    1985-09-01

    TRIO is a code developed at CEA to investigate general incompressible 2D and 3D viscous flows. Two calculations are presented: the lid driven cubic cavity at Re=400; steady state (velocity and temperature field) of a LMFBR hot plenum, carried out in order to prepare the calculation of a cold shock consecutive to a reactor scram. 8 refs., 26 figs.

  5. Transoptr-a second order beam transport design code with automatic internal optimization and general constraints

    International Nuclear Information System (INIS)

    Heighway, E.A.

    1980-07-01

    A second order beam transport design code with parametric optimization is described. The code analyzes the transport of charged particle beams through a user defined magnet system. The magnet system parameters are varied (within user defined limits) until the properties of the transported beam and/or the system transport matrix match those properties requested by the user. The code uses matrix formalism to represent the transport elements and optimization is achieved using the variable metric method. Any constraints that can be expressed algebraically may be included by the user as part of his design. Instruction in the use of the program is given. (auth)

  6. The theta/gamma discrete phase code occuring during the hippocampal phase precession may be a more general brain coding scheme.

    Science.gov (United States)

    Lisman, John

    2005-01-01

    In the hippocampus, oscillations in the theta and gamma frequency range occur together and interact in several ways, indicating that they are part of a common functional system. It is argued that these oscillations form a coding scheme that is used in the hippocampus to organize the readout from long-term memory of the discrete sequence of upcoming places, as cued by current position. This readout of place cells has been analyzed in several ways. First, plots of the theta phase of spikes vs. position on a track show a systematic progression of phase as rats run through a place field. This is termed the phase precession. Second, two cells with nearby place fields have a systematic difference in phase, as indicated by a cross-correlation having a peak with a temporal offset that is a significant fraction of a theta cycle. Third, several different decoding algorithms demonstrate the information content of theta phase in predicting the animal's position. It appears that small phase differences corresponding to jitter within a gamma cycle do not carry information. This evidence, together with the finding that principle cells fire preferentially at a given gamma phase, supports the concept of theta/gamma coding: a given place is encoded by the spatial pattern of neurons that fire in a given gamma cycle (the exact timing within a gamma cycle being unimportant); sequential places are encoded in sequential gamma subcycles of the theta cycle (i.e., with different discrete theta phase). It appears that this general form of coding is not restricted to readout of information from long-term memory in the hippocampus because similar patterns of theta/gamma oscillations have been observed in multiple brain regions, including regions involved in working memory and sensory integration. It is suggested that dual oscillations serve a general function: the encoding of multiple units of information (items) in a way that preserves their serial order. The relationship of such coding to

  7. A general-purpose framework to simulate musculoskeletal system of human body: using a motion tracking approach.

    Science.gov (United States)

    Ehsani, Hossein; Rostami, Mostafa; Gudarzi, Mohammad

    2016-02-01

    Computation of muscle force patterns that produce specified movements of muscle-actuated dynamic models is an important and challenging problem. This problem is an undetermined one, and then a proper optimization is required to calculate muscle forces. The purpose of this paper is to develop a general model for calculating all muscle activation and force patterns in an arbitrary human body movement. For this aim, the equations of a multibody system forward dynamics, which is considered for skeletal system of the human body model, is derived using Lagrange-Euler formulation. Next, muscle contraction dynamics is added to this model and forward dynamics of an arbitrary musculoskeletal system is obtained. For optimization purpose, the obtained model is used in computed muscle control algorithm, and a closed-loop system for tracking desired motions is derived. Finally, a popular sport exercise, biceps curl, is simulated by using this algorithm and the validity of the obtained results is evaluated via EMG signals.

  8. Assessment of general public perceptions toward traditional medicines used for aphrodisiac purpose in state of Penang, Malaysia.

    Science.gov (United States)

    Hassali, Mohamed Azmi; Saleem, Fahad; Shafie, Asrul Akmal; Al-Qazaz, Harith Khalid; Farooqui, Maryam; Aljadhey, Hisham; Atif, Muhammad; Masood, Imran

    2012-11-01

    The study aims to evaluate general public perceptions regarding the use of Traditional and Complementary Medicines (TCM) for aphrodisiac purposes. A questionnaire based, cross-sectional study was undertaken. Respondents were selected in the state of Penang, Malaysia. A total of 392 respondents were included in the study. Descriptive statistics were used for data analysis. Chi Square/Fischer Exact tests were used where appropriate. Out of 392 respondents, 150 (38.26%) reported using specific Traditional medicines for aphrodisiac purposes. Most respondents (46.94%) agreed that aphrodisiac medicines were easily available t. Moreover, 40.31% of the respondents reported that traditional aphrodisiac medicines were cheaper than modern (prescription) medicines. This study highlights limited public knowledge regarding the use of traditional aphrodisiac medicine. Healthcare professionals should be aware of informal TCM usage when prescribing allopathic medicines. Copyright © 2012 Elsevier Ltd. All rights reserved.

  9. Analysis of Coded FHSS Systems with Multiple Access Interference over Generalized Fading Channels

    Directory of Open Access Journals (Sweden)

    Salam A. Zummo

    2009-02-01

    Full Text Available We study the effect of interference on the performance of coded FHSS systems. This is achieved by modeling the physical channel in these systems as a block fading channel. In the derivation of the bit error probability over Nakagami fading channels, we use the exact statistics of the multiple access interference (MAI in FHSS systems. Due to the mathematically intractable expression of the Rician distribution, we use the Gaussian approximation to derive the error probability of coded FHSS over Rician fading channel. The effect of pilot-aided channel estimation is studied for Rician fading channels using the Gaussian approximation. From this, the optimal hopping rate in coded FHSS is approximated. Results show that the performance loss due to interference increases as the hopping rate decreases.

  10. Effects of detector-source distance and detector bias voltage variations on time resolution of general purpose plastic scintillation detectors.

    Science.gov (United States)

    Ermis, E E; Celiktas, C

    2012-12-01

    Effects of source-detector distance and the detector bias voltage variations on time resolution of a general purpose plastic scintillation detector such as BC400 were investigated. (133)Ba and (207)Bi calibration sources with and without collimator were used in the present work. Optimum source-detector distance and bias voltage values were determined for the best time resolution by using leading edge timing method. Effect of the collimator usage on time resolution was also investigated. Copyright © 2012 Elsevier Ltd. All rights reserved.

  11. Application of a general-purpose scintigraphic scanner to transverse-section (tomographic) gamma-ray imaging

    International Nuclear Information System (INIS)

    Bradstock, P.A.; Milward, R.C.

    1976-01-01

    The paper describes the recent application of a general-purpose commercial scintigraphic scanner to transverse-section radioisotope tomography. The principle of the method is to obtain the distribution of radioactive material in a thin transverse slice of the body or brain, from a mathematical reconstruction using the measured transverse projections of the activity within that slice. The usefulness of the radioisotope section-scanning technique for clinical diagnosis, as evidenced from one year's use of the machine at the Midland Centre for Neurology and Neurosurgery, Birmingham, U.K., is briefly discussed. (orig.) [de

  12. General-purpose heat source project and space nuclear safety and fuels program. Progress reportt, January 1980

    International Nuclear Information System (INIS)

    Maraman, W.J.

    1980-04-01

    This formal monthly report covers the studies related to the use of 238 PuO 2 in radioisotopic power systems carried out for the Advanced Nuclear Systems and Projects Division of the Los Alamos Scientific Laboratory. The two programs involved are the general-purpose heat source development and space nuclear safety and fuels. Most of the studies discussed here are of a continuing nature. Results and conclusions described may change as the work continues. Published reference to the results cited in this report should not be made without the explicit permission of the person in charge of the work

  13. Prometheus: the implementation of clinical coding schemes in French routine general practice

    Directory of Open Access Journals (Sweden)

    Laurent Letrilliart

    2006-09-01

    Conclusions Coding health problems on a routine basis proved to be feasible. However, this process can be used on a more widespread basis and linked to other management data only if physicians are specially trained and rewarded, and the software incorporates large terminologies mapped with classifications.

  14. Using supervised machine learning to code policy issues: Can classifiers generalize across contexts?

    NARCIS (Netherlands)

    Burscher, B.; Vliegenthart, R.; de Vreese, C.H.

    2015-01-01

    Content analysis of political communication usually covers large amounts of material and makes the study of dynamics in issue salience a costly enterprise. In this article, we present a supervised machine learning approach for the automatic coding of policy issues, which we apply to news articles

  15. Multi­-Threaded Algorithms for General purpose Graphics Processor Units in the ATLAS High Level Trigger

    CERN Document Server

    Conde Mui\\~no, Patricia; The ATLAS collaboration

    2016-01-01

    General purpose Graphics Processor Units (GPGPU) are being evaluated for possible future inclusion in an upgraded ATLAS High Level Trigger farm. We have developed a demonstrator including GPGPU implementations of Inner Detector and Muon tracking and Calorimeter clustering within the ATLAS software framework. ATLAS is a general purpose particle physics experiment located on the LHC collider at CERN. The ATLAS Trigger system consists of two levels, with level 1 implemented in hardware and the High Level Trigger implemented in software running on a farm of commodity CPU. The High Level Trigger reduces the trigger rate from the 100 kHz level 1 acceptance rate to 1 kHz for recording, requiring an average per­-event processing time of ~250 ms for this task. The selection in the high level trigger is based on reconstructing tracks in the Inner Detector and Muon Spectrometer and clusters of energy deposited in the Calorimeter. Performing this reconstruction within the available farm resources presents a significant ...

  16. Generalized Alamouti Codes for Trading Quality of Service against Data Rate in MIMO UMTS

    Directory of Open Access Journals (Sweden)

    Rupp Markus

    2004-01-01

    Full Text Available New space-time block coding schemes for multiple transmit and receive antennas are proposed. First, the well-known Alamouti scheme is extended to transmit antennas achieving high transmit diversity. Many receiver details are worked out for four and eight transmit antennas. Further, solutions for arbitrary, even numbers of transmit antennas are presented achieving decoding advantages due to orthogonalization properties while preserving high diversity. In a final step, such extended Alamouti and BLAST schemes are combined, offering a continuous trade-off between quality of service (QoS and data rate. Due to the simplicity of the coding schemes, they are very well suited to operate under UMTS with only very moderate modifications in the existing standard. The number of supported antennas at transmitter alone is a sufficient knowledge to select the most appropriate scheme. While the proposed schemes are motivated by utilization in UMTS, they are not restricted to this standard.

  17. Languages for Specific Academic Purposes or Languages for General Academic Purposes? A Critical Reappraisal of a Key Issue for Language Provision in Higher Education

    Science.gov (United States)

    Krekeler, Christian

    2013-01-01

    The debate about the subject specificity of university language tuition has been going on for decades; it has mostly been discussed in the context of English for Academic Purposes. This paper considers the case for disciplinary specificity with regard to languages other than English. Few, if any, developed curricula, syllabuses, suitable textbooks…

  18. Screening tests in toxicity or drug effect studies with use of centrifichem general-purpose spectrophotometeric analyzer

    International Nuclear Information System (INIS)

    Nagy, B.; Bercz, J.P.

    1986-01-01

    CentrifiChem System 400 general-purpose spectrophotometric analyzer which can process simultaneously 30 samples and reads the reactions within milliseconds was used for toxicity studies. Organic and inorganic chemicals were screened for inhibitory action of the hydrolytic activity of sarcoplasmic reticulum (SR) Ca,Mg-ATPase and that of the sacrolemmal (SL) Na,K-ATPase, or mitochondrial ATPase (M). SR and SL were prepared from rabbit muscles, Na,K-ATPase from pig kidneys, M from pig hearts. Pseudosubstrates of paranitrophenyl phosphate and 2,4-dinitrophenyl phosphate, both proven high energy phosphate substitutes for ATPase coupled ion transfer were used. The reaction rates were followed spectrophotometrically at 405 nm measuring the accumulation of yellow nitrophenolate ions. The reported calcium transfer coupling ratio to hydrolysis of 2:1 was ascertained with use of 45 Ca in case of SR. Inhibition constants (pI) on SR, SL, and M for the pseudosubstrate hydrolysis will be given for over 20 chemicals tested. The applicability of the system to general toxicity testing and to general cardio-effective drug screening will be presented

  19. General-purpose heat source: Research and development program, radioisotope thermoelectric generator/thin fragment impact test

    International Nuclear Information System (INIS)

    Reimus, M.A.H.; Hinckley, J.E.

    1996-11-01

    The general-purpose heat source provides power for space missions by transmitting the heat of 238 Pu decay to an array of thermoelectric elements in a radioisotope thermoelectric generator (RTG). Because the potential for a launch abort or return from orbit exists for any space mission, the heat source response to credible accident scenarios is being evaluated. This test was designed to provide information on the response of a loaded RTG to impact by a fragment similar to the type of fragment produced by breakup of the spacecraft propulsion module system. The results of this test indicated that impact by a thin aluminum fragment traveling at 306 m/s may result in significant damage to the converter housing, failure of one fueled clad, and release of a small quantity of fuel

  20. Effects of detector–source distance and detector bias voltage variations on time resolution of general purpose plastic scintillation detectors

    International Nuclear Information System (INIS)

    Ermis, E.E.; Celiktas, C.

    2012-01-01

    Effects of source-detector distance and the detector bias voltage variations on time resolution of a general purpose plastic scintillation detector such as BC400 were investigated. 133 Ba and 207 Bi calibration sources with and without collimator were used in the present work. Optimum source-detector distance and bias voltage values were determined for the best time resolution by using leading edge timing method. Effect of the collimator usage on time resolution was also investigated. - Highlights: ► Effect of the source-detector distance on time spectra was investigated. ► Effect of the detector bias voltage variations on time spectra was examined. ► Optimum detector–source distance was determined for the best time resolution. ► Optimum detector bias voltage was determined for the best time resolution. ► 133 Ba and 207 Bi radioisotopes were used.

  1. Evaluation of Aqueous and Powder Processing Techniques for Production of Pu-238-Fueled General Purpose Heat Sources

    Energy Technology Data Exchange (ETDEWEB)

    2008-06-01

    This report evaluates alternative processes that could be used to produce Pu-238 fueled General Purpose Heat Sources (GPHS) for radioisotope thermoelectric generators (RTG). Fabricating GPHSs with the current process has remained essentially unchanged since its development in the 1970s. Meanwhile, 30 years of technological advancements have been made in the fields of chemistry, manufacturing, ceramics, and control systems. At the Department of Energy’s request, alternate manufacturing methods were compared to current methods to determine if alternative fabrication processes could reduce the hazards, especially the production of respirable fines, while producing an equivalent GPHS product. An expert committee performed the evaluation with input from four national laboratories experienced in Pu-238 handling.

  2. Duplication of complete dentures using general-purpose handheld optical scanner and 3-dimensional printer: Introduction and clinical considerations.

    Science.gov (United States)

    Kurahashi, Kosuke; Matsuda, Takashi; Goto, Takaharu; Ishida, Yuichi; Ito, Teruaki; Ichikawa, Tetsuo

    2017-01-01

    To introduce a new clinical procedure for fabricating duplicates of complete dentures by bite pressure impression using digital technology, and to discuss its clinical significance. The denture is placed on a rotary table and the 3-dimensional form of the denture is digitized using a general-purpose handheld optical scanner. The duplicate denture is made of polylactic acid by a 3-dimensional printer using the 3-dimensional data. This procedure has the advantages of wasting less material, employing less human power, decreasing treatment time at the chair side, lowering the rates of contamination, and being readily fabricated at the time of the treatment visit. Copyright © 2016 Japan Prosthodontic Society. Published by Elsevier Ltd. All rights reserved.

  3. Huemul: a two dimensional multigroup collision probability code for general geometries

    International Nuclear Information System (INIS)

    Calabrese, C.R.; Grant, C.R.

    1990-01-01

    The control rod calculation and the necessity of having a 2-D transport code able to calculate geometries as different as pool reactor or power reactor control rods resulted in the development of a new tool according to these requirements. This new tool permits a 2-D spatial representation, and the calculation mesh is formed by circumferential arcs segments not necessarily parallel to the coordinate axis. It includes the possibility of considering boundary conditions in the form of an albedo matrix (J + /J - ) as well as different external currents for each face of the model. It also allows an arbitrary number of energy groups compatible with computer limitations. These possibilities make HUEMUL a useful tool for a great variety of control rod or supercell type calculations. HUEMUL has been tested with copper activity measurements performed in the Canadian D2O facility ZED-2 with stainless-steel adjuster rods obtaining a very good agreement (better than 2%). Also manganese activity measurements in RA-2 pool reactor were used to compare calculated and measured values inside a MTR fuel element calculations, (better than 1%). Comparisons with results from the WIMS code for a light water cell with 3% enriched UO 2 has also shown a very good agreement in fluxes and multiplication constant (better than 1.5% in fluxes and 50 pcm in k-infinity). (Author) [es

  4. ACUTRI: a computer code for assessing doses to the general public due to acute tritium releases

    Energy Technology Data Exchange (ETDEWEB)

    Yokoyama, Sumi; Noguchi, Hiroshi [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment; Ryufuku, Susumu; Sasaki, Toshihisa; Kurosawa, Naohiro [Visible Information Center, Inc., Tokai, Ibaraki (Japan)

    2002-11-01

    Tritium, which is used as a fuel of a D-T burning fusion reactor, is the most important radionuclide for the safety assessment of a nuclear fusion experimental reactor such as ITER. Thus, a computer code, ACUTRI, which calculates the radiological impact of tritium released accidentally to the atmosphere, has been developed, aiming to be of use in a discussion of licensing of a fusion experimental reactor and an environmental safety evaluation method in Japan. ACUTRI calculates an individual tritium dose based on transfer models specific to tritium in the environment and ICRP dose models. In this calculation it is also possible to analyze statistically on meteorology in the same way as a conventional dose assessment method according to the meteorological guide of the Nuclear Safety Commission of Japan. A Gaussian plume model is used for calculating the atmospheric dispersion of tritium gas (HT) and/or tritiated water (HTO). The environmental pathway model in ACUTRI considers the following internal exposures: inhalation from a primary plume (HT and/or HTO) released from the facilities and inhalation from a secondary plume (HTO) reemitted from the ground following deposition of HT and HTO. This report describes an outline of the ACUTRI code, a user guide and the results of test calculation. (author)

  5. Heuristic simulation of nuclear systems on a supercomputer using the HAL-1987 general-purpose production-rule analysis system

    International Nuclear Information System (INIS)

    Ragheb, M.; Gvillo, D.; Makowitz, H.

    1987-01-01

    HAL-1987 is a general-purpose tool for the construction of production-rule analysis systems. It uses the rule-based paradigm from the part of artificial intelligence concerned with knowledge engineering. It uses backward-chaining and forward-chaining in an antecedent-consequent logic, and is programmed in Portable Standard Lisp (PSL). The inference engine is flexible and accommodates general additions and modifications to the knowledge base. The system is used in coupled symbolic-procedural programming adaptive methodologies for stochastic simulations. In Monte Carlo simulations of particle transport, the system considers the pre-processing of the input data to the simulation and adaptively controls the variance reduction process as the simulation progresses. This is accomplished through the use of a knowledge base of rules which encompass the user's expertise in the variance reduction process. It is also applied to the construction of model-based systems for monitoring, fault-diagnosis and crisis-alert in engineering devices, particularly in the field of nuclear reactor safety analysis

  6. MOCUM: A two-dimensional method of characteristics code based on constructive solid geometry and unstructured meshing for general geometries

    International Nuclear Information System (INIS)

    Yang Xue; Satvat, Nader

    2012-01-01

    Highlight: ► A two-dimensional numerical code based on the method of characteristics is developed. ► The complex arbitrary geometries are represented by constructive solid geometry and decomposed by unstructured meshing. ► Excellent agreement between Monte Carlo and the developed code is observed. ► High efficiency is achieved by parallel computing. - Abstract: A transport theory code MOCUM based on the method of characteristics as the flux solver with an advanced general geometry processor has been developed for two-dimensional rectangular and hexagonal lattice and full core neutronics modeling. In the code, the core structure is represented by the constructive solid geometry that uses regularized Boolean operations to build complex geometries from simple polygons. Arbitrary-precision arithmetic is also used in the process of building geometry objects to eliminate the round-off error from the commonly used double precision numbers. Then, the constructed core frame will be decomposed and refined into a Conforming Delaunay Triangulation to ensure the quality of the meshes. The code is fully parallelized using OpenMP and is verified and validated by various benchmarks representing rectangular, hexagonal, plate type and CANDU reactor geometries. Compared with Monte Carlo and deterministic reference solution, MOCUM results are highly accurate. The mentioned characteristics of the MOCUM make it a perfect tool for high fidelity full core calculation for current and GenIV reactor core designs. The detailed representation of reactor physics parameters can enhance the safety margins with acceptable confidence levels, which lead to more economically optimized designs.

  7. Majorana fermion codes

    International Nuclear Information System (INIS)

    Bravyi, Sergey; Terhal, Barbara M; Leemhuis, Bernhard

    2010-01-01

    We initiate the study of Majorana fermion codes (MFCs). These codes can be viewed as extensions of Kitaev's one-dimensional (1D) model of unpaired Majorana fermions in quantum wires to higher spatial dimensions and interacting fermions. The purpose of MFCs is to protect quantum information against low-weight fermionic errors, that is, operators acting on sufficiently small subsets of fermionic modes. We examine to what extent MFCs can surpass qubit stabilizer codes in terms of their stability properties. A general construction of 2D MFCs is proposed that combines topological protection based on a macroscopic code distance with protection based on fermionic parity conservation. Finally, we use MFCs to show how to transform any qubit stabilizer code to a weakly self-dual CSS code.

  8. ACUTRI a computer code for assessing doses to the general public due to acute tritium releases

    CERN Document Server

    Yokoyama, S; Noguchi, H; Ryufuku, S; Sasaki, T

    2002-01-01

    Tritium, which is used as a fuel of a D-T burning fusion reactor, is the most important radionuclide for the safety assessment of a nuclear fusion experimental reactor such as ITER. Thus, a computer code, ACUTRI, which calculates the radiological impact of tritium released accidentally to the atmosphere, has been developed, aiming to be of use in a discussion of licensing of a fusion experimental reactor and an environmental safety evaluation method in Japan. ACUTRI calculates an individual tritium dose based on transfer models specific to tritium in the environment and ICRP dose models. In this calculation it is also possible to analyze statistically on meteorology in the same way as a conventional dose assessment method according to the meteorological guide of the Nuclear Safety Commission of Japan. A Gaussian plume model is used for calculating the atmospheric dispersion of tritium gas (HT) and/or tritiated water (HTO). The environmental pathway model in ACUTRI considers the following internal exposures: i...

  9. In vivo dosimetry in intraoperative electron radiotherapy: microMOSFETs, radiochromic films and a general-purpose linac.

    Science.gov (United States)

    López-Tarjuelo, Juan; Bouché-Babiloni, Ana; Morillo-Macías, Virginia; de Marco-Blancas, Noelia; Santos-Serra, Agustín; Quirós-Higueras, Juan David; Ferrer-Albiach, Carlos

    2014-10-01

    In vivo dosimetry is desirable for the verification, recording, and eventual correction of treatment in intraoperative electron radiotherapy (IOERT). Our aim is to share our experience of metal oxide semiconductor field-effect transistors (MOSFETs) and radiochromic films with patients undergoing IOERT using a general-purpose linac. We used MOSFETs inserted into sterile bronchus catheters and radiochromic films that were cut, digitized, and sterilized by means of gas plasma. In all, 59 measurements were taken from 27 patients involving 15 primary tumors (seven breast and eight non-breast tumors) and 12 relapses. Data were subjected to an outliers' analysis and classified according to their compatibility with the relevant doses. Associations were sought regarding the type of detector, breast and non-breast irradiation, and the radiation oncologist's assessment of the difficulty of detector placement. At the same time, 19 measurements were carried out at the tumor bed with both detectors. MOSFET measurements ([Formula: see text]  = 93.5 %, sD  =  6.5 %) were not significantly shifted from film measurements ([Formula: see text]  =  96.0 %, sD  =  5.5 %; p  =  0.109), and no associations were found (p = 0.526, p = 0.295,  and p = 0.501, respectively). As regards measurements performed at the tumor bed with both detectors, MOSFET measurements ([Formula: see text]  =  95.0 %, sD  =  5.4 % were not significantly shifted from film measurements ([Formula: see text]  =  96.4 %, sD  =  5.0 %; p  =  0.363). In vivo dosimetry can produce satisfactory results at every studied location with a general-purpose linac. Detector choice should depend on user factors, not on the detector performance itself. Surgical team collaboration is crucial to success.

  10. SafeNet: a methodology for integrating general-purpose unsafe devices in safe-robot rehabilitation systems.

    Science.gov (United States)

    Vicentini, Federico; Pedrocchi, Nicola; Malosio, Matteo; Molinari Tosatti, Lorenzo

    2014-09-01

    Robot-assisted neurorehabilitation often involves networked systems of sensors ("sensory rooms") and powerful devices in physical interaction with weak users. Safety is unquestionably a primary concern. Some lightweight robot platforms and devices designed on purpose include safety properties using redundant sensors or intrinsic safety design (e.g. compliance and backdrivability, limited exchange of energy). Nonetheless, the entire "sensory room" shall be required to be fail-safe and safely monitored as a system at large. Yet, sensor capabilities and control algorithms used in functional therapies require, in general, frequent updates or re-configurations, making a safety-grade release of such devices hardly sustainable in cost-effectiveness and development time. As such, promising integrated platforms for human-in-the-loop therapies could not find clinical application and manufacturing support because of lacking in the maintenance of global fail-safe properties. Under the general context of cross-machinery safety standards, the paper presents a methodology called SafeNet for helping in extending the safety rate of Human Robot Interaction (HRI) systems using unsafe components, including sensors and controllers. SafeNet considers, in fact, the robotic system as a device at large and applies the principles of functional safety (as in ISO 13489-1) through a set of architectural procedures and implementation rules. The enabled capability of monitoring a network of unsafe devices through redundant computational nodes, allows the usage of any custom sensors and algorithms, usually planned and assembled at therapy planning-time rather than at platform design-time. A case study is presented with an actual implementation of the proposed methodology. A specific architectural solution is applied to an example of robot-assisted upper-limb rehabilitation with online motion tracking. Copyright © 2014 Elsevier Ireland Ltd. All rights reserved.

  11. A computational code for resolution of general compartment models applied to internal dosimetry

    Energy Technology Data Exchange (ETDEWEB)

    Claro, Thiago R.; Todo, Alberto S., E-mail: claro@usp.br, E-mail: astodo@ipen.br [Instituto de Pesquisas Energeticas e Nucleares (IPEN/CNEN-SP), Sao Paulo, SP (Brazil)

    2011-07-01

    The dose resulting from internal contamination can be estimated with the use of biokinetic models combined with experimental results obtained from bio analysis and the knowledge of the incorporation time. The biokinetics models can be represented by a set of compartments expressing the transportation, retention and elimination of radionuclides from the body. The ICRP publications, number 66, 78 and 100, present compartmental models for the respiratory tract, gastrointestinal tract and for systemic distribution for an array of radionuclides of interest for the radiological protection. The objective of this work is to develop a computational code for designing, visualization and resolution of compartmental models of any nature. There are available four different techniques for the resolution of system of differential equations, including semi-analytical and numerical methods. The software was developed in C{ne} programming, using a Microsoft Access database and XML standards for file exchange with other applications. Compartmental models for uranium, thorium and iodine radionuclides were generated for the validation of the CBT software. The models were subsequently solved by SSID software and the results compared with the values published in the issue 78 of ICRP. In all cases the system is in accordance with the values published by ICRP. (author)

  12. A computational code for resolution of general compartment models applied to internal dosimetry

    International Nuclear Information System (INIS)

    Claro, Thiago R.; Todo, Alberto S.

    2011-01-01

    The dose resulting from internal contamination can be estimated with the use of biokinetic models combined with experimental results obtained from bio analysis and the knowledge of the incorporation time. The biokinetics models can be represented by a set of compartments expressing the transportation, retention and elimination of radionuclides from the body. The ICRP publications, number 66, 78 and 100, present compartmental models for the respiratory tract, gastrointestinal tract and for systemic distribution for an array of radionuclides of interest for the radiological protection. The objective of this work is to develop a computational code for designing, visualization and resolution of compartmental models of any nature. There are available four different techniques for the resolution of system of differential equations, including semi-analytical and numerical methods. The software was developed in C≠ programming, using a Microsoft Access database and XML standards for file exchange with other applications. Compartmental models for uranium, thorium and iodine radionuclides were generated for the validation of the CBT software. The models were subsequently solved by SSID software and the results compared with the values published in the issue 78 of ICRP. In all cases the system is in accordance with the values published by ICRP. (author)

  13. Deposition, characterization, and in vivo performance of parylene coating on general-purpose silicone for examining potential biocompatible surface modifications

    International Nuclear Information System (INIS)

    Chou, Chia-Man; Shiao, Chiao-Ju; Chung, Chi-Jen; He, Ju-Liang

    2013-01-01

    In this study, a thorough investigation of parylene coatings was conducted, as follows: microstructure (i.e., X-ray diffractometer (XRD) and cold field emission scanning electron microscope (FESEM)), mechanical property (i.e., pencil hardness and cross-cut adhesion test), surface property (i.e., water contact angle measurement, IR, and X-ray photoelectron spectroscopy (XPS)), and biocompatibility tests (i.e., fibroblast cell culture, platelet adhesion, and animal studies). The results revealed that parylene, a crystalline and brittle coating, exhibited satisfactory film adhesion and relative hydrophobicity, thereby contributing to its effective barrier properties. Fibroblast cell culturing on the parylene-deposited specimen demonstrated improved cell proliferation and equivalent to or superior blood compatibility than that of the medical-grade silicone (currently used clinically). In the animal study, parylene coatings exhibited similar subcutaneous inflammatory reactions compared with the medical-grade silicone. Both in vitro and in vivo tests demonstrated the satisfactory biocompatibility of parylene coatings. - Highlights: • A complete investigation to identify the characteristics of parylene coatings on general-purpose silicones. • Microstructures, surface properties and mechanical properties of parylene coatings were examined. • In vitro (Cell culture, platelet adhesion) tests and animal studies revealed satisfactory biocompatibility. • An alternative of medical-grade silicones is expected to be obtained

  14. General-purpose computer networks and resource sharing in ERDA. Volume 3. Remote resource-sharing experience and findings

    Energy Technology Data Exchange (ETDEWEB)

    1977-07-15

    The investigation focused on heterogeneous networks in which a variety of dissimilar computers and operating systems were interconnected nationwide. Homogeneous networks, such as MFE net and SACNET, were not considered since they could not be used for general purpose resource sharing. Issues of privacy and security are of concern in any network activity. However, consideration of privacy and security of sensitive data arise to a much lesser degree in unclassified scientific research than in areas involving personal or proprietary information. Therefore, the existing mechanisms at individual sites for protecting sensitive data were relied on, and no new protection mechanisms to prevent infringement of privacy and security were attempted. Further development of ERDA networking will need to incorporate additional mechanisms to prevent infringement of privacy. The investigation itself furnishes an excellent example of computational resource sharing through a heterogeneous network. More than twenty persons, representing seven ERDA computing sites, made extensive use of both ERDA and non-ERDA computers in coordinating, compiling, and formatting the data which constitute the bulk of this report. Volume 3 analyzes the benefits and barriers encountered in actual resource sharing experience, and provides case histories of typical applications.

  15. Architecture of a general purpose embedded Slow-Control Adapter ASIC for future high-energy physics experiments

    Science.gov (United States)

    Gabrielli, Alessandro; Loddo, Flavio; Ranieri, Antonio; De Robertis, Giuseppe

    2008-10-01

    This work is aimed at defining the architecture of a new digital ASIC, namely Slow-Control Adapter (SCA), which will be designed in a commercial 130-nm CMOS technology. This chip will be embedded within a high-speed data acquisition optical link (GBT) to control and monitor the front-end electronics in future high-energy physics experiments. The GBT link provides a transparent transport layer between the SCA and control electronics in the counting room. The proposed SCA supports a variety of common bus protocols to interface with end-user general-purpose electronics. Between the GBT and the SCA a standard 100 Mb/s IEEE-802.3 compatible protocol will be implemented. This standard protocol allows off-line tests of the prototypes using commercial components that support the same standard. The project is justified because embedded applications in modern large HEP experiments require particular care to assure the lowest possible power consumption, still offering the highest reliability demanded by very large particle detectors.

  16. Deposition, characterization, and in vivo performance of parylene coating on general-purpose silicone for examining potential biocompatible surface modifications

    Energy Technology Data Exchange (ETDEWEB)

    Chou, Chia-Man [Division of Pediatric Surgery, Department of Surgery, Taichung Veterans General Hospital, 160, Sec. 3, Taichung Port Rd., Taichung 40705, Taiwan, ROC (China); Department of Medicine, National Yang-Ming University, 155, Sec. 2, Linong Street, Taipei 11221, Taiwan, ROC (China); Shiao, Chiao-Ju [Department of Materials Science and Engineering, Feng Chia University, 100, Wen-Hwa Rd., Taichung 40724, Taiwan, ROC (China); Chung, Chi-Jen, E-mail: cjchung@seed.net.tw [Department of Dental Technology and Materials Science, Central Taiwan University of Science and Technology, 666 Buzih Rd., Beitun District, Taichung 40601, Taiwan, ROC (China); He, Ju-Liang [Department of Materials Science and Engineering, Feng Chia University, 100, Wen-Hwa Rd., Taichung 40724, Taiwan, ROC (China)

    2013-12-31

    In this study, a thorough investigation of parylene coatings was conducted, as follows: microstructure (i.e., X-ray diffractometer (XRD) and cold field emission scanning electron microscope (FESEM)), mechanical property (i.e., pencil hardness and cross-cut adhesion test), surface property (i.e., water contact angle measurement, IR, and X-ray photoelectron spectroscopy (XPS)), and biocompatibility tests (i.e., fibroblast cell culture, platelet adhesion, and animal studies). The results revealed that parylene, a crystalline and brittle coating, exhibited satisfactory film adhesion and relative hydrophobicity, thereby contributing to its effective barrier properties. Fibroblast cell culturing on the parylene-deposited specimen demonstrated improved cell proliferation and equivalent to or superior blood compatibility than that of the medical-grade silicone (currently used clinically). In the animal study, parylene coatings exhibited similar subcutaneous inflammatory reactions compared with the medical-grade silicone. Both in vitro and in vivo tests demonstrated the satisfactory biocompatibility of parylene coatings. - Highlights: • A complete investigation to identify the characteristics of parylene coatings on general-purpose silicones. • Microstructures, surface properties and mechanical properties of parylene coatings were examined. • In vitro (Cell culture, platelet adhesion) tests and animal studies revealed satisfactory biocompatibility. • An alternative of medical-grade silicones is expected to be obtained.

  17. Development of a general-purpose method for cell purification using Cre/loxP-mediated recombination.

    Science.gov (United States)

    Kuroki, Shunsuke; Akiyoshi, Mika; Ideguchi, Ko; Kitano, Satsuki; Miyachi, Hitoshi; Hirose, Michiko; Mise, Nathan; Abe, Kuniya; Ogura, Atsuo; Tachibana, Makoto

    2015-06-01

    A mammalian body is composed of more than 200 different types of cells. The purification of a certain cell type from tissues/organs enables a wide variety of studies. One popular cell purification method is immunological isolation, using antibodies against specific cell surface antigens. However, this is not a general-purpose method, since suitable antigens have not been found in certain cell types, including embryonic gonadal somatic cells and Sertoli cells. To address this issue, we established a knock-in mouse line, named R26 KI, designed to express the human cell surface antigen hCD271 through Cre/loxP-mediated recombination. First, we used the R26 Kl mouse line to purify embryonic gonadal somatic cells. Gonadal somatic cells were purified from the R26 KI; Nr5a1-Cre-transgenic (tg) embryos almost equally as efficiently as from Nr5a1-hCD271-tg embryos. Second, we used the R26 KI mouse line to purify Sertoli cells successfully from R26 KI; Amh-Cre-tg testes. In summary, we propose that the R26 KI mouse line is a powerful tool for the purification of various cell types. © 2015 Wiley Periodicals, Inc.

  18. STICK: Spike Time Interval Computational Kernel, a Framework for General Purpose Computation Using Neurons, Precise Timing, Delays, and Synchrony.

    Science.gov (United States)

    Lagorce, Xavier; Benosman, Ryad

    2015-11-01

    There has been significant research over the past two decades in developing new platforms for spiking neural computation. Current neural computers are primarily developed to mimic biology. They use neural networks, which can be trained to perform specific tasks to mainly solve pattern recognition problems. These machines can do more than simulate biology; they allow us to rethink our current paradigm of computation. The ultimate goal is to develop brain-inspired general purpose computation architectures that can breach the current bottleneck introduced by the von Neumann architecture. This work proposes a new framework for such a machine. We show that the use of neuron-like units with precise timing representation, synaptic diversity, and temporal delays allows us to set a complete, scalable compact computation framework. The framework provides both linear and nonlinear operations, allowing us to represent and solve any function. We show usability in solving real use cases from simple differential equations to sets of nonlinear differential equations leading to chaotic attractors.

  19. Architecture of a general purpose embedded Slow-Control Adapter ASIC for future high-energy physics experiments

    International Nuclear Information System (INIS)

    Gabrielli, Alessandro; Loddo, Flavio; Ranieri, Antonio; De Robertis, Giuseppe

    2008-01-01

    This work is aimed at defining the architecture of a new digital ASIC, namely Slow-Control Adapter (SCA), which will be designed in a commercial 130-nm CMOS technology. This chip will be embedded within a high-speed data acquisition optical link (GBT) to control and monitor the front-end electronics in future high-energy physics experiments. The GBT link provides a transparent transport layer between the SCA and control electronics in the counting room. The proposed SCA supports a variety of common bus protocols to interface with end-user general-purpose electronics. Between the GBT and the SCA a standard 100 Mb/s IEEE-802.3 compatible protocol will be implemented. This standard protocol allows off-line tests of the prototypes using commercial components that support the same standard. The project is justified because embedded applications in modern large HEP experiments require particular care to assure the lowest possible power consumption, still offering the highest reliability demanded by very large particle detectors.

  20. Design and Deployment of a General Purpose, Open Source LoRa to Wi-Fi Hub and Data Logger

    Science.gov (United States)

    DeBell, T. C.; Udell, C.; Kwon, M.; Selker, J. S.; Lopez Alcala, J. M.

    2017-12-01

    Methods and technologies facilitating internet connectivity and near-real-time status updates for in site environmental sensor data are of increasing interest in Earth Science. However, Open Source, Do-It-Yourself technologies that enable plug and play functionality for web-connected sensors and devices remain largely inaccessible for typical researchers in our community. The Openly Published Environmental Sensing Lab at Oregon State University (OPEnS Lab) constructed an Open Source 900 MHz Long Range Radio (LoRa) receiver hub with SD card data logger, Ethernet and Wi-Fi shield, and 3D printed enclosure that dynamically uploads transmissions from multiple wirelessly-connected environmental sensing devices. Data transmissions may be received from devices up to 20km away. The hub time-stamps, saves to SD card, and uploads all transmissions to a Google Drive spreadsheet to be accessed in near-real-time by researchers and GeoVisualization applications (such as Arc GIS) for access, visualization, and analysis. This research expands the possibilities of scientific observation of our Earth, transforming the technology, methods, and culture by combining open-source development and cutting edge technology. This poster details our methods and evaluates the application of using 3D printing, Arduino Integrated Development Environment (IDE), Adafruit's Open-Hardware Feather development boards, and the WIZNET5500 Ethernet shield for designing this open-source, general purpose LoRa to Wi-Fi data logger.

  1. Cafe Variome: general-purpose software for making genotype-phenotype data discoverable in restricted or open access contexts.

    Science.gov (United States)

    Lancaster, Owen; Beck, Tim; Atlan, David; Swertz, Morris; Thangavelu, Dhiwagaran; Veal, Colin; Dalgleish, Raymond; Brookes, Anthony J

    2015-10-01

    Biomedical data sharing is desirable, but problematic. Data "discovery" approaches-which establish the existence rather than the substance of data-precisely connect data owners with data seekers, and thereby promote data sharing. Cafe Variome (http://www.cafevariome.org) was therefore designed to provide a general-purpose, Web-based, data discovery tool that can be quickly installed by any genotype-phenotype data owner, or network of data owners, to make safe or sensitive content appropriately discoverable. Data fields or content of any type can be accommodated, from simple ID and label fields through to extensive genotype and phenotype details based on ontologies. The system provides a "shop window" in front of data, with main interfaces being a simple search box and a powerful "query-builder" that enable very elaborate queries to be formulated. After a successful search, counts of records are reported grouped by "openAccess" (data may be directly accessed), "linkedAccess" (a source link is provided), and "restrictedAccess" (facilitated data requests and subsequent provision of approved records). An administrator interface provides a wide range of options for system configuration, enabling highly customized single-site or federated networks to be established. Current uses include rare disease data discovery, patient matchmaking, and a Beacon Web service. © 2015 WILEY PERIODICALS, INC.

  2. A k-distribution-based radiation code and its computational optimization for an atmospheric general circulation model

    International Nuclear Information System (INIS)

    Sekiguchi, Miho; Nakajima, Teruyuki

    2008-01-01

    The gas absorption process scheme in the broadband radiative transfer code 'mstrn8', which is used to calculate atmospheric radiative transfer efficiently in a general circulation model, is improved. Three major improvements are made. The first is an update of the database of line absorption parameters and the continuum absorption model. The second is a change to the definition of the selection rule for gas absorption used to choose which absorption bands to include. The last is an upgrade of the optimization method used to decrease the number of quadrature points used for numerical integration in the correlated k-distribution approach, thereby realizing higher computational efficiency without losing accuracy. The new radiation package termed 'mstrnX' computes radiation fluxes and heating rates with errors less than 0.6 W/m 2 and 0.3 K/day, respectively, through the troposphere and the lower stratosphere for any standard AFGL atmospheres. A serious cold bias problem of an atmospheric general circulation model using the ancestor code 'mstrn8' is almost solved by the upgrade to 'mstrnX'

  3. A general multiblock Euler code for propulsion integration. Volume 1: Theory document

    Science.gov (United States)

    Chen, H. C.; Su, T. Y.; Kao, T. J.

    1991-01-01

    A general multiblock Euler solver was developed for the analysis of flow fields over geometrically complex configurations either in free air or in a wind tunnel. In this approach, the external space around a complex configuration was divided into a number of topologically simple blocks, so that surface-fitted grids and an efficient flow solution algorithm could be easily applied in each block. The computational grid in each block is generated using a combination of algebraic and elliptic methods. A grid generation/flow solver interface program was developed to facilitate the establishment of block-to-block relations and the boundary conditions for each block. The flow solver utilizes a finite volume formulation and an explicit time stepping scheme to solve the Euler equations. A multiblock version of the multigrid method was developed to accelerate the convergence of the calculations. The generality of the method was demonstrated through the analysis of two complex configurations at various flow conditions. Results were compared to available test data. Two accompanying volumes, user manuals for the preparation of multi-block grids (vol. 2) and for the Euler flow solver (vol. 3), provide information on input data format and program execution.

  4. In vivo dosimetry in intraoperative electron radiotherapy. microMOSFETs, radiochromic films and a general-purpose linac

    Energy Technology Data Exchange (ETDEWEB)

    Lopez-Tarjuelo, Juan; Marco-Blancas, Noelia de; Santos-Serra, Agustin; Quiros-Higueras, Juan David [Consorcio Hospitalario Provincial de Castellon, Servicio de Radiofisica y Proteccion Radiologica, Castellon de la Plana (Spain); Bouche-Babiloni, Ana; Morillo-Macias, Virginia; Ferrer-Albiach, Carlos [Consorcio Hospitalario Provincial de Castellon, Servicio de Oncologia Radioterapica, Castellon de la Plana (Spain)

    2014-11-15

    In vivo dosimetry is desirable for the verification, recording, and eventual correction of treatment in intraoperative electron radiotherapy (IOERT). Our aim is to share our experience of metal oxide semiconductor field-effect transistors (MOSFETs) and radiochromic films with patients undergoing IOERT using a general-purpose linac. We used MOSFETs inserted into sterile bronchus catheters and radiochromic films that were cut, digitized, and sterilized by means of gas plasma. In all, 59 measurements were taken from 27 patients involving 15 primary tumors (seven breast and eight non-breast tumors) and 12 relapses. Data were subjected to an outliers' analysis and classified according to their compatibility with the relevant doses. Associations were sought regarding the type of detector, breast and non-breast irradiation, and the radiation oncologist's assessment of the difficulty of detector placement. At the same time, 19 measurements were carried out at the tumor bed with both detectors. MOSFET measurements (D = 93.5 %, s{sub D} = 6.5 %) were not significantly shifted from film measurements (D = 96.0 %, s{sub D} = 5.5 %; p = 0.109), and no associations were found (p = 0.526, p = 0.295, and p = 0.501, respectively). As regards measurements performed at the tumor bed with both detectors, MOSFET measurements (D = 95.0 %, s{sub D} = 5.4 %) were not significantly shifted from film measurements (D = 96.4 %, s{sub D} = 5.0 %; p = 0.363). In vivo dosimetry can produce satisfactory results at every studied location with a general-purpose linac. Detector choice should depend on user factors, not on the detector performance itself. Surgical team collaboration is crucial to success. (orig.) [German] Die In-vivo-Dosimetrie ist wuenschenswert fuer die Ueberpruefung, Registrierung und die eventuelle Korrektur der Behandlungen in der IOERT (''Intraoperative Electron Radiation Therapy''). Unser Ziel ist die Veroeffentlichung unserer Erfahrungen beim

  5. Creep properties of forged 2219 T6 aluminum alloy shell of general-purpose heat source-radioisotope thermoelectric generator

    International Nuclear Information System (INIS)

    Hammond, J.P.

    1981-12-01

    The shell (2219 T6 aluminum forging) of the General Purpose Heat Source-Radioisotope Thermoelectric Generator was designed to retain the generator under sufficient elastic stress to secure it during space flight. A major concern was the extent to which the elastic stress would relax by creep. To determine acceptability of the shell construction material, the following proof tests simulating service were performed: 600 h of testing at 270 0 C under 24.1 MPa stress followed by 10,000 h of storage at 177 0 C under 55.1 MPa, both on the ground; and 10,000 h of flight in space at 270 0 C under 34.4 MPa stress. Additionally, systematic creep testing was performed at 177 and 260 0 C to establish creep design curves. The creep tests performed at 177 0 C revealed comparatively large amounts of primary creep followed by small amounts of secondary creep. The early creep is believed to be abetted by unstable substructures that are annealed out during testing at this temperature. The creep tests performed at 270 0 C showed normal primary creep followed by large amounts of secondary creep. Duplicate proof tests simulating the ground exposure conditions gave results that were in good agreement. The proof test simulating space flight at 270 0 C gave 0.11% primary creep followed by 0.59% secondary creep. About 10% of the second-stage creep was caused by four or five instantaneous strains, which began at the 4500-h mark. One or two of these strain bursts, occurred in each of several other tests at 177 and 260 0 C but were assessed as very moderate in magnitude. The effect is attributable to a slightly microsegregated condition remaining from the original cast structure

  6. Development and validation of a general-purpose ASIC chip for the control of switched reluctance machines

    International Nuclear Information System (INIS)

    Chen Haijin; Lu Shengli; Shi Longxing

    2009-01-01

    A general-purpose application specific integrated circuit (ASIC) chip for the control of switched reluctance machines (SRMs) was designed and validated to fill the gap between the microcontroller capability and the controller requirements of high performance switched reluctance drive (SRD) systems. It can be used for the control of SRM running either in low speed or in high-speed, i.e., either in chopped current control (CCC) mode or in angular position control (APC) mode. Main functions of the chip include filtering and cycle calculation of rotor angular position signals, commutation logic according to rotor cycle and turn-on/turn-off angles (θ on /θ off ), controllable pulse width modulation (PWM) waveforms generation, chopping control with adjustable delay time, and commutation control with adjustable delay time. All the control parameters of the chip are set online by the microcontroller through a serial peripheral interface (SPI). The chip has been designed with the standard cell based design methodology, and implemented in the central semiconductor manufacturing corporation (CSMC) 0.5 μm complementary metal-oxide-semiconductor (CMOS) process technology. After a successful automatic test equipment (ATE) test using the Nextest's Maverick test system, the chip was further validated through an experimental three-phase 6/2-pole SRD system. Both the ATE test and experimental validation results show that the chip can meet the control requirements of high performance SRD systems, and simplify the controller construction. For a resolution of 0.36 deg. (electrical degree), the chip's maximum processable frequency of the rotor angular position signals is 10 kHz, which is 300,000 rev/min when a three-phase 6/2-pole SRM is concerned

  7. General Purpose Ground Forces: What Purpose?

    National Research Council Canada - National Science Library

    Challis, Dan

    1993-01-01

    "New World Order," a phrase uttered frequently by former President George Bush during and after the Persian Gulf War, no longer connotes the optimism of America's global view at the end of Desert Storm...

  8. "Why Do I Have to Take This Course?": How Academic Advisers Can Help Students Find Personal Meaning and Purpose in General Education

    Science.gov (United States)

    Kirk-Kuwaye, Michael; Sano-Franchini, Dominic

    2015-01-01

    For a variety of reasons, student engagement in general education continues to be a challenge. Perhaps one way to increase engagement is to connect general education with a deep student need: finding meaning and purpose in their lives or exploring what some have called "big questions." Recent scholarship has defined these clusters of…

  9. The Treatment of Polysemy and Homonymy in Monolingual General-purpose Dictionaries with Special Reference to Isichazamazwi SesiNdebele

    Directory of Open Access Journals (Sweden)

    Eventhough Ndlovu

    2011-10-01

    Full Text Available

    ABSTRACT: This article focuses on the treatment of polysemy and homonymy in general-purpose monolingual dictionaries with special reference to Isichazamazwi SesiNdebele. It was found that there are some inconsistencies in the treatment of polysemous and homonymous entries in this dictionary. The article shows that an overreliance on one criterion, particularly etymology, to distinguish polysemy and homonymy is often misleading and unreliable. Polysemy itself has its own inherent complexities, among these being the problem of determining the exact number of meanings of a polysemous lemma. When the meanings of a polysemous lemma are listed, the central or primary meaning, which is not always easily ascertainable, should come first. A holistic approach is proposed to distinguish polysemy and homonymy, which entails the use of the following criteria: etymology, relatedness vs unrelatedness of meaning, componential analysis, the identification of the central or core meaning and the test of ambiguity. Whatever results are obtained from a particular criterion, these findings must be compared with those of other criteria, and verified against native speakers' intuitive knowledge and introspective judgements.

    OPSOMMING: Die behandeling van polisemie en homonimie in eentalige algemene woordeboeke met spesiale verwysing na Isichazamazwi SesiNdebele. Hierdie artikel fokus op die behandeling van polisemie en homonimie in algemene eentalige woordeboeke met spesiale verwysing na Isichazamazwi SesiNdebele. Daar is vasgestel dat daar 'n aantal inkonsekwensies in die behandeling van poliseme en homonieme inskrywings in hierdie woordeboek is. Die artikel toon dat 'n te groot steun op een kriterium, veral etimologie, om polisemie en homonimie te onderskei, dikwels misleidend en onbetroubaar is. Polisemie self het sy eie inherente gekompliseerdhede waarvan sommige die probleem is om die presiese aantal betekenisse van 'n poliseme lemma te bepaal. Wanneer

  10. Codex general standard for irradiated foods and recommended international code of practice for the operation of radiation facilities used for the treatment of foods

    International Nuclear Information System (INIS)

    1990-06-01

    The FAO/WHO Codex Alimentarius Commission was established to implement the Joint FAO/WHO Food Standards Programme. The purpose of this programme is to protect the health of consumers and to ensure fair practices in the food trade. At its 15th session, held in July 1983, the Commission adopted a Codex General Standard for Irradiated Foods and a Recommended International Code of Practice for the Operation of Radiation Facilities used for the Treatment of Foods. This Standard takes into account the recommendations and conclusions of the Joint FAO/IAEA/WHO Expert Committees convened to evaluate all available data concerning the various aspects of food irradiation. This Standard refers only to those aspects which relate to the processing of foods by ionising energy. The Standard recognizes that the process of food irradiation has been established as safe for general application to an overall average level of absorbed dose of 10 KGy. The latter value shold not be regarded as a toxicological upper limit above which irradiated foods become unsafe; it is simply the level at or below which safety has been established. The Standard provides certain mandatory provisions concerning the facilities used and for the control of the process in the irradiation plants. The present Standard requires that shipping documents accompanying irradiated foods moving in trade should indicate the fact of irradiation. The labelling of prepackaged irradiated foods intended for direct sale to the consumer is not covered in this Standard

  11. Codex general standard for irradiated foods and recommended international code of practice for the operation of radiation facilities used for the treatment of foods

    International Nuclear Information System (INIS)

    1984-01-01

    The FAO/WHO Codex Alimentarius Commission was established to implement the Joint FAO/WHO Food Standards Programme. The purpose of this programme is to protect the health of consumers and to ensure fair practices in the food trade. At its 15th session, held in July 1983, the Commission adopted a Codex General Standard for Irradiated Foods and a Recommended International Code of Practice for the Operation of Radiation Facilities used for the Treatment of Foods. This Standard takes into account the recommendations and conclusions of the Joint FAO/IAEA/WHO Expert Committees convened to evaluate all available data concerning the various aspects of food irradiation. This Standard refers only to those aspects which relate to the processing of foods by ionising energy. The Standard recognizes that the process of food irradiation has been established as safe for general application to an overall average level of absorbed dose of 10 kGy. The latter value should not be regarded as a toxicological upper limit above which irradiated foods become unsafe; it is simply the level at or below which safety has been established. The Standard provides certain mandatory provisions concerning the facilities used and for the control of the process in the irradiation plants. The present Standard requires that shipping documents accompanying irradiated foods moving in trade should indicate the fact of irradiation. The labelling of prepackaged irradiated foods intended for direct sale to the consumer is not covered in this Standard

  12. Introduction of SCIENCE code package

    International Nuclear Information System (INIS)

    Lu Haoliang; Li Jinggang; Zhu Ya'nan; Bai Ning

    2012-01-01

    The SCIENCE code package is a set of neutronics tools based on 2D assembly calculations and 3D core calculations. It is made up of APOLLO2F, SMART and SQUALE and used to perform the nuclear design and loading pattern analysis for the reactors on operation or under construction of China Guangdong Nuclear Power Group. The purpose of paper is to briefly present the physical and numerical models used in each computation codes of the SCIENCE code pack age, including the description of the general structure of the code package, the coupling relationship of APOLLO2-F transport lattice code and SMART core nodal code, and the SQUALE code used for processing the core maps. (authors)

  13. Coding Partitions

    Directory of Open Access Journals (Sweden)

    Fabio Burderi

    2007-05-01

    Full Text Available Motivated by the study of decipherability conditions for codes weaker than Unique Decipherability (UD, we introduce the notion of coding partition. Such a notion generalizes that of UD code and, for codes that are not UD, allows to recover the ``unique decipherability" at the level of the classes of the partition. By tacking into account the natural order between the partitions, we define the characteristic partition of a code X as the finest coding partition of X. This leads to introduce the canonical decomposition of a code in at most one unambiguouscomponent and other (if any totally ambiguouscomponents. In the case the code is finite, we give an algorithm for computing its canonical partition. This, in particular, allows to decide whether a given partition of a finite code X is a coding partition. This last problem is then approached in the case the code is a rational set. We prove its decidability under the hypothesis that the partition contains a finite number of classes and each class is a rational set. Moreover we conjecture that the canonical partition satisfies such a hypothesis. Finally we consider also some relationships between coding partitions and varieties of codes.

  14. Online retrieval of patient information by asynchronous communication between general purpose computer and stand-alone personal computer

    International Nuclear Information System (INIS)

    Tsutsumi, Reiko; Takahashi, Kazuei; Sato, Toshiko; Komatani, Akio; Yamaguchi, Koichi

    1988-01-01

    Asynchronous communication was made between host (FACOM M-340) and personal computer (OLIBETTIE S-2250) to get patient's information required for RIA test registration. The retrieval system consists of a keyboad input of six numeric codes, patient's ID, and a real time reply containing six parameters for the patient. Their identified parameters are patient's name, sex, date of birth (include area), department, and out- or inpatient. Linking this program to RIA registration program for individual patient, then, operator can input name of RIA test requested. Our simple retrieval program made a useful data network between different types of host and stand-alone personal computers, and enabled us accurate and labor-saving registration for RIA test. (author)

  15. Synthesizing Certified Code

    Science.gov (United States)

    Whalen, Michael; Schumann, Johann; Fischer, Bernd

    2002-01-01

    Code certification is a lightweight approach to demonstrate software quality on a formal level. Its basic idea is to require producers to provide formal proofs that their code satisfies certain quality properties. These proofs serve as certificates which can be checked independently. Since code certification uses the same underlying technology as program verification, it also requires many detailed annotations (e.g., loop invariants) to make the proofs possible. However, manually adding theses annotations to the code is time-consuming and error-prone. We address this problem by combining code certification with automatic program synthesis. We propose an approach to generate simultaneously, from a high-level specification, code and all annotations required to certify generated code. Here, we describe a certification extension of AUTOBAYES, a synthesis tool which automatically generates complex data analysis programs from compact specifications. AUTOBAYES contains sufficient high-level domain knowledge to generate detailed annotations. This allows us to use a general-purpose verification condition generator to produce a set of proof obligations in first-order logic. The obligations are then discharged using the automated theorem E-SETHEO. We demonstrate our approach by certifying operator safety for a generated iterative data classification program without manual annotation of the code.

  16. Insurance billing and coding.

    Science.gov (United States)

    Napier, Rebecca H; Bruelheide, Lori S; Demann, Eric T K; Haug, Richard H

    2008-07-01

    The purpose of this article is to highlight the importance of understanding various numeric and alpha-numeric codes for accurately billing dental and medically related services to private pay or third-party insurance carriers. In the United States, common dental terminology (CDT) codes are most commonly used by dentists to submit claims, whereas current procedural terminology (CPT) and International Classification of Diseases, Ninth Revision, Clinical Modification (ICD.9.CM) codes are more commonly used by physicians to bill for their services. The CPT and ICD.9.CM coding systems complement each other in that CPT codes provide the procedure and service information and ICD.9.CM codes provide the reason or rationale for a particular procedure or service. These codes are more commonly used for "medical necessity" determinations, and general dentists and specialists who routinely perform care, including trauma-related care, biopsies, and dental treatment as a result of or in anticipation of a cancer-related treatment, are likely to use these codes. Claim submissions for care provided can be completed electronically or by means of paper forms.

  17. International Classification of Primary Care-2 coding of primary care data at the general out-patients' clinic of General Hospital, Lagos, Nigeria.

    Science.gov (United States)

    Olagundoye, Olawunmi Abimbola; van Boven, Kees; van Weel, Chris

    2016-01-01

    Primary care serves as an integral part of the health systems of nations especially the African continent. It is the portal of entry for nearly all patients into the health care system. Paucity of accurate data for health statistics remains a challenge in the most parts of Africa because of inadequate technical manpower and infrastructure. Inadequate quality of data systems contributes to inaccurate data. A simple-to-use classification system such as the International Classification of Primary Care (ICPC) may be a solution to this problem at the primary care level. To apply ICPC-2 for secondary coding of reasons for encounter (RfE), problems managed and processes of care in a Nigerian primary care setting. Furthermore, to analyze the value of selected presented symptoms as predictors of the most common diagnoses encountered in the study setting. Content analysis of randomly selected patients' paper records for data collection at the end of clinic sessions conducted by family physicians at the general out-patients' clinics. Contents of clinical consultations were secondarily coded with the ICPC-2 and recorded into excel spreadsheets with fields for sociodemographic data such as age, sex, occupation, religion, and ICPC elements of an encounter: RfE/complaints, diagnoses/problems, and interventions/processes of care. Four hundred and one encounters considered in this study yielded 915 RfEs, 546 diagnoses, and 1221 processes. This implies an average of 2.3 RfE, 1.4 diagnoses, and 3.0 processes per encounter. The top 10 RfE, diagnoses/common illnesses, and processes were determined. Through the determination of the probability of the occurrence of certain diseases beginning with a RfE/complaint, the top five diagnoses that resulted from each of the top five RfE were also obtained. The top five RfE were: headache, fever, pain general/multiple sites, visual disturbance other and abdominal pain/cramps general. The top five diagnoses were: Malaria, hypertension

  18. 77 FR 18899 - To Modify Duty-Free Treatment Under the Generalized System of Preferences and For Other Purposes

    Science.gov (United States)

    2012-03-29

    ... Generalized System of Preferences (GSP) if such country fails to act in good faith in recognizing as binding or in enforcing arbitral awards in favor of United States citizens or a corporation, partnership, or... faith in enforcing arbitral awards in favor of United States citizens or a corporation, partnership, or...

  19. General-purpose parallel algorithm based on CUDA for source pencils' deployment of large γ irradiator

    International Nuclear Information System (INIS)

    Yang Lei; Gong Xueyu; Wang Ling

    2013-01-01

    Combined with standard mathematical model for evaluating quality of deploying results, a new high-performance parallel algorithm for source pencils' deployment was obtained by using parallel plant growth simulation algorithm which was completely parallelized with CUDA execute model, and the corresponding code can run on GPU. Based on such work, several instances in various scales were used to test the new version of algorithm. The results show that, based on the advantage of old versions. the performance of new one is improved more than 500 times comparing with the CPU version, and also 30 times with the CPU plus GPU hybrid version. The computation time of new version is less than ten minutes for the irradiator of which the activity is less than 111 PBq. For a single GTX275 GPU, the maximum computing power of new version is no more than 167 PBq as well as the computation time is no more than 25 minutes, and for multiple GPUs, the power can be improved more. Overall, the new version of algorithm running on GPU can satisfy the requirement of source pencils' deployment of any domestic irradiator, and it is of high competitiveness. (authors)

  20. Text messaging among residents and faculty in a university general surgery residency program: prevalence, purpose, and patient care.

    Science.gov (United States)

    Shah, Dhruvil R; Galante, Joseph M; Bold, Richard J; Canter, Robert J; Martinez, Steve R

    2013-01-01

    There is little information about the use of text messaging (texting) devices among resident and faculty physicians for patient-related care (PRC). To determine the prevalence, frequency, purpose, and concerns regarding texting among resident and attending surgeons and to identify factors associated with PRC texting. E-mail survey. University medical center and its affiliated hospitals. Surgery resident and attending staff. Prevalence, frequency, purpose, and concerns regarding patient-related care text messaging. Overall, 73 (65%) surveyed physicians responded, including 45 resident (66%) and 28 attending surgeons (62%). All respondents owned a texting device. Majority of surgery residents (88%) and attendings (71%) texted residents, whereas only 59% of residents and 65% of attendings texted other faculty. Most resident to resident text occurred at a frequency of 3-5 times/d (43%) compared with most attending to resident texts, which occurred 1-2 times/d (33%). Most resident to attending (25%) and attending to attending (30%) texts occurred 1-2 times/d. Among those that texted, PRC was the most frequently reported purpose for resident to resident (46%), resident to attending (64%), attending to resident (82%), and attending to other attending staff (60%) texting. Texting was the most preferred method to communicate about routine PRC (47% of residents vs 44% of attendings). Age (OR: 0.86, 95% CI: 0.79-0.95; p = 0.003), but not sex, specialty/clinical rotation, academic rank, or postgraduate year (PGY) level predicted PRC texting. Most resident and attending staff surveyed utilize texting, mostly for PRC. Texting was preferred for communicating routine PRC information. Our data may facilitate the development of guidelines for the appropriate use of PRC texting. Copyright © 2013 Association of Program Directors in Surgery. Published by Elsevier Inc. All rights reserved.

  1. Optimal source coding, removable noise elimination, and natural coordinate system construction for general vector sources using replicator neural networks

    Science.gov (United States)

    Hecht-Nielsen, Robert

    1997-04-01

    A new universal one-chart smooth manifold model for vector information sources is introduced. Natural coordinates (a particular type of chart) for such data manifolds are then defined. Uniformly quantized natural coordinates form an optimal vector quantization code for a general vector source. Replicator neural networks (a specialized type of multilayer perceptron with three hidden layers) are the introduced. As properly configured examples of replicator networks approach minimum mean squared error (e.g., via training and architecture adjustment using randomly chosen vectors from the source), these networks automatically develop a mapping which, in the limit, produces natural coordinates for arbitrary source vectors. The new concept of removable noise (a noise model applicable to a wide variety of real-world noise processes) is then discussed. Replicator neural networks, when configured to approach minimum mean squared reconstruction error (e.g., via training and architecture adjustment on randomly chosen examples from a vector source, each with randomly chosen additive removable noise contamination), in the limit eliminate removable noise and produce natural coordinates for the data vector portions of the noise-corrupted source vectors. Consideration regarding selection of the dimension of a data manifold source model and the training/configuration of replicator neural networks are discussed.

  2. Patient Dashboard: the use of a colour-coded computerised clinical reminder in Whanganui regional general practices.

    Science.gov (United States)

    McMenamin, John; Nicholson, Rick; Leech, Ken

    2011-12-01

    Clinical reminders have been shown to help general practice achieve an increase in some preventive care items, especially if they identify a patient's eligibility for the target item, prompt clinicians at the right time, provide a fast link to management tools and facilitate clinical recording. WRPHO has introduced the Patient Dashboard clinical reminder and monitored its impact on health targets. This paper reports the impact of a computerised colour-coded clinical reminder on achieving agreed health targets in Whanganui regional practices. Patient Dashboard was developed from previous versions in Auckland and Northland and provided to Whanganui regional practices with Primary Health Organisation (PHO) support. The Dashboard was linked with existing and new clinical management tools which automatically updated clinical records. Data from practices was pooled by Whanganui Regional Primary Health Organisation and target achievement rates reported over 15 months. Over the initial 15 months of Patient Dashboard use, recording of smoking status increased from 74% to 82% and of alcohol use from 15% to 47%. Screening for diabetes increased from 62% to 74%, cardiovascular risk assessment from 20% to 43%, cervical screening from 71% to 79%, and breast screening from 60% to 80%. Patient Dashboard was associated with increased performance indicators both for those targets which were part of a PHO programme and for targets without additional support.

  3. A study of longwave radiation codes for climate studies: Validation with ARM observations and tests in general circulation models

    International Nuclear Information System (INIS)

    Ellingson, R.G.; Baer, F.

    1993-01-01

    This report summarizes the activities of our group to meet our stated objectives. The report is divided into sections entitled: Radiation Model Testing Activities, General Circulation Model Testing Activities, Science Team Activities, and Publications, Presentations and Meetings. The section on Science Team Activities summarizes our participation with the science team to further advance the observation and modeling programs. Appendix A lists graduate students supported, and post-doctoral appointments during the project. Reports on the activities during each of the first two years are included as Appendix B. Significant progress has been made in: determining the ability of line-by-line radiation models to calculate the downward longwave flux at the surface; determining the uncertainties in calculated the downwelling radiance and flux at the surface associated with the use of different proposed profiling techniques; intercomparing clear-sky radiance and flux observations with calculations from radiation codes from different climate models; determining the uncertainties associated with estimating N* from surface longwave flux observations; and determining the sensitivity of model calculations to different formulations of the effects of finite sized clouds

  4. General Purpose Force Capability; the Challenge of Versatility and Achieving Balance Along the Widest Possible Spectrum of Conflict

    Science.gov (United States)

    2010-04-01

    STRATEGY 16 Ir 16 Balance 17 The Imbalance between Traditional and Irregul 17 Finding the Proper Balance 18 CHAPTER 3 - DEFINING 20 Introduction 20...wider struggle for control and support of the contested c ir warfare as well as W 41      4. IRREGULAR WARFARE – HISTORICAL CONTEXT AND CURRENT...Congressional Research Service, Washington, D.C., July 20, 2009. R 1984. homas E. The Gamble: General David Petraeus and the American Military

  5. "I Love Knowing That What I'm Doing Has Purpose": Male Instrumentalists Who Choose to Teach Elementary General Music

    Science.gov (United States)

    Shouldice, Heather Nelson

    2017-01-01

    Few entering music education students plan to teach in an elementary general music (EGM) setting, and the majority of those who teach EGM are female and have vocalist backgrounds. The purpose of this collective case study was to examine six male instrumentalists' decisions to pursue a career in EGM teaching rather than secondary instrumental…

  6. Comparison of General Purpose Heat Source testing with the ANSI N43.6-1977 (R 1989) sealed source standard

    International Nuclear Information System (INIS)

    Grigsby, C.O.

    1998-01-01

    This analysis provides a comparison of the testing of Radioisotope Thermoelectric Generators (RTGs) and RTG components with the testing requirements of ANSI N43.6-1977 (R1989) ''Sealed Radioactive Sources, Categorization''. The purpose of this comparison is to demonstrate that the RTGs meet or exceed the requirements of the ANSI standard, and thus can be excluded from the radioactive inventory of the Chemistry and Metallurgy Research (CMR) building in Los Alamos per Attachment 1 of DOE STD 1027-92. The approach used in this analysis is as follows: (1) describe the ANSI sealed source classification methodology; (2) develop sealed source performance requirements for the RTG and/or RTG components based on criteria from the accident analysis for CMR; (3) compare the existing RTG or RTG component test data to the CMR requirements; and (4) determine the appropriate ANSI classification for the RTG and/or RTG components based on CMR performance requirements. The CMR requirements for treating RTGs as sealed sources are derived from the radiotoxicity of the isotope ( 238 P7) and amount (13 kg) of radioactive material contained in the RTG. The accident analysis for the CMR BIO identifies the bounding accidents as wing-wide fire, explosion and earthquake. These accident scenarios set the requirements for RTGs or RTG components stored within the CMR

  7. KENO-V code

    International Nuclear Information System (INIS)

    Cramer, S.N.

    1984-01-01

    The KENO-V code is the current release of the Oak Ridge multigroup Monte Carlo criticality code development. The original KENO, with 16 group Hansen-Roach cross sections and P 1 scattering, was one ot the first multigroup Monte Carlo codes and it and its successors have always been a much-used research tool for criticality studies. KENO-V is able to accept large neutron cross section libraries (a 218 group set is distributed with the code) and has a general P/sub N/ scattering capability. A supergroup feature allows execution of large problems on small computers, but at the expense of increased calculation time and system input/output operations. This supergroup feature is activated automatically by the code in a manner which utilizes as much computer memory as is available. The primary purpose of KENO-V is to calculate the system k/sub eff/, from small bare critical assemblies to large reflected arrays of differing fissile and moderator elements. In this respect KENO-V neither has nor requires the many options and sophisticated biasing techniques of general Monte Carlo codes

  8. On the Representation of Aquifer Compressibility in General Subsurface Flow Codes: How an Alternate Definition of Aquifer Compressibility Matches Results from the Groundwater Flow Equation

    Science.gov (United States)

    Birdsell, D.; Karra, S.; Rajaram, H.

    2017-12-01

    The governing equations for subsurface flow codes in deformable porous media are derived from the fluid mass balance equation. One class of these codes, which we call general subsurface flow (GSF) codes, does not explicitly track the motion of the solid porous media but does accept general constitutive relations for porosity, density, and fluid flux. Examples of GSF codes include PFLOTRAN, FEHM, STOMP, and TOUGH2. Meanwhile, analytical and numerical solutions based on the groundwater flow equation have assumed forms for porosity, density, and fluid flux. We review the derivation of the groundwater flow equation, which uses the form of Darcy's equation that accounts for the velocity of fluids with respect to solids and defines the soil matrix compressibility accordingly. We then show how GSF codes have a different governing equation if they use the form of Darcy's equation that is written only in terms of fluid velocity. The difference is seen in the porosity change, which is part of the specific storage term in the groundwater flow equation. We propose an alternative definition of soil matrix compressibility to correct for the untracked solid velocity. Simulation results show significantly less error for our new compressibility definition than the traditional compressibility when compared to analytical solutions from the groundwater literature. For example, the error in one calculation for a pumped sandstone aquifer goes from 940 to <70 Pa when the new compressibility is used. Code users and developers need to be aware of assumptions in the governing equations and constitutive relations in subsurface flow codes, and our newly-proposed compressibility function should be incorporated into GSF codes.

  9. ANDOSE: a computer code for calculating annual doses to man from routine releases of LWR effluents for the purpose of evaluating compliance with JAEC's guide for doseobjectives

    International Nuclear Information System (INIS)

    Iijima, Toshinori; Shiraishi, Tadao

    1979-10-01

    For environmental doses from routine releases of LWRs effluents to meet the Criterion 'As Low As is Practicable (ALAP)', Japan Atomic Energy Commission (JAEC) established a series of guides, the first for 'Dose Objectives' (May 1975), the second for models and parameters for calculating the environmental doses to compare with the 'Dose Objectives' (September 1976), and the third providing onsite meteorological programs, statistics of the data obtained and atmospheric dispersion models (June 1977). JAERI has developed a computer code, designated as ANDOSE, for calculating annual releases of radioactive gaseous and liquid effluents and, then, total body doses and thyroid doses to individuals around sites on the basis of these guides. The total body doses are from radioactive noble gases as well as from radioactive materials taken with marine food. For the calculation of thyroid doses are taken into account exposure pathways via inhalation and ingestion of leafy vegetables, cow's milk and marine food. The age-specific thyroid doses are evaluated. The doses are summed up when multisource or multisite conditions need to be evaluated (Nuclear Safety Bureau's requirement). In the present report, are described source-term models, environmental transport models and dose models used in the code, of which most are provided in the guides but some are complemented by the authors, the functions of ANDOSE and the manual for users of the code. The program lists and the latter two guides mentioned above are included in the appendices. (author)

  10. A new multi-dimensional general relativistic neutrino hydrodynamics code for core-collapse supernovae. IV. The neutrino signal

    Energy Technology Data Exchange (ETDEWEB)

    Müller, Bernhard [Monash Center for Astrophysics, School of Mathematical Sciences, Building 28, Monash University, Victoria 3800 (Australia); Janka, Hans-Thomas, E-mail: bernhard.mueller@monash.edu, E-mail: bjmuellr@mpa-garching.mpg.de, E-mail: thj@mpa-garching.mpg.de [Max-Planck-Institut für Astrophysik, Karl-Schwarzschild-Str. 1, D-85748 Garching (Germany)

    2014-06-10

    Considering six general relativistic, two-dimensional (2D) supernova (SN) explosion models of progenitor stars between 8.1 and 27 M {sub ☉}, we systematically analyze the properties of the neutrino emission from core collapse and bounce to the post-explosion phase. The models were computed with the VERTEX-COCONUT code, using three-flavor, energy-dependent neutrino transport in the ray-by-ray-plus approximation. Our results confirm the close similarity of the mean energies, (E), of ν-bar {sub e} and heavy-lepton neutrinos and even their crossing during the accretion phase for stars with M ≳ 10 M {sub ☉} as observed in previous 1D and 2D simulations with state-of-the-art neutrino transport. We establish a roughly linear scaling of 〈E{sub ν-bar{sub e}}〉 with the proto-neutron star (PNS) mass, which holds in time as well as for different progenitors. Convection inside the PNS affects the neutrino emission on the 10%-20% level, and accretion continuing beyond the onset of the explosion prevents the abrupt drop of the neutrino luminosities seen in artificially exploded 1D models. We demonstrate that a wavelet-based time-frequency analysis of SN neutrino signals in IceCube will offer sensitive diagnostics for the SN core dynamics up to at least ∼10 kpc distance. Strong, narrow-band signal modulations indicate quasi-periodic shock sloshing motions due to the standing accretion shock instability (SASI), and the frequency evolution of such 'SASI neutrino chirps' reveals shock expansion or contraction. The onset of the explosion is accompanied by a shift of the modulation frequency below 40-50 Hz, and post-explosion, episodic accretion downflows will be signaled by activity intervals stretching over an extended frequency range in the wavelet spectrogram.

  11. Call for consistent coding in diabetes mellitus using the Royal College of General Practitioners and NHS pragmatic classification of diabetes

    Directory of Open Access Journals (Sweden)

    Simon de Lusignan

    2013-03-01

    Full Text Available Background The prevalence of diabetes is increasing with growing levels of obesity and an aging population. New practical guidelines for diabetes provide an applicable classification. Inconsistent coding of diabetes hampers the use of computerised disease registers for quality improvement, and limits the monitoring of disease trends.Objective To develop a consensus set of codes that should be used when recording diabetes diagnostic data.Methods The consensus approach was hierarchical, with a preference for diagnostic/disorder codes, to define each type of diabetes and non-diabetic hyperglycaemia, which were listed as being completely, partially or not readily mapped to available codes. The practical classification divides diabetes into type 1 (T1DM, type 2 (T2DM, genetic, other, unclassified and non-diabetic fasting hyperglycaemia. We mapped the classification to Read version 2, Clinical Terms version 3 and SNOMED CT.Results T1DMand T2DM were completely mapped to appropriate codes. However, in other areas only partial mapping is possible. Genetics is a fast-moving field and there were considerable gaps in the available labels for genetic conditions; what the classification calls ‘other’ the coding system labels ‘secondary’ diabetes. The biggest gap was the lack of a code for diabetes where the type of diabetes was uncertain. Notwithstanding these limitations we were able to develop a consensus list.Conclusions It is a challenge to develop codes that readily map to contemporary clinical concepts. However, clinicians should adopt the standard recommended codes; and audit the quality of their existing records.

  12. Area-based cell colony surviving fraction evaluation: A novel fully automatic approach using general-purpose acquisition hardware.

    Science.gov (United States)

    Militello, Carmelo; Rundo, Leonardo; Conti, Vincenzo; Minafra, Luigi; Cammarata, Francesco Paolo; Mauri, Giancarlo; Gilardi, Maria Carla; Porcino, Nunziatina

    2017-10-01

    The current methodology for the Surviving Fraction (SF) measurement in clonogenic assay, which is a technique to study the anti-proliferative effect of treatments on cell cultures, involves manual counting of cell colony forming units. This procedure is operator-dependent and error-prone. Moreover, the identification of the exact colony number is often not feasible due to the high growth rate leading to the adjacent colony merging. As a matter of fact, conventional assessment does not deal with the colony size, which is generally correlated with the delivered radiation dose or the administered cytotoxic agent. Considering that the Area Covered by Colony (ACC) is proportional to the colony number and size as well as to the growth rate, we propose a novel fully automatic approach exploiting Circle Hough Transform, to automatically detect the wells in the plate, and local adaptive thresholding, which calculates the percentage of ACC for the SF quantification. This measurement relies just on this covering percentage and does not consider the colony number, preventing inconsistencies due to intra- and inter-operator variability. To evaluate the accuracy of the proposed approach, we compared the SFs obtained by our automatic ACC-based method against the conventional counting procedure. The achieved results (r = 0.9791 and r = 0.9682 on MCF7 and MCF10A cells, respectively) showed values highly correlated with the measurements using the traditional approach based on colony number alone. The proposed computer-assisted methodology could be integrated in laboratory practice as an expert system for the SF evaluation in clonogenic assays. Copyright © 2017 Elsevier Ltd. All rights reserved.

  13. To develop a dynamic model of a collector loop for purpose of improved control of solar heating and cooling. Final technical report. [TRNSYS code

    Energy Technology Data Exchange (ETDEWEB)

    Herczfeld, P R; Fischl, R

    1980-01-01

    The program objectives were to (1) assess the feasibility of using the TRNSYS computer code for solar heating and cooling control studies and modify it wherever possible, and (2) develop a new dynamic model of the solar collector which reflects the performance of the collector under transient conditions. Also, the sensitivity of the performance of this model to the various system parameters such as collector time constants, flow rates, turn-on and turn-off temperature set points, solar insolation, etc., was studied. Results are presented and discussed. (WHK)

  14. Summary of computational support and general documentation for computer code (GENTREE) used in Office of Nuclear Waste Isolation Pilot Salt Site Selection Project

    International Nuclear Information System (INIS)

    Beatty, J.A.; Younker, J.L.; Rousseau, W.F.; Elayat, H.A.

    1983-01-01

    A Decision Tree Computer Model was adapted for the purposes of a Pilot Salt Site Selection Project conducted by the Office of Nuclear Waste Isolation (ONWI). A deterministic computer model was developed to structure the site selection problem with submodels reflecting the five major outcome categories (Cost, Safety, Delay, Environment, Community Impact) to be evaluated in the decision process. Time-saving modifications were made in the tree code as part of the effort. In addition, format changes allowed retention of information items which are valuable in directing future research and in isolation of key variabilities in the Site Selection Decision Model. The deterministic code was linked to the modified tree code and the entire program was transferred to the ONWI-VAX computer for future use by the ONWI project

  15. Bilingual Language Control and General Purpose Cognitive Control among Individuals with Bilingual Aphasia: Evidence Based on Negative Priming and Flanker Tasks

    Science.gov (United States)

    Dash, Tanya; Kar, Bhoomika R.

    2014-01-01

    Background. Bilingualism results in an added advantage with respect to cognitive control. The interaction between bilingual language control and general purpose cognitive control systems can also be understood by studying executive control among individuals with bilingual aphasia. Objectives. The current study examined the subcomponents of cognitive control in bilingual aphasia. A case study approach was used to investigate whether cognitive control and language control are two separate systems and how factors related to bilingualism interact with control processes. Methods. Four individuals with bilingual aphasia performed a language background questionnaire, picture description task, and two experimental tasks (nonlinguistic negative priming task and linguistic and nonlinguistic versions of flanker task). Results. A descriptive approach was used to analyse the data using reaction time and accuracy measures. The cumulative distribution function plots were used to visualize the variations in performance across conditions. The results highlight the distinction between general purpose cognitive control and bilingual language control mechanisms. Conclusion. All participants showed predominant use of the reactive control mechanism to compensate for the limited resources system. Independent yet interactive systems for bilingual language control and general purpose cognitive control were postulated based on the experimental data derived from individuals with bilingual aphasia. PMID:24982591

  16. Bilingual language control and general purpose cognitive control among individuals with bilingual aphasia: evidence based on negative priming and flanker tasks.

    Science.gov (United States)

    Dash, Tanya; Kar, Bhoomika R

    2014-01-01

    Bilingualism results in an added advantage with respect to cognitive control. The interaction between bilingual language control and general purpose cognitive control systems can also be understood by studying executive control among individuals with bilingual aphasia. objectives: The current study examined the subcomponents of cognitive control in bilingual aphasia. A case study approach was used to investigate whether cognitive control and language control are two separate systems and how factors related to bilingualism interact with control processes. Four individuals with bilingual aphasia performed a language background questionnaire, picture description task, and two experimental tasks (nonlinguistic negative priming task and linguistic and nonlinguistic versions of flanker task). A descriptive approach was used to analyse the data using reaction time and accuracy measures. The cumulative distribution function plots were used to visualize the variations in performance across conditions. The results highlight the distinction between general purpose cognitive control and bilingual language control mechanisms. All participants showed predominant use of the reactive control mechanism to compensate for the limited resources system. Independent yet interactive systems for bilingual language control and general purpose cognitive control were postulated based on the experimental data derived from individuals with bilingual aphasia.

  17. Tree Coding of Bilevel Images

    DEFF Research Database (Denmark)

    Martins, Bo; Forchhammer, Søren

    1998-01-01

    Presently, sequential tree coders are the best general purpose bilevel image coders and the best coders of halftoned images. The current ISO standard, Joint Bilevel Image Experts Group (JBIG), is a good example. A sequential tree coder encodes the data by feeding estimates of conditional...... is one order of magnitude slower than JBIG, obtains excellent and highly robust compression performance. A multipass free tree coding scheme produces superior compression results for all test images. A multipass free template coding scheme produces significantly better results than JBIG for difficult...... images such as halftones. By utilizing randomized subsampling in the template selection, the speed becomes acceptable for practical image coding...

  18. Pelletron general purpose scattering chamber

    International Nuclear Information System (INIS)

    Chatterjee, A.; Kailas, S.; Kerekette, S.S.; Navin, A.; Kumar, Suresh

    1993-01-01

    A medium sized stainless steel scattering chamber has been constructed for nuclear scattering and reaction experiments at the 14UD pelletron accelerator facility. It has been so designed that several types of detectors, varying from small sized silicon surface barrier detectors to medium sized gas detectors and NaI detectors can be conveniently positioned inside the chamber for detection of charged particles. The chamber has been planned to perform the following types of experiments : angular distributions of elastically scattered particles, fission fragments and other charged particles, angular correlations for charged particles e.g. protons, alphas and fission fragments. (author). 2 figs

  19. General purpose nuclear irradiation chamber

    International Nuclear Information System (INIS)

    Nurul Fadzlin Hasbullah; Nuurul Iffah Che Omar; Nahrul Khair Alang Md Rashid; Jaafar Abdullah

    2013-01-01

    Nuclear technology has found a great need for use in medicine, industry, and research. Smoke detectors in our homes, medical treatments and new varieties of plants by irradiating its seeds are just a few examples of the benefits of nuclear technology. Portable neutron source such as Californium-252, available at Industrial Technology Division (BTI/ PAT), Malaysian Nuclear Agency, has a 2.645 year half-life. However, 252 Cf is known to emit gamma radiation from the source. Thus, this chamber aims to provide a proper gamma shielding for samples to distinguish the use of mixed neutron with gamma-rays or pure neutron radiation. The chamber is compatible to be used with other portable neutron sources such as 241 Am-Be as well as the reactor TRIGA PUSPATI for higher neutron dose. This chamber was designed through a collaborative effort of Kulliyyah Engineering, IIUM with the Industrial Technology Division (BTI) team, Malaysian Nuclear Agency. (Author)

  20. Validity of silhouette showcards as a measure of body size and obesity in a population in the African region: A practical research tool for general-purpose surveys.

    Science.gov (United States)

    Yepes, Maryam; Viswanathan, Barathi; Bovet, Pascal; Maurer, Jürgen

    2015-01-01

    The purpose of this study is to validate the Pulvers silhouette showcard as a measure of weight status in a population in the African region. This tool is particularly beneficial when scarce resources do not allow for direct anthropometric measurements due to limited survey time or lack of measurement technology in face-to-face general-purpose surveys or in mailed, online, or mobile device-based surveys. A cross-sectional study was conducted in the Republic of Seychelles with a sample of 1240 adults. We compared self-reported body sizes measured by Pulvers' silhouette showcards to four measurements of body size and adiposity: body mass index (BMI), body fat percent measured, waist circumference, and waist to height ratio. The accuracy of silhouettes as an obesity indicator was examined using sex-specific receiver operator curve (ROC) analysis and the reliability of this tool to detect socioeconomic gradients in obesity was compared to BMI-based measurements. Our study supports silhouette body size showcards as a valid and reliable survey tool to measure self-reported body size and adiposity in an African population. The mean correlation coefficients of self-reported silhouettes with measured BMI were 0.80 in men and 0.81 in women (P general-purpose surveys of obesity in social sciences, where limited resources do not allow for direct anthropometric measurements.

  1. Development of a General-Purpose Analysis System Based on a Programmable Fluid Processor Final Report CRADA No. TC-2027-01

    Energy Technology Data Exchange (ETDEWEB)

    McConaghy, C. F. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Gascoyne, P. R. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2017-09-28

    The purpose ofthis project was to develop a general-purpose analysis system based on a programmable fluid processor (PFP). The PFP is an array of electrodes surrounded by fluid reservoirs and injectors. Injected droplets of various reagents are manjpulated and combined on the array by Dielectrophoretic (DEP) forces. The goal was to create a small handheld device that could accomplish the tasks currently undertaken by much larger, time consuming, manual manipulation in the lab. The entire effo1t was funded by DARPA under the Bio-Flips program. MD Anderson Cancer Center was the PI for the DARPA effort. The Bio-Flips program was a 3- year program that ran from September 2000 to September 2003. The CRADA was somewhat behind the Bi-Flips program running from June 2001 to June 2004 with a no cost extension to September 2004.

  2. Performance of the Research Animal Holding Facility (RAHF) and General Purpose Work Station (GPWS) and other hardware in the microgravity environment

    Science.gov (United States)

    Hogan, Robert P.; Dalton, Bonnie P.

    1991-01-01

    This paper discusses the performance of the Research Animal Holding Facility (RAHF) and General Purpose Work Station (GPWS) plus other associated hardware during the recent flight of Spacelab Life Sciences 1 (SLS-1). The RAHF was developed to provide proper housing (food, water, temperature control, lighting and waste management) for up to 24 rodents during flights on the Spacelab. The GPWS was designed to contain particulates and toxic chemicals generated during plant and animal handling and dissection/fixation activities during space flights. A history of the hardware development involves as well as the redesign activities prior to the actual flight are discussed.

  3. A boundary-Fitted Coordinate Code for General Two-Dimensional Regions with Obstacles and Boundary Intrusions.

    Science.gov (United States)

    1983-03-01

    values of these functions on the two sides of the slits. The acceleration parameters for the iteration at each point are in the field array WACC (I,J...code will calculate a locally optimum value at each point in the field, these values being placed in the field array WACC . This calculation is...changes in x and y, are calculated by calling subroutine ERROR.) The acceleration parameter is placed in the field 65 array WACC . The addition to the

  4. A Monte Carlo burnup code linking MCNP and REBUS

    International Nuclear Information System (INIS)

    Hanan, N.A.; Olson, A.P.; Pond, R.B.; Matos, J.E.

    1998-01-01

    The REBUS-3 burnup code, used in the anl RERTR Program, is a very general code that uses diffusion theory (DIF3D) to obtain the fluxes required for reactor burnup analyses. Diffusion theory works well for most reactors. However, to include the effects of exact geometry and strong absorbers that are difficult to model using diffusion theory, a Monte Carlo method is required. MCNP, a general-purpose, generalized-geometry, time-dependent, Monte Carlo transport code, is the most widely used Monte Carlo code. This paper presents a linking of the MCNP code and the REBUS burnup code to perform these difficult analyses. The linked code will permit the use of the full capabilities of REBUS which include non-equilibrium and equilibrium burnup analyses. Results of burnup analyses using this new linked code are also presented. (author)

  5. A Monte Carlo burnup code linking MCNP and REBUS

    International Nuclear Information System (INIS)

    Hanan, N. A.

    1998-01-01

    The REBUS-3 burnup code, used in the ANL RERTR Program, is a very general code that uses diffusion theory (DIF3D) to obtain the fluxes required for reactor burnup analyses. Diffusion theory works well for most reactors. However, to include the effects of exact geometry and strong absorbers that are difficult to model using diffusion theory, a Monte Carlo method is required. MCNP, a general-purpose, generalized-geometry, time-dependent, Monte Carlo transport code, is the most widely used Monte Carlo code. This paper presents a linking of the MCNP code and the REBUS burnup code to perform these difficult burnup analyses. The linked code will permit the use of the full capabilities of REBUS which include non-equilibrium and equilibrium burnup analyses. Results of burnup analyses using this new linked code are also presented

  6. The time-resolved and extreme conditions XAS (TEXAS) facility at the European Synchrotron Radiation Facility: the general-purpose EXAFS bending-magnet beamline BM23

    Energy Technology Data Exchange (ETDEWEB)

    Mathon, O., E-mail: mathon@esrf.fr; Beteva, A.; Borrel, J.; Bugnazet, D.; Gatla, S.; Hino, R.; Kantor, I.; Mairs, T. [European Synchrotron Radiation Facility, CS 40220, 38043 Grenoble Cedex 9 (France); Munoz, M. [European Synchrotron Radiation Facility, CS 40220, 38043 Grenoble Cedex 9 (France); Université Joseph Fourier, 1381 rue de la Piscine, BP 53, 38041 Grenoble Cedex 9 (France); Pasternak, S.; Perrin, F.; Pascarelli, S. [European Synchrotron Radiation Facility, CS 40220, 38043 Grenoble Cedex 9 (France)

    2015-10-17

    BM23 is the general-purpose EXAFS bending-magnet beamline at the ESRF, replacing the former BM29 beamline in the framework of the ESRF upgrade. Its mission is to serve the whole XAS user community by providing access to a basic service in addition to the many specialized instruments available at the ESRF. BM23 offers high-signal-to-noise ratio EXAFS in a large energy range (5–75 keV), continuous energy scanning for quick-EXAFS on the second timescale and a micro-XAS station delivering a spot size of 4 µm × 4 µm FWHM. BM23 is the general-purpose EXAFS bending-magnet beamline at the ESRF, replacing the former BM29 beamline in the framework of the ESRF upgrade. Its mission is to serve the whole XAS user community by providing access to a basic service in addition to the many specialized instruments available at the ESRF. BM23 offers high signal-to-noise ratio EXAFS in a large energy range (5–75 keV), continuous energy scanning for quick-EXAFS on the second timescale and a micro-XAS station delivering a spot size of 4 µm × 4 µm FWHM. It is a user-friendly facility featuring a high degree of automation, online EXAFS data reduction and a flexible sample environment.

  7. DIDACTIC PRINCIPLES AND PSYCHOLOGICAL CHARACTERISTICS IN DEFINITION OF QUALITY OF SOFTWARE TOOLS FOR EDUCATIONAL PURPOSE IN THE GENERAL EDUCATIONAL ENVIRONMENT OF UKRAINE

    Directory of Open Access Journals (Sweden)

    Maryna V. Pirko

    2011-02-01

    Full Text Available The fundamental feature of economy of postindustrial society is the knowledge that represents the basic source of competitive advantage. In the article the circle of didactic, psychological indicators in researches of problems of achievement of a high degree of quality of education and educational services is considered and described. The attention is paid to pedagogical requirements of the given period which are a standard substantiation in orientations for quality estimation of software tools for educational purpose of the general educational environment in Ukraine. The scheme of internal model of maintenance of quality of software tools for educational purpose is considered, the aspects integrated by internal model of quality of software for educational purpose are listed. The article describes the directions of researches in the conditions of formation of the global international educational environment and uniform information space of  education system taking into account the growth of availability of educational services. It is specified the main principles in the organization of pedagogical software tools.

  8. ABCXYZ: vector potential (A) and magnetic field (B) code (C) for Cartesian (XYZ) geometry using general current elements

    International Nuclear Information System (INIS)

    Anderson, D.V.; Breazeal, J.; Finan, C.H.; Johnston, B.M.

    1976-01-01

    ABCXYZ is a computer code for obtaining the Cartesian components of the vector potential and the magnetic field on an observed grid from an arrangement of current-carrying wires. Arbitrary combinations of straight line segments, arcs, and loops are allowed in the specification of the currents. Arbitrary positions and orientations of the current-carrying elements are also allowed. Specification of the wire diameter permits the computation of well-defined fields, even in the interiors of the conductors. An optical feature generates magnetic field lines. Extensive graphical and printed output is available to the user including contour, grid-line, and field-line plots. 12 figures, 1 table

  9. New code of conduct

    CERN Multimedia

    Laëtitia Pedroso

    2010-01-01

    During his talk to the staff at the beginning of the year, the Director-General mentioned that a new code of conduct was being drawn up. What exactly is it and what is its purpose? Anne-Sylvie Catherin, Head of the Human Resources (HR) Department, talked to us about the whys and wherefores of the project.   Drawing by Georges Boixader from the cartoon strip “The World of Particles” by Brian Southworth. A code of conduct is a general framework laying down the behaviour expected of all members of an organisation's personnel. “CERN is one of the very few international organisations that don’t yet have one", explains Anne-Sylvie Catherin. “We have been thinking about introducing a code of conduct for a long time but lacked the necessary resources until now”. The call for a code of conduct has come from different sources within the Laboratory. “The Equal Opportunities Advisory Panel (read also the "Equal opportuni...

  10. Uplink Coding

    Science.gov (United States)

    Andrews, Ken; Divsalar, Dariush; Dolinar, Sam; Moision, Bruce; Hamkins, Jon; Pollara, Fabrizio

    2007-01-01

    This slide presentation reviews the objectives, meeting goals and overall NASA goals for the NASA Data Standards Working Group. The presentation includes information on the technical progress surrounding the objective, short LDPC codes, and the general results on the Pu-Pw tradeoff.

  11. Network Coding

    Indian Academy of Sciences (India)

    Home; Journals; Resonance – Journal of Science Education; Volume 15; Issue 7. Network Coding. K V Rashmi Nihar B Shah P Vijay Kumar. General Article Volume 15 Issue 7 July 2010 pp 604-621. Fulltext. Click here to view fulltext PDF. Permanent link: https://www.ias.ac.in/article/fulltext/reso/015/07/0604-0621 ...

  12. Expander Codes

    Indian Academy of Sciences (India)

    Home; Journals; Resonance – Journal of Science Education; Volume 10; Issue 1. Expander Codes - The Sipser–Spielman Construction. Priti Shankar. General Article Volume 10 ... Author Affiliations. Priti Shankar1. Department of Computer Science and Automation, Indian Institute of Science Bangalore 560 012, India.

  13. Development, dissemination, and applications of a new terminological resource, the Q-Code taxonomy for professional aspects of general practice/family medicine.

    Science.gov (United States)

    Jamoulle, Marc; Resnick, Melissa; Grosjean, Julien; Ittoo, Ashwin; Cardillo, Elena; Vander Stichele, Robert; Darmoni, Stefan; Vanmeerbeek, Marc

    2018-12-01

    While documentation of clinical aspects of General Practice/Family Medicine (GP/FM) is assured by the International Classification of Primary Care (ICPC), there is no taxonomy for the professional aspects (context and management) of GP/FM. To present the development, dissemination, applications, and resulting face validity of the Q-Codes taxonomy specifically designed to describe contextual features of GP/FM, proposed as an extension to the ICPC. The Q-Codes taxonomy was developed from Lamberts' seminal idea for indexing contextual content (1987) by a multi-disciplinary team of knowledge engineers, linguists and general practitioners, through a qualitative and iterative analysis of 1702 abstracts from six GP/FM conferences using Atlas.ti software. A total of 182 concepts, called Q-Codes, representing professional aspects of GP/FM were identified and organized in a taxonomy. Dissemination: The taxonomy is published as an online terminological resource, using semantic web techniques and web ontology language (OWL) ( http://www.hetop.eu/Q ). Each Q-Code is identified with a unique resource identifier (URI), and provided with preferred terms, and scope notes in ten languages (Portuguese, Spanish, English, French, Dutch, Korean, Vietnamese, Turkish, Georgian, German) and search filters for MEDLINE and web searches. This taxonomy has already been used to support queries in bibliographic databases (e.g., MEDLINE), to facilitate indexing of grey literature in GP/FM as congress abstracts, master theses, websites and as an educational tool in vocational teaching, Conclusions: The rapidly growing list of practical applications provides face-validity for the usefulness of this freely available new terminological resource.

  14. The time-resolved and extreme conditions XAS (TEXAS) facility at the European Synchrotron Radiation Facility: the general-purpose EXAFS bending-magnet beamline BM23.

    Science.gov (United States)

    Mathon, O; Beteva, A; Borrel, J; Bugnazet, D; Gatla, S; Hino, R; Kantor, I; Mairs, T; Munoz, M; Pasternak, S; Perrin, F; Pascarelli, S

    2015-11-01

    BM23 is the general-purpose EXAFS bending-magnet beamline at the ESRF, replacing the former BM29 beamline in the framework of the ESRF upgrade. Its mission is to serve the whole XAS user community by providing access to a basic service in addition to the many specialized instruments available at the ESRF. BM23 offers high signal-to-noise ratio EXAFS in a large energy range (5-75 keV), continuous energy scanning for quick-EXAFS on the second timescale and a micro-XAS station delivering a spot size of 4 µm × 4 µm FWHM. It is a user-friendly facility featuring a high degree of automation, online EXAFS data reduction and a flexible sample environment.

  15. The new versatile general purpose surface-muon instrument (GPS) based on silicon photomultipliers for μSR measurements on a continuous-wave beam.

    Science.gov (United States)

    Amato, A; Luetkens, H; Sedlak, K; Stoykov, A; Scheuermann, R; Elender, M; Raselli, A; Graf, D

    2017-09-01

    We report on the design and commissioning of a new spectrometer for muon-spin relaxation/rotation studies installed at the Swiss Muon Source (SμS) of the Paul Scherrer Institute (PSI, Switzerland). This new instrument is essentially a new design and replaces the old general-purpose surface-muon (GPS) instrument that has been for long the workhorse of the μSR user facility at PSI. By making use of muon and positron detectors made of plastic scintillators read out by silicon photomultipliers, a time resolution of the complete instrument of about 160 ps (standard deviation) could be achieved. In addition, the absence of light guides, which are needed in traditionally built μSR instrument to deliver the scintillation light to photomultiplier tubes located outside magnetic fields applied, allowed us to design a compact instrument with a detector set covering an increased solid angle compared with the old GPS.

  16. High Precision Thermal, Structural and Optical Analysis of an External Occulter Using a Common Model and the General Purpose Multi-Physics Analysis Tool Cielo

    Science.gov (United States)

    Hoff, Claus; Cady, Eric; Chainyk, Mike; Kissil, Andrew; Levine, Marie; Moore, Greg

    2011-01-01

    The efficient simulation of multidisciplinary thermo-opto-mechanical effects in precision deployable systems has for years been limited by numerical toolsets that do not necessarily share the same finite element basis, level of mesh discretization, data formats, or compute platforms. Cielo, a general purpose integrated modeling tool funded by the Jet Propulsion Laboratory and the Exoplanet Exploration Program, addresses shortcomings in the current state of the art via features that enable the use of a single, common model for thermal, structural and optical aberration analysis, producing results of greater accuracy, without the need for results interpolation or mapping. This paper will highlight some of these advances, and will demonstrate them within the context of detailed external occulter analyses, focusing on in-plane deformations of the petal edges for both steady-state and transient conditions, with subsequent optical performance metrics including intensity distributions at the pupil and image plane.

  17. LSENS: A General Chemical Kinetics and Sensitivity Analysis Code for homogeneous gas-phase reactions. Part 1: Theory and numerical solution procedures

    Science.gov (United States)

    Radhakrishnan, Krishnan

    1994-01-01

    LSENS, the Lewis General Chemical Kinetics and Sensitivity Analysis Code, has been developed for solving complex, homogeneous, gas-phase chemical kinetics problems and contains sensitivity analysis for a variety of problems, including nonisothermal situations. This report is part 1 of a series of three reference publications that describe LENS, provide a detailed guide to its usage, and present many example problems. Part 1 derives the governing equations and describes the numerical solution procedures for the types of problems that can be solved. The accuracy and efficiency of LSENS are examined by means of various test problems, and comparisons with other methods and codes are presented. LSENS is a flexible, convenient, accurate, and efficient solver for chemical reaction problems such as static system; steady, one-dimensional, inviscid flow; reaction behind incident shock wave, including boundary layer correction; and perfectly stirred (highly backmixed) reactor. In addition, the chemical equilibrium state can be computed for the following assigned states: temperature and pressure, enthalpy and pressure, temperature and volume, and internal energy and volume. For static problems the code computes the sensitivity coefficients of the dependent variables and their temporal derivatives with respect to the initial values of the dependent variables and/or the three rate coefficient parameters of the chemical reactions.

  18. General Electromagnetic Model for the Analysis of Complex Systems (GEMACS) Computer Code Documentation (Version 3). Volume 3. Part 2.

    Science.gov (United States)

    1983-09-01

    F.P. PX /AMPZIJ/ REFH /AMPZIJ/ REFV /AI4PZIJ/ * RHOX /AI4PZIJ/ RHOY /At4PZIJ/ RHOZ /AI4PZIJ/ S A-ZJ SA /AMPZIJ/ SALP /AMPZIJ/ 6. CALLING ROUTINE: FLDDRV...US3NG ALGORITHM 72 COMPUTE P- YES .~:*:.~~ USING* *. 1. NAME: PLAINT (GTD) ] 2. PURPOSE: To determine if a ray traveling from a given source loca...determine if a source ray reflection from plate MP occurs. If a ray traveling from the source image location in the reflected ray direction passes through

  19. Coded Network Function Virtualization

    DEFF Research Database (Denmark)

    Al-Shuwaili, A.; Simone, O.; Kliewer, J.

    2016-01-01

    Network function virtualization (NFV) prescribes the instantiation of network functions on general-purpose network devices, such as servers and switches. While yielding a more flexible and cost-effective network architecture, NFV is potentially limited by the fact that commercial off......-the-shelf hardware is less reliable than the dedicated network elements used in conventional cellular deployments. The typical solution for this problem is to duplicate network functions across geographically distributed hardware in order to ensure diversity. In contrast, this letter proposes to leverage channel...... coding in order to enhance the robustness on NFV to hardware failure. The proposed approach targets the network function of uplink channel decoding, and builds on the algebraic structure of the encoded data frames in order to perform in-network coding on the signals to be processed at different servers...

  20. Benchmarking and scaling studies of pseudospectral code Tarang ...

    Indian Academy of Sciences (India)

    Tarang is a general-purpose pseudospectral parallel code for simulating flows involving fluids, magnetohydrodynamics, and Rayleigh–Bénard convection in turbulence and instability regimes. In this paper we present code validation and benchmarking results of Tarang. We performed our simulations on 10243, 20483, and ...

  1. 48 CFR 1.105-1 - Publication and code arrangement.

    Science.gov (United States)

    2010-10-01

    ... 48 Federal Acquisition Regulations System 1 2010-10-01 2010-10-01 false Publication and code arrangement. 1.105-1 Section 1.105-1 Federal Acquisition Regulations System FEDERAL ACQUISITION REGULATION GENERAL FEDERAL ACQUISITION REGULATIONS SYSTEM Purpose, Authority, Issuance 1.105-1 Publication and code...

  2. Evaluation of general-purpose collimators against high-resolution collimators with resolution recovery with a view to reducing radiation dose in myocardial perfusion SPECT: A preliminary phantom study.

    Science.gov (United States)

    Armstrong, Ian S; Saint, Kimberley J; Tonge, Christine M; Arumugam, Parthiban

    2017-04-01

    There is a growing focus on reducing radiation dose to patients undergoing myocardial perfusion imaging. This preliminary phantom study aims to evaluate the use of general-purpose collimators with resolution recovery (RR) to allow a reduction in patient radiation dose. Images of a cardiac torso phantom with inferior and anterior wall defects were acquired on a GE Infinia and Siemens Symbia T6 using both high-resolution and general-purpose collimators. Imaging time, a surrogate for administered activity, was reduced between 35% and 40% with general-purpose collimators to match the counts acquired with high-resolution collimators. Images were reconstructed with RR with and without attenuation correction. Two pixel sizes were also investigated. Defect contrast was measured. Defect contrast on general-purpose images was superior or comparable to the high-resolution collimators on both systems despite the reduced imaging time. Infinia general-purpose images required a smaller pixel size to be used to maintain defect contrast, while Symbia T6 general-purpose images did not require a change in pixel size to that used for standard myocardial perfusion SPECT. This study suggests that general-purpose collimators with RR offer a potential for substantial dose reductions while providing similar or better image quality to images acquired using high-resolution collimators.

  3. Adapting machine learning techniques to censored time-to-event health record data: A general-purpose approach using inverse probability of censoring weighting.

    Science.gov (United States)

    Vock, David M; Wolfson, Julian; Bandyopadhyay, Sunayan; Adomavicius, Gediminas; Johnson, Paul E; Vazquez-Benitez, Gabriela; O'Connor, Patrick J

    2016-06-01

    Models for predicting the probability of experiencing various health outcomes or adverse events over a certain time frame (e.g., having a heart attack in the next 5years) based on individual patient characteristics are important tools for managing patient care. Electronic health data (EHD) are appealing sources of training data because they provide access to large amounts of rich individual-level data from present-day patient populations. However, because EHD are derived by extracting information from administrative and clinical databases, some fraction of subjects will not be under observation for the entire time frame over which one wants to make predictions; this loss to follow-up is often due to disenrollment from the health system. For subjects without complete follow-up, whether or not they experienced the adverse event is unknown, and in statistical terms the event time is said to be right-censored. Most machine learning approaches to the problem have been relatively ad hoc; for example, common approaches for handling observations in which the event status is unknown include (1) discarding those observations, (2) treating them as non-events, (3) splitting those observations into two observations: one where the event occurs and one where the event does not. In this paper, we present a general-purpose approach to account for right-censored outcomes using inverse probability of censoring weighting (IPCW). We illustrate how IPCW can easily be incorporated into a number of existing machine learning algorithms used to mine big health care data including Bayesian networks, k-nearest neighbors, decision trees, and generalized additive models. We then show that our approach leads to better calibrated predictions than the three ad hoc approaches when applied to predicting the 5-year risk of experiencing a cardiovascular adverse event, using EHD from a large U.S. Midwestern healthcare system. Copyright © 2016 Elsevier Inc. All rights reserved.

  4. ProtDCal: A program to compute general-purpose-numerical descriptors for sequences and 3D-structures of proteins.

    Science.gov (United States)

    Ruiz-Blanco, Yasser B; Paz, Waldo; Green, James; Marrero-Ponce, Yovani

    2015-05-16

    software is intended to provide a useful tool for general-purpose encoding of protein sequences and structures for applications is protein classification, similarity analyses and function prediction.

  5. Real-Time and Real-Fast Performance of General-Purpose and Real-Time Operating Systems in Multithreaded Physical Simulation of Complex Mechanical Systems

    Directory of Open Access Journals (Sweden)

    Carlos Garre

    2014-01-01

    Full Text Available Physical simulation is a valuable tool in many fields of engineering for the tasks of design, prototyping, and testing. General-purpose operating systems (GPOS are designed for real-fast tasks, such as offline simulation of complex physical models that should finish as soon as possible. Interfacing hardware at a given rate (as in a hardware-in-the-loop test requires instead maximizing time determinism, for which real-time operating systems (RTOS are designed. In this paper, real-fast and real-time performance of RTOS and GPOS are compared when simulating models of high complexity with large time steps. This type of applications is usually present in the automotive industry and requires a good trade-off between real-fast and real-time performance. The performance of an RTOS and a GPOS is compared by running a tire model scalable on the number of degrees-of-freedom and parallel threads. The benchmark shows that the GPOS present better performance in real-fast runs but worse in real-time due to nonexplicit task switches and to the latency associated with interprocess communication (IPC and task switch.

  6. Towards anatomic scale agent-based modeling with a massively parallel spatially explicit general-purpose model of enteric tissue (SEGMEnT_HPC).

    Science.gov (United States)

    Cockrell, Robert Chase; Christley, Scott; Chang, Eugene; An, Gary

    2015-01-01

    Perhaps the greatest challenge currently facing the biomedical research community is the ability to integrate highly detailed cellular and molecular mechanisms to represent clinical disease states as a pathway to engineer effective therapeutics. This is particularly evident in the representation of organ-level pathophysiology in terms of abnormal tissue structure, which, through histology, remains a mainstay in disease diagnosis and staging. As such, being able to generate anatomic scale simulations is a highly desirable goal. While computational limitations have previously constrained the size and scope of multi-scale computational models, advances in the capacity and availability of high-performance computing (HPC) resources have greatly expanded the ability of computational models of biological systems to achieve anatomic, clinically relevant scale. Diseases of the intestinal tract are exemplary examples of pathophysiological processes that manifest at multiple scales of spatial resolution, with structural abnormalities present at the microscopic, macroscopic and organ-levels. In this paper, we describe a novel, massively parallel computational model of the gut, the Spatially Explicitly General-purpose Model of Enteric Tissue_HPC (SEGMEnT_HPC), which extends an existing model of the gut epithelium, SEGMEnT, in order to create cell-for-cell anatomic scale simulations. We present an example implementation of SEGMEnT_HPC that simulates the pathogenesis of ileal pouchitis, and important clinical entity that affects patients following remedial surgery for ulcerative colitis.

  7. Improved detection of sentinel lymph nodes in SPECT/CT images acquired using a low- to medium-energy general-purpose collimator.

    Science.gov (United States)

    Yoneyama, Hiroto; Tsushima, Hiroyuki; Kobayashi, Masato; Onoguchi, Masahisa; Nakajima, Kenichi; Kinuya, Seigo

    2014-01-01

    The use of the low-energy high-resolution (LEHR) collimator for lymphoscintigraphy causes the appearance of star-shaped artifacts at injection sites. The aim of this study was to confirm whether the lower resolution of the low- to medium-energy general-purpose (LMEGP) collimator is compensated by decrease in the degree of septal penetration and the reduction in star-shaped artifacts. A total of 106 female patients with breast cancer, diagnosed by biopsy, were enrolled in this study. Tc phytate (37 MBq, 1 mCi) was injected around the tumor, and planar and SPECT/CT images were obtained after 3 to 4 hours. When sentinel lymph nodes (SLNs) could not be identified from planar and SPECT/CT images by using the LEHR collimator, we repeated the study with the LMEGP collimator. Planar imaging performed using the LEHR and LEHR + LMEGP collimators positively identified SLNs in 96.2% (102/106) and 99.1% (105/106) of the patients, respectively. Using combination of planar and SPECT/CT imaging with the LEHR and LEHR + LMEGP collimators, SLNs were positively identified in 97.2% (103/106) and 100% (106/106) of the patients, respectively. The LMEGP collimator provided better results than the LEHR collimator because of the lower degree of septal penetration. The use of the LMEGP collimator improved SLN detection.

  8. Code Cactus; Code Cactus

    Energy Technology Data Exchange (ETDEWEB)

    Fajeau, M; Nguyen, L T; Saunier, J [Commissariat a l' Energie Atomique, Centre d' Etudes Nucleaires de Saclay, 91 - Gif-sur-Yvette (France)

    1966-09-01

    This code handles the following problems: -1) Analysis of thermal experiments on a water loop at high or low pressure; steady state or transient behavior; -2) Analysis of thermal and hydrodynamic behavior of water-cooled and moderated reactors, at either high or low pressure, with boiling permitted; fuel elements are assumed to be flat plates: - Flowrate in parallel channels coupled or not by conduction across plates, with conditions of pressure drops or flowrate, variable or not with respect to time is given; the power can be coupled to reactor kinetics calculation or supplied by the code user. The code, containing a schematic representation of safety rod behavior, is a one dimensional, multi-channel code, and has as its complement (FLID), a one-channel, two-dimensional code. (authors) [French] Ce code permet de traiter les problemes ci-dessous: 1. Depouillement d'essais thermiques sur boucle a eau, haute ou basse pression, en regime permanent ou transitoire; 2. Etudes thermiques et hydrauliques de reacteurs a eau, a plaques, a haute ou basse pression, ebullition permise: - repartition entre canaux paralleles, couples on non par conduction a travers plaques, pour des conditions de debit ou de pertes de charge imposees, variables ou non dans le temps; - la puissance peut etre couplee a la neutronique et une representation schematique des actions de securite est prevue. Ce code (Cactus) a une dimension d'espace et plusieurs canaux, a pour complement Flid qui traite l'etude d'un seul canal a deux dimensions. (auteurs)

  9. IBM Demonstrates a General-Purpose, High-Performance, High-Availability Cloud-Hosted Data Distribution System With Live GOES-16 Weather Satellite Data

    Science.gov (United States)

    Snyder, P. L.; Brown, V. W.

    2017-12-01

    IBM has created a general purpose, data-agnostic solution that provides high performance, low data latency, high availability, scalability, and persistent access to the captured data, regardless of source or type. This capability is hosted on commercially available cloud environments and uses much faster, more efficient, reliable, and secure data transfer protocols than the more typically used FTP. The design incorporates completely redundant data paths at every level, including at the cloud data center level, in order to provide the highest assurance of data availability to the data consumers. IBM has been successful in building and testing a Proof of Concept instance on our IBM Cloud platform to receive and disseminate actual GOES-16 data as it is being downlinked. This solution leverages the inherent benefits of a cloud infrastructure configured and tuned for continuous, stable, high-speed data dissemination to data consumers worldwide at the downlink rate. It also is designed to ingest data from multiple simultaneous sources and disseminate data to multiple consumers. Nearly linear scalability is achieved by adding servers and storage.The IBM Proof of Concept system has been tested with our partners to achieve in excess of 5 Gigabits/second over public internet infrastructure. In tests with live GOES-16 data, the system routinely achieved 2.5 Gigabits/second pass-through to The Weather Company from the University of Wisconsin-Madison SSEC. Simulated data was also transferred from the Cooperative Institute for Climate and Satellites — North Carolina to The Weather Company, as well. The storage node allocated to our Proof of Concept system as tested was sized at 480 Terabytes of RAID protected disk as a worst case sizing to accommodate the data from four GOES-16 class satellites for 30 days in a circular buffer. This shows that an abundance of performance and capacity headroom exists in the IBM design that can be applied to additional missions.

  10. Development of ADINA-J-integral code

    International Nuclear Information System (INIS)

    Kurihara, Ryoichi

    1988-07-01

    A general purpose finite element program ADINA (Automatic Dynamic Incremental Nonlinear Analysis), which was developed by Bathe et al., was revised to be able to calculate the J- and J-integral. This report introduced the numerical method to add this capability to the code, and the evaluation of the revised ADINA-J code by using a few of examples of the J estimation model, i.e. a compact tension specimen, a center cracked panel subjected to dynamic load, and a thick shell cylinder having inner axial crack subjected to thermal load. The evaluation testified the function of the revised code. (author)

  11. GPScheDVS: A New Paradigm of the Autonomous CPU Speed Control for Commodity-OS-based General-Purpose Mobile Computers with a DVS-friendly Task Scheduling

    OpenAIRE

    Kim, Sookyoung

    2008-01-01

    This dissertation studies the problem of increasing battery life-time and reducing CPU heat dissipation without degrading system performance in commodity-OS-based general-purpose (GP) mobile computers using the dynamic voltage scaling (DVS) function of modern CPUs. The dissertation especially focuses on the impact of task scheduling on the effectiveness of DVS in achieving this goal. The task scheduling mechanism used in most contemporary general-purpose operating systems (GPOS) prioritizes t...

  12. Gauge color codes

    DEFF Research Database (Denmark)

    Bombin Palomo, Hector

    2015-01-01

    Color codes are topological stabilizer codes with unusual transversality properties. Here I show that their group of transversal gates is optimal and only depends on the spatial dimension, not the local geometry. I also introduce a generalized, subsystem version of color codes. In 3D they allow...

  13. MCNP code

    International Nuclear Information System (INIS)

    Cramer, S.N.

    1984-01-01

    The MCNP code is the major Monte Carlo coupled neutron-photon transport research tool at the Los Alamos National Laboratory, and it represents the most extensive Monte Carlo development program in the United States which is available in the public domain. The present code is the direct descendent of the original Monte Carlo work of Fermi, von Neumaum, and Ulam at Los Alamos in the 1940s. Development has continued uninterrupted since that time, and the current version of MCNP (or its predecessors) has always included state-of-the-art methods in the Monte Carlo simulation of radiation transport, basic cross section data, geometry capability, variance reduction, and estimation procedures. The authors of the present code have oriented its development toward general user application. The documentation, though extensive, is presented in a clear and simple manner with many examples, illustrations, and sample problems. In addition to providing the desired results, the output listings give a a wealth of detailed information (some optional) concerning each state of the calculation. The code system is continually updated to take advantage of advances in computer hardware and software, including interactive modes of operation, diagnostic interrupts and restarts, and a variety of graphical and video aids

  14. FERRET data analysis code

    International Nuclear Information System (INIS)

    Schmittroth, F.

    1979-09-01

    A documentation of the FERRET data analysis code is given. The code provides a way to combine related measurements and calculations in a consistent evaluation. Basically a very general least-squares code, it is oriented towards problems frequently encountered in nuclear data and reactor physics. A strong emphasis is on the proper treatment of uncertainties and correlations and in providing quantitative uncertainty estimates. Documentation includes a review of the method, structure of the code, input formats, and examples

  15. Enhancing QR Code Security

    OpenAIRE

    Zhang, Linfan; Zheng, Shuang

    2015-01-01

    Quick Response code opens possibility to convey data in a unique way yet insufficient prevention and protection might lead into QR code being exploited on behalf of attackers. This thesis starts by presenting a general introduction of background and stating two problems regarding QR code security, which followed by a comprehensive research on both QR code itself and related issues. From the research a solution taking advantages of cloud and cryptography together with an implementation come af...

  16. Panda code

    International Nuclear Information System (INIS)

    Altomare, S.; Minton, G.

    1975-02-01

    PANDA is a new two-group one-dimensional (slab/cylinder) neutron diffusion code designed to replace and extend the FAB series. PANDA allows for the nonlinear effects of xenon, enthalpy and Doppler. Fuel depletion is allowed. PANDA has a completely general search facility which will seek criticality, maximize reactivity, or minimize peaking. Any single parameter may be varied in a search. PANDA is written in FORTRAN IV, and as such is nearly machine independent. However, PANDA has been written with the present limitations of the Westinghouse CDC-6600 system in mind. Most computation loops are very short, and the code is less than half the useful 6600 memory size so that two jobs can reside in the core at once. (auth)

  17. Further Generalisations of Twisted Gabidulin Codes

    DEFF Research Database (Denmark)

    Puchinger, Sven; Rosenkilde, Johan Sebastian Heesemann; Sheekey, John

    2017-01-01

    We present a new family of maximum rank distance (MRD) codes. The new class contains codes that are neither equivalent to a generalised Gabidulin nor to a twisted Gabidulin code, the only two known general constructions of linear MRD codes.......We present a new family of maximum rank distance (MRD) codes. The new class contains codes that are neither equivalent to a generalised Gabidulin nor to a twisted Gabidulin code, the only two known general constructions of linear MRD codes....

  18. PENGEOM-A general-purpose geometry package for Monte Carlo simulation of radiation transport in material systems defined by quadric surfaces

    Science.gov (United States)

    Almansa, Julio; Salvat-Pujol, Francesc; Díaz-Londoño, Gloria; Carnicer, Artur; Lallena, Antonio M.; Salvat, Francesc

    2016-02-01

    The Fortran subroutine package PENGEOM provides a complete set of tools to handle quadric geometries in Monte Carlo simulations of radiation transport. The material structure where radiation propagates is assumed to consist of homogeneous bodies limited by quadric surfaces. The PENGEOM subroutines (a subset of the PENELOPE code) track particles through the material structure, independently of the details of the physics models adopted to describe the interactions. Although these subroutines are designed for detailed simulations of photon and electron transport, where all individual interactions are simulated sequentially, they can also be used in mixed (class II) schemes for simulating the transport of high-energy charged particles, where the effect of soft interactions is described by the random-hinge method. The definition of the geometry and the details of the tracking algorithm are tailored to optimize simulation speed. The use of fuzzy quadric surfaces minimizes the impact of round-off errors. The provided software includes a Java graphical user interface for editing and debugging the geometry definition file and for visualizing the material structure. Images of the structure are generated by using the tracking subroutines and, hence, they describe the geometry actually passed to the simulation code.

  19. QR CODES IN EDUCATION AND COMMUNICATION

    Directory of Open Access Journals (Sweden)

    Gurhan DURAK

    2016-04-01

    Full Text Available Technological advances brought applications of innovations to education. Conventional education increasingly flourishes with new technologies accompanied by more learner active environments. In this continuum, there are learners preferring self-learning. Traditional learning materials yield attractive, motivating and technologically enhanced learning materials. The QR (Quick Response Codes are one of these innovations. The aim of this study is to redesign a lesson unit supported with QR Codes and to get the learner views about the redesigned material. For this purpose, the redesigned lesson unit was delivered to 15 learners in Balıkesir University in the academic year of 2013-2014. The learners were asked to study the material. The learners who had smart phones and Internet access were chosen for the study. To provide sectional diversity, three groups were created. The group learners were from Faculty of Education, Faculty of Science and Literature and Faculty of Engineering. After the semi-structured interviews were held, the learners were asked about their pre-knowledge about QR Codes, QR Codes’ contribution to learning, difficulties with using QR Codes about and design issues. Descriptive data analysis was used in the study. The findings were interpreted on the basis of Theory of Diffusion of Innovations and Theory of Uses and Gratifications. After the research, the themes found were awareness of QR Code, types of QR Codes and applications, contributions to learning, and proliferation of QR Codes. Generally, the learners participating in the study reported that they were aware of QR Codes; that they could use the QR Codes; and that using QR Codes in education was useful. They also expressed that such features as visual elements, attractiveness and direct routing had positive impact on learning. In addition, they generally mentioned that they did not have any difficulty using QR Codes; that they liked the design; and that the content should

  20. Bi-level image compression with tree coding

    DEFF Research Database (Denmark)

    Martins, Bo; Forchhammer, Søren

    1996-01-01

    Presently, tree coders are the best bi-level image coders. The current ISO standard, JBIG, is a good example. By organising code length calculations properly a vast number of possible models (trees) can be investigated within reasonable time prior to generating code. Three general-purpose coders...... are constructed by this principle. A multi-pass free tree coding scheme produces superior compression results for all test images. A multi-pass fast free template coding scheme produces much better results than JBIG for difficult images, such as halftonings. Rissanen's algorithm `Context' is presented in a new...