WorldWideScience

Sample records for dynamics analysis code

  1. Convolutional Sparse Coding for Static and Dynamic Images Analysis

    Directory of Open Access Journals (Sweden)

    B. A. Knyazev

    2014-01-01

    Full Text Available The objective of this work is to improve performance of static and dynamic objects recognition. For this purpose a new image representation model and a transformation algorithm are proposed. It is examined and illustrated that limitations of previous methods make it difficult to achieve this objective. Static images, specifically handwritten digits of the widely used MNIST dataset, are the primary focus of this work. Nevertheless, preliminary qualitative results of image sequences analysis based on the suggested model are presented.A general analytical form of the Gabor function, often employed to generate filters, is described and discussed. In this research, this description is required for computing parameters of responses returned by our algorithm. The recursive convolution operator is introduced, which allows extracting free shape features of visual objects. The developed parametric representation model is compared with sparse coding based on energy function minimization.In the experimental part of this work, errors of estimating the parameters of responses are determined. Also, parameters statistics and their correlation coefficients for more than 106 responses extracted from the MNIST dataset are calculated. It is demonstrated that these data correspond well with previous research studies on Gabor filters as well as with works on visual cortex primary cells of mammals, in which similar responses were observed. A comparative test of the developed model with three other approaches is conducted; speed and accuracy scores of handwritten digits classification are presented. A support vector machine with a linear or radial basic function is used for classification of images and their representations while principal component analysis is used in some cases to prepare data beforehand. High accuracy is not attained due to the specific difficulties of combining our model with a support vector machine (a 3.99% error rate. However, another method is

  2. Application of computational fluid dynamics methods to improve thermal hydraulic code analysis

    Science.gov (United States)

    Sentell, Dennis Shannon, Jr.

    A computational fluid dynamics code is used to model the primary natural circulation loop of a proposed small modular reactor for comparison to experimental data and best-estimate thermal-hydraulic code results. Recent advances in computational fluid dynamics code modeling capabilities make them attractive alternatives to the current conservative approach of coupled best-estimate thermal hydraulic codes and uncertainty evaluations. The results from a computational fluid dynamics analysis are benchmarked against the experimental test results of a 1:3 length, 1:254 volume, full pressure and full temperature scale small modular reactor during steady-state power operations and during a depressurization transient. A comparative evaluation of the experimental data, the thermal hydraulic code results and the computational fluid dynamics code results provides an opportunity to validate the best-estimate thermal hydraulic code's treatment of a natural circulation loop and provide insights into expanded use of the computational fluid dynamics code in future designs and operations. Additionally, a sensitivity analysis is conducted to determine those physical phenomena most impactful on operations of the proposed reactor's natural circulation loop. The combination of the comparative evaluation and sensitivity analysis provides the resources for increased confidence in model developments for natural circulation loops and provides for reliability improvements of the thermal hydraulic code.

  3. Sensitivity Analysis and Uncertainty Quantification for the LAMMPS Molecular Dynamics Simulation Code

    Energy Technology Data Exchange (ETDEWEB)

    Picard, Richard Roy [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Bhat, Kabekode Ghanasham [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2017-07-18

    We examine sensitivity analysis and uncertainty quantification for molecular dynamics simulation. Extreme (large or small) output values for the LAMMPS code often occur at the boundaries of input regions, and uncertainties in those boundary values are overlooked by common SA methods. Similarly, input values for which code outputs are consistent with calibration data can also occur near boundaries. Upon applying approaches in the literature for imprecise probabilities (IPs), much more realistic results are obtained than for the complacent application of standard SA and code calibration.

  4. Validation of a plant dynamics code for 4S - Test analysis of natural circulation behavior

    Energy Technology Data Exchange (ETDEWEB)

    Sebe, F.; Horie, H.; Matsumiya, H. [Toshiba Corporation, 8 Shinsugita-Cho, Isogo-Ku, Yokohama, 235-8523 (Japan); Fanning, T. H. [Argonne National Laboratory, 9700 S Cass Ave, Argonne, IL 60439 (United States)

    2012-07-01

    A plant transient dynamics code for a sodium-cooled fast reactor was developed by Toshiba. The code is used to evaluate the safety performance of Super-Safe, Small, and Simple reactor (4S) for Anticipated Operational Occurrences (AOOs), Design Basis Accident (DBA) and Beyond DBA (BDBA). The code is currently undergoing verification and validation (V and V). As one of the validation, test analysis of the Shutdown Heat Removal Test (SHRT)-17 performed in the Experimental Breeder Reactor (EBR)-II was conducted. The SHRT-17 is protected loss of flow test. The purpose of this validation is to confirm capability of the code to simulate natural circulation behavior of the plant. As a result, good agreements are shown between the analytical results and the measured data which were available from instrumented subassembly. The detailed validation result of the natural circulation behavior is described in this paper. (authors)

  5. Dynamical analysis of innovative core designs facing unprotected transients with the MAT5 DYN code

    Energy Technology Data Exchange (ETDEWEB)

    Darmet, G.; Massara, S. [EDF R and D, 1 avenue du general de Gaulle, 92140 Clamart (France)

    2012-07-01

    Since 2007, advanced Sodium-cooled Fast Reactors (SFR) are investigated by CEA, AREVA and EDF in the framework of a joint French collaboration. A prototype called ASTRID, sets out to demonstrate progress made in SFR technology, is due to operate in the years 2020's. The modeling of unprotected transients by computer codes is one of the key safety issues in the design approach to such SFR systems. For that purpose, the activity on CATHARE, which is the reference code for the transient analysis of ASTRID, has been strengthened during last years by CEA. In the meantime, EDF has developed a simplified and multi-channel code, named MAT5 DYN, to analyze and validate innovative core designs facing protected and unprotected transients. First, the paper consists in a description of MAT5 DYN: a code based on the existing code MAT4 DYN including major improvements on geometry description and physical modeling. Second, two core designs based on the CFV core design developed at CEA are presented. Then, the dynamic response of those heterogeneous cores is analyzed during unprotected loss of flow (ULOF) transient and unprotected transient of power (UTOP). The results highlight the importance of the low void core effect specific to the CFV design. Such an effect, when combined with a sufficient primary pump halving time and an optimized cooling group scheme, allows to delay (or, possibly, avoid) the sodium boiling onset during ULOF accidents. (authors)

  6. The NEST Dry-Run Mode: Efficient Dynamic Analysis of Neuronal Network Simulation Code

    Directory of Open Access Journals (Sweden)

    Susanne Kunkel

    2017-06-01

    Full Text Available NEST is a simulator for spiking neuronal networks that commits to a general purpose approach: It allows for high flexibility in the design of network models, and its applications range from small-scale simulations on laptops to brain-scale simulations on supercomputers. Hence, developers need to test their code for various use cases and ensure that changes to code do not impair scalability. However, running a full set of benchmarks on a supercomputer takes up precious compute-time resources and can entail long queuing times. Here, we present the NEST dry-run mode, which enables comprehensive dynamic code analysis without requiring access to high-performance computing facilities. A dry-run simulation is carried out by a single process, which performs all simulation steps except communication as if it was part of a parallel environment with many processes. We show that measurements of memory usage and runtime of neuronal network simulations closely match the corresponding dry-run data. Furthermore, we demonstrate the successful application of the dry-run mode in the areas of profiling and performance modeling.

  7. An evaluation and analysis of three dynamic watershed acidification codes (MAGIC, ETD, and ILWAS)

    Energy Technology Data Exchange (ETDEWEB)

    Jenne, E.A.; Eary, L.E.; Vail, L.W.; Girvin, D.C.; Liebetrau, A.M.; Hibler, L.F.; Miley, T.B.; Monsour, M.J.

    1989-01-01

    The US Environmental Protection Agency is currently using the dynamic watershed acidification codes MAGIC, ILWAS, and ETD to assess the potential future impact of the acidic deposition on surface water quality by simulating watershed acid neutralization processes. The reliability of forecasts made with these codes is of considerable concern. The present study evaluates the process formulations (i.e., conceptual and numerical representation of atmospheric, hydrologic geochemical and biogeochemical processes), compares their approaches to calculating acid neutralizing capacity (ANC), and estimates the relative effects (sensitivity) of perturbations in the input data on selected output variables for each code. Input data were drawn from three Adirondack (upstate New York) watersheds: Panther Lake, Clear Pond, and Woods Lake. Code calibration was performed by the developers of the codes. Conclusions focus on summarizing the adequacy of process formulations, differences in ANC simulation among codes and recommendations for further research to increase forecast reliability. 87 refs., 11 figs., 77 tabs.

  8. Code-to-Code Validation and Application of a Building Dynamic Simulation Tool for the Building Energy Performance Analysis

    Directory of Open Access Journals (Sweden)

    Annamaria Buonomano

    2016-04-01

    Full Text Available In this paper details about the results of a code-to-code validation procedure of an in-house developed building simulation model, called DETECt, are reported. The tool was developed for research purposes in order to carry out dynamic building energy performance and parametric analyses by taking into account new building envelope integrated technologies, novel construction materials and innovative energy saving strategies. The reliability and accuracy of DETECt was appropriately tested by means of the standard BESTEST validation procedure. In the paper, details of this validation process are accurately described. A good agreement between the obtained results and all the reference data of the BESTEST qualification cases is achieved. In particular, the obtained results vs. standard BESTEST output are always within the provided ranges of confidence. In addition, several test cases output obtained by DETECt (e.g., dynamic profiles of indoor air and building surfaces temperature and heat fluxes and spatial trends of temperature across walls are provided.

  9. Development of a computer code for dynamic analysis of the primary circuit of advanced reactors

    Energy Technology Data Exchange (ETDEWEB)

    Rocha, Jussie Soares da; Lira, Carlos A.B.O.; Magalhaes, Mardson A. de Sa, E-mail: cabol@ufpe.b [Universidade Federal de Pernambuco (DEN/UFPE), Recife, PE (Brazil). Dept. de Energia Nuclear

    2011-07-01

    Currently, advanced reactors are being developed, seeking for enhanced safety, better performance and low environmental impacts. Reactor designs must follow several steps and numerous tests before a conceptual project could be certified. In this sense, computational tools become indispensable in the preparation of such projects. Thus, this study aimed at the development of a computational tool for thermal-hydraulic analysis by coupling two computer codes to evaluate the influence of transients caused by pressure variations and flow surges in the region of the primary circuit of IRIS reactor between the core and the pressurizer. For the simulation, it was used a situation of 'insurge', characterized by the entry of water in the pressurizer, due to the expansion of the refrigerant in the primary circuit. This expansion was represented by a pressure disturbance in step form, through the block 'step' of SIMULINK, thus enabling the transient startup. The results showed that the dynamic tool, obtained through the coupling of the codes, generated very satisfactory responses within model limitations, preserving the most important phenomena in the process. (author)

  10. Modification of the SAS4A Safety Analysis Code for Integration with the ADAPT Discrete Dynamic Event Tree Framework.

    Energy Technology Data Exchange (ETDEWEB)

    Jankovsky, Zachary Kyle [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Denman, Matthew R. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2017-05-01

    It is difficult to assess the consequences of a transient in a sodium-cooled fast reactor (SFR) using traditional probabilistic risk assessment (PRA) methods, as numerous safety-related sys- tems have passive characteristics. Often there is significant dependence on the value of con- tinuous stochastic parameters rather than binary success/failure determinations. One form of dynamic PRA uses a system simulator to represent the progression of a transient, tracking events through time in a discrete dynamic event tree (DDET). In order to function in a DDET environment, a simulator must have characteristics that make it amenable to changing physical parameters midway through the analysis. The SAS4A SFR system analysis code did not have these characteristics as received. This report describes the code modifications made to allow dynamic operation as well as the linking to a Sandia DDET driver code. A test case is briefly described to demonstrate the utility of the changes.

  11. Multi-code analysis of scrape-off layer filament dynamics in MAST

    DEFF Research Database (Denmark)

    Militello, F.; Walkden, N. R.; Farley, T.

    2016-01-01

    Four numerical codes are employed to investigate the dynamics of scrape-off layer filaments in tokamak relevant conditions. Experimental measurements were taken in the MAST device using visual camera imaging, which allows the evaluation of the perpendicular size and velocity of the filaments...

  12. Analysis and compensation for code Doppler effect of BDS II signal under high dynamics

    Science.gov (United States)

    Ouyang, Xiaofeng; Zeng, Fangling

    2016-01-01

    In high dynamic circumstances, the acquisition of BDS (BeiDou Navigation Satellite System) signal would be affected by the pseudo-code Doppler. The pseudo-code frequency shift is more prominent and complex when BOC modulation has been adopted by BDS-II, but is not yet involved in current compensation algorithm. In addition, the most frequently used code Doppler compensation algorithm is modifying the sampling rate or local bit rate, which not only increases the complexity of the acquisition and tracking, but also is barely realizable for the hardware receiver to modify the local frequency. Therefore, this paper proposes a code Doppler compensation method based on double estimator receiver, which simultaneously controls NCO delay of code tracking loop and subcarrier tracking loop to compensate for pseudo-code frequency shift. The simulation and test are implemented with BDS-II BOC signal. The test results demonstrate that the proposed algorithm can effectively compensate for pseudo-code Doppler of BOC signal and has detection probability 3dB higher than the uncompensated situation when the false alarm rate is under 0.01 and the coherent integration time is 1ms.

  13. Multi-code analysis of scrape-off layer filament dynamics in MAST

    Science.gov (United States)

    Militello, F.; Walkden, N. R.; Farley, T.; Gracias, W. A.; Olsen, J.; Riva, F.; Easy, L.; Fedorczak, N.; Lupelli, I.; Madsen, J.; Nielsen, A. H.; Ricci, P.; Tamain, P.; Young, J.

    2016-11-01

    Four numerical codes are employed to investigate the dynamics of scrape-off layer filaments in tokamak relevant conditions. Experimental measurements were taken in the MAST device using visual camera imaging, which allows the evaluation of the perpendicular size and velocity of the filaments, as well as the combination of density and temperature associated with the perturbation. A new algorithm based on the light emission integrated along the field lines associated with the position of the filament is developed to ensure that it is properly detected and tracked. The filaments are found to have velocities of the order of 1~\\text{km}~{{\\text{s}}-1} , a perpendicular diameter of around 2-3 cm and a density amplitude 2-3.5 times the background plasma. 3D and 2D numerical codes (the STORM module of BOUT++, GBS, HESEL and TOKAM3X) are used to reproduce the motion of the observed filaments with the purpose of validating the codes and of better understanding the experimental data. Good agreement is found between the 3D codes. The seeded filament simulations are also able to reproduce the dynamics observed in experiments with accuracy up to the experimental errorbar levels. In addition, the numerical results showed that filaments characterised by similar size and light emission intensity can have quite different dynamics if the pressure perturbation is distributed differently between density and temperature components. As an additional benefit, several observations on the dynamics of the filaments in the presence of evolving temperature fields were made and led to a better understanding of the behaviour of these coherent structures.

  14. Visualization and analysis of microtubule dynamics using dual color-coded display of plus-end labels.

    Directory of Open Access Journals (Sweden)

    Amy K Garrison

    Full Text Available Investigating spatial and temporal control of microtubule dynamics in live cells is critical to understanding cell morphogenesis in development and disease. Tracking fluorescently labeled plus-end-tracking proteins over time has become a widely used method to study microtubule assembly. Here, we report a complementary approach that uses only two images of these labels to visualize and analyze microtubule dynamics at any given time. Using a simple color-coding scheme, labeled plus-ends from two sequential images are pseudocolored with different colors and then merged to display color-coded ends. Based on object recognition algorithms, these colored ends can be identified and segregated into dynamic groups corresponding to four events, including growth, rescue, catastrophe, and pause. Further analysis yields not only their spatial distribution throughout the cell but also provides measurements such as growth rate and direction for each labeled end. We have validated the method by comparing our results with ground-truth data derived from manual analysis as well as with data obtained using the tracking method. In addition, we have confirmed color-coded representation of different dynamic events by analyzing their history and fate. Finally, we have demonstrated the use of the method to investigate microtubule assembly in cells and provided guidance in selecting optimal image acquisition conditions. Thus, this simple computer vision method offers a unique and quantitative approach to study spatial regulation of microtubule dynamics in cells.

  15. PRONTO3D users` instructions: A transient dynamic code for nonlinear structural analysis

    Energy Technology Data Exchange (ETDEWEB)

    Attaway, S.W.; Mello, F.J.; Heinstein, M.W.; Swegle, J.W.; Ratner, J.A. [Sandia National Labs., Albuquerque, NM (United States); Zadoks, R.I. [Univ. of Texas, El Paso, TX (United States)

    1998-06-01

    This report provides an updated set of users` instructions for PRONTO3D. PRONTO3D is a three-dimensional, transient, solid dynamics code for analyzing large deformations of highly nonlinear materials subjected to extremely high strain rates. This Lagrangian finite element program uses an explicit time integration operator to integrate the equations of motion. Eight-node, uniform strain, hexahedral elements and four-node, quadrilateral, uniform strain shells are used in the finite element formulation. An adaptive time step control algorithm is used to improve stability and performance in plasticity problems. Hourglass distortions can be eliminated without disturbing the finite element solution using either the Flanagan-Belytschko hourglass control scheme or an assumed strain hourglass control scheme. All constitutive models in PRONTO3D are cast in an unrotated configuration defined using the rotation determined from the polar decomposition of the deformation gradient. A robust contact algorithm allows for the impact and interaction of deforming contact surfaces of quite general geometry. The Smooth Particle Hydrodynamics method has been embedded into PRONTO3D using the contact algorithm to couple it with the finite element method.

  16. Automated Facial Action Coding System for dynamic analysis of facial expressions in neuropsychiatric disorders.

    Science.gov (United States)

    Hamm, Jihun; Kohler, Christian G; Gur, Ruben C; Verma, Ragini

    2011-09-15

    Facial expression is widely used to evaluate emotional impairment in neuropsychiatric disorders. Ekman and Friesen's Facial Action Coding System (FACS) encodes movements of individual facial muscles from distinct momentary changes in facial appearance. Unlike facial expression ratings based on categorization of expressions into prototypical emotions (happiness, sadness, anger, fear, disgust, etc.), FACS can encode ambiguous and subtle expressions, and therefore is potentially more suitable for analyzing the small differences in facial affect. However, FACS rating requires extensive training, and is time consuming and subjective thus prone to bias. To overcome these limitations, we developed an automated FACS based on advanced computer science technology. The system automatically tracks faces in a video, extracts geometric and texture features, and produces temporal profiles of each facial muscle movement. These profiles are quantified to compute frequencies of single and combined Action Units (AUs) in videos, and they can facilitate a statistical study of large populations in disorders known to impact facial expression. We derived quantitative measures of flat and inappropriate facial affect automatically from temporal AU profiles. Applicability of the automated FACS was illustrated in a pilot study, by applying it to data of videos from eight schizophrenia patients and controls. We created temporal AU profiles that provided rich information on the dynamics of facial muscle movements for each subject. The quantitative measures of flatness and inappropriateness showed clear differences between patients and the controls, highlighting their potential in automatic and objective quantification of symptom severity. Copyright © 2011 Elsevier B.V. All rights reserved.

  17. Improvement of level-1 PSA computer code package - Modeling and analysis for dynamic reliability of nuclear power plants

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Chang Hoon; Baek, Sang Yeup; Shin, In Sup; Moon, Shin Myung; Moon, Jae Phil; Koo, Hoon Young; Kim, Ju Shin [Seoul National University, Seoul (Korea, Republic of); Hong, Jung Sik [Seoul National Polytechnology University, Seoul (Korea, Republic of); Lim, Tae Jin [Soongsil University, Seoul (Korea, Republic of)

    1996-08-01

    The objective of this project is to develop a methodology of the dynamic reliability analysis for NPP. The first year`s research was focused on developing a procedure for analyzing failure data of running components and a simulator for estimating the reliability of series-parallel structures. The second year`s research was concentrated on estimating the lifetime distribution and PM effect of a component from its failure data in various cases, and the lifetime distribution of a system with a particular structure. Computer codes for performing these jobs were also developed. The objectives of the third year`s research is to develop models for analyzing special failure types (CCFs, Standby redundant structure) that were nor considered in the first two years, and to complete a methodology of the dynamic reliability analysis for nuclear power plants. The analysis of failure data of components and related researches for supporting the simulator must be preceded for providing proper input to the simulator. Thus this research is divided into three major parts. 1. Analysis of the time dependent life distribution and the PM effect. 2. Development of a simulator for system reliability analysis. 3. Related researches for supporting the simulator : accelerated simulation analytic approach using PH-type distribution, analysis for dynamic repair effects. 154 refs., 5 tabs., 87 figs. (author)

  18. Mathematical model and computer code for the analysis of advanced fast reactor dynamics

    Energy Technology Data Exchange (ETDEWEB)

    Schukin, N.V. (Moscow Engineering Physics Inst. (Russian Federation)); Korsun, A.S. (Moscow Engineering Physics Inst. (Russian Federation)); Vitruk, S.G. (Moscow Engineering Physics Inst. (Russian Federation)); Zimin, V.G. (Moscow Engineering Physics Inst. (Russian Federation)); Romanin, S.D. (Moscow Engineering Physics Inst. (Russian Federation))

    1993-04-01

    Efficient algorithms for mathematical modeling of 3-D neutron kinetics and thermal hydraulics are described. The model and appropriate computer code make it possible to analyze a variety of transient events ranging from normal operational states to catastrophic accident excursions. To verify the code, a number of calculations of different kind of transients was carried out. The results of the calculations show that the model and the computer code could be used for conceptual design of advanced liquid metal reactors. The detailed description of calculations of TOP WS accident is presented. (orig./DG)

  19. Beam-dynamics codes used at DARHT

    Energy Technology Data Exchange (ETDEWEB)

    Ekdahl, Jr., Carl August [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2017-02-01

    Several beam simulation codes are used to help gain a better understanding of beam dynamics in the DARHT LIAs. The most notable of these fall into the following categories: for beam production – Tricomp Trak orbit tracking code, LSP Particle in cell (PIC) code, for beam transport and acceleration – XTR static envelope and centroid code, LAMDA time-resolved envelope and centroid code, LSP-Slice PIC code, for coasting-beam transport to target – LAMDA time-resolved envelope code, LSP-Slice PIC code. These codes are also being used to inform the design of Scorpius.

  20. TRACK The New Beam Dynamics Code

    CERN Document Server

    Mustapha, Brahim; Ostroumov, Peter; Schnirman-Lessner, Eliane

    2005-01-01

    The new ray-tracing code TRACK was developed* to fulfill the special requirements of the RIA accelerator systems. The RIA lattice includes an ECR ion source, a LEBT containing a MHB and a RFQ followed by three SC linac sections separated by two stripping stations with appropriate magnetic transport systems. No available beam dynamics code meet all the necessary requirements for an end-to-end simulation of the RIA driver linac. The latest version of TRACK was used for end-to-end simulations of the RIA driver including errors and beam loss analysis.** In addition to the standard capabilities, the code includes the following new features: i) multiple charge states ii) realistic stripper model; ii) static and dynamic errors iii) automatic steering to correct for misalignments iv) detailed beam-loss analysis; v) parallel computing to perform large scale simulations. Although primarily developed for simulations of the RIA machine, TRACK is a general beam dynamics code. Currently it is being used for the design and ...

  1. Efficient Quantum Private Communication Based on Dynamic Control Code Sequence

    Science.gov (United States)

    Cao, Zheng-Wen; Feng, Xiao-Yi; Peng, Jin-Ye; Zeng, Gui-Hua; Qi, Jin

    2016-12-01

    Based on chaos and quantum properties, we propose a quantum private communication scheme with dynamic control code sequence. The initial sequence is obtained via chaotic systems, and the control code sequence is derived by grouping, XOR and extracting. A shift cycle algorithm is designed to enable the dynamic change of control code sequence. Analysis shows that transmission efficiency could reach 100 % with high dynamics and security.

  2. Efficient Quantum Private Communication Based on Dynamic Control Code Sequence

    Science.gov (United States)

    Cao, Zheng-Wen; Feng, Xiao-Yi; Peng, Jin-Ye; Zeng, Gui-Hua; Qi, Jin

    2017-04-01

    Based on chaos and quantum properties, we propose a quantum private communication scheme with dynamic control code sequence. The initial sequence is obtained via chaotic systems, and the control code sequence is derived by grouping, XOR and extracting. A shift cycle algorithm is designed to enable the dynamic change of control code sequence. Analysis shows that transmission efficiency could reach 100 % with high dynamics and security.

  3. The NATA code: Theory and analysis, volume 1. [user manuals (computer programming) - gas dynamics, wind tunnels

    Science.gov (United States)

    Bade, W. L.; Yos, J. M.

    1975-01-01

    A computer program for calculating quasi-one-dimensional gas flow in axisymmetric and two-dimensional nozzles and rectangular channels is presented. Flow is assumed to start from a state of thermochemical equilibrium at a high temperature in an upstream reservoir. The program provides solutions based on frozen chemistry, chemical equilibrium, and nonequilibrium flow with finite reaction rates. Electronic nonequilibrium effects can be included using a two-temperature model. An approximate laminar boundary layer calculation is given for the shear and heat flux on the nozzle wall. Boundary layer displacement effects on the inviscid flow are considered also. Chemical equilibrium and transport property calculations are provided by subroutines. The code contains precoded thermochemical, chemical kinetic, and transport cross section data for high-temperature air, CO2-N2-Ar mixtures, helium, and argon. It provides calculations of the stagnation conditions on axisymmetric or two-dimensional models, and of the conditions on the flat surface of a blunt wedge. The primary purpose of the code is to describe the flow conditions and test conditions in electric arc heated wind tunnels.

  4. Non-reacting Flow Analysis from Combustor Inlet to Outlet using Computational Fluid Dynamics Code

    Directory of Open Access Journals (Sweden)

    G. Ananda Reddy

    2004-10-01

    Full Text Available This paper describes non-reacting flow analysis of a gas turbine combustion system. The method is based on the solution of Navier-Strokes equations using generalised non-orthogonal coordinate system. The turbulence effects are modelled through the renormalisation group k-E model. The method has been applied to a practical gas turbine combustor. The combustionsystem includes swirler vane passages, fuel nozzles, rotor bleed, customer bleed, air-blast atomiser, swirl cone, and all holes in primary , dilution , dome, flare, and cooling ring. Thetotal geometry has been created using the pre-processors GAMBIT and CATIA, and the meshing has been done using GAMBIT, and the analysis carried out in a FLUENT solver. The interaction between the diffuser and the combustor external flows plays a key role in controlling the pressure loss, air flow distribution around the combustor liner, durability, and stability. The aero gas turbine combustor designs are generally guided by experimental methods and past experience; however, experimental methods are inherently slow, costly, especially at hightemperature engine-operating conditions. These drawbacks and the growing need to understand the complex flow-field phenomenon involved, have led to the development of a numericalmodel for predicting flow in the gas turbine combustor. These models are used to optimise the design of the combustor and its subcomponents, and reduce cost, time, and the number ofexperiments.

  5. Combustion chamber analysis code

    Science.gov (United States)

    Przekwas, A. J.; Lai, Y. G.; Krishnan, A.; Avva, R. K.; Giridharan, M. G.

    1993-05-01

    A three-dimensional, time dependent, Favre averaged, finite volume Navier-Stokes code has been developed to model compressible and incompressible flows (with and without chemical reactions) in liquid rocket engines. The code has a non-staggered formulation with generalized body-fitted-coordinates (BFC) capability. Higher order differencing methodologies such as MUSCL and Osher-Chakravarthy schemes are available. Turbulent flows can be modeled using any of the five turbulent models present in the code. A two-phase, two-liquid, Lagrangian spray model has been incorporated into the code. Chemical equilibrium and finite rate reaction models are available to model chemically reacting flows. The discrete ordinate method is used to model effects of thermal radiation. The code has been validated extensively against benchmark experimental data and has been applied to model flows in several propulsion system components of the SSME and the STME.

  6. Dynamic Index Coding for Wireless Broadcast Networks

    CERN Document Server

    Neely, Michael J; Zhang, Zhen

    2011-01-01

    We consider a wireless broadcast station that transmits packets to multiple users. The packet requests for each user may overlap, and some users may already have certain packets. This presents a problem of broadcasting in the presence of side information, and is a generalization of the well known (and unsolved) index coding problem of information theory. Rather than achieving the full capacity region, we develop a code-constrained capacity region, which restricts attention to a pre-specified set of coding actions. We develop a dynamic max-weight algorithm that allows for random packet arrivals and supports any traffic inside the code-constrained capacity region. Further, we provide a simple set of codes based on cycles in the underlying demand graph. We show these codes are optimal for a class of broadcast relay problems.

  7. The reactor dynamics code DYN3D

    Energy Technology Data Exchange (ETDEWEB)

    Kliem, Soeren; Bilodid, Yuri; Fridman, Emil; Baier, Silvio; Grahn, Alexander; Gommlich, Andre; Nikitin, Evgeny; Rohde, Ulrich [Helmholtz-Zentrum Dresden-Rossendorf e.V., Dresden (Germany)

    2016-05-15

    The article provides an overview on the code DYN3D which is a three-dimensional core model for steady-state, dynamic and depletion calculations in reactor cores with quadratic or hexagonal fuel assembly geometry being developed by the Helmholtz-Zentrum Dresden-Rossendorf for more than 20 years. The current paper gives an overview on the basic DYN3D models and the available code couplings. The verification and validation status is shortly outlined. The paper concludes with the current developments of the DYN3D code. For more detailed information the reader is referred to the publications cited in the corresponding chapters.

  8. Source Code Generator Based on Dynamic Frames

    Directory of Open Access Journals (Sweden)

    Danijel Radošević

    2011-06-01

    Full Text Available Normal 0 21 false false false HR X-NONE X-NONE MicrosoftInternetExplorer4 /* Style Definitions */ table.MsoNormalTable {mso-style-name:"Obična tablica"; mso-tstyle-rowband-size:0; mso-tstyle-colband-size:0; mso-style-noshow:yes; mso-style-priority:99; mso-style-qformat:yes; mso-style-parent:""; mso-padding-alt:0cm 5.4pt 0cm 5.4pt; mso-para-margin:0cm; mso-para-margin-bottom:.0001pt; mso-pagination:widow-orphan; font-size:11.0pt; font-family:"Calibri","sans-serif"; mso-ascii-font-family:Calibri; mso-ascii-theme-font:minor-latin; mso-fareast-font-family:"Times New Roman"; mso-fareast-theme-font:minor-fareast; mso-hansi-font-family:Calibri; mso-hansi-theme-font:minor-latin; mso-bidi-font-family:"Times New Roman"; mso-bidi-theme-font:minor-bidi;} This paper presents the model of source code generator based on dynamic frames. The model is named as the SCT model because if its three basic components: Specification (S, which describes the application characteristics, Configuration (C, which describes the rules for building applications, and Templates (T, which refer to application building blocks. The process of code generation dynamically creates XML frames containing all building elements (S, C ant T until final code is produced. This approach is compared to existing XVCL frames based model for source code generating. The SCT model is described by both XML syntax and the appropriate graphical elements. The SCT model is aimed to build complete applications, not just skeletons. The main advantages of the presented model are its textual and graphic description, a fully configurable generator, and the reduced overhead of the generated source code. The presented SCT model is shown on development of web application example in order to demonstrate its features and justify our design choices.

  9. An Analysis of Syndrome Coding

    Science.gov (United States)

    Amiruzzaman, Md; Abdullah-Al-Wadud, M.; Chung, Yoojin

    In this paper a detail analysis is presented based on BCH syndrome coding for covert channel data hiding methods. The experimented technique is nothing but a syndrome coding algorithm with a coset based approach, analyzed results are showing that the examined method has more flexibility to choose coset, also providing less modification distortion caused by data hiding. Analyzed method presented by clear mathematical way. As it is mathematical equation dependent, hence analyzed results are showing that the analyzed method has fast computation ability and find perfect roots for modification.

  10. Experimental methodology for computational fluid dynamics code validation

    Energy Technology Data Exchange (ETDEWEB)

    Aeschliman, D.P.; Oberkampf, W.L.

    1997-09-01

    Validation of Computational Fluid Dynamics (CFD) codes is an essential element of the code development process. Typically, CFD code validation is accomplished through comparison of computed results to previously published experimental data that were obtained for some other purpose, unrelated to code validation. As a result, it is a near certainty that not all of the information required by the code, particularly the boundary conditions, will be available. The common approach is therefore unsatisfactory, and a different method is required. This paper describes a methodology developed specifically for experimental validation of CFD codes. The methodology requires teamwork and cooperation between code developers and experimentalists throughout the validation process, and takes advantage of certain synergisms between CFD and experiment. The methodology employs a novel uncertainty analysis technique which helps to define the experimental plan for code validation wind tunnel experiments, and to distinguish between and quantify various types of experimental error. The methodology is demonstrated with an example of surface pressure measurements over a model of varying geometrical complexity in laminar, hypersonic, near perfect gas, 3-dimensional flow.

  11. Braneworld dynamics with the BraneCode

    CERN Document Server

    Martín, J; Frolov, A V; Peloso, M; Kofman, L A; Martin, Johannes; Felder, Gary N.; Frolov, Andrei V.; Peloso, Marco; Kofman, Lev

    2003-01-01

    We give a full nonlinear numerical treatment of time-dependent 5d braneworld geometry, which is determined self-consistently by potentials for the scalar field in the bulk and at two orbifold branes, supplemented by boundary conditions at the branes. We describe the BraneCode, an algorithm which we designed to solve the dynamical equations numerically. We applied the BraneCode to braneworld models and found several novel phenomena of the brane dynamics. Starting with static warped geometry with de Sitter branes, we found numerically that this configuration is often unstable due to a tachyonic mass of the radion during inflation. If the model admits other static configurations with lower values of de Sitter curvature, this effect causes a violent re-structuring towards them, flattening the branes, which appears as a lowering of the 4d effective cosmological constant. Braneworld dynamics can often lead to brane collisions. We found that in the presence of the bulk scalar field, the 5d geometry between colliding...

  12. Implementing peridynamics within a molecular dynamics code.

    Energy Technology Data Exchange (ETDEWEB)

    Lehoucq, Richard B.; Silling, Stewart Andrew; Plimpton, Steven James; Parks, Michael L.

    2007-12-01

    Peridynamics (PD) is a continuum theory that employs a nonlocal model to describe material properties. In this context, nonlocal means that continuum points separated by a finite distance may exert force upon each other. A meshless method results when PD is discretized with material behavior approximated as a collection of interacting particles. This paper describes how PD can be implemented within a molecular dynamics (MD) framework, and provides details of an efficient implementation. This adds a computational mechanics capability to an MD code, enabling simulations at mesoscopic or even macroscopic length and time scales.

  13. Analysis of Void Fraction Distribution and Departure from Nucleate Boiling in Single Subchannel and Bundle Geometries Using Subchannel, System, and Computational Fluid Dynamics Codes

    Directory of Open Access Journals (Sweden)

    Taewan Kim

    2012-01-01

    Full Text Available In order to assess the accuracy and validity of subchannel, system, and computational fluid dynamics codes, the Paul Scherrer Institut has participated in the OECD/NRC PSBT benchmark with the thermal-hydraulic system code TRACE5.0 developed by US NRC, the subchannel code FLICA4 developed by CEA, and the computational fluid dynamic code STAR-CD developed by CD-adapco. The PSBT benchmark consists of a series of void distribution exercises and departure from nucleate boiling exercises. The results reveal that the prediction by the subchannel code FLICA4 agrees with the experimental data reasonably well in both steady-state and transient conditions. The analyses of single-subchannel experiments by means of the computational fluid dynamic code STAR-CD with the CD-adapco boiling model indicate that the prediction of the void fraction has no significant discrepancy from the experiments. The analyses with TRACE point out the necessity to perform additional assessment of the subcooled boiling model and bulk condensation model of TRACE.

  14. Dynamic Reverse Code Generation for Backward Execution

    DEFF Research Database (Denmark)

    Lee, Jooyong

    2007-01-01

    . In this paper, we present a method to generate reverse code, so that backtracking can be performed by executing reverse code. The novelty of our work is that we generate reverse code on-the-fly, while running a debugger, which makes it possible to apply the method even to debugging multi-threaded programs....

  15. Flow Analysis of Code Customizations

    DEFF Research Database (Denmark)

    Hessellund, Anders; Sestoft, Peter

    2008-01-01

    Inconsistency between metadata and code customizations is a major concern in modern, configurable enterprise systems. The increasing reliance on metadata, in the form of XML files, and code customizations, in the form of Java files, has led to a hybrid development platform. The expected consisten...

  16. Analysis of run-to-run variation of bar-coded pyrosequencing for evaluating bacterial community shifts and individual taxa dynamics.

    Science.gov (United States)

    Ge, Yuan; Schimel, Joshua P; Holden, Patricia A

    2014-01-01

    Bar-coded pyrosequencing has been increasingly used due to its fine taxonomic resolution and high throughput. Yet, concerns arise regarding the reproducibility of bar-coded pyrosequencing. We evaluated the run-to-run variation of bar-coded pyrosequencing in detecting bacterial community shifts and taxa dynamics. Our results demonstrate that pyrosequencing is reproducible in evaluating community shifts within a run, but not between runs. Also, the reproducibility of pyrosequencing in detecting individual taxa increased as a function of taxa abundance. Based on our findings: (1) for studies with modest sequencing depth, it is doubtful that data from different pyrosequencing runs can be considered comparable; (2) if multiple pyrosequencing runs are needed to increase the sequencing depth, additional sequencing efforts should be applied to all samples, rather than to selected samples; (3) if pyrosequencing is used for estimating bacterial population dynamics, only the abundant taxa should be considered; (4) for less-abundant taxa, the sequencing depth should be increased to ensure an accurate evaluation of taxon variation trends across samples.

  17. Accurate discrimination of conserved coding and non-coding regions through multiple indicators of evolutionary dynamics

    Directory of Open Access Journals (Sweden)

    Pesole Graziano

    2009-09-01

    Full Text Available Abstract Background The conservation of sequences between related genomes has long been recognised as an indication of functional significance and recognition of sequence homology is one of the principal approaches used in the annotation of newly sequenced genomes. In the context of recent findings that the number non-coding transcripts in higher organisms is likely to be much higher than previously imagined, discrimination between conserved coding and non-coding sequences is a topic of considerable interest. Additionally, it should be considered desirable to discriminate between coding and non-coding conserved sequences without recourse to the use of sequence similarity searches of protein databases as such approaches exclude the identification of novel conserved proteins without characterized homologs and may be influenced by the presence in databases of sequences which are erroneously annotated as coding. Results Here we present a machine learning-based approach for the discrimination of conserved coding sequences. Our method calculates various statistics related to the evolutionary dynamics of two aligned sequences. These features are considered by a Support Vector Machine which designates the alignment coding or non-coding with an associated probability score. Conclusion We show that our approach is both sensitive and accurate with respect to comparable methods and illustrate several situations in which it may be applied, including the identification of conserved coding regions in genome sequences and the discrimination of coding from non-coding cDNA sequences.

  18. Automatic code generation from the OMT-based dynamic model

    Energy Technology Data Exchange (ETDEWEB)

    Ali, J.; Tanaka, J.

    1996-12-31

    The OMT object-oriented software development methodology suggests creating three models of the system, i.e., object model, dynamic model and functional model. We have developed a system that automatically generates implementation code from the dynamic model. The system first represents the dynamic model as a table and then generates executable Java language code from it. We used inheritance for super-substate relationships. We considered that transitions relate to states in a state diagram exactly as operations relate to classes in an object diagram. In the generated code, each state in the state diagram becomes a class and each event on a state becomes an operation on the corresponding class. The system is implemented and can generate executable code for any state diagram. This makes the role of the dynamic model more significant and the job of designers even simpler.

  19. Wind turbine control systems: Dynamic model development using system identification and the fast structural dynamics code

    Energy Technology Data Exchange (ETDEWEB)

    Stuart, J.G.; Wright, A.D.; Butterfield, C.P.

    1996-10-01

    Mitigating the effects of damaging wind turbine loads and responses extends the lifetime of the turbine and, consequently, reduces the associated Cost of Energy (COE). Active control of aerodynamic devices is one option for achieving wind turbine load mitigation. Generally speaking, control system design and analysis requires a reasonable dynamic model of {open_quotes}plant,{close_quotes} (i.e., the system being controlled). This paper extends the wind turbine aileron control research, previously conducted at the National Wind Technology Center (NWTC), by presenting a more detailed development of the wind turbine dynamic model. In prior research, active aileron control designs were implemented in an existing wind turbine structural dynamics code, FAST (Fatigue, Aerodynamics, Structures, and Turbulence). In this paper, the FAST code is used, in conjunction with system identification, to generate a wind turbine dynamic model for use in active aileron control system design. The FAST code is described and an overview of the system identification technique is presented. An aileron control case study is used to demonstrate this modeling technique. The results of the case study are then used to propose ideas for generalizing this technique for creating dynamic models for other wind turbine control applications.

  20. Improvement of Basic Fluid Dynamics Models for the COMPASS Code

    Science.gov (United States)

    Zhang, Shuai; Morita, Koji; Shirakawa, Noriyuki; Yamamoto, Yuichi

    The COMPASS code is a new next generation safety analysis code to provide local information for various key phenomena in core disruptive accidents of sodium-cooled fast reactors, which is based on the moving particle semi-implicit (MPS) method. In this study, improvement of basic fluid dynamics models for the COMPASS code was carried out and verified with fundamental verification calculations. A fully implicit pressure solution algorithm was introduced to improve the numerical stability of MPS simulations. With a newly developed free surface model, numerical difficulty caused by poor pressure solutions is overcome by involving free surface particles in the pressure Poisson equation. In addition, applicability of the MPS method to interactions between fluid and multi-solid bodies was investigated in comparison with dam-break experiments with solid balls. It was found that the PISO algorithm and free surface model makes simulation with the passively moving solid model stable numerically. The characteristic behavior of solid balls was successfully reproduced by the present numerical simulations.

  1. A Case for Dynamic Reverse-code Generation to Debug Non-deterministic Programs

    Directory of Open Access Journals (Sweden)

    Jooyong Yi

    2013-09-01

    Full Text Available Backtracking (i.e., reverse execution helps the user of a debugger to naturally think backwards along the execution path of a program, and thinking backwards makes it easy to locate the origin of a bug. So far backtracking has been implemented mostly by state saving or by checkpointing. These implementations, however, inherently do not scale. Meanwhile, a more recent backtracking method based on reverse-code generation seems promising because executing reverse code can restore the previous states of a program without state saving. In the literature, there can be found two methods that generate reverse code: (a static reverse-code generation that pre-generates reverse code through static analysis before starting a debugging session, and (b dynamic reverse-code generation that generates reverse code by applying dynamic analysis on the fly during a debugging session. In particular, we espoused the latter one in our previous work to accommodate non-determinism of a program caused by e.g., multi-threading. To demonstrate the usefulness of our dynamic reverse-code generation, this article presents a case study of various backtracking methods including ours. We compare the memory usage of various backtracking methods in a simple but nontrivial example, a bounded-buffer program. In the case of non-deterministic programs such as this bounded-buffer program, our dynamic reverse-code generation outperforms the existing backtracking methods in terms of memory efficiency.

  2. Expert system interaction with existing analysis codes

    Energy Technology Data Exchange (ETDEWEB)

    Ransom, V.H.; Fink, R.K.; Bertch, W.J.; Callow, R.A.

    1986-01-01

    Coupling expert systems with existing engineering analysis codes is a promising area in the field of artificial intelligence. The added intelligence can provide for easier and less costly use of the code and also reduce the potential for code misuse. This paper will discuss the methods available to allow interaction between an expert system and a large analysis code running on a mainframe. Concluding remarks will identify potential areas of expert system application with specific areas that are being considered in a current research program. The difficulty of interaction between an analysis code and an expert system is due to the incompatibility between the FORTRAN environment used for the analysis code and the AI environment used for the expert system. Three methods, excluding file transfer techniques, are discussed to help overcome this incompatibility. The first method is linking the FORTRAN routines to the LISP environment on the same computer. Various LISP dialects available on mainframes and their interlanguage communication capabilities are discussed. The second method involves network interaction between a LISP machine and a mainframe computer. Comparisons between the linking method and networking are noted. The third method involves the use of an expert system tool that is campatible with a FORTRAN environment. Several available tools are discussed. With the interaction methods identified, several potential application areas are considered. Selection of the specific areas that will be developed for the pilot project and applied to a thermal-hydraulic energy analysis code are noted.

  3. Spectral Analysis Code: PARAS SPEC

    CERN Document Server

    Chaturvedi, Priyanka; Anandarao, B G

    2016-01-01

    The light emitted from the stellar photosphere serves as a unique signature for the nature of stars. The behaviour of these stellar lines depend upon the surface temperature, mass, evolutionary status and chemical composition of the star. With the advent of high-resolution spectrographs coupled with medium to large aperture telescopes around the globe, there is plenty of high-resolution and high signal-to-noise ratio data available to the astronomy community. Apart from radial velocity (RV) studies, such data offer us the unique opportunity to study chemical composition and atmospheric properties of the star. The procedure used to derive these parameters must be automated and well adaptable to data available from any high-resolution spectrograph. We hereby present an IDL code, PARAS SPEC, which was primary designed to handle high-resolution spectroscopy data from PARAS spectrograph coupled with the 1.2~m telescope at Mt. Abu, India. This code is designed to adapt with data from other spectrographs as well. Th...

  4. Confinement Vessel Dynamic Analysis

    Energy Technology Data Exchange (ETDEWEB)

    R. Robert Stevens; Stephen P. Rojas

    1999-08-01

    A series of hydrodynamic and structural analyses of a spherical confinement vessel has been performed. The analyses used a hydrodynamic code to estimate the dynamic blast pressures at the vessel's internal surfaces caused by the detonation of a mass of high explosive, then used those blast pressures as applied loads in an explicit finite element model to simulate the vessel's structural response. Numerous load cases were considered. Particular attention was paid to the bolted port connections and the O-ring pressure seals. The analysis methods and results are discussed, and comparisons to experimental results are made.

  5. Sensory Coding with Dynamically Competitive Networks

    CERN Document Server

    Rabinovich, M I; Volkovskii, A R; Abarbanel, Henry D I; Laurent, G; Abarbanel, Henry D I

    1999-01-01

    Studies of insect olfactory processing indicate that odors are represented by rich spatio-temporal patterns of neural activity. These patterns are very difficult to predict a priori, yet they are stimulus specific and reliable upon repeated stimulation with the same input. We formulate here a theoretical framework in which we can interpret these experimental results. We propose a paradigm of ``dynamic competition'' in which inputs (odors) are represented by internally competing neural assemblies. Each pattern is the result of dynamical motion within the network and does not involve a ``winner'' among competing possibilities. The model produces spatio-temporal patterns with strong resemblance to those observed experimentally and possesses many of the general features one desires for pattern classifiers: large information capacity, reliability, specific responses to specific inputs, and reduced sensitivity to initial conditions or influence of noise. This form of neural processing may thus describe the organiza...

  6. Salinas - An implicit finite element structural dynamics code developed for massively parallel platforms

    Energy Technology Data Exchange (ETDEWEB)

    BHARDWAJ, MANLJ K.; REESE,GARTH M.; DRIESSEN,BRIAN; ALVIN,KENNETH F.; DAY,DAVID M.

    2000-04-06

    As computational needs for structural finite element analysis increase, a robust implicit structural dynamics code is needed which can handle millions of degrees of freedom in the model and produce results with quick turn around time. A parallel code is needed to avoid limitations of serial platforms. Salinas is an implicit structural dynamics code specifically designed for massively parallel platforms. It computes the structural response of very large complex structures and provides solutions faster than any existing serial machine. This paper gives a current status of Salinas and uses demonstration problems to show Salinas' performance.

  7. Electromagnetic Gun Circuit Analysis Code (EGCAC)

    Science.gov (United States)

    Rolader, Glenn E.; Thornhill, Lindsey D.; Batteh, Jad H.; Scanlon, James J., III

    1993-01-01

    This paper describes a system engineering code that simulates the performance of a railgun/power supply system. The code, named EGCAC (Electromagnetic Gun Circuit Analysis Code), accounts for many performance degrading effects including viscous drag on the armature, viscous drag on the gas being pushed in front of the projectile, entrained gas that must be accelerated in front of the projectile, time-dependent rail resistance, armature resistance, system resistance, and ablation drag. EGCAC has been utilized to predict railgun performance up to a velocity of approximately 4 km/s for experiments at several laboratories. In this paper, the theory of EGCAC is described, and sample calculations are presented.

  8. Dynamic Contingency Analysis Tool

    Energy Technology Data Exchange (ETDEWEB)

    2016-01-14

    The Dynamic Contingency Analysis Tool (DCAT) is an open-platform and publicly available methodology to help develop applications that aim to improve the capabilities of power system planning engineers to assess the impact and likelihood of extreme contingencies and potential cascading events across their systems and interconnections. Outputs from the DCAT will help find mitigation solutions to reduce the risk of cascading outages in technically sound and effective ways. The current prototype DCAT implementation has been developed as a Python code that accesses the simulation functions of the Siemens PSS�E planning tool (PSS/E). It has the following features: It uses a hybrid dynamic and steady-state approach to simulating the cascading outage sequences that includes fast dynamic and slower steady-state events. It integrates dynamic models with protection scheme models for generation, transmission, and load. It models special protection systems (SPSs)/remedial action schemes (RASs) and automatic and manual corrective actions. Overall, the DCAT attempts to bridge multiple gaps in cascading-outage analysis in a single, unique prototype tool capable of automatically simulating and analyzing cascading sequences in real systems using multiprocessor computers.While the DCAT has been implemented using PSS/E in Phase I of the study, other commercial software packages with similar capabilities can be used within the DCAT framework.

  9. Dynamic Contingency Analysis Tool

    Energy Technology Data Exchange (ETDEWEB)

    2016-01-14

    The Dynamic Contingency Analysis Tool (DCAT) is an open-platform and publicly available methodology to help develop applications that aim to improve the capabilities of power system planning engineers to assess the impact and likelihood of extreme contingencies and potential cascading events across their systems and interconnections. Outputs from the DCAT will help find mitigation solutions to reduce the risk of cascading outages in technically sound and effective ways. The current prototype DCAT implementation has been developed as a Python code that accesses the simulation functions of the Siemens PSS/E planning tool (PSS/E). It has the following features: It uses a hybrid dynamic and steady-state approach to simulating the cascading outage sequences that includes fast dynamic and slower steady-state events. It integrates dynamic models with protection scheme models for generation, transmission, and load. It models special protection systems (SPSs)/remedial action schemes (RASs) and automatic and manual corrective actions. Overall, the DCAT attempts to bridge multiple gaps in cascading-outage analysis in a single, unique prototype tool capable of automatically simulating and analyzing cascading sequences in real systems using multiprocessor computers.While the DCAT has been implemented using PSS/E in Phase I of the study, other commercial software packages with similar capabilities can be used within the DCAT framework.

  10. Population Code Dynamics in Categorical Perception.

    Science.gov (United States)

    Tajima, Chihiro I; Tajima, Satohiro; Koida, Kowa; Komatsu, Hidehiko; Aihara, Kazuyuki; Suzuki, Hideyuki

    2016-03-03

    Categorical perception is a ubiquitous function in sensory information processing, and is reported to have important influences on the recognition of presented and/or memorized stimuli. However, such complex interactions among categorical perception and other aspects of sensory processing have not been explained well in a unified manner. Here, we propose a recurrent neural network model to process categorical information of stimuli, which approximately realizes a hierarchical Bayesian estimation on stimuli. The model accounts for a wide variety of neurophysiological and cognitive phenomena in a consistent framework. In particular, the reported complexity of categorical effects, including (i) task-dependent modulation of neural response, (ii) clustering of neural population representation, (iii) temporal evolution of perceptual color memory, and (iv) a non-uniform discrimination threshold, are explained as different aspects of a single model. Moreover, we directly examine key model behaviors in the monkey visual cortex by analyzing neural population dynamics during categorization and discrimination of color stimuli. We find that the categorical task causes temporally-evolving biases in the neuronal population representations toward the focal colors, which supports the proposed model. These results suggest that categorical perception can be achieved by recurrent neural dynamics that approximates optimal probabilistic inference in the changing environment.

  11. On the Organizational Dynamics of the Genetic Code

    KAUST Repository

    Zhang, Zhang

    2011-06-07

    The organization of the canonical genetic code needs to be thoroughly illuminated. Here we reorder the four nucleotides—adenine, thymine, guanine and cytosine—according to their emergence in evolution, and apply the organizational rules to devising an algebraic representation for the canonical genetic code. Under a framework of the devised code, we quantify codon and amino acid usages from a large collection of 917 prokaryotic genome sequences, and associate the usages with its intrinsic structure and classification schemes as well as amino acid physicochemical properties. Our results show that the algebraic representation of the code is structurally equivalent to a content-centric organization of the code and that codon and amino acid usages under different classification schemes were correlated closely with GC content, implying a set of rules governing composition dynamics across a wide variety of prokaryotic genome sequences. These results also indicate that codons and amino acids are not randomly allocated in the code, where the six-fold degenerate codons and their amino acids have important balancing roles for error minimization. Therefore, the content-centric code is of great usefulness in deciphering its hitherto unknown regularities as well as the dynamics of nucleotide, codon, and amino acid compositions.

  12. On the organizational dynamics of the genetic code.

    Science.gov (United States)

    Zhang, Zhang; Yu, Jun

    2011-04-01

    The organization of the canonical genetic code needs to be thoroughly illuminated. Here we reorder the four nucleotides-adenine, thymine, guanine and cytosine-according to their emergence in evolution, and apply the organizational rules to devising an algebraic representation for the canonical genetic code. Under a framework of the devised code, we quantify codon and amino acid usages from a large collection of 917 prokaryotic genome sequences, and associate the usages with its intrinsic structure and classification schemes as well as amino acid physicochemical properties. Our results show that the algebraic representation of the code is structurally equivalent to a content-centric organization of the code and that codon and amino acid usages under different classification schemes were correlated closely with GC content, implying a set of rules governing composition dynamics across a wide variety of prokaryotic genome sequences. These results also indicate that codons and amino acids are not randomly allocated in the code, where the six-fold degenerate codons and their amino acids have important balancing roles for error minimization. Therefore, the content-centric code is of great usefulness in deciphering its hitherto unknown regularities as well as the dynamics of nucleotide, codon, and amino acid compositions.

  13. Superimposed Code Theorectic Analysis of DNA Codes and DNA Computing

    Science.gov (United States)

    2010-03-01

    Bounds for DNA Codes Based on Fibonacci Ensembles of DNA Sequences ”, 2008 IEEE Proceedings of International Symposium on Information Theory, pp. 2292...5, June 2008, pp. 525-34. 32 28. A. Macula, et al., “Random Coding Bounds for DNA Codes Based on Fibonacci Ensembles of DNA Sequences ”, 2008...combinatorial method of bio-memory design and detection that encodes item or process information as numerical sequences represented in DNA. ComDMem is a

  14. Parallel Beam Dynamics Code Development for High Intensity Cyclotron

    Institute of Scientific and Technical Information of China (English)

    2008-01-01

    <正>1 Parallel PIC algorithm Self field solver is the key part of a high intensity beam dynamic PIC code which usually adopts the P-M (Particle-Mesh) method to solve the space charge. The P-M method is composed of four major

  15. Code Shift: Grid Specifications and Dynamic Wind Turbine Models

    DEFF Research Database (Denmark)

    Ackermann, Thomas; Ellis, Abraham; Fortmann, Jens

    2013-01-01

    Grid codes (GCs) and dynamic wind turbine (WT) models are key tools to allow increasing renewable energy penetration without challenging security of supply. In this article, the state of the art and the further development of both tools are discussed, focusing on the European and North American...

  16. Sending policies in dynamic wireless mesh using network coding

    DEFF Research Database (Denmark)

    Pandi, Sreekrishna; Fitzek, Frank; Pihl, Jeppe;

    2015-01-01

    This paper demonstrates the quick prototyping capabilities of the Python-Kodo library for network coding based performance evaluation and investigates the problem of data redundancy in a network coded wireless mesh with opportunistic overhearing. By means of several wireless meshed architectures...... of appropriate relays. Finally, various sending policies that can be employed by the nodes in order to improve the overall transmission efficiency in a dynamic wireless mesh network are discussed and their performance is analysed on the constructed simulation setup....... simulated on the constructed test-bed, the advantage of network coding over state of the art routing schemes and the challenges of this new technology are shown. By providing maximum control of the network coding parameters and the simulation environment to the user, the test-bed facilitates quick...

  17. Development of multi-physics code systems based on the reactor dynamics code DYN3D

    Energy Technology Data Exchange (ETDEWEB)

    Kliem, Soeren; Gommlich, Andre; Grahn, Alexander; Rohde, Ulrich [Helmholtz-Zentrum Dresden-Rossendorf e.V., Dresden (Germany); Schuetze, Jochen [ANSYS Germany GmbH, Darmstadt (Germany); Frank, Thomas [ANSYS Germany GmbH, Otterfing (Germany); Gomez Torres, Armando M.; Sanchez Espinoza, Victor Hugo [Karlsruher Institut fuer Technologie (KIT), Eggenstein-Leopoldshafen (Germany)

    2011-07-15

    The reactor dynamics code DYN3D has been coupled with the CFD code ANSYS CFX and the 3D thermal hydraulic core model FLICA4. In the coupling with ANSYS CFX, DYN3D calculates the neutron kinetics and the fuel behavior including the heat transfer to the coolant. The physical data interface between the codes is the volumetric heat release rate into the coolant. In the coupling with FLICA4 only the neutron kinetics module of DYN3D is used. Fluid dynamics and related transport phenomena in the reactor's coolant and fuel behavior is calculated by FLICA4. The correctness of the coupling of DYN3D with both thermal hydraulic codes was verified by the calculation of different test problems. These test problems were set-up in such a way that comparison with the DYN3D stand-alone code was possible. This included steady-state and transient calculations of a mini-core consisting of nine real-size PWR fuel assemblies with ANSYS CFX/DYN3D as well as mini-core and a full core steady-state calculation using FLICA4/DYN3D. (orig.)

  18. New Codes for Ambient Seismic Noise Analysis

    Science.gov (United States)

    Duret, F.; Mooney, W. D.; Detweiler, S.

    2007-12-01

    In order to determine a velocity model of the crust, scientists generally use earthquakes recorded by seismic stations. However earthquakes do not occur continuously and most are too weak to be useful. When no event is recorded, a waveform is generally considered to be noise. This noise, however, is not useless and carries a wealth of information. Thus, ambient seismic noise analysis is an inverse method of investigating the Earth's interior. Until recently, this technique was quite difficult to apply, as it requires significant computing capacities. In early 2007, however, a team led by Gregory Benson and Mike Ritzwoller from UC Boulder published a paper describing a new method for extracting group and phase velocities from those waveforms. The analysis consisting of recovering Green functions between a pair of stations, is composed of four steps: 1) single station data preparation, 2) cross-correlation and stacking, 3) quality control and data selection and 4) dispersion measurements. At the USGS, we developed a set of ready-to-use computing codes for analyzing waveforms to run the ambient noise analysis of Benson et al. (2007). Our main contribution to the analysis technique was to fully automate the process. The computation codes were written in Fortran 90 and the automation scripts were written in Perl. Furthermore, some operations were run with SAC. Our choices of programming language offer an opportunity to adapt our codes to the major platforms. The codes were developed under Linux but are meant to be adapted to Mac OS X and Windows platforms. The codes have been tested on Southern California data and our results compare nicely with those from the UC Boulder team. Next, we plan to apply our codes to Indonesian data, so that we might take advantage of newly upgraded seismic stations in that region.

  19. Static Code Analysis with Gitlab-CI

    CERN Document Server

    Datko, Szymon Tomasz

    2016-01-01

    Static Code Analysis is a simple but efficient way to ensure that application’s source code is free from known flaws and security vulnerabilities. Although such analysis tools are often coming with more advanced code editors, there are a lot of people who prefer less complicated environments. The easiest solution would involve education – where to get and how to use the aforementioned tools. However, counting on the manual usage of such tools still does not guarantee their actual usage. On the other hand, reducing the required effort, according to the idea “setup once, use anytime without sweat” seems like a more promising approach. In this paper, the approach to automate code scanning, within the existing CERN’s Gitlab installation, is described. For realization of that project, the Gitlab-CI service (the “CI” stands for "Continuous Integration"), with Docker assistance, was employed to provide a variety of static code analysers for different programming languages. This document covers the gene...

  20. Dynamic Analysis of Shells

    Directory of Open Access Journals (Sweden)

    Charles R. Steele

    1995-01-01

    Full Text Available Shell structures are indispensable in virtually every industry. However, in the design, analysis, fabrication, and maintenance of such structures, there are many pitfalls leading to various forms of disaster. The experience gained by engineers over some 200 years of disasters and brushes with disaster is expressed in the extensive archival literature, national codes, and procedural documentation found in larger companies. However, the advantage of the richness in the behavior of shells is that the way is always open for innovation. In this survey, we present a broad overview of the dynamic response of shell structures. The intention is to provide an understanding of the basic themes behind the detailed codes and stimulate, not restrict, positive innovation. Such understanding is also crucial for the correct computation of shell structures by any computer code. The physics dictates that the thin shell structure offers a challenge for analysis and computation. Shell response can be generally categorized by states of extension, inextensional bending, edge bending, and edge transverse shear. Simple estimates for the magnitudes of stress, deformation, and resonance in the extensional and inextensional states are provided by ring response. Several shell examples demonstrate the different states and combinations. For excitation frequency above the extensional resonance, such as in impact and acoustic excitation, a fine mesh is needed over the entire shell surface. For this range, modal and implicit methods are of limited value. The example of a sphere impacting a rigid surface shows that plastic unloading occurs continuously. Thus, there are no short cuts; the complete material behavior must be included.

  1. Transient Ejector Analysis (TEA) code user's guide

    Science.gov (United States)

    Drummond, Colin K.

    1993-01-01

    A FORTRAN computer program for the semi analytic prediction of unsteady thrust augmenting ejector performance has been developed, based on a theoretical analysis for ejectors. That analysis blends classic self-similar turbulent jet descriptions with control-volume mixing region elements. Division of the ejector into an inlet, diffuser, and mixing region allowed flexibility in the modeling of the physics for each region. In particular, the inlet and diffuser analyses are simplified by a quasi-steady-analysis, justified by the assumption that pressure is the forcing function in those regions. Only the mixing region is assumed to be dominated by viscous effects. The present work provides an overview of the code structure, a description of the required input and output data file formats, and the results for a test case. Since there are limitations to the code for applications outside the bounds of the test case, the user should consider TEA as a research code (not as a production code), designed specifically as an implementation of the proposed ejector theory. Program error flags are discussed, and some diagnostic routines are presented.

  2. Sandia National Laboratories analysis code data base

    Energy Technology Data Exchange (ETDEWEB)

    Peterson, C.W.

    1994-11-01

    Sandia National Laboratories, mission is to solve important problems in the areas of national defense, energy security, environmental integrity, and industrial technology. The Laboratories` strategy for accomplishing this mission is to conduct research to provide an understanding of the important physical phenomena underlying any problem, and then to construct validated computational models of the phenomena which can be used as tools to solve the problem. In the course of implementing this strategy, Sandia`s technical staff has produced a wide variety of numerical problem-solving tools which they use regularly in the design, analysis, performance prediction, and optimization of Sandia components, systems and manufacturing processes. This report provides the relevant technical and accessibility data on the numerical codes used at Sandia, including information on the technical competency or capability area that each code addresses, code ``ownership`` and release status, and references describing the physical models and numerical implementation.

  3. Development of Tritium Permeation Analysis Code (TPAC)

    Energy Technology Data Exchange (ETDEWEB)

    Eung S. Kim; Chang H. Oh; Mike Patterson

    2010-10-01

    Idaho National Laboratory developed the Tritium Permeation Analysis Code (TPAC) for tritium permeation in the Very High Temperature Gas Cooled Reactor (VHTR). All the component models in the VHTR were developed and were embedded into the MATHLAB SIMULINK package with a Graphic User Interface. The governing equations of the nuclear ternary reaction and thermal neutron capture reactions from impurities in helium and graphite core, reflector, and control rods were implemented. The TPAC code was verified using analytical solutions for the tritium birth rate from the ternary fission, the birth rate from 3He, and the birth rate from 10B. This paper also provides comparisons of the TPAC with the existing other codes. A VHTR reference design was selected for tritium permeation study from the reference design to the nuclear-assisted hydrogen production plant and some sensitivity study results are presented based on the HTGR outlet temperature of 750 degrees C.

  4. Development of dynamic simulation code for fuel cycle fusion reactor

    Energy Technology Data Exchange (ETDEWEB)

    Aoki, Isao; Seki, Yasushi [Department of Fusion Engineering Research, Naka Fusion Research Establishment, Japan Atomic Energy Research Institute, Naka, Ibaraki (Japan); Sasaki, Makoto; Shintani, Kiyonori; Kim, Yeong-Chan

    1999-02-01

    A dynamic simulation code for fuel cycle of a fusion experimental reactor has been developed. The code follows the fuel inventory change with time in the plasma chamber and the fuel cycle system during 2 days pulse operation cycles. The time dependence of the fuel inventory distribution is evaluated considering the fuel burn and exhaust in the plasma chamber, purification and supply functions. For each subsystem of the plasma chamber and the fuel cycle system, the fuel inventory equation is written based on the equation of state considering the fuel burn and the function of exhaust, purification, and supply. The processing constants of subsystem for steady states were taken from the values in the ITER Conceptual Design Activity (CDA) report. Using this code, the time dependence of the fuel supply and inventory depending on the burn state and subsystem processing functions are shown. (author)

  5. Development of computer code in PNC, 8; Development and using of Dynamic Analysis Program for Base Isolation System (DAP-BS)

    Energy Technology Data Exchange (ETDEWEB)

    Ohhira, Mitsuru (Power Reactor and Nuclear Fuel Development Corp., Tokyo (Japan))

    1990-12-01

    Private buildings applied base isolation system, are on the practical stage now. So, under Construction and Maintenance Management Office, we are doing an application study of base isolation system to nuclear fuel facilities. On the process of this study, we have developed Dynamic Analysis Program-Base Isolation System (DAP-BS) which is able to run a 32-bit personal computer. Using this program, we can analyze a 3-dimensional structure, and evaluate the various properties of base isolation parts that are divided into maximum 16 blocks. And from the results of some simulation analyses, we thought that DAP-BS had good reliability and marketability. So, we put DAP-BS on the market. (author).

  6. Analysis of SMA Hybrid Composite Structures using Commercial Codes

    Science.gov (United States)

    Turner, Travis L.; Patel, Hemant D.

    2004-01-01

    A thermomechanical model for shape memory alloy (SMA) actuators and SMA hybrid composite (SMAHC) structures has been recently implemented in the commercial finite element codes MSC.Nastran and ABAQUS. The model may be easily implemented in any code that has the capability for analysis of laminated composite structures with temperature dependent material properties. The model is also relatively easy to use and requires input of only fundamental engineering properties. A brief description of the model is presented, followed by discussion of implementation and usage in the commercial codes. Results are presented from static and dynamic analysis of SMAHC beams of two types; a beam clamped at each end and a cantilevered beam. Nonlinear static (post-buckling) and random response analyses are demonstrated for the first specimen. Static deflection (shape) control is demonstrated for the cantilevered beam. Approaches for modeling SMAHC material systems with embedded SMA in ribbon and small round wire product forms are demonstrated and compared. The results from the commercial codes are compared to those from a research code as validation of the commercial implementations; excellent correlation is achieved in all cases.

  7. Fractal methods in image analysis and coding

    OpenAIRE

    Neary, David

    2001-01-01

    In this thesis we present an overview of image processing techniques which use fractal methods in some way. We show how these fields relate to each other, and examine various aspects of fractal methods in each area. The three principal fields of image processing and analysis th a t we examine are texture classification, image segmentation and image coding. In the area of texture classification, we examine fractal dimension estimators, comparing these methods to other methods in use, a...

  8. Methodology for computational fluid dynamics code verification/validation

    Energy Technology Data Exchange (ETDEWEB)

    Oberkampf, W.L.; Blottner, F.G.; Aeschliman, D.P.

    1995-07-01

    The issues of verification, calibration, and validation of computational fluid dynamics (CFD) codes has been receiving increasing levels of attention in the research literature and in engineering technology. Both CFD researchers and users of CFD codes are asking more critical and detailed questions concerning the accuracy, range of applicability, reliability and robustness of CFD codes and their predictions. This is a welcomed trend because it demonstrates that CFD is maturing from a research tool to the world of impacting engineering hardware and system design. In this environment, the broad issue of code quality assurance becomes paramount. However, the philosophy and methodology of building confidence in CFD code predictions has proven to be more difficult than many expected. A wide variety of physical modeling errors and discretization errors are discussed. Here, discretization errors refer to all errors caused by conversion of the original partial differential equations to algebraic equations, and their solution. Boundary conditions for both the partial differential equations and the discretized equations will be discussed. Contrasts are drawn between the assumptions and actual use of numerical method consistency and stability. Comments are also made concerning the existence and uniqueness of solutions for both the partial differential equations and the discrete equations. Various techniques are suggested for the detection and estimation of errors caused by physical modeling and discretization of the partial differential equations.

  9. Parallelization and performance tuning of molecular dynamics code with OpenMP

    Institute of Scientific and Technical Information of China (English)

    2006-01-01

    An OpenMP approach was proposed to parallelize the sequential molecular dynamics (MD) code on shared memory machines. When a code is converted from the sequential form to the parallel form, data dependence is a main problem. A traditional sequential molecular dynamics code is anatomized to find the data dependence segments in it, and the two different methods, i.e. , recover method and backward mapping method were used to eliminate those data dependencies in order to realize the parallelization of this sequential MD code. The performance of the parallelized MD code was analyzed by using some performance analysis tools. The results of the test show that the computing size of this code increases sharply form 1 million atoms before parallelization to 20 million atoms after parallelization, and the wall clock during computing is reduced largely. Some hot-spots in this code are found and optimized by improved algorithm. The efficiency of parallel computing is 30% higher than that of before, and the calculation time is saved and larger scale calculation problems are solved.

  10. AnDa:a Dynamic Analysis System for Malicious Code%AnDa:恶意代码动态分析系统

    Institute of Scientific and Technical Information of China (English)

    任伟; 柳坤; 周金

    2014-01-01

    近年来,移动终端崛起迅速,其功能已扩展到商务应用等领域,与用户的隐私、财产等信息关系紧密。静态监控已无法满足人们对应用软件安全使用的需求,采用动态监控沙盒分析可以实时监控应用程序,具有速度快、准确性好、安全性高、可行性强的特性。针对Android平台下恶意软件在后台获取用户隐私信息,如获取用户数据并发送到网络端、拦截和窥探用户电话和短信等问题,提出一套采用动态检测沙盒分析技术记录Android恶意软件敏感行为的方案及系统--AnDa,详细描述了该系统总体设计和关键技术,实现了对访问电话、短信、位置信息、手机SIM卡信息等行为的实时监控,并在虚拟机和实体机上测试了AnDa系统。该作品采用动态监控沙盒分析技术,实现了在Android平台下软件动态监控和行为分析,并且实现了对Android框架层API的Java Method Hook和常见的恶意软件特征的有效监控。它可以在Android 4.0以上的设备上使用,可以根据监控到的应用软件恶意行为信息,判定所属恶意软件的类型,使得更加迅速发现新型病毒和更加隐蔽的病毒模型,从而更好地保护手机以及个人重要的数据,极大地提高了安全性。%Recently, mobile terminals have been extended to business applications rapidly, and have been more closely related to user privacy and property. As static monitoring cannot guarantee software security, the analysis of dynamic monitoring sandbox can realize real-time monitoring in a faster, more accurate, safer, and high feasible manner. The problem of privacy leakage exists in Android platform malware, such as accessing user data and exposing them to networks, or intercepting and spying on phone calls and short text messages. Thus, this article proposes a solution system called AnDa, which records sensitive behavior of Android malwares using dynamic

  11. On the use of ZIP codes and ZIP code tabulation areas (ZCTAs for the spatial analysis of epidemiological data

    Directory of Open Access Journals (Sweden)

    Matisziw Timothy C

    2006-12-01

    Full Text Available Abstract Background While the use of spatially referenced data for the analysis of epidemiological data is growing, issues associated with selecting the appropriate geographic unit of analysis are also emerging. A particularly problematic unit is the ZIP code. Lacking standardization and highly dynamic in structure, the use of ZIP codes and ZIP code tabulation areas (ZCTA for the spatial analysis of disease present a unique challenge to researchers. Problems associated with these units for detecting spatial patterns of disease are explored. Results A brief review of ZIP codes and their spatial representation is conducted. Though frequently represented as polygons to facilitate analysis, ZIP codes are actually defined at a narrower spatial resolution reflecting the street addresses they serve. This research shows that their generalization as continuous regions is an imposed structure that can have serious implications in the interpretation of research results. ZIP codes areas and Census defined ZCTAs, two commonly used polygonal representations of ZIP code address ranges, are examined in an effort to identify the spatial statistical sensitivities that emerge given differences in how these representations are defined. Here, comparative analysis focuses on the detection of patterns of prostate cancer in New York State. Of particular interest for studies utilizing local, spatial statistical tests, is that differences in the topological structures of ZIP code areas and ZCTAs give rise to different spatial patterns of disease. These differences are related to the different methodologies used in the generalization of ZIP code information. Given the difficulty associated with generating ZIP code boundaries, both ZIP code areas and ZCTAs contain numerous representational errors which can have a significant impact on spatial analysis. While the use of ZIP code polygons for spatial analysis is relatively straightforward, ZCTA representations contain

  12. A proposed methodology for computational fluid dynamics code verification, calibration, and validation

    Science.gov (United States)

    Aeschliman, D. P.; Oberkampf, W. L.; Blottner, F. G.

    Verification, calibration, and validation (VCV) of Computational Fluid Dynamics (CFD) codes is an essential element of the code development process. The exact manner in which code VCV activities are planned and conducted, however, is critically important. It is suggested that the way in which code validation, in particular, is often conducted--by comparison to published experimental data obtained for other purposes--is in general difficult and unsatisfactory, and that a different approach is required. This paper describes a proposed methodology for CFD code VCV that meets the technical requirements and is philosophically consistent with code development needs. The proposed methodology stresses teamwork and cooperation between code developers and experimentalists throughout the VCV process, and takes advantage of certain synergisms between CFD and experiment. A novel approach to uncertainty analysis is described which can both distinguish between and quantify various types of experimental error, and whose attributes are used to help define an appropriate experimental design for code VCV experiments. The methodology is demonstrated with an example of laminar, hypersonic, near perfect gas, 3-dimensional flow over a sliced sphere/cone of varying geometrical complexity.

  13. A Dynamic Programming Approach To Length-Limited Huffman Coding

    CERN Document Server

    Golin, Mordecai

    2008-01-01

    The ``state-of-the-art'' in Length Limited Huffman Coding algorithms is the $\\Theta(ND)$-time, $\\Theta(N)$-space one of Hirschberg and Larmore, where $D\\le N$ is the length restriction on the code. This is a very clever, very problem specific, technique. In this note we show that there is a simple Dynamic-Programming (DP) method that solves the problem with the same time and space bounds. The fact that there was an $\\Theta(ND)$ time DP algorithm was previously known; it is a straightforward DP with the Monge property (which permits an order of magnitude speedup). It was not interesting, though, because it also required $\\Theta(ND)$ space. The main result of this paper is the technique developed for reducing the space. It is quite simple and applicable to many other problems modeled by DPs with the Monge property. We illustrate this with examples from web-proxy design and wireless mobile paging.

  14. Mechanic: The MPI/HDF code framework for dynamical astronomy

    Science.gov (United States)

    Słonina, Mariusz; Goździewski, Krzysztof; Migaszewski, Cezary

    2015-01-01

    We introduce the Mechanic, a new open-source code framework. It is designed to reduce the development effort of scientific applications by providing unified API (Application Programming Interface) for configuration, data storage and task management. The communication layer is based on the well-established Message Passing Interface (MPI) standard, which is widely used on variety of parallel computers and CPU-clusters. The data storage is performed within the Hierarchical Data Format (HDF5). The design of the code follows core-module approach which allows to reduce the user’s codebase and makes it portable for single- and multi-CPU environments. The framework may be used in a local user’s environment, without administrative access to the cluster, under the PBS or Slurm job schedulers. It may become a helper tool for a wide range of astronomical applications, particularly focused on processing large data sets, such as dynamical studies of long-term orbital evolution of planetary systems with Monte Carlo methods, dynamical maps or evolutionary algorithms. It has been already applied in numerical experiments conducted for Kepler-11 (Migaszewski et al., 2012) and νOctantis planetary systems (Goździewski et al., 2013). In this paper we describe the basics of the framework, including code listings for the implementation of a sample user’s module. The code is illustrated on a model Hamiltonian introduced by (Froeschlé et al., 2000) presenting the Arnold diffusion. The Arnold web is shown with the help of the MEGNO (Mean Exponential Growth of Nearby Orbits) fast indicator (Goździewski et al., 2008a) applied onto symplectic SABAn integrators family (Laskar and Robutel, 2001).

  15. Code Verification of the HIGRAD Computational Fluid Dynamics Solver

    Energy Technology Data Exchange (ETDEWEB)

    Van Buren, Kendra L. [Los Alamos National Laboratory; Canfield, Jesse M. [Los Alamos National Laboratory; Hemez, Francois M. [Los Alamos National Laboratory; Sauer, Jeremy A. [Los Alamos National Laboratory

    2012-05-04

    The purpose of this report is to outline code and solution verification activities applied to HIGRAD, a Computational Fluid Dynamics (CFD) solver of the compressible Navier-Stokes equations developed at the Los Alamos National Laboratory, and used to simulate various phenomena such as the propagation of wildfires and atmospheric hydrodynamics. Code verification efforts, as described in this report, are an important first step to establish the credibility of numerical simulations. They provide evidence that the mathematical formulation is properly implemented without significant mistakes that would adversely impact the application of interest. Highly accurate analytical solutions are derived for four code verification test problems that exercise different aspects of the code. These test problems are referred to as: (i) the quiet start, (ii) the passive advection, (iii) the passive diffusion, and (iv) the piston-like problem. These problems are simulated using HIGRAD with different levels of mesh discretization and the numerical solutions are compared to their analytical counterparts. In addition, the rates of convergence are estimated to verify the numerical performance of the solver. The first three test problems produce numerical approximations as expected. The fourth test problem (piston-like) indicates the extent to which the code is able to simulate a 'mild' discontinuity, which is a condition that would typically be better handled by a Lagrangian formulation. The current investigation concludes that the numerical implementation of the solver performs as expected. The quality of solutions is sufficient to provide credible simulations of fluid flows around wind turbines. The main caveat associated to these findings is the low coverage provided by these four problems, and somewhat limited verification activities. A more comprehensive evaluation of HIGRAD may be beneficial for future studies.

  16. Understanding Code Patterns - Analysis, Interpretation & Measurement

    CERN Document Server

    Dundas, Jitesh

    2011-01-01

    This research paper aims to find, analyze and understand code patterns in any software system and measure its quality by defining standards and proposing a formula for the same. Every code that is written can be divided into different code segments, each having its own impact on the overall system. We can analyze these code segments to get the code quality. The measures used in this paper include Lines of Code, Number of calls made by a module, Execution time, the system knowledge of user and developers, the use of generalization, inheritance, reusability and other object-oriented concepts. The entire software code is divided into code snippets, based on the logic that they implement. Each of these code snippets has an impact. This measure is called Impact Factor and is valued by the software developer and/or other system stakeholders. Efficiency = (Code Area / Execution Time) * Qr

  17. Dynamic Texture Coding using Modified Haar Wavelet with CUDA Premanand

    Directory of Open Access Journals (Sweden)

    P Ghadekar

    2015-03-01

    Full Text Available Texture is an image having repetition of patterns. There are two types, static and dynamic texture. Static texture is an image having repetitions of patterns in the spatial domain. Dynamic texture is number of frames having repetitions in spatial and temporal domain. This paper introduces a novel method for dynamic texture coding to achieve higher compression ratio of dynamic texture using 2D-modified Haar wavelet transform. The dynamic texture video contains high redundant parts in spatial and temporal domain. Redundant parts can be removed to achieve high compression ratios with better visual quality. The modified Haar wavelet is used to exploit spatial and temporal correlations amongst the pixels. The YCbCr color model is used to exploit chromatic components as HVS is less sensitive to chrominance. To decrease the time complexity of algorithm parallel programming is done using CUDA (Compute Unified Device Architecture. GPU contains the number of cores as compared to CPU, which is utilized to reduce the time complexity of algorithms.

  18. Digital Image Analysis for Detechip Code Determination

    Directory of Open Access Journals (Sweden)

    Marcus Lyon

    2012-08-01

    Full Text Available DETECHIP® is a molecular sensing array used for identification of a large variety of substances. Previous methodology for the analysis of DETECHIP® used human vision to distinguish color changes induced by the presence of the analyte of interest. This paper describes several analysis techniques using digital images of DETECHIP® . Both a digital camera and flatbed desktop photo scanner were used to obtain Jpeg images. Color information within these digital images was obtained through the measurement of redgreen-blue (RGB values using software such as GIMP, Photoshop and ImageJ. Several different techniques were used to evaluate these color changes. It was determined that the flatbed scanner produced in the clearest and more reproducible images. Furthermore, codes obtained using a macro written for use within ImageJ showed improved consistency versus pervious methods.

  19. Learning to Estimate Dynamical State with Probabilistic Population Codes.

    Directory of Open Access Journals (Sweden)

    Joseph G Makin

    2015-11-01

    Full Text Available Tracking moving objects, including one's own body, is a fundamental ability of higher organisms, playing a central role in many perceptual and motor tasks. While it is unknown how the brain learns to follow and predict the dynamics of objects, it is known that this process of state estimation can be learned purely from the statistics of noisy observations. When the dynamics are simply linear with additive Gaussian noise, the optimal solution is the well known Kalman filter (KF, the parameters of which can be learned via latent-variable density estimation (the EM algorithm. The brain does not, however, directly manipulate matrices and vectors, but instead appears to represent probability distributions with the firing rates of population of neurons, "probabilistic population codes." We show that a recurrent neural network-a modified form of an exponential family harmonium (EFH-that takes a linear probabilistic population code as input can learn, without supervision, to estimate the state of a linear dynamical system. After observing a series of population responses (spike counts to the position of a moving object, the network learns to represent the velocity of the object and forms nearly optimal predictions about the position at the next time-step. This result builds on our previous work showing that a similar network can learn to perform multisensory integration and coordinate transformations for static stimuli. The receptive fields of the trained network also make qualitative predictions about the developing and learning brain: tuning gradually emerges for higher-order dynamical states not explicitly present in the inputs, appearing as delayed tuning for the lower-order states.

  20. Porting of a serial molecular dynamics code on MIMD platforms

    Energy Technology Data Exchange (ETDEWEB)

    Celino, M. [ENEA Centro Ricerche Casaccia, S. Maria di Galeria, RM (Italy). HPCN Project

    1999-07-01

    A molecular dynamics (MD) code, utilized for the study of atomistic models of metallic systems has been parallelized for MIMD (multiple instructions multiple data) parallel platforms by means of the parallel virtual machine (PVM) message passing library. Since the parallelization implies modifications of the sequential algorithms, these are described from the point of view of the statistical mechanical theory. Furthermore, techniques and parallelization strategies utilized and the MD parallel code are described in detail. Benchmarks on several MIMD platforms (IBM SP1, SP2, Cray T3D, cluster of workstations) allow performances evaluation of the code versus the different characteristics of the parallel platforms. [Italian] Un codice seriale di dinamica molecolare (MD) utilizzato per lo studio di modelli atomici di materiali metallici e' stato parallelizzato per piattaforme parallele MIMD (multiple instructions multiple data) utilizzando librerie del parallel virtual machine (PVM). Poiche' l'operazione di parallelizzazione ha implicato la modifica degli algoritmi seriali del codice, questi vengono descritti ripercorrendo i concetti fondamentali della meccanica statistica. Inoltre sono presentate le tecniche e le strategie di parallelizzazione utilizzate descrivendo in dettaglio il codice parallelo di MD: Risultati di benchmark su diverse piattaforme MIMD (IBM SP1, SP2, Cray T3D, cluster of workstations) permettono di analizzare le performances del codice in funzione delle differenti caratteristiche delle piattaforme parallele.

  1. Error-correction coding and decoding bounds, codes, decoders, analysis and applications

    CERN Document Server

    Tomlinson, Martin; Ambroze, Marcel A; Ahmed, Mohammed; Jibril, Mubarak

    2017-01-01

    This book discusses both the theory and practical applications of self-correcting data, commonly known as error-correcting codes. The applications included demonstrate the importance of these codes in a wide range of everyday technologies, from smartphones to secure communications and transactions. Written in a readily understandable style, the book presents the authors’ twenty-five years of research organized into five parts: Part I is concerned with the theoretical performance attainable by using error correcting codes to achieve communications efficiency in digital communications systems. Part II explores the construction of error-correcting codes and explains the different families of codes and how they are designed. Techniques are described for producing the very best codes. Part III addresses the analysis of low-density parity-check (LDPC) codes, primarily to calculate their stopping sets and low-weight codeword spectrum which determines the performance of these codes. Part IV deals with decoders desi...

  2. SPH code for dynamical and chemical evolution of disk galaxies

    CERN Document Server

    Berczik, P

    1998-01-01

    The problem of chemical and dynamical evolution of galaxies is one of the most attracting and complex problems of modern astrophysics. Within the framework of the given work the standard dynamic Smoothed Particle Hydrodynamics (SPH) code (Monaghan J.J. 1992, ARAA, 30, 543) is noticeably expanded. Our investigation concernes with the changes and incorporation of new ideas into the algorithmic inclusion of Star Formation (SF) and Super Novae (SN) explosions in SPH (Berczik P. & Kravchuk S.G., 1996, ApSpSci, 245, 27). The proposed energy criterion for definition of a place and efficiency of SF results in the successfully explain Star Formation History (SFH) in isolated galaxies of different types. On the base of original ideas we expand a code in a more realistic way of the description of effects of return of a hot, chemical enriched gas in Interstellar Matter (ISM). In addition to the account of SNII, we offer the self-agreed account of SNIa and PN. This allows to describe not only the ISM content of $ O^{1...

  3. Dynamic Model on the Transmission of Malicious Codes in Network

    Directory of Open Access Journals (Sweden)

    Bimal Kumar Mishra

    2013-08-01

    Full Text Available This paper introduces differential susceptible e-epidemic model S_i IR (susceptible class-1 for virus (S1 - susceptible class-2 for worms (S2 -susceptible class-3 for Trojan horse (S3 – infectious (I – recovered (R for the transmission of malicious codes in a computer network. We derive the formula for reproduction number (R0 to study the spread of malicious codes in computer network. We show that the Infectious free equilibrium is globally asymptotically stable and endemic equilibrium is locally asymptotically sable when reproduction number is less than one. Also an analysis has been made on the effect of antivirus software in the infectious nodes. Numerical methods are employed to solve and simulate the system of equations developed.

  4. Analysis of quantum error-correcting codes: Symplectic lattice codes and toric codes

    Science.gov (United States)

    Harrington, James William

    Quantum information theory is concerned with identifying how quantum mechanical resources (such as entangled quantum states) can be utilized for a number of information processing tasks, including data storage, computation, communication, and cryptography. Efficient quantum algorithms and protocols have been developed for performing some tasks (e.g. , factoring large numbers, securely communicating over a public channel, and simulating quantum mechanical systems) that appear to be very difficult with just classical resources. In addition to identifying the separation between classical and quantum computational power, much of the theoretical focus in this field over the last decade has been concerned with finding novel ways of encoding quantum information that are robust against errors, which is an important step toward building practical quantum information processing devices. In this thesis I present some results on the quantum error-correcting properties of oscillator codes (also described as symplectic lattice codes) and toric codes. Any harmonic oscillator system (such as a mode of light) can be encoded with quantum information via symplectic lattice codes that are robust against shifts in the system's continuous quantum variables. I show the existence of lattice codes whose achievable rates match the one-shot coherent information over the Gaussian quantum channel. Also, I construct a family of symplectic self-dual lattices and search for optimal encodings of quantum information distributed between several oscillators. Toric codes provide encodings of quantum information into two-dimensional spin lattices that are robust against local clusters of errors and which require only local quantum operations for error correction. Numerical simulations of this system under various error models provide a calculation of the accuracy threshold for quantum memory using toric codes, which can be related to phase transitions in certain condensed matter models. I also present

  5. SINFAC - SYSTEMS IMPROVED NUMERICAL FLUIDS ANALYSIS CODE

    Science.gov (United States)

    Costello, F. A.

    1994-01-01

    The Systems Improved Numerical Fluids Analysis Code, SINFAC, consists of additional routines added to the April 1983 revision of SINDA, a general thermal analyzer program. The purpose of the additional routines is to allow for the modeling of active heat transfer loops. The modeler can simulate the steady-state and pseudo-transient operations of 16 different heat transfer loop components including radiators, evaporators, condensers, mechanical pumps, reservoirs and many types of valves and fittings. In addition, the program contains a property analysis routine that can be used to compute the thermodynamic properties of 20 different refrigerants. SINFAC can simulate the response to transient boundary conditions. SINFAC was first developed as a method for computing the steady-state performance of two phase systems. It was then modified using CNFRWD, SINDA's explicit time-integration scheme, to accommodate transient thermal models. However, SINFAC cannot simulate pressure drops due to time-dependent fluid acceleration, transient boil-out, or transient fill-up, except in the accumulator. SINFAC also requires the user to be familiar with SINDA. The solution procedure used by SINFAC is similar to that which an engineer would use to solve a system manually. The solution to a system requires the determination of all of the outlet conditions of each component such as the flow rate, pressure, and enthalpy. To obtain these values, the user first estimates the inlet conditions to the first component of the system, then computes the outlet conditions from the data supplied by the manufacturer of the first component. The user then estimates the temperature at the outlet of the third component and computes the corresponding flow resistance of the second component. With the flow resistance of the second component, the user computes the conditions down stream, namely the inlet conditions of the third. The computations follow for the rest of the system, back to the first component

  6. Dynamic clustering of distributed source coding in wireless sensor networks

    Institute of Scientific and Technical Information of China (English)

    LIU Bing

    2009-01-01

    There are correlations of data in adjacent sensor nodes in wireless sensor networks (WSNs). Distributed source coding (DSC) is an idea to improve the energy efficiency in WSNs by compressing the sensor data with correlations to others. When utilizing the DSC, the network architecture that, deciding which nodes to transmit the side information and which nodes to compress according to the correlations, influences the compression efficiency significantly. Comparing with former schemes that have no adaptations, a dynamic clustering scheme is presented in this article, with which the network is partitioned to clusters adaptive to the topology and the degree of correlations. The simulation indicates that the proposed scheme has higher efficiency than static clustering schemes.

  7. Mechanic: the MPI/HDF code framework for dynamical astronomy

    CERN Document Server

    Słonina, Mariusz; Migaszewski, Cezary

    2014-01-01

    We introduce the Mechanic, a new open-source code framework. It is designed to reduce the development effort of scientific applications by providing unified API (Application Programming Interface) for configuration, data storage and task management. The communication layer is based on the well-established Message Passing Interface (MPI) standard, which is widely used on variety of parallel computers and CPU-clusters. The data storage is performed within the Hierarchical Data Format (HDF5). The design of the code follows em core-module approach which allows to reduce the user's codebase and makes it portable for single- and multi-CPU environments. The framework may be used in a local user's environment, without administrative access to the cluster, under the PBS or Slurm job schedulers. It may become a helper tool for a wide range of astronomical applications, particularly focused on processing large data sets, such as dynamical studies of long-term orbital evolution of planetary systems with Monte Carlo metho...

  8. Users' guide to CACECO containment analysis code. [LMFBR

    Energy Technology Data Exchange (ETDEWEB)

    Peak, R.D.

    1979-06-01

    The CACECO containment analysis code was developed to predict the thermodynamic responses of LMFBR containment facilities to a variety of accidents. The code is included in the National Energy Software Center Library at Argonne National Laboratory as Program No. 762. This users' guide describes the CACECO code and its data input requirements. The code description covers the many mathematical models used and the approximations used in their solution. The descriptions are detailed to the extent that the user can modify the code to suit his unique needs, and, indeed, the reader is urged to consider code modification acceptable.

  9. Overhead analysis of universal concatenated quantum codes

    Science.gov (United States)

    Chamberland, Christopher; Jochym-O'Connor, Tomas; Laflamme, Raymond

    2017-02-01

    We analyze the resource overhead of recently proposed methods for universal fault-tolerant quantum computation using concatenated codes. Namely, we examine the concatenation of the 7-qubit Steane code with the 15-qubit Reed-Muller code, which allows for the construction of the 49- and 105-qubit codes that do not require the need for magic state distillation for universality. We compute a lower bound for the adversarial noise threshold of the 105-qubit code and find it to be 8.33 ×10-6. We obtain a depolarizing noise threshold for the 49-qubit code of 9.69 ×10-4 which is competitive with the 105-qubit threshold result of 1.28 ×10-3 . We then provide lower bounds on the resource requirements of the 49- and 105-qubit codes and compare them with the surface code implementation of a logical T gate using magic state distillation. For the sampled input error rates and noise model, we find that the surface code achieves a smaller overhead compared to our concatenated schemes.

  10. Development of a dynamic coupled hydro-geomechanical code and its application to induced seismicity

    Science.gov (United States)

    Miah, Md Mamun

    This research describes the importance of a hydro-geomechanical coupling in the geologic sub-surface environment from fluid injection at geothermal plants, large-scale geological CO2 sequestration for climate mitigation, enhanced oil recovery, and hydraulic fracturing during wells construction in the oil and gas industries. A sequential computational code is developed to capture the multiphysics interaction behavior by linking a flow simulation code TOUGH2 and a geomechanics modeling code PyLith. Numerical formulation of each code is discussed to demonstrate their modeling capabilities. The computational framework involves sequential coupling, and solution of two sub-problems- fluid flow through fractured and porous media and reservoir geomechanics. For each time step of flow calculation, pressure field is passed to the geomechanics code to compute effective stress field and fault slips. A simplified permeability model is implemented in the code that accounts for the permeability of porous and saturated rocks subject to confining stresses. The accuracy of the TOUGH-PyLith coupled simulator is tested by simulating Terzaghi's 1D consolidation problem. The modeling capability of coupled poroelasticity is validated by benchmarking it against Mandel's problem. The code is used to simulate both quasi-static and dynamic earthquake nucleation and slip distribution on a fault from the combined effect of far field tectonic loading and fluid injection by using an appropriate fault constitutive friction model. Results from the quasi-static induced earthquake simulations show a delayed response in earthquake nucleation. This is attributed to the increased total stress in the domain and not accounting for pressure on the fault. However, this issue is resolved in the final chapter in simulating a single event earthquake dynamic rupture. Simulation results show that fluid pressure has a positive effect on slip nucleation and subsequent crack propagation. This is confirmed by

  11. The importance of including dynamic soil-structure interaction into wind turbine simulation codes

    DEFF Research Database (Denmark)

    Damgaard, Mads; Andersen, Lars Vabbersgaard; Ibsen, Lars Bo

    2014-01-01

    is examined. The optimal order of the models is determined and implemented into the aeroelastic code HAWC2, where the dynamic response of a 5.0 MW wind turbine is evaluated. In contrast to the fore-aft vibrations, the inclusion of soil-structure interaction is shown to be critical for the side-side vibrations......A rigorous numerical model, describing a wind turbine structure and subsoil, may contain thousands of degrees of freedom, making the approach computationally inefficient for fast time domain analysis. In order to meet the requirements of real-time calculations, the dynamic impedance...

  12. The analysis of thermal-hydraulic models in MELCOR code

    Energy Technology Data Exchange (ETDEWEB)

    Kim, M. H.; Hur, C.; Kim, D. K.; Cho, H. J. [POhang Univ., of Science and TECHnology, Pohang (Korea, Republic of)

    1996-07-15

    The objective of the present work is to verify the prediction and analysis capability of MELCOR code about the progression of severe accidents in light water reactor and also to evaluate appropriateness of thermal-hydraulic models used in MELCOR code. Comparing the results of experiment and calculation with MELCOR code is carried out to achieve the above objective. Specially, the comparison between the CORA-13 experiment and the MELCOR code calculation was performed.

  13. LPNORM: A linear programming normative analysis code

    Science.gov (United States)

    de Caritat, Patrice; Bloch, John; Hutcheon, Ian

    1994-04-01

    The computer code LPNORM implements the mathematical method of linear programming to calculate the mineralogical makeup of mineral mixtures, such as rock, sediment, or soil samples, from their bulk geochemical composition and from the mineralogical (or geochemical) composition of the contained minerals. This method simultaneously solves the set of linear equations governing the distribution of oxides into these minerals, subject to an objective function and a set of basic constraints. LPNORM allows the user to specify what minerals will be considered for normative analysis, what their composition is (in terms of mineral formula or geochemical composition), and whether to maximize mineral abundances, minimize slack variables (oxides that can not be accounted for), or do both at once in the objective function. Independent knowledge about the abundance of one or several of the minerals in the sample can be entered as additional equality or inequality constraints. Trial-and-error approach enables the user to "optimize" the composition of one or a few of the contained minerals. Results of comparative tests, highlighting the efficiency, as well as the shortcomings, of LPNORM are presented.

  14. OPR1000 RCP Flow Coastdown Analysis using SPACE Code

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Dong-Hyuk; Kim, Seyun [KHNP CRI, Daejeon (Korea, Republic of)

    2016-10-15

    The Korean nuclear industry developed a thermal-hydraulic analysis code for the safety analysis of PWRs, named SPACE(Safety and Performance Analysis Code for Nuclear Power Plant). Current loss of flow transient analysis of OPR1000 uses COAST code to calculate transient RCS(Reactor Coolant System) flow. The COAST code calculates RCS loop flow using pump performance curves and RCP(Reactor Coolant Pump) inertia. In this paper, SPACE code is used to reproduce RCS flowrates calculated by COAST code. The loss of flow transient is transient initiated by reduction of forced reactor coolant circulation. Typical loss of flow transients are complete loss of flow(CLOF) and locked rotor(LR). OPR1000 RCP flow coastdown analysis was performed using SPACE using simplified nodalization. Complete loss of flow(4 RCP trip) was analyzed. The results show good agreement with those from COAST code, which is CE code for calculating RCS flow during loss of flow transients. Through this study, we confirmed that SPACE code can be used instead of COAST code for RCP flow coastdown analysis.

  15. Parallelization of Subchannel Analysis Code MATRA

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Seongjin; Hwang, Daehyun; Kwon, Hyouk [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)

    2014-05-15

    A stand-alone calculation of MATRA code used up pertinent computing time for the thermal margin calculations while a relatively considerable time is needed to solve the whole core pin-by-pin problems. In addition, it is strongly required to improve the computation speed of the MATRA code to satisfy the overall performance of the multi-physics coupling calculations. Therefore, a parallel approach to improve and optimize the computability of the MATRA code is proposed and verified in this study. The parallel algorithm is embodied in the MATRA code using the MPI communication method and the modification of the previous code structure was minimized. An improvement is confirmed by comparing the results between the single and multiple processor algorithms. The speedup and efficiency are also evaluated when increasing the number of processors. The parallel algorithm was implemented to the subchannel code MATRA using the MPI. The performance of the parallel algorithm was verified by comparing the results with those from the MATRA with the single processor. It is also noticed that the performance of the MATRA code was greatly improved by implementing the parallel algorithm for the 1/8 core and whole core problems.

  16. The importance of including dynamic soil-structure interaction into wind turbine simulation codes

    DEFF Research Database (Denmark)

    Damgaard, Mads; Andersen, Lars Vabbersgaard; Ibsen, Lars Bo

    2014-01-01

    A rigorous numerical model, describing a wind turbine structure and subsoil, may contain thousands of degrees of freedom, making the approach computationally inefficient for fast time domain analysis. In order to meet the requirements of real-time calculations, the dynamic impedance of the founda......A rigorous numerical model, describing a wind turbine structure and subsoil, may contain thousands of degrees of freedom, making the approach computationally inefficient for fast time domain analysis. In order to meet the requirements of real-time calculations, the dynamic impedance...... is examined. The optimal order of the models is determined and implemented into the aeroelastic code HAWC2, where the dynamic response of a 5.0 MW wind turbine is evaluated. In contrast to the fore-aft vibrations, the inclusion of soil-structure interaction is shown to be critical for the side-side vibrations...

  17. Computer code for the atomistic simulation of lattice defects and dynamics. [COMENT code

    Energy Technology Data Exchange (ETDEWEB)

    Schiffgens, J.O.; Graves, N.J.; Oster, C.A.

    1980-04-01

    This document has been prepared to satisfy the need for a detailed, up-to-date description of a computer code that can be used to simulate phenomena on an atomistic level. COMENT was written in FORTRAN IV and COMPASS (CDC assembly language) to solve the classical equations of motion for a large number of atoms interacting according to a given force law, and to perform the desired ancillary analysis of the resulting data. COMENT is a dual-purpose intended to describe static defect configurations as well as the detailed motion of atoms in a crystal lattice. It can be used to simulate the effect of temperature, impurities, and pre-existing defects on radiation-induced defect production mechanisms, defect migration, and defect stability.

  18. Dynamics in Epistasis Analysis.

    Science.gov (United States)

    Awdeh, Aseel; Phenix, Hilary; Kaern, Mads; Perkins, Theodore

    2017-01-16

    Finding regulatory relationships between genes, including the direction and nature of influence between them, is a fundamental challenge in the field of molecular genetics. One classical approach to this problem is epistasis analysis. Broadly speaking, epistasis analysis infers the regulatory relationships between a pair of genes in a genetic pathway by considering the patterns of change in an observable trait resulting from single and double deletion of genes. While classical epistasis analysis has yielded deep insights on numerous genetic pathways, it is not without limitations. Here, we explore the possibility of dynamic epistasis analysis, in which, in addition to performing genetic perturbations of a pathway, we drive the pathway by a time-varying upstream signal. We explore the theoretical power of dynamical epistasis analysis by conducting an identifiability analysis of Boolean models of genetic pathways, comparing static and dynamic approaches. We find that even relatively simple input dynamics greatly increases the power of epistasis analysis to discriminate alternative network structures. Further, we explore the question of experiment design, and show that a subset of short time-varying signals, which we call dynamic primitives, allow maximum discriminative power with a reduced number of experiments.

  19. ALEGRA -- A massively parallel h-adaptive code for solid dynamics

    Energy Technology Data Exchange (ETDEWEB)

    Summers, R.M.; Wong, M.K.; Boucheron, E.A.; Weatherby, J.R. [Sandia National Labs., Albuquerque, NM (United States)

    1997-12-31

    ALEGRA is a multi-material, arbitrary-Lagrangian-Eulerian (ALE) code for solid dynamics designed to run on massively parallel (MP) computers. It combines the features of modern Eulerian shock codes, such as CTH, with modern Lagrangian structural analysis codes using an unstructured grid. ALEGRA is being developed for use on the teraflop supercomputers to conduct advanced three-dimensional (3D) simulations of shock phenomena important to a variety of systems. ALEGRA was designed with the Single Program Multiple Data (SPMD) paradigm, in which the mesh is decomposed into sub-meshes so that each processor gets a single sub-mesh with approximately the same number of elements. Using this approach the authors have been able to produce a single code that can scale from one processor to thousands of processors. A current major effort is to develop efficient, high precision simulation capabilities for ALEGRA, without the computational cost of using a global highly resolved mesh, through flexible, robust h-adaptivity of finite elements. H-adaptivity is the dynamic refinement of the mesh by subdividing elements, thus changing the characteristic element size and reducing numerical error. The authors are working on several major technical challenges that must be met to make effective use of HAMMER on MP computers.

  20. A preliminary uncertainty analysis of phenomenological inputs employed in MAAP code using the SAUNA system

    Energy Technology Data Exchange (ETDEWEB)

    Park, S. H.; Park, S. Y.; Kim, K. R.; Ahn, K. I. [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)

    2009-10-15

    Uncertainty analysis is an essential element of safety analysis of nuclear power plants, and especially on the increase as an essential methodology of safety assessment by computer codes. Recently, these efforts have been stepped up to apply the uncertainty methodology in severe accident analysis and PSA Level 2. From this point of view, a statistical sampling-based MAAP-specific platform for a severe accident uncertainty analysis, SAUNA, is being developed in KAERI. Its main purpose is to execute many simulations that are employed for uncertainty analysis. For its efficient implementation, the SAUNA system is composed of three related modules: Firstly, a module for preparing a statistical sampling matrix, secondly, a module for the dynamic linking between code and samples for code simulation, and thirdly, a postprocessing module for further analysis of the code simulation results. The main objective of this paper is to introduce the main functions of the SAUNA system and its example of implementation.

  1. Challenges on innovations of newly-developed safety analysis codes

    Energy Technology Data Exchange (ETDEWEB)

    Yang, Yanhua [Shanghai Jiao Tong Univ. (China). School of Nuclear Science and Engineering; Zhang, Hao [State Nuclear Power Software Development Center, Beijing (China). Beijing Future Science and Technology City

    2016-05-15

    With the development of safety analysis method, the safety analysis codes meet more challenges. Three challenges are presented in this paper, which are mathematic model, code design and user interface. Combined with the self-reliance safety analysis code named COSINE, the ways of meeting these requirements are suggested, that is to develop multi-phases, multi-fields and multi-dimension models, to adopt object-oriented code design ideal and to improve the way of modeling, calculation control and data post-processing in the user interface.

  2. Manometer Behavior Analysis using CATHENA, RELAP and GOTHIC Codes

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Yang Hoon; Han, Kee Soo; Moon, Bok Ja; Jang, Misuk [Nuclear Engineering Service and Solution Co. Ltd., Daejeon (Korea, Republic of)

    2014-05-15

    In this presentation, simple thermal hydraulic behavior is analyzed using three codes to show the possibility of using alternative codes. We established three models of simple u-tube manometer using three different codes. CATHENA (Canadian Algorithm for Thermal hydraulic Network Analysis), RELAP (Reactor Excursion and Leak Analysis Program), GOTHIC (Generation of Thermal Hydraulic Information for Containments) are used for this analysis. CATHENA and RELAP are widely used codes for the analysis of system behavior of CANDU and PWR. And GOTHIC code also has been widely used for the analysis of thermal hydraulic behavior in the containment system. In this paper, the internal behavior of u-tube manometer was analyzed using 3 codes, CATHENA, RELAP and GOTHIC. The general transient behavior is similar among 3 codes. However, the behavior simulated using GOTHIC shows some different trend compared with the results from the other 2 codes at the end of the transient. It would be resulted from the use of different physical model in GOTHIC, which is specialized for the multi-phase thermal hydraulic behavior analysis of containment system unlike the other two codes.

  3. Dynamic Coding of Signed Quantities in Cortical Feedback Circuits

    Directory of Open Access Journals (Sweden)

    Dana eBallard

    2012-08-01

    Full Text Available In the early sensory and motor areas of the cortex, individual neurons transmit information about specific sensory features via a peaked response. This concept has been crystallized as `labeled lines,' to denote that axons communicate the specific properties of their sensory or motor parent cell. Such cells also can be characterized as being polarized, that is, as representing a signed quantity that is either positive or negative. We show in a model simulation that there are two important consequences when learning receptive fields using such signed codings in circuits that subtract different inputs. The first is that, in feedback circuits using labeled lines, such arithmetic operations need to be distributed across multiple distinct pathways. The second consequence is that such pathways must be necessarily dynamic, i.e. that synapses can grow and retract when forming receptive fields. The model monitors the breaking and growing of new circuit connections when their synapses need to change polarities and predicts that the rate of such changes should be inversely correlated with the progress of receptive field formation.

  4. Transcriptional dynamics reveal critical roles for non-coding RNAs in the immediate-early response.

    Directory of Open Access Journals (Sweden)

    Stuart Aitken

    2015-04-01

    Full Text Available The immediate-early response mediates cell fate in response to a variety of extracellular stimuli and is dysregulated in many cancers. However, the specificity of the response across stimuli and cell types, and the roles of non-coding RNAs are not well understood. Using a large collection of densely-sampled time series expression data we have examined the induction of the immediate-early response in unparalleled detail, across cell types and stimuli. We exploit cap analysis of gene expression (CAGE time series datasets to directly measure promoter activities over time. Using a novel analysis method for time series data we identify transcripts with expression patterns that closely resemble the dynamics of known immediate-early genes (IEGs and this enables a comprehensive comparative study of these genes and their chromatin state. Surprisingly, these data suggest that the earliest transcriptional responses often involve promoters generating non-coding RNAs, many of which are produced in advance of canonical protein-coding IEGs. IEGs are known to be capable of induction without de novo protein synthesis. Consistent with this, we find that the response of both protein-coding and non-coding RNA IEGs can be explained by their transcriptionally poised, permissive chromatin state prior to stimulation. We also explore the function of non-coding RNAs in the attenuation of the immediate early response in a small RNA sequencing dataset matched to the CAGE data: We identify a novel set of microRNAs responsible for the attenuation of the IEG response in an estrogen receptor positive cancer cell line. Our computational statistical method is well suited to meta-analyses as there is no requirement for transcripts to pass thresholds for significant differential expression between time points, and it is agnostic to the number of time points per dataset.

  5. Implementation of the Resonance Analysis Code SAMMY

    Institute of Scientific and Technical Information of China (English)

    2002-01-01

    The multi-level multi-channel R-matrix SAMMY code is used for making the resonance parameters,which was developed by Oak Ridge National Laboratory (ORNL), and widely used around the USA(ORELA, KAPL, LANL, TUNL...) and around the world (Belgium, Japan, France, Bulgaria, etc.).Thecode SAMMY is an important program to CNDC.

  6. Accelerator-driven transmutation reactor analysis code system (ATRAS)

    Energy Technology Data Exchange (ETDEWEB)

    Sasa, Toshinobu; Tsujimoto, Kazufumi; Takizuka, Takakazu; Takano, Hideki [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment

    1999-03-01

    JAERI is proceeding a design study of the hybrid type minor actinide transmutation system which mainly consist of an intense proton accelerator and a fast subcritical core. Neutronics and burnup characteristics of the accelerator-driven system is important from a view point of the maintenance of subcriticality and energy balance during the system operation. To determine those characteristics accurately, it is necessary to involve reactions at high-energy region, which are not treated on ordinary reactor analysis codes. The authors developed a code system named ATRAS to analyze the neutronics and burnup characteristics of accelerator-driven subcritical reactor systems. ATRAS has a function of burnup analysis taking account of the effect of spallation neutron source. ATRAS consists of a spallation analysis code, a neutron transport codes and a burnup analysis code. Utility programs for fuel exchange, pre-processing and post-processing are also incorporated. (author)

  7. GASFLOW: A Computational Fluid Dynamics Code for Gases, Aerosols, and Combustion, Volume 3: Assessment Manual

    Energy Technology Data Exchange (ETDEWEB)

    Müller, C.; Hughes, E. D.; Niederauer, G. F.; Wilkening, H.; Travis, J. R.; Spore, J. W.; Royl, P.; Baumann, W.

    1998-10-01

    Los Alamos National Laboratory (LANL) and Forschungszentrum Karlsruhe (FzK) are developing GASFLOW, a three-dimensional (3D) fluid dynamics field code as a best- estimate tool to characterize local phenomena within a flow field. Examples of 3D phenomena include circulation patterns; flow stratification; hydrogen distribution mixing and stratification; combustion and flame propagation; effects of noncondensable gas distribution on local condensation and evaporation; and aerosol entrainment, transport, and deposition. An analysis with GASFLOW will result in a prediction of the gas composition and discrete particle distribution in space and time throughout the facility and the resulting pressure and temperature loadings on the walls and internal structures with or without combustion. A major application of GASFLOW is for predicting the transport, mixing, and combustion of hydrogen and other gases in nuclear reactor containment and other facilities. It has been applied to situations involving transporting and distributing combustible gas mixtures. It has been used to study gas dynamic behavior in low-speed, buoyancy-driven flows, as well as sonic flows or diffusion dominated flows; and during chemically reacting flows, including deflagrations. The effects of controlling such mixtures by safety systems can be analyzed. The code version described in this manual is designated GASFLOW 2.1, which combines previous versions of the United States Nuclear Regulatory Commission code HMS (for Hydrogen Mixing Studies) and the Department of Energy and FzK versions of GASFLOW. The code was written in standard Fortran 90. This manual comprises three volumes. Volume I describes the governing physical equations and computational model. Volume II describes how to use the code to set up a model geometry, specify gas species and material properties, define initial and boundary conditions, and specify different outputs, especially graphical displays. Sample problems are included. Volume

  8. Axisymmetric Plume Simulations with NASA's DSMC Analysis Code

    Science.gov (United States)

    Stewart, B. D.; Lumpkin, F. E., III

    2012-01-01

    A comparison of axisymmetric Direct Simulation Monte Carlo (DSMC) Analysis Code (DAC) results to analytic and Computational Fluid Dynamics (CFD) solutions in the near continuum regime and to 3D DAC solutions in the rarefied regime for expansion plumes into a vacuum is performed to investigate the validity of the newest DAC axisymmetric implementation. This new implementation, based on the standard DSMC axisymmetric approach where the representative molecules are allowed to move in all three dimensions but are rotated back to the plane of symmetry by the end of the move step, has been fully integrated into the 3D-based DAC code and therefore retains all of DAC s features, such as being able to compute flow over complex geometries and to model chemistry. Axisymmetric DAC results for a spherically symmetric isentropic expansion are in very good agreement with a source flow analytic solution in the continuum regime and show departure from equilibrium downstream of the estimated breakdown location. Axisymmetric density contours also compare favorably against CFD results for the R1E thruster while temperature contours depart from equilibrium very rapidly away from the estimated breakdown surface. Finally, axisymmetric and 3D DAC results are in very good agreement over the entire plume region and, as expected, this new axisymmetric implementation shows a significant reduction in computer resources required to achieve accurate simulations for this problem over the 3D simulations.

  9. A novel dynamic frame rate control algorithm for H.264 low-bit-rate video coding

    Institute of Scientific and Technical Information of China (English)

    Yang Jing; Fang Xiangzhong

    2007-01-01

    The goal of this paper is to improve human visual perceptual quality as well as coding efficiency of H.264 video at low bit rate conditions by adaptively adjusting the number of skipped frames. The encoding frames ale selected according to the motion activity of each frame and the motion accumulation of successive frames. The motion activity analysis is based on the statistics of motion vectors and with consideration of the characteristics of H. 264 coding standard. A prediction model of motion accumulation is proposed to reduce complex computation of motion estimation. The dynamic encoding frame rate control algorithm is applied to both the frame level and the GOB (Group of Macroblocks) level. Simulation is done to compare the performance of JM76 with the proposed frame level scheme and GOB level scheme.

  10. Sending policies in dynamic wireless mesh using network coding

    DEFF Research Database (Denmark)

    Pandi, Sreekrishna; Fitzek, Frank; Pihl, Jeppe

    2015-01-01

    This paper demonstrates the quick prototyping capabilities of the Python-Kodo library for network coding based performance evaluation and investigates the problem of data redundancy in a network coded wireless mesh with opportunistic overhearing. By means of several wireless meshed architectures ...

  11. Core-seis: a code for LMFBR core seismic analysis

    Energy Technology Data Exchange (ETDEWEB)

    Chellapandi, P.; Ravi, R.; Chetal, S.C.; Bhoje, S.B. [Indira Gandhi Centre for Atomic Research, Kalpakkam (India). Reactor Group

    1995-12-31

    This paper deals with a computer code CORE-SEIS specially developed for seismic analysis of LMFBR core configurations. For demonstrating the prediction capability of the code, results are presented for one of the MONJU reactor core mock ups which deals with a cluster of 37 subassemblies kept in water. (author). 3 refs., 7 figs., 2 tabs.

  12. Dependency graph for code analysis on emerging architectures

    Energy Technology Data Exchange (ETDEWEB)

    Shashkov, Mikhail Jurievich [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Lipnikov, Konstantin [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2017-08-08

    Direct acyclic dependency (DAG) graph is becoming the standard for modern multi-physics codes.The ideal DAG is the true block-scheme of a multi-physics code. Therefore, it is the convenient object for insitu analysis of the cost of computations and algorithmic bottlenecks related to statistical frequent data motion and dymanical machine state.

  13. Source Code Analysis Laboratory (SCALe) for Energy Delivery Systems

    Science.gov (United States)

    2010-12-01

    and standar - dizing them through the ISO/IEC process should eliminate many of the problems encountered at the NIST SATE and also increase the...view the source code, using both structured and unstruct ons of secure coding rules discovered. However, manua than automated analysis, and the

  14. Extreme genetic code optimality from a molecular dynamics calculation of amino acid polar requirement

    Science.gov (United States)

    Butler, Thomas; Goldenfeld, Nigel; Mathew, Damien; Luthey-Schulten, Zaida

    2009-06-01

    A molecular dynamics calculation of the amino acid polar requirement is used to score the canonical genetic code. Monte Carlo simulation shows that this computational polar requirement has been optimized by the canonical genetic code, an order of magnitude more than any previously known measure, effectively ruling out a vertical evolution dynamics. The sensitivity of the optimization to the precise metric used in code scoring is consistent with code evolution having proceeded through the communal dynamics of statistical proteins using horizontal gene transfer, as recently proposed. The extreme optimization of the genetic code therefore strongly supports the idea that the genetic code evolved from a communal state of life prior to the last universal common ancestor.

  15. Extreme genetic code optimality from a molecular dynamics calculation of amino acid polar requirement.

    Science.gov (United States)

    Butler, Thomas; Goldenfeld, Nigel; Mathew, Damien; Luthey-Schulten, Zaida

    2009-06-01

    A molecular dynamics calculation of the amino acid polar requirement is used to score the canonical genetic code. Monte Carlo simulation shows that this computational polar requirement has been optimized by the canonical genetic code, an order of magnitude more than any previously known measure, effectively ruling out a vertical evolution dynamics. The sensitivity of the optimization to the precise metric used in code scoring is consistent with code evolution having proceeded through the communal dynamics of statistical proteins using horizontal gene transfer, as recently proposed. The extreme optimization of the genetic code therefore strongly supports the idea that the genetic code evolved from a communal state of life prior to the last universal common ancestor.

  16. Performance analysis of adaptive turbo coded modulation with time delay

    Institute of Scientific and Technical Information of China (English)

    伍守豪; 宋文涛; 罗汉文

    2004-01-01

    The method of data fitting is applied to obtain the BER expression for turbo coded modulation, and a fitting mathematical model is proposed, which resolves the problem that there is no exact BER expression for turbo coded modulation in performance analysis. With the time delay consideration, the performance of BER of adaptive turbo coded modulation is analyzed and simulated. The results show that adaptive turbo coded modulation is very sensitive to time delay. In order to meet the target BER requirement, the total time delay should be less than 0. 001/fD.

  17. Improvement of QR Code Recognition Based on Pillbox Filter Analysis

    OpenAIRE

    Jia-Shing Sheu; Kai-Chung Teng

    2013-01-01

    The objective of this paper is to perform the innovation design for improving the recognition of a captured QR code image with blur through the Pillbox filter analysis. QR code images can be captured by digital video cameras. Many factors contribute to QR code decoding failure, such as the low quality of the image. Focus is an important factor that affects the quality of the image. This study discusses the out-of-focus QR code image and aims to improve the recognition of the conte...

  18. Structural dynamics analysis

    Science.gov (United States)

    Housner, J. M.; Anderson, M.; Belvin, W.; Horner, G.

    1985-01-01

    Dynamic analysis of large space antenna systems must treat the deployment as well as vibration and control of the deployed antenna. Candidate computer programs for deployment dynamics, and issues and needs for future program developments are reviewed. Some results for mast and hoop deployment are also presented. Modeling of complex antenna geometry with conventional finite element methods and with repetitive exact elements is considered. Analytical comparisons with experimental results for a 15 meter hoop/column antenna revealed the importance of accurate structural properties including nonlinear joints. Slackening of cables in this antenna is also a consideration. The technology of designing actively damped structures through analytical optimization is discussed and results are presented.

  19. A Case for Dynamic Reverse-code Generation

    DEFF Research Database (Denmark)

    Lee, Jooyong

    2007-01-01

    Backtracking (i.e. reverse execution) helps the user of a debugger to naturally think backwards along the execution path of a program, and thinking backwards makes it easy to locate the origin of a bug. So far backtracking has been implemented mostly by state saving or by checkpointing....... These implementations, however, inherently do not scale. As has often been said, the ultimate solution for backtracking is to use reverse code: executing the reverse code restores the previous states of a program. In our earlier work, we presented a method to generate reverse code on the fly while running a debugger...

  20. Performance Analysis of 3-Dimensional Turbo Codes

    CERN Document Server

    Rosnes, Eirik

    2011-01-01

    In this work, we consider the minimum distance properties and convergence thresholds of 3-dimensional turbo codes (3D-TCs), recently introduced by Berrou et al.. Here, we consider binary 3D-TCs while the original work of Berrou et al. considered double-binary codes. In the first part of the paper, the minimum distance properties are analyzed from an ensemble perspective, both in the finite-length regime and in the asymptotic case of large block lengths. In particular, we analyze the asymptotic weight distribution of 3D-TCs and show numerically that their typical minimum distance dmin may, depending on the specific parameters, asymptotically grow linearly with the block length, i.e., the 3D-TC ensemble is asymptotically good for some parameters. In the second part of the paper, we derive some useful upper bounds on the dmin when using quadratic permutation polynomial (QPP) interleavers with a quadratic inverse. Furthermore, we give examples of interleaver lengths where an upper bound appears to be tight. The b...

  1. Dynamically Translating Binary Code for Multi-Threaded Programs Using Shared Code Cache

    Institute of Scientific and Technical Information of China (English)

    Chia-Lun Liu; Jiunn-Yeu Chen; Wuu Yang; Wei-Chung Hsu

    2014-01-01

    mc2llvm is a process-level ARM-to-x86 binary translator developed in our lab in the past several years. Currently, it is able to emulate single-threaded programs. We extend mc2llvm to emulate multi-threaded programs. Our main task is to reconstruct its architecture for multi-threaded programs. Register mapping, code cache management, and address mapping in mc2llvm have all been modified. In addition, to further speed up the emulation, we collect hot paths, aggressively optimize and generate code for them at run time. Additional threads are used to alleviate the overhead. Thus, when the same hot path is walked through again, the corresponding optimized native code will be executed instead. In our experiments, our system is 8.8X faster than QEMU (quick emulator) on average when emulating the specified benchmarks with 8 guest threads.

  2. SLSF loop handling system. Volume III. AISC code evaluations and analysis of critical attachments. [LMFBR

    Energy Technology Data Exchange (ETDEWEB)

    Ahmed, H.; Cowie, A.; Malek, R. A.; Rafer, A.; Ma, D.; Tebo, F.

    1978-10-01

    SLSF loop handling system was analyzed for deadweight and postulated dynamic loading conditions using a linear elastic static equivalent method of stress analysis. Stress computations of Cradle and critical attachments per AISC Code guidelines are presented. HFEF is credited with in-depth review of initial phase of work.

  3. Improvement of QR Code Recognition Based on Pillbox Filter Analysis

    Directory of Open Access Journals (Sweden)

    Jia-Shing Sheu

    2013-04-01

    Full Text Available The objective of this paper is to perform the innovation design for improving the recognition of a captured QR code image with blur through the Pillbox filter analysis. QR code images can be captured by digital video cameras. Many factors contribute to QR code decoding failure, such as the low quality of the image. Focus is an important factor that affects the quality of the image. This study discusses the out-of-focus QR code image and aims to improve the recognition of the contents in the QR code image. Many studies have used the pillbox filter (circular averaging filter method to simulate an out-of-focus image. This method is also used in this investigation to improve the recognition of a captured QR code image. A blurred QR code image is separated into nine levels. In the experiment, four different quantitative approaches are used to reconstruct and decode an out-of-focus QR code image. These nine reconstructed QR code images using methods are then compared. The final experimental results indicate improvements in identification.

  4. Code manual for CONTAIN 2.0: A computer code for nuclear reactor containment analysis

    Energy Technology Data Exchange (ETDEWEB)

    Murata, K.K.; Williams, D.C.; Griffith, R.O.; Gido, R.G.; Tadios, E.L.; Davis, F.J.; Martinez, G.M.; Washington, K.E. [Sandia National Labs., Albuquerque, NM (United States); Tills, J. [J. Tills and Associates, Inc., Sandia Park, NM (United States)

    1997-12-01

    The CONTAIN 2.0 computer code is an integrated analysis tool used for predicting the physical conditions, chemical compositions, and distributions of radiological materials inside a containment building following the release of material from the primary system in a light-water reactor accident. It can also predict the source term to the environment. CONTAIN 2.0 is intended to replace the earlier CONTAIN 1.12, which was released in 1991. The purpose of this Code Manual is to provide full documentation of the features and models in CONTAIN 2.0. Besides complete descriptions of the models, this Code Manual provides a complete description of the input and output from the code. CONTAIN 2.0 is a highly flexible and modular code that can run problems that are either quite simple or highly complex. An important aspect of CONTAIN is that the interactions among thermal-hydraulic phenomena, aerosol behavior, and fission product behavior are taken into account. The code includes atmospheric models for steam/air thermodynamics, intercell flows, condensation/evaporation on structures and aerosols, aerosol behavior, and gas combustion. It also includes models for reactor cavity phenomena such as core-concrete interactions and coolant pool boiling. Heat conduction in structures, fission product decay and transport, radioactive decay heating, and the thermal-hydraulic and fission product decontamination effects of engineered safety features are also modeled. To the extent possible, the best available models for severe accident phenomena have been incorporated into CONTAIN, but it is intrinsic to the nature of accident analysis that significant uncertainty exists regarding numerous phenomena. In those cases, sensitivity studies can be performed with CONTAIN by means of user-specified input parameters. Thus, the code can be viewed as a tool designed to assist the knowledge reactor safety analyst in evaluating the consequences of specific modeling assumptions.

  5. Code Coupling for Multi-Dimensional Core Transient Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Park, Jin-Woo; Park, Guen-Tae; Park, Min-Ho; Ryu, Seok-Hee; Um, Kil-Sup; Lee Jae-Il [KEPCO NF, Daejeon (Korea, Republic of)

    2015-05-15

    After the CEA ejection, the nuclear power of the reactor dramatically increases in an exponential behavior until the Doppler effect becomes important and turns the reactivity balance and power down to lower levels. Although this happens in a very short period of time, only few seconds, the energy generated can be very significant and cause fuel failures. The current safety analysis methodology which is based on overly conservative assumptions with the point kinetics model results in quite adverse consequences. Thus, KEPCO Nuclear Fuel(KNF) is developing the multi-dimensional safety analysis methodology to mitigate the consequences of the single CEA ejection accident. For this purpose, three-dimensional core neutron kinetics code ASTRA, sub-channel analysis code THALES, and fuel performance analysis code FROST, which have transient calculation performance, were coupled using message passing interface (MPI). This paper presents the methodology used for code coupling and the preliminary simulation results with the coupled code system (CHASER). Multi-dimensional core transient analysis code system, CHASER, has been developed and it was applied to simulate a single CEA ejection accident. CHASER gave a good prediction of multi-dimensional core transient behaviors during transient. In the near future, the multi-dimension CEA ejection analysis methodology using CHASER is planning to be developed. CHASER is expected to be a useful tool to gain safety margin for reactivity initiated accidents (RIAs), such as a single CEA ejection accident.

  6. Source Authentication for Code Dissemination Supporting Dynamic Packet Size in Wireless Sensor Networks

    Directory of Open Access Journals (Sweden)

    Daehee Kim

    2016-07-01

    Full Text Available Code dissemination in wireless sensor networks (WSNs is a procedure for distributing a new code image over the air in order to update programs. Due to the fact that WSNs are mostly deployed in unattended and hostile environments, secure code dissemination ensuring authenticity and integrity is essential. Recent works on dynamic packet size control in WSNs allow enhancing the energy efficiency of code dissemination by dynamically changing the packet size on the basis of link quality. However, the authentication tokens attached by the base station become useless in the next hop where the packet size can vary according to the link quality of the next hop. In this paper, we propose three source authentication schemes for code dissemination supporting dynamic packet size. Compared to traditional source authentication schemes such as μTESLA and digital signatures, our schemes provide secure source authentication under the environment, where the packet size changes in each hop, with smaller energy consumption.

  7. CRITICA: coding region identification tool invoking comparative analysis

    Science.gov (United States)

    Badger, J. H.; Olsen, G. J.; Woese, C. R. (Principal Investigator)

    1999-01-01

    Gene recognition is essential to understanding existing and future DNA sequence data. CRITICA (Coding Region Identification Tool Invoking Comparative Analysis) is a suite of programs for identifying likely protein-coding sequences in DNA by combining comparative analysis of DNA sequences with more common noncomparative methods. In the comparative component of the analysis, regions of DNA are aligned with related sequences from the DNA databases; if the translation of the aligned sequences has greater amino acid identity than expected for the observed percentage nucleotide identity, this is interpreted as evidence for coding. CRITICA also incorporates noncomparative information derived from the relative frequencies of hexanucleotides in coding frames versus other contexts (i.e., dicodon bias). The dicodon usage information is derived by iterative analysis of the data, such that CRITICA is not dependent on the existence or accuracy of coding sequence annotations in the databases. This independence makes the method particularly well suited for the analysis of novel genomes. CRITICA was tested by analyzing the available Salmonella typhimurium DNA sequences. Its predictions were compared with the DNA sequence annotations and with the predictions of GenMark. CRITICA proved to be more accurate than GenMark, and moreover, many of its predictions that would seem to be errors instead reflect problems in the sequence databases. The source code of CRITICA is freely available by anonymous FTP (rdp.life.uiuc.edu in/pub/critica) and on the World Wide Web (http:/(/)rdpwww.life.uiuc.edu).

  8. CRITICA: coding region identification tool invoking comparative analysis

    Science.gov (United States)

    Badger, J. H.; Olsen, G. J.; Woese, C. R. (Principal Investigator)

    1999-01-01

    Gene recognition is essential to understanding existing and future DNA sequence data. CRITICA (Coding Region Identification Tool Invoking Comparative Analysis) is a suite of programs for identifying likely protein-coding sequences in DNA by combining comparative analysis of DNA sequences with more common noncomparative methods. In the comparative component of the analysis, regions of DNA are aligned with related sequences from the DNA databases; if the translation of the aligned sequences has greater amino acid identity than expected for the observed percentage nucleotide identity, this is interpreted as evidence for coding. CRITICA also incorporates noncomparative information derived from the relative frequencies of hexanucleotides in coding frames versus other contexts (i.e., dicodon bias). The dicodon usage information is derived by iterative analysis of the data, such that CRITICA is not dependent on the existence or accuracy of coding sequence annotations in the databases. This independence makes the method particularly well suited for the analysis of novel genomes. CRITICA was tested by analyzing the available Salmonella typhimurium DNA sequences. Its predictions were compared with the DNA sequence annotations and with the predictions of GenMark. CRITICA proved to be more accurate than GenMark, and moreover, many of its predictions that would seem to be errors instead reflect problems in the sequence databases. The source code of CRITICA is freely available by anonymous FTP (rdp.life.uiuc.edu in/pub/critica) and on the World Wide Web (http:/(/)rdpwww.life.uiuc.edu).

  9. Software and codes for analysis of concentrating solar power technologies.

    Energy Technology Data Exchange (ETDEWEB)

    Ho, Clifford Kuofei

    2008-12-01

    This report presents a review and evaluation of software and codes that have been used to support Sandia National Laboratories concentrating solar power (CSP) program. Additional software packages developed by other institutions and companies that can potentially improve Sandia's analysis capabilities in the CSP program are also evaluated. The software and codes are grouped according to specific CSP technologies: power tower systems, linear concentrator systems, and dish/engine systems. A description of each code is presented with regard to each specific CSP technology, along with details regarding availability, maintenance, and references. A summary of all the codes is then presented with recommendations regarding the use and retention of the codes. A description of probabilistic methods for uncertainty and sensitivity analyses of concentrating solar power technologies is also provided.

  10. Benchmarking Of Improved DPAC Transient Deflagration Analysis Code

    Energy Technology Data Exchange (ETDEWEB)

    Laurinat, James E.; Hensel, Steve J.

    2013-03-21

    The transient deflagration code DPAC (Deflagration Pressure Analysis Code) has been upgraded for use in modeling hydrogen deflagration transients. The upgraded code is benchmarked using data from vented hydrogen deflagration tests conducted at the HYDRO-SC Test Facility at the University of Pisa. DPAC originally was written to calculate peak deflagration pressures for deflagrations in radioactive waste storage tanks and process facilities at the Savannah River Site. Upgrades include the addition of a laminar flame speed correlation for hydrogen deflagrations and a mechanistic model for turbulent flame propagation, incorporation of inertial effects during venting, and inclusion of the effect of water vapor condensation on vessel walls. In addition, DPAC has been coupled with CEA, a NASA combustion chemistry code. The deflagration tests are modeled as end-to-end deflagrations. The improved DPAC code successfully predicts both the peak pressures during the deflagration tests and the times at which the pressure peaks.

  11. Analysis of Iterated Hard Decision Decoding of Product Codes with Reed-Solomon Component Codes

    DEFF Research Database (Denmark)

    Justesen, Jørn; Høholdt, Tom

    2007-01-01

    Products of Reed-Solomon codes are important in applications because they offer a combination of large blocks, low decoding complexity, and good performance. A recent result on random graphs can be used to show that with high probability a large number of errors can be corrected by iterating...... minimum distance decoding. We present an analysis related to density evolution which gives the exact asymptotic value of the decoding threshold and also provides a closed form approximation to the distribution of errors in each step of the decoding of finite length codes....

  12. Guide to Using Onionskin Analysis Code (U)

    Energy Technology Data Exchange (ETDEWEB)

    Fugate, Michael Lynn [Los Alamos National Lab. (LANL), Los Alamos, NM (United States). Statistical Sciences Group; Morzinski, Jerome Arthur [Los Alamos National Lab. (LANL), Los Alamos, NM (United States). Statistical Sciences Group

    2016-09-15

    This document is a guide to using R-code written for the purpose of analyzing onionskin experiments. We expect the user to be very familiar with statistical methods and the R programming language. For more details about onionskin experiments and the statistical methods mentioned in this document see Storlie, Fugate, et al. (2013). Engineers at LANL experiment with detonators and high explosives to assess performance. The experimental unit, called an onionskin, is a hemisphere consisting of a detonator and a booster pellet surrounded by explosive material. When the detonator explodes, a streak camera mounted above the pole of the hemisphere records when the shock wave arrives at the surface. The output from the camera is a two-dimensional image that is transformed into a curve that shows the arrival time as a function of polar angle. The statistical challenge is to characterize a baseline population of arrival time curves and to compare the baseline curves to curves from a new, so-called, test series. The hope is that the new test series of curves is statistically similar to the baseline population.

  13. Characteristic Analysis of Fire Modeling Codes

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Yoon Hwan; Yang, Joon Eon [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of); Kim, Jong Hoon [Kyeongmin College, Ujeongbu (Korea, Republic of)

    2004-04-15

    This report documents and compares key features of four zone models: CFAST, COMPBRN IIIE, MAGIC and the Fire Induced Vulnerability Evaluation (FIVE) methodology. CFAST and MAGIC handle multi-compartment, multi-fire problems, using many equations; COMPBRN and FIVE handle single compartment, single fire source problems, using simpler equation. The increased rigor of the formulation of CFAST and MAGIC does not mean that these codes are more accurate in every domain; for instance, the FIVE methodology uses a single zone approximation with a plume/ceiling jet sublayer, while the other models use a two-zone treatment without a plume/ceiling jet sublayer. Comparisons with enclosure fire data indicate that inclusion of plume/ceiling jet sublayer temperatures is more conservative, and generally more accurate than neglecting them. Adding a plume/ceiling jet sublayer to the two-zone models should be relatively straightforward, but it has not been done yet for any of the two-zone models. Such an improvement is in progress for MAGIC.

  14. A predictive coding framework for rapid neural dynamics during sentence-level language comprehension

    NARCIS (Netherlands)

    Lewis, A.G.; Bastiaansen, M.C.M.

    2015-01-01

    There is a growing literature investigating the relationship between oscillatory neural dynamics measured using electroencephalography (EEG) and/or magnetoencephalography (MEG), and sentence-level language comprehension. Recent proposals have suggested a strong link between predictive coding account

  15. PROSA-1: a probabilistic response-surface analysis code. [LMFBR

    Energy Technology Data Exchange (ETDEWEB)

    Vaurio, J. K.; Mueller, C.

    1978-06-01

    Techniques for probabilistic response-surface analysis have been developed to obtain the probability distributions of the consequences of postulated nuclear-reactor accidents. The uncertainties of the consequences are caused by the variability of the system and model input parameters used in the accident analysis. Probability distributions are assigned to the input parameters, and parameter values are systematically chosen from these distributions. These input parameters are then used in deterministic consequence analyses performed by mechanistic accident-analysis codes. The results of these deterministic consequence analyses are used to generate the coefficients for analytical functions that approximate the consequences in terms of the selected input parameters. These approximating functions are used to generate the probability distributions of the consequences with random sampling being used to obtain values for the accident parameters from their distributions. A computer code PROSA has been developed for implementing the probabilistic response-surface technique. Special features of the code generate or treat sensitivities, statistical moments of the input and output variables, regionwise response surfaces, correlated input parameters, and conditional distributions. The code can also be used for calculating important distributions of the input parameters. The use of the code is illustrated in conjunction with the fast-running accident-analysis code SACO to provide probability studies of LMFBR hypothetical core-disruptive accidents. However, the methods and the programming are general and not limited to such applications.

  16. Code conversion for system design and safety analysis of NSSS

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Hae Cho; Kim, Young Tae; Choi, Young Gil; Kim, Hee Kyung [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    1996-01-01

    This report describes overall project works related to conversion, installation and validation of computer codes which are used in NSSS design and safety analysis of nuclear power plants. Domain/os computer codes for system safety analysis are installed and validated on Apollo DN10000, and then Apollo version are converted and installed again on HP9000/700 series with appropriate validation. Also, COOLII and COAST which are cyber version computer codes are converted into versions of Apollo DN10000 and HP9000/700, and installed with validation. This report details whole processes of work involved in the computer code conversion and installation, as well as software verification and validation results which are attached to this report. 12 refs., 8 figs. (author)

  17. Phonemic Coding Might Result From Sensory-Motor Coupling Dynamics

    OpenAIRE

    2002-01-01

    Human sound systems are invariably phonemically coded. Furthermore, phoneme inventories follow very particular tendancies. To explain these phenomena, there existed so far three kinds of approaches : ``Chomskyan''/cognitive innatism, morpho-perceptual innatism and the more recent approach of ``language as a complex cultural system which adapts under the pressure of efficient communication''. The two first approaches are clearly not satisfying, while the third, even if ...

  18. Nonlinear dynamic analysis of sandwich panels

    Science.gov (United States)

    Lush, A. M.

    1984-01-01

    Two analytical techniques applicable to large deflection dynamic response calculations for pressure loaded composite sandwich panels are demonstrated. One technique utilizes finite element modeling with a single equivalent layer representing the face sheets and core. The other technique utilizes the modal analysis computer code DEPROP which was recently modified to include transverse shear deformation in a core layer. The example problem consists of a simply supported rectangular sandwich panel. Included are comparisons of linear and nonlinear static response calculations, in addition to dynamic response calculations.

  19. Parallelization of a beam dynamics code and first large scale radio frequency quadrupole simulations

    Directory of Open Access Journals (Sweden)

    J. Xu

    2007-01-01

    Full Text Available The design and operation support of hadron (proton and heavy-ion linear accelerators require substantial use of beam dynamics simulation tools. The beam dynamics code TRACK has been originally developed at Argonne National Laboratory (ANL to fulfill the special requirements of the rare isotope accelerator (RIA accelerator systems. From the beginning, the code has been developed to make it useful in the three stages of a linear accelerator project, namely, the design, commissioning, and operation of the machine. To realize this concept, the code has unique features such as end-to-end simulations from the ion source to the final beam destination and automatic procedures for tuning of a multiple charge state heavy-ion beam. The TRACK code has become a general beam dynamics code for hadron linacs and has found wide applications worldwide. Until recently, the code has remained serial except for a simple parallelization used for the simulation of multiple seeds to study the machine errors. To speed up computation, the TRACK Poisson solver has been parallelized. This paper discusses different parallel models for solving the Poisson equation with the primary goal to extend the scalability of the code onto 1024 and more processors of the new generation of supercomputers known as BlueGene (BG/L. Domain decomposition techniques have been adapted and incorporated into the parallel version of the TRACK code. To demonstrate the new capabilities of the parallelized TRACK code, the dynamics of a 45 mA proton beam represented by 10^{8} particles has been simulated through the 325 MHz radio frequency quadrupole and initial accelerator section of the proposed FNAL proton driver. The results show the benefits and advantages of large-scale parallel computing in beam dynamics simulations.

  20. Recent applications of the transonic wing analysis computer code, TWING

    Science.gov (United States)

    Subramanian, N. R.; Holst, T. L.; Thomas, S. D.

    1982-01-01

    An evaluation of the transonic-wing-analysis computer code TWING is given. TWING utilizes a fully implicit approximate factorization iteration scheme to solve the full potential equation in conservative form. A numerical elliptic-solver grid-generation scheme is used to generate the required finite-difference mesh. Several wing configurations were analyzed, and the limits of applicability of this code was evaluated. Comparisons of computed results were made with available experimental data. Results indicate that the code is robust, accurate (when significant viscous effects are not present), and efficient. TWING generally produces solutions an order of magnitude faster than other conservative full potential codes using successive-line overrelaxation. The present method is applicable to a wide range of isolated wing configurations including high-aspect-ratio transport wings and low-aspect-ratio, high-sweep, fighter configurations.

  1. QR Codes in the Library: Are They Worth the Effort? Analysis of a QR Code Pilot Project

    OpenAIRE

    Wilson, Andrew M

    2012-01-01

    The literature is filled with potential uses for Quick Response (QR) codes in the library. Setting, but few library QR code projects have publicized usage statistics. A pilot project carried out in the Eda Kuhn Loeb Music Library of the Harvard College Library sought to determine whether library patrons actually understand and use QR codes. Results and analysis of the pilot project are provided, attempting to answer the question as to whether QR codes are worth the effort for libraries.

  2. Analysis of LAPAN-IPB image lossless compression using differential pulse code modulation and huffman coding

    Science.gov (United States)

    Hakim, P. R.; Permala, R.

    2017-01-01

    LAPAN-A3/IPB satellite is the latest Indonesian experimental microsatellite with remote sensing and earth surveillance missions. The satellite has three optical payloads, which are multispectral push-broom imager, digital matrix camera and video camera. To increase data transmission efficiency, the multispectral imager data can be compressed using either lossy or lossless compression method. This paper aims to analyze Differential Pulse Code Modulation (DPCM) method and Huffman coding that are used in LAPAN-IPB satellite image lossless compression. Based on several simulation and analysis that have been done, current LAPAN-IPB lossless compression algorithm has moderate performance. There are several aspects that can be improved from current configuration, which are the type of DPCM code used, the type of Huffman entropy-coding scheme, and the use of sub-image compression method. The key result of this research shows that at least two neighboring pixels should be used for DPCM calculation to increase compression performance. Meanwhile, varying Huffman tables with sub-image approach could also increase the performance if on-board computer can support for more complicated algorithm. These results can be used as references in designing Payload Data Handling System (PDHS) for an upcoming LAPAN-A4 satellite.

  3. EVAPRED - A CODE FOR FATIGUE ANALYSIS OPTIMIZATION

    Directory of Open Access Journals (Sweden)

    Dorin LOZICI-BRÎNZEI

    2010-03-01

    Full Text Available The fatigue can be, in fact, defined as: “failure under a repeated or otherwise varying load, which never reaches a level sufficient to cause failure in a single application”.Physical testing is clearly unrealistic for every design component. In most applications, fatigue-safe life design requires the prediction of the component fatigue life that accounts for predicted service loads and materials. The primary tool for both understanding and being able to predict and avoid fatigue has proven to be the finite element analysis (FEA. Computer-aided engineering (CAE programs use three major methods to determine the total fatigue life: Stress life (SN, Strain life (EN and Fracture Mechanics (FM. FEA can predict stress concentration areas and can help design engineers to predict how long their designs are likely to last before experiencing the onset of fatigue.

  4. Quasi-3d aerodynamic code for analyzing dynamic flap response

    DEFF Research Database (Denmark)

    Ramos García, Néstor

    frequencies and oscillation amplitudes, and generally a good agreement is obtained. The capability of the code to simulate a trailing edge flap under steady or unsteady flow conditions has been proven. A parametric study on rotational effects induced by Coriolis and centrifugal forces in the boundary layer...... is modeled using a panel method whereas the viscous part is modeled by using the integral form of the the laminar and turbulent boundary layer equations and with extensions for 3-D rotational effects. Laminar to turbulent transition can be forced with a boundary layer trip or computed with a modified e9...

  5. FARO base case post-test analysis by COMETA code

    Energy Technology Data Exchange (ETDEWEB)

    Annunziato, A.; Addabbo, C. [Joint Research Centre, Ispra (Italy)

    1995-09-01

    The paper analyzes the COMETA (Core Melt Thermal-Hydraulic Analysis) post test calculations of FARO Test L-11, the so-called Base Case Test. The FARO Facility, located at JRC Ispra, is used to simulate the consequences of Severe Accidents in Nuclear Power Plants under a variety of conditions. The COMETA Code has a 6 equations two phase flow field and a 3 phases corium field: the jet, the droplets and the fused-debris bed. The analysis shown that the code is able to pick-up all the major phenomena occurring during the fuel-coolant interaction pre-mixing phase.

  6. StarFinder: A code for stellar field analysis

    Science.gov (United States)

    Diolaiti, Emiliano; Bendinelli, Orazio; Bonaccini, Domenico; Close, Laird M.; Currie, Doug G.; Parmeggiani, Gianluigi

    2000-11-01

    StarFinder is an IDL code for the deep analysis of stellar fields, designed for Adaptive Optics well-sampled images with high and low Strehl ratio. The Point Spread Function is extracted directly from the frame, to take into account the actual structure of the instrumental response and the atmospheric effects. The code is written in IDL language and organized in the form of a self-contained widget-based application, provided with a series of tools for data visualization and analysis. A description of the method and some applications to Adaptive Optics data are presented.

  7. Analysis of isoplanatic high resolution stellar fields by Starfinder code

    CERN Document Server

    Diolaiti, E; Bonaccini, D; Close, L M; Currie, D; Parmeggiani, G

    2000-01-01

    We describe a new code for the deep analysis of stellar fields, designed for Adaptive Optics Nyquist-sampled images with high and low Strehl ratio. The Point Spread Function is extracted directly from the image frame, to take into account the actual structure of the instrumental response and the atmospheric effects. The code is written in IDL language and organized in the form of a self-contained widget-based application, provided with a series of tools for data visualization and analysis. A description of the method and some applications to AO data are presented.

  8. Phantom of RAMSES (POR): A new Milgromian dynamics N-body code

    CERN Document Server

    Lüghausen, Fabian; Kroupa, Pavel

    2014-01-01

    Since its first formulation in 1983, Milgromian dynamics (MOND) has been very successful in predicting the gravitational potential of galaxies from the distribution of baryons alone, including general scaling relations and detailed rotation curves of large statistical samples of individual galaxies covering a large range of masses and sizes. Most predictions however rely on static models, and only a handful of N-body codes have been developed over the years to investigate the consequences of the Milgromian framework for the dynamics of complex evolving dynamical systems. In this work, we present a new Milgromian N-body code, which is a customized version of the RAMSES code (Teyssier 2002) and thus comes with all its features: it includes particles and gas dynamics, and importantly allows for high spatial resolution of complex systems due to the adaptive mesh refinement (AMR) technique. It further allows the direct comparison between Milgromian simulations and standard Newtonian simulations with dark matter pa...

  9. Choreographer Pre-Testing Code Analysis and Operational Testing.

    Energy Technology Data Exchange (ETDEWEB)

    Fritz, David J. [Sandia National Laboratories (SNL-CA), Livermore, CA (United States); Harrison, Christopher B. [Sandia National Laboratories (SNL-CA), Livermore, CA (United States); Perr, C. W. [Sandia National Laboratories (SNL-CA), Livermore, CA (United States); Hurd, Steven A [Sandia National Laboratories (SNL-CA), Livermore, CA (United States)

    2014-07-01

    Choreographer is a "moving target defense system", designed to protect against attacks aimed at IP addresses without corresponding domain name system (DNS) lookups. It coordinates actions between a DNS server and a Network Address Translation (NAT) device to regularly change which publicly available IP addresses' traffic will be routed to the protected device versus routed to a honeypot. More details about how Choreographer operates can be found in Section 2: Introducing Choreographer. Operational considerations for the successful deployment of Choreographer can be found in Section 3. The Testing & Evaluation (T&E) for Choreographer involved 3 phases: Pre-testing, Code Analysis, and Operational Testing. Pre-testing, described in Section 4, involved installing and configuring an instance of Choreographer and verifying it would operate as expected for a simple use case. Our findings were that it was simple and straightforward to prepare a system for a Choreographer installation as well as configure Choreographer to work in a representative environment. Code Analysis, described in Section 5, consisted of running a static code analyzer (HP Fortify) and conducting dynamic analysis tests using the Valgrind instrumentation framework. Choreographer performed well, such that only a few errors that might possibly be problematic in a given operating situation were identified. Operational Testing, described in Section 6, involved operating Choreographer in a representative environment created through EmulyticsTM . Depending upon the amount of server resources dedicated to Choreographer vis-á-vis the amount of client traffic handled, Choreographer had varying degrees of operational success. In an environment with a poorly resourced Choreographer server and as few as 50-100 clients, Choreographer failed to properly route traffic over half the time. Yet, with a well-resourced server, Choreographer handled over 1000 clients without missrouting. Choreographer

  10. Adaptive uniform grayscale coded aperture design for high dynamic range compressive spectral imaging

    Science.gov (United States)

    Diaz, Nelson; Rueda, Hoover; Arguello, Henry

    2016-05-01

    Imaging spectroscopy is an important area with many applications in surveillance, agriculture and medicine. The disadvantage of conventional spectroscopy techniques is that they collect the whole datacube. In contrast, compressive spectral imaging systems capture snapshot compressive projections, which are the input of reconstruction algorithms to yield the underlying datacube. Common compressive spectral imagers use coded apertures to perform the coded projections. The coded apertures are the key elements in these imagers since they define the sensing matrix of the system. The proper design of the coded aperture entries leads to a good quality in the reconstruction. In addition, the compressive measurements are prone to saturation due to the limited dynamic range of the sensor, hence the design of coded apertures must consider saturation. The saturation errors in compressive measurements are unbounded and compressive sensing recovery algorithms only provide solutions for bounded noise or bounded with high probability. In this paper it is proposed the design of uniform adaptive grayscale coded apertures (UAGCA) to improve the dynamic range of the estimated spectral images by reducing the saturation levels. The saturation is attenuated between snapshots using an adaptive filter which updates the entries of the grayscale coded aperture based on the previous snapshots. The coded apertures are optimized in terms of transmittance and number of grayscale levels. The advantage of the proposed method is the efficient use of the dynamic range of the image sensor. Extensive simulations show improvements in the image reconstruction of the proposed method compared with grayscale coded apertures (UGCA) and adaptive block-unblock coded apertures (ABCA) in up to 10 dB.

  11. Diversity Analysis of Bit-Interleaved Coded Multiple Beamforming

    CERN Document Server

    Park, Hong Ju

    2008-01-01

    In this paper, diversity analysis of bit-interleaved coded multiple beamforming (BICMB) is extended to the case of general spatial interleavers, removing a condition on their previously known design criteria and quantifying the resulting diversity order. The diversity order is determined by a parameter Qmax which is inherited from the convolutional code and the spatial de-multiplexer used in BICMB. We introduce a method to find this parameter by employing a transfer function approach as in finding the weight spectrum of a convolutional code. By using this method, several Qmax's are shown and verified to be identical with the results from a computer searching program tracing paths on the trellis. The diversity analysis and the method to find the parameter are supported by simulation results.

  12. The Aster code; Code Aster

    Energy Technology Data Exchange (ETDEWEB)

    Delbecq, J.M

    1999-07-01

    The Aster code is a 2D or 3D finite-element calculation code for structures developed by the R and D direction of Electricite de France (EdF). This dossier presents a complete overview of the characteristics and uses of the Aster code: introduction of version 4; the context of Aster (organisation of the code development, versions, systems and interfaces, development tools, quality assurance, independent validation); static mechanics (linear thermo-elasticity, Euler buckling, cables, Zarka-Casier method); non-linear mechanics (materials behaviour, big deformations, specific loads, unloading and loss of load proportionality indicators, global algorithm, contact and friction); rupture mechanics (G energy restitution level, restitution level in thermo-elasto-plasticity, 3D local energy restitution level, KI and KII stress intensity factors, calculation of limit loads for structures), specific treatments (fatigue, rupture, wear, error estimation); meshes and models (mesh generation, modeling, loads and boundary conditions, links between different modeling processes, resolution of linear systems, display of results etc..); vibration mechanics (modal and harmonic analysis, dynamics with shocks, direct transient dynamics, seismic analysis and aleatory dynamics, non-linear dynamics, dynamical sub-structuring); fluid-structure interactions (internal acoustics, mass, rigidity and damping); linear and non-linear thermal analysis; steels and metal industry (structure transformations); coupled problems (internal chaining, internal thermo-hydro-mechanical coupling, chaining with other codes); products and services. (J.S.)

  13. Multiparticle dynamics in the E-phi tracking code ESME

    Energy Technology Data Exchange (ETDEWEB)

    James A. MacLachlan

    2002-06-21

    ESME has developed over a twenty year period from its origins as a program for modeling rf gymnastics to a rather general facility for that fraction of beam dynamics of synchrotrons and storage rings which can be properly treated in the two dimensional longitudinal phase space. The features of this program which serve particularly for multiparticle calculations are described, some underling principles are noted, and illustrative results are given.

  14. COOL: A code for Dynamic Monte Carlo Simulation of molecular dynamics

    Science.gov (United States)

    Barletta, Paolo

    2012-02-01

    Cool is a program to simulate evaporative and sympathetic cooling for a mixture of two gases co-trapped in an harmonic potential. The collisions involved are assumed to be exclusively elastic, and losses are due to evaporation from the trap. Each particle is followed individually in its trajectory, consequently properties such as spatial densities or energy distributions can be readily evaluated. The code can be used sequentially, by employing one output as input for another run. The code can be easily generalised to describe more complicated processes, such as the inclusion of inelastic collisions, or the possible presence of more than two species in the trap. New version program summaryProgram title: COOL Catalogue identifier: AEHJ_v2_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEHJ_v2_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 1 097 733 No. of bytes in distributed program, including test data, etc.: 18 425 722 Distribution format: tar.gz Programming language: C++ Computer: Desktop Operating system: Linux RAM: 500 Mbytes Classification: 16.7, 23 Catalogue identifier of previous version: AEHJ_v1_0 Journal reference of previous version: Comput. Phys. Comm. 182 (2011) 388 Does the new version supersede the previous version?: Yes Nature of problem: Simulation of the sympathetic process occurring for two molecular gases co-trapped in a deep optical trap. Solution method: The Direct Simulation Monte Carlo method exploits the decoupling, over a short time period, of the inter-particle interaction from the trapping potential. The particle dynamics is thus exclusively driven by the external optical field. The rare inter-particle collisions are considered with an acceptance/rejection mechanism, that is, by comparing a random number to the collisional probability

  15. Development of a subchannel analysis code MATRA (Ver. {alpha})

    Energy Technology Data Exchange (ETDEWEB)

    Yoo, Y. J.; Hwang, D. H

    1998-04-01

    A subchannel analysis code MATRA-{alpha}, an interim version of MATRA, has been developed to be run on an IBM PC or HP WS based on the existing CDC CYBER mainframe version of COBRA-IV-I. This MATRA code is a thermal-hydraulic analysis code based on the subchannel approach for calculating the enthalpy and flow distribution in fuel assemblies and reactor cores for both steady-state and transient conditions. MATRA-{alpha} has been provided with an improved structure, various functions, and models to give the more convenient user environment and to increase the code accuracy, various functions, and models to give the more convenient user environment and to increase the code accuracy. Among them, the pressure drop model has been improved to be applied to non-square-lattice rod arrays, and the lateral transport models between adjacent subchannels have been improved to increase the accuracy in predicting two-phase flow phenomena. Also included in this report are the detailed instructions for input data preparation and for auxiliary pre-processors to serve as a guide to those who want to use MATRA-{alpha}. In addition, we compared the predictions of MATRA-{alpha} with the experimental data on the flow and enthalpy distribution in three sample rod-bundle cases to evaluate the performance of MATRA-{alpha}. All the results revealed that the prediction of MATRA-{alpha} were better than those of COBRA-IV-I. (author). 16 refs., 1 tab., 13 figs.

  16. NOAA/DOE CWP structural analysis package. [CWPFLY, CWPEXT, COTEC, and XOTEC codes

    Energy Technology Data Exchange (ETDEWEB)

    Pompa, J.A.; Lunz, D.F. (eds.)

    1979-09-01

    The theoretical development and computer code user's manual for analysis of the Ocean Thermal Energy Conversion (OTEC) plant cold water pipe (CWP) are presented. The analysis of the CWP includes coupled platform/CWP loadngs and dynamic responses. This report with the exception of the Introduction and Appendix F was orginally published as Hydronautics, Inc., Technical Report No. 7825-2 (by Barr, Chang, and Thasanatorn) in November 1978. A detailed theoretical development of the equations describing the coupled platform/CWP system and preliminary validation efforts are described. The appendices encompass a complete user's manual, describing the inputs, outputs and operation of the four component programs, and detail changes and updates implemented since the original release of the code by Hydronautics. The code itself is available through NOAA's Office of Ocean Technology and Engineering Services.

  17. 76 FR 56413 - Building Energy Codes Cost Analysis

    Science.gov (United States)

    2011-09-13

    ... analysis using an appropriate building energy estimation tool. DOE intends to use the EnergyPlus \\3...-revision codes, two prototype buildings would be analyzed--one that exactly complies with the pre-revision... in the following sections. \\4\\ ``Exactly complies'' means that the prototype complies with...

  18. Course and Research Analysis Using a Coded Classification System.

    Science.gov (United States)

    Lochstet, Gwenn S.

    1997-01-01

    A system of course analysis was developed and used to code and compare faculty research, courses, and library materials in the Mathematics, Physics, and Statistics departments of the University of South Carolina. The purpose is to provide a guide in determining how well the library's collection supports the academic needs of these departments. (10…

  19. Global Analysis of Nonlinear Dynamics

    CERN Document Server

    Luo, Albert

    2012-01-01

    Global Analysis of Nonlinear Dynamics collects chapters on recent developments in global analysis of non-linear dynamical systems with a particular emphasis on cell mapping methods developed by Professor C.S. Hsu of the University of California, Berkeley. This collection of contributions prepared by a diverse group of internationally recognized researchers is intended to stimulate interests in global analysis of complex and high-dimensional nonlinear dynamical systems, whose global properties are largely unexplored at this time. This book also: Presents recent developments in global analysis of non-linear dynamical systems Provides in-depth considerations and extensions of cell mapping methods Adopts an inclusive style accessible to non-specialists and graduate students Global Analysis of Nonlinear Dynamics is an ideal reference for the community of nonlinear dynamics in different disciplines including engineering, applied mathematics, meteorology, life science, computational science, and medicine.  

  20. A Generic Top-Down Dynamic-Programming Approach to Prefix-Free Coding

    CERN Document Server

    Golin, Mordecai; Yu, Jiajin

    2008-01-01

    Given a probability distribution over a set of n words to be transmitted, the Huffman Coding problem is to find a minimal-cost prefix free code for transmitting those words. The basic Huffman coding problem can be solved in O(n log n) time but variations are more difficult. One of the standard techniques for solving these variations utilizes a top-down dynamic programming approach. In this paper we show that this approach is amenable to dynamic programming speedup techniques, permitting a speedup of an order of magnitude for many algorithms in the literature for such variations as mixed radix, reserved length and one-ended coding. These speedups are immediate implications of a general structural property that permits batching together the calculation of many DP entries.

  1. Verification of the Mimir-N2 Joyo plant dynamic code

    Energy Technology Data Exchange (ETDEWEB)

    Yoshida, Akihiro [Japan Nuclear Cycle Development Inst., Oarai, Ibaraki (Japan). Oarai Engineering Center; Kuroha, Takaya [Nuclear Energy System Inc., Tokyo (Japan)

    2002-06-01

    Passive safety systems at Experimental Fast Reactor JOYO were studied to demonstrate the inherent safety of MOX fueled sodium cooled fast breeder reactors. The Mimir-N2 analysis code, developed to analyze JOYO plant kinetics, was selected as the standard code. To increase the reliability, Mimir-N2 was modified based on data from plant characteristics and natural circulation tests in JOYO. JOYO operational data suggest that the burn-up dependency of the power reactivity coefficient could be due to the reactivity shift caused by decrease of fuel pellet thermal expansion in the axial direction. Based on the relationship between the measured power reactivity coefficient and the core averaged burn-up, burn-up dependency was estimated and introduced to the Mimir-N2 model. This brought good correspondence between calculated and measured values for a step reactivity response test. Calculated plant parameters including power range neutron monitor response and fuel subassembly outlet coolant temperature corresponded to measured values. Mimir-N2 could simulate plant dynamics such as the perturbations due to core support plate thermal expansion. (author)

  2. The development of an intelligent interface to a computational fluid dynamics flow-solver code

    Science.gov (United States)

    Williams, Anthony D.

    1988-01-01

    Researchers at NASA Lewis are currently developing an 'intelligent' interface to aid in the development and use of large, computational fluid dynamics flow-solver codes for studying the internal fluid behavior of aerospace propulsion systems. This paper discusses the requirements, design, and implementation of an intelligent interface to Proteus, a general purpose, three-dimensional, Navier-Stokes flow solver. The interface is called PROTAIS to denote its introduction of artificial intelligence (AI) concepts to the Proteus code.

  3. Room Heat-Up Analysis with GOTHIC code

    Energy Technology Data Exchange (ETDEWEB)

    Jimenez, G.; Olza, J. M.

    2010-07-01

    The GOTHIC{sup T}M computer code is a state-of-the art program for modeling multiphase, multicomponent fluid flow. GOTHIC is rapidly becoming the industry-standard code for performing both containment design basis accident (DBA) analyses and analyses to support equipment qualification. GOTHIC has a flexible nodding structure that allows both lumped parameter and 3-D modeling capabilities. Multidimensional analysis capabilities greatly enhance the study of noncondensable gases and stratification and permit the calculation of flow field details within any given volume.

  4. FARO and KROTOS code simulation and analysis at JRC Ispra

    Energy Technology Data Exchange (ETDEWEB)

    Annunziato, A.; Yerkess, A.; Addabbo, C. [European Commission-Joint Research Centre, Inst. for Systems, Informatics and Safety, 21020 Ispra (Italy)

    1998-01-01

    The paper summarizes relevant results from the pre and post test calculations of fuel coolant interaction and quenching tests performed in the FARO and KROTOS test facilities. The main analytical tools adopted at JRC Ispra are the COMETA and the TEXAS codes. COMETA pre and post test calculations of FARO Test L-20 as well as an application of the code to KROTOS test facility are presented. The analysis provides the need to account for H{sub 2} generation models into the pre-mixing calculations. In addition salient results from the application of TEXAS to FARO and KROTOS tests are shown. (author)

  5. Analysis of LMFBR containment response to an HCDA using a multifield Eulerian code. [MICE code

    Energy Technology Data Exchange (ETDEWEB)

    Chu, H.Y.; Chang, Y.W.

    1977-01-01

    During a hypothetical core disruptive accident (HCDA), a core meltdown may cause the fuel cladding to rupture and the fuel fragments to penetrate into the sodium coolant. The heat in the molten fuel may cause the liquid sodium to boil, changing its phase. The interactions between materials are so complicated that a single-material model with homogenized material properties is not adequate. In order to analyze the above phenomena more realistically, a Multifield Implicit Continuous-Fluid Eulerian containment code (MICE) is being developed at Argonne National Laboratory (ANL) to solve the multifield fluid-flow problems in which the interpenetrations of materials, heat transfer, and phase changes are considered in the analysis. The hydrodynamics of the MICE code is based upon the implicit multifield (IMF) method developed by Harlow and Amsden. A partial donor-cell formulation is used for the calculation of the convective fluxes to minimize the truncation errors, while the Newton-Raphson method is used for the numerical iterations. An implicit treatment of the mass convection together with the equation of state for each material enables the method to be applicable to both compressible and incompressible flows. A partial implicit treatment of the momentum-exchange functions allows the coupling drag forces between two material fields to range from very weak to those strong enough to tie the fields completely. The differential equations and exchange functions used in the MICE code, and the treatment of the fluid and structure interactions as well as the numerical procedure are described. Two sample calculations are given to illustrate the present capability of the MICE code.

  6. Issues in computational fluid dynamics code verification and validation

    Energy Technology Data Exchange (ETDEWEB)

    Oberkampf, W.L.; Blottner, F.G.

    1997-09-01

    A broad range of mathematical modeling errors of fluid flow physics and numerical approximation errors are addressed in computational fluid dynamics (CFD). It is strongly believed that if CFD is to have a major impact on the design of engineering hardware and flight systems, the level of confidence in complex simulations must substantially improve. To better understand the present limitations of CFD simulations, a wide variety of physical modeling, discretization, and solution errors are identified and discussed. Here, discretization and solution errors refer to all errors caused by conversion of the original partial differential, or integral, conservation equations representing the physical process, to algebraic equations and their solution on a computer. The impact of boundary conditions on the solution of the partial differential equations and their discrete representation will also be discussed. Throughout the article, clear distinctions are made between the analytical mathematical models of fluid dynamics and the numerical models. Lax`s Equivalence Theorem and its frailties in practical CFD solutions are pointed out. Distinctions are also made between the existence and uniqueness of solutions to the partial differential equations as opposed to the discrete equations. Two techniques are briefly discussed for the detection and quantification of certain types of discretization and grid resolution errors.

  7. Starfinder a code for crowded stellar fields analysis

    CERN Document Server

    Diolaiti, E; Bonaccini, D; Close, L M; Currie, D; Parmeggiani, G

    1999-01-01

    Starfinder is an IDL code for the deep analysis of stellar fields, designed for well-sampled images with high and low Strehl factor. An important feature is represented by the possibility to measure the anisoplanatic effect in wide-field Adaptive Optics observations and exploit this knowledge to improve the analysis of the observed field. A description of the method and applications to real AO data are presented.

  8. A Multi-Phase Chemo-Dynamical SPH Code for Galaxy Evolution

    OpenAIRE

    Berczik, P.; Hensler, G.; Theis, Ch.; Spurzem, R.

    2003-01-01

    In this paper we present some test results of our newly developed Multi-Phase Chemo-Dynamical Smoothed Particle Hydrodynamics (MP- CD-SPH) code for galaxy evolution. At first, we present a test of the ``pure'' hydro SPH part of the code. Then we describe and test the multi-phase description of the gaseous components of the interstellar matter. In this second part we also compare our condensation and evaporation description with the results of a previous 2d multi-phase hydrodynamic mesh code.

  9. Development and assessment of best estimate integrated safety analysis code

    Energy Technology Data Exchange (ETDEWEB)

    Chung, Bub Dong; Lee, Young Jin; Hwang, Moon Kyu (and others)

    2007-03-15

    Improvement of the integrated safety analysis code MARS3.0 has been carried out and a multi-D safety analysis application system has been established. Iterative matrix solver and parallel processing algorithm have been introduced, and a LINUX version has been generated to enable MARS to run in cluster PCs. MARS variables and sub-routines have been reformed and modularised to simplify code maintenance. Model uncertainty analyses have been performed for THTF, FLECHT, NEPTUN, and LOFT experiments as well as APR1400 plant. Participations in international cooperation research projects such as OECD BEMUSE, SETH, PKL, BFBT, and TMI-2 have been actively pursued as part of code assessment efforts. The assessment, evaluation and experimental data obtained through international cooperation projects have been registered and maintained in the T/H Databank. Multi-D analyses of APR1400 LBLOCA, DVI Break, SLB, and SGTR have been carried out as a part of application efforts in multi-D safety analysis. GUI based 3D input generator has been developed for user convenience. Operation of the MARS Users Group (MUG) was continued and through MUG, the technology has been transferred to 24 organisations. A set of 4 volumes of user manuals has been compiled and the correction reports for the code errors reported during MARS development have been published.

  10. MATHEMATICAL FRAMEWORK FOR THE ANALYSIS OF DYNAMC STOCHASTIC SYSTEMS WITH THE RAVEN CODE

    Energy Technology Data Exchange (ETDEWEB)

    C. Rabiti; D. Mandelli; J. Cogliati; R. Kinoshita

    2013-05-01

    RAVEN (Reactor Analysis and Virtual control Environment) is a software code under development at Idaho National Laboratory aimed at performing probabilistic risk assessment and uncertainty quantification using RELAP-7, for which it acts also as a simulation controller. In this paper we will present the equations characterizing a dynamic stochastic system and we will then discuss the behavior of each stochastic term and how it is accounted for in the RAVEN software design. Moreover we will present preliminary results of the implementation.

  11. Statistical mechanics analysis of LDPC coding in MIMO Gaussian channels

    Energy Technology Data Exchange (ETDEWEB)

    Alamino, Roberto C; Saad, David [Neural Computing Research Group, Aston University, Birmingham B4 7ET (United Kingdom)

    2007-10-12

    Using analytical methods of statistical mechanics, we analyse the typical behaviour of a multiple-input multiple-output (MIMO) Gaussian channel with binary inputs under low-density parity-check (LDPC) network coding and joint decoding. The saddle point equations for the replica symmetric solution are found in particular realizations of this channel, including a small and large number of transmitters and receivers. In particular, we examine the cases of a single transmitter, a single receiver and symmetric and asymmetric interference. Both dynamical and thermodynamical transitions from the ferromagnetic solution of perfect decoding to a non-ferromagnetic solution are identified for the cases considered, marking the practical and theoretical limits of the system under the current coding scheme. Numerical results are provided, showing the typical level of improvement/deterioration achieved with respect to the single transmitter/receiver result, for the various cases.

  12. Flight Dynamics Analysis Branch 2005 Technical Highlights

    Science.gov (United States)

    2005-01-01

    This report summarizes the major activities and accomplishments carried out by the Flight Dynamics Analysis Branch (FDAB), Code 595, in support of flight projects and technology development initiatives in Fiscal Year (FY) 2005. The report is intended to serve as a summary of the type of support carried out by the FDAB, as well as a concise reference of key accomplishments and mission experience derived from the various mission support roles. The primary focus of the FDAB is to provide expertise in the disciplines of flight dynamics including spacecraft navigation (autonomous and ground based); spacecraft trajectory design and maneuver planning; attitude analysis; attitude determination and sensor calibration; and attitude control subsystem (ACS) analysis and design. The FDAB currently provides support for missions and technology development projects involving NASA, other government agencies, academia, and private industry.

  13. Dynamic divisive normalization predicts time-varying value coding in decision-related circuits.

    Science.gov (United States)

    Louie, Kenway; LoFaro, Thomas; Webb, Ryan; Glimcher, Paul W

    2014-11-26

    Normalization is a widespread neural computation, mediating divisive gain control in sensory processing and implementing a context-dependent value code in decision-related frontal and parietal cortices. Although decision-making is a dynamic process with complex temporal characteristics, most models of normalization are time-independent and little is known about the dynamic interaction of normalization and choice. Here, we show that a simple differential equation model of normalization explains the characteristic phasic-sustained pattern of cortical decision activity and predicts specific normalization dynamics: value coding during initial transients, time-varying value modulation, and delayed onset of contextual information. Empirically, we observe these predicted dynamics in saccade-related neurons in monkey lateral intraparietal cortex. Furthermore, such models naturally incorporate a time-weighted average of past activity, implementing an intrinsic reference-dependence in value coding. These results suggest that a single network mechanism can explain both transient and sustained decision activity, emphasizing the importance of a dynamic view of normalization in neural coding.

  14. Phantom of RAMSES (POR): A new Milgromian dynamicsN-body code

    Science.gov (United States)

    Lüghausen, Fabian; Famaey, Benoit; Kroupa, Pavel

    2015-02-01

    Since its first formulation in 1983, Milgromian dynamics (MOND) has been very successful in predicting the gravitational potential of galaxies from the distribution of baryons alone, including general scaling relations and detailed rotation curves of large statistical samples of individual galaxies covering a large range of masses and sizes. Most predictions however rely on static models, and only a handful of N-body codes have been developed over the years to investigate the consequences of the Milgromian framework for the dynamics of complex evolving dynamical systems. In this work, we present a new Milgromian N-body code, which is a customized version of the RAMSES code (Teyssier 2002) and thus comes with all its features: it includes particles and gas dynamics, and importantly allows for high spatial resolution of complex systems due to the adaptive mesh refinement (AMR) technique. It further allows the direct comparison between Milgromian simulations and standard Newtonian simulations with dark matter particles. We provide basic tests of this customized code and demonstrate its performance by presenting N-body computations of dark-matter-free spherical equilibrium models as well as dark-matter-free disk galaxies in Milgromian dynamics.

  15. Sandia National Laboratories environmental fluid dynamics code. Marine Hydrokinetic Module User's Manual

    Energy Technology Data Exchange (ETDEWEB)

    James, Scott Carlton; Roberts, Jesse D

    2014-03-01

    This document describes the marine hydrokinetic (MHK) input file and subroutines for the Sandia National Laboratories Environmental Fluid Dynamics Code (SNL-EFDC), which is a combined hydrodynamic, sediment transport, and water quality model based on the Environmental Fluid Dynamics Code (EFDC) developed by John Hamrick [1], formerly sponsored by the U.S. Environmental Protection Agency, and now maintained by Tetra Tech, Inc. SNL-EFDC has been previously enhanced with the incorporation of the SEDZLJ sediment dynamics model developed by Ziegler, Lick, and Jones [2-4]. SNL-EFDC has also been upgraded to more accurately simulate algae growth with specific application to optimizing biomass in an open-channel raceway for biofuels production [5]. A detailed description of the input file containing data describing the MHK device/array is provided, along with a description of the MHK FORTRAN routine. Both a theoretical description of the MHK dynamics as incorporated into SNL-EFDC and an explanation of the source code are provided. This user manual is meant to be used in conjunction with the original EFDC [6] and sediment dynamics SNL-EFDC manuals [7]. Through this document, the authors provide information for users who wish to model the effects of an MHK device (or array of devices) on a flow system with EFDC and who also seek a clear understanding of the source code, which is available from staff in the Water Power Technologies Department at Sandia National Laboratories, Albuquerque, New Mexico.

  16. High-Performance Java Codes for Computational Fluid Dynamics

    Science.gov (United States)

    Riley, Christopher; Chatterjee, Siddhartha; Biswas, Rupak; Biegel, Bryan (Technical Monitor)

    2001-01-01

    The computational science community is reluctant to write large-scale computationally -intensive applications in Java due to concerns over Java's poor performance, despite the claimed software engineering advantages of its object-oriented features. Naive Java implementations of numerical algorithms can perform poorly compared to corresponding Fortran or C implementations. To achieve high performance, Java applications must be designed with good performance as a primary goal. This paper presents the object-oriented design and implementation of two real-world applications from the field of Computational Fluid Dynamics (CFD): a finite-volume fluid flow solver (LAURA, from NASA Langley Research Center), and an unstructured mesh adaptation algorithm (2D_TAG, from NASA Ames Research Center). This work builds on our previous experience with the design of high-performance numerical libraries in Java. We examine the performance of the applications using the currently available Java infrastructure and show that the Java version of the flow solver LAURA performs almost within a factor of 2 of the original procedural version. Our Java version of the mesh adaptation algorithm 2D_TAG performs within a factor of 1.5 of its original procedural version on certain platforms. Our results demonstrate that object-oriented software design principles are not necessarily inimical to high performance.

  17. Simulation of IST Turbomachinery Power-Neutral Tests with the ANL Plant Dynamics Code

    Energy Technology Data Exchange (ETDEWEB)

    Moisseytsev, A. [Argonne National Lab. (ANL), Argonne, IL (United States); Sienicki, J. J. [Argonne National Lab. (ANL), Argonne, IL (United States)

    2016-12-13

    The validation of the Plant Dynamics Code (PDC) developed at Argonne National Laboratory (ANL) for the steady-state and transient analysis of supercritical carbon dioxide (sCO2) systems has been continued with new test data from the Naval Nuclear Laboratory (operated by Bechtel Marine Propulsion Corporation) Integrated System Test (IST). Although data from three runs were provided to ANL, only two of the data sets were analyzed and described in this report. The common feature of these tests is the power-neutral operation of the turbine-compressor shaft, where no external power through the alternator was provided during the tests. Instead, the shaft speed was allowed to change dictated by the power balance between the turbine, the compressor, and the power losses in the shaft. The new test data turned out to be important for code validation for several reasons. First, the power-neutral operation of the shaft allows validation of the shaft dynamics equations in asynchronous mode, when the shaft is disconnected from the grid. Second, the shaft speed control with the compressor recirculation (CR) valve not only allows for testing the code control logic itself, but it also serves as a good test for validation of both the compressor surge control and the turbine bypass control actions, since the effect of the CR action on the loop conditions is similar for both of these controls. Third, the varying compressor-inlet temperature change test allows validation of the transient response of the precooler, a shell-and-tube heat exchanger. The first transient simulation of the compressor-inlet temperature variation Test 64661 showed a much slower calculated response of the precooler in the calculations than the test data. Further investigation revealed an error in calculating the heat exchanger tube mass for the PDC dynamic equations that resulted in a slower change in the tube wall temperature than measured. The transient calculations for both tests were done in two steps. The

  18. ls1 mardyn: The massively parallel molecular dynamics code for large systems

    CERN Document Server

    Niethammer, Christoph; Bernreuther, Martin; Buchholz, Martin; Eckhardt, Wolfgang; Heinecke, Alexander; Werth, Stephan; Bungartz, Hans-Joachim; Glass, Colin W; Hasse, Hans; Vrabec, Jadran; Horsch, Martin

    2014-01-01

    The molecular dynamics simulation code ls1 mardyn is presented. It is a highly scalable code, optimized for massively parallel execution on supercomputing architectures, and currently holds the world record for the largest molecular simulation with over four trillion particles. It enables the application of pair potentials to length and time scales which were previously out of scope for molecular dynamics simulation. With an efficient dynamic load balancing scheme, it delivers high scalability even for challenging heterogeneous configurations. Presently, multi-center rigid potential models based on Lennard-Jones sites, point charges and higher-order polarities are supported. Due to its modular design, ls1 mardyn can be extended to new physical models, methods, and algorithms, allowing future users to tailor it to suit their respective needs. Possible applications include scenarios with complex geometries, e.g. for fluids at interfaces, as well as non-equilibrium molecular dynamics simulation of heat and mass ...

  19. Serial Concatenation of RS Codes with Kite Codes: Performance Analysis, Iterative Decoding and Design

    CERN Document Server

    Ma, Xiao; Bai, Baoming; Zhang, Xiaoyi

    2011-01-01

    In this paper, we propose a new ensemble of rateless forward error correction (FEC) codes. The proposed codes are serially concatenated codes with Reed-Solomon (RS) codes as outer codes and Kite codes as inner codes. The inner Kite codes are a special class of prefix rateless low-density parity-check (PRLDPC) codes, which can generate potentially infinite (or as many as required) random-like parity-check bits. The employment of RS codes as outer codes not only lowers down error-floors but also ensures (with high probability) the correctness of successfully decoded codewords. In addition to the conventional two-stage decoding, iterative decoding between the inner code and the outer code are also implemented to improve the performance further. The performance of the Kite codes under maximum likelihood (ML) decoding is analyzed by applying a refined Divsalar bound to the ensemble weight enumerating functions (WEF). We propose a simulation-based optimization method as well as density evolution (DE) using Gaussian...

  20. High-dynamic range compressive spectral imaging by grayscale coded aperture adaptive filtering

    Directory of Open Access Journals (Sweden)

    Nelson Eduardo Diaz

    2015-12-01

    Full Text Available The coded aperture snapshot spectral imaging system (CASSI is an imaging architecture which senses the three dimensional informa-tion of a scene with two dimensional (2D focal plane array (FPA coded projection measurements. A reconstruction algorithm takes advantage of the compressive measurements sparsity to recover the underlying 3D data cube. Traditionally, CASSI uses block-un-block coded apertures (BCA to spatially modulate the light. In CASSI the quality of the reconstructed images depends on the design of these coded apertures and the FPA dynamic range. This work presents a new CASSI architecture based on grayscaled coded apertu-res (GCA which reduce the FPA saturation and increase the dynamic range of the reconstructed images. The set of GCA is calculated in a real-time adaptive manner exploiting the information from the FPA compressive measurements. Extensive simulations show the attained improvement in the quality of the reconstructed images when GCA are employed.  In addition, a comparison between traditional coded apertures and GCA is realized with respect to noise tolerance.

  1. Performance Analysis for Dispensing Mechanism of Active Code

    Institute of Scientific and Technical Information of China (English)

    2003-01-01

    The dispensing mechanism of active code is a key technology in an active network. Conventional capsule and programmable switch approaches have their own shortcomings. The DCCAN(distributed code caching for active network) mechanism presented in this paper overcomes these shortcomings. In this paper, capsule and programmable switch approaches are introduced, and their shortcomings are analyzed. The principle of the DCCAN mechanism is described. The theory analysis in transmit width based on the DCCAN mechanism and capsule approach are described. The theory analysis shows that the DCCAN mechanism has many good characteristics and can improve the efficiency of an active network. Key factors which affect the transmit width based on the DCCAN mechanism are discussed. The using condition of the DCCAN mechanism is also discussed.

  2. Parallelization of the molecular dynamics code GROMOS87 for distributed memory parallel architectures

    NARCIS (Netherlands)

    Green, DG; Meacham, KE; vanHoesel, F; Hertzberger, B; Serazzi, G

    1995-01-01

    This paper describes the techniques and methodologies employed during parallelization of the Molecular Dynamics (MD) code GROMOS87, with the specific requirement that the program run efficiently on a range of distributed-memory parallel platforms. We discuss the preliminary results of our parallel

  3. When are network coding based dynamic multi-homing techniques beneficial?

    DEFF Research Database (Denmark)

    Pereira, Carlos; Aguiar, Ana; Roetter, Daniel Enrique Lucani

    2016-01-01

    high resiliency under time-varying channel conditions. This paper seeks to explore the parameter space and identify the operating regions where dynamic coded policies offer most improvement over static ones in terms of energy consumption and channel utilization. We leverage meta-heuristics to find...

  4. Supporting Dynamic Adaptive Streaming over HTTP in Wireless Meshed Networks using Random Linear Network Coding

    DEFF Research Database (Denmark)

    Hundebøll, Martin; Pedersen, Morten Videbæk; Roetter, Daniel Enrique Lucani

    2014-01-01

    This work studies the potential and impact of the FRANC network coding protocol for delivering high quality Dynamic Adaptive Streaming over HTTP (DASH) in wireless networks. Although DASH aims to tailor the video quality rate based on the available throughput to the destination, it relies...

  5. IllinoisGRMHD: An Open-Source, User-Friendly GRMHD Code for Dynamical Spacetimes

    CERN Document Server

    Etienne, Zachariah B; Haas, Roland; Moesta, Philipp; Shapiro, Stuart L

    2015-01-01

    In the extreme violence of merger and mass accretion, compact objects like black holes and neutron stars are thought to launch some of the most luminous outbursts of electromagnetic and gravitational wave energy in the Universe. Modeling these systems realistically is a central problem in theoretical astrophysics, but has proven extremely challenging, requiring the development of numerical relativity codes that solve Einstein's equations for the spacetime, coupled to the equations of general relativistic (ideal) magnetohydrodynamics (GRMHD) for the magnetized fluids. Over the past decade, the Illinois Numerical Relativity (ILNR) Group's dynamical spacetime, GRMHD code has proven itself as one of the most robust and reliable tools for theoretical modeling of such GRMHD phenomena. Despite the code's outstanding reputation, it was written "by experts and for experts" of the code, with a steep learning curve that would severely hinder community adoption if it were open-sourced. Here we present IllinoisGRMHD, whic...

  6. Coupling a system code with computational fluid dynamics for the simulation of complex coolant reactivity effects

    Energy Technology Data Exchange (ETDEWEB)

    Bertolotto, D.

    2011-11-15

    The current doctoral research is focused on the development and validation of a coupled computational tool, to combine the advantages of computational fluid dynamics (CFD) in analyzing complex flow fields and of state-of-the-art system codes employed for nuclear power plant (NPP) simulations. Such a tool can considerably enhance the analysis of NPP transient behavior, e.g. in the case of pressurized water reactor (PWR) accident scenarios such as Main Steam Line Break (MSLB) and boron dilution, in which strong coolant flow asymmetries and multi-dimensional mixing effects strongly influence the reactivity of the reactor core, as described in Chap. 1. To start with, a literature review on code coupling is presented in Chap. 2, together with the corresponding ongoing projects in the international community. Special reference is made to the framework in which this research has been carried out, i.e. the Paul Scherrer Institute's (PSI) project STARS (Steady-state and Transient Analysis Research for the Swiss reactors). In particular, the codes chosen for the coupling, i.e. the CFD code ANSYS CFX V11.0 and the system code US-NRC TRACE V5.0, are part of the STARS codes system. Their main features are also described in Chap. 2. The development of the coupled tool, named CFX/TRACE from the names of the two constitutive codes, has proven to be a complex and broad-based task, and therefore constraints had to be put on the target requirements, while keeping in mind a certain modularity to allow future extensions to be made with minimal efforts. After careful consideration, the coupling was defined to be on-line, parallel and with non-overlapping domains connected by an interface, which was developed through the Parallel Virtual Machines (PVM) software, as described in Chap. 3. Moreover, two numerical coupling schemes were implemented and tested: a sequential explicit scheme and a sequential semi-implicit scheme. Finally, it was decided that the coupling would be single

  7. LACEwING: A New Moving Group Analysis Code

    Science.gov (United States)

    Riedel, Adric R.; Blunt, Sarah C.; Lambrides, Erini L.; Rice, Emily L.; Cruz, Kelle L.; Faherty, Jacqueline K.

    2017-03-01

    We present a new nearby young moving group (NYMG) kinematic membership analysis code, LocAting Constituent mEmbers In Nearby Groups (LACEwING), a new Catalog of Suspected Nearby Young Stars, a new list of bona fide members of moving groups, and a kinematic traceback code. LACEwING is a convergence-style algorithm with carefully vetted membership statistics based on a large numerical simulation of the Solar Neighborhood. Given spatial and kinematic information on stars, LACEwING calculates membership probabilities in 13 NYMGs and three open clusters within 100 pc. In addition to describing the inputs, methods, and products of the code, we provide comparisons of LACEwING to other popular kinematic moving group membership identification codes. As a proof of concept, we use LACEwING to reconsider the membership of 930 stellar systems in the Solar Neighborhood (within 100 pc) that have reported measurable lithium equivalent widths. We quantify the evidence in support of a population of young stars not attached to any NYMGs, which is a possible sign of new as-yet-undiscovered groups or of a field population of young stars.

  8. Dynamic Event Tree Analysis Through RAVEN

    Energy Technology Data Exchange (ETDEWEB)

    A. Alfonsi; C. Rabiti; D. Mandelli; J. Cogliati; R. A. Kinoshita; A. Naviglio

    2013-09-01

    Conventional Event-Tree (ET) based methodologies are extensively used as tools to perform reliability and safety assessment of complex and critical engineering systems. One of the disadvantages of these methods is that timing/sequencing of events and system dynamics is not explicitly accounted for in the analysis. In order to overcome these limitations several techniques, also know as Dynamic Probabilistic Risk Assessment (D-PRA), have been developed. Monte-Carlo (MC) and Dynamic Event Tree (DET) are two of the most widely used D-PRA methodologies to perform safety assessment of Nuclear Power Plants (NPP). In the past two years, the Idaho National Laboratory (INL) has developed its own tool to perform Dynamic PRA: RAVEN (Reactor Analysis and Virtual control ENvironment). RAVEN has been designed in a high modular and pluggable way in order to enable easy integration of different programming languages (i.e., C++, Python) and coupling with other application including the ones based on the MOOSE framework, developed by INL as well. RAVEN performs two main tasks: 1) control logic driver for the new Thermo-Hydraulic code RELAP-7 and 2) post-processing tool. In the first task, RAVEN acts as a deterministic controller in which the set of control logic laws (user defined) monitors the RELAP-7 simulation and controls the activation of specific systems. Moreover, RAVEN also models stochastic events, such as components failures, and performs uncertainty quantification. Such stochastic modeling is employed by using both MC and DET algorithms. In the second task, RAVEN processes the large amount of data generated by RELAP-7 using data-mining based algorithms. This paper focuses on the first task and shows how it is possible to perform the analysis of dynamic stochastic systems using the newly developed RAVEN DET capability. As an example, the Dynamic PRA analysis, using Dynamic Event Tree, of a simplified pressurized water reactor for a Station Black-Out scenario is presented.

  9. Industrial Computer Codes

    Science.gov (United States)

    Shapiro, Wilbur

    1996-01-01

    This is an overview of new and updated industrial codes for seal design and testing. GCYLT (gas cylindrical seals -- turbulent), SPIRALI (spiral-groove seals -- incompressible), KTK (knife to knife) Labyrinth Seal Code, and DYSEAL (dynamic seal analysis) are covered. CGYLT uses G-factors for Poiseuille and Couette turbulence coefficients. SPIRALI is updated to include turbulence and inertia, but maintains the narrow groove theory. KTK labyrinth seal code handles straight or stepped seals. And DYSEAL provides dynamics for the seal geometry.

  10. A proposed framework for computational fluid dynamics code calibration/validation

    Energy Technology Data Exchange (ETDEWEB)

    Oberkampf, W.L.

    1993-12-31

    The paper reviews the terminology and methodology that have been introduced during the last several years for building confidence n the predictions from Computational Fluid Dynamics (CID) codes. Code validation terminology developed for nuclear reactor analyses and aerospace applications is reviewed and evaluated. Currently used terminology such as ``calibrated code,`` ``validated code,`` and a ``validation experiment`` is discussed along with the shortcomings and criticisms of these terms. A new framework is proposed for building confidence in CFD code predictions that overcomes some of the difficulties of past procedures and delineates the causes of uncertainty in CFD predictions. Building on previous work, new definitions of code verification and calibration are proposed. These definitions provide more specific requirements for the knowledge level of the flow physics involved and the solution accuracy of the given partial differential equations. As part of the proposed framework, categories are also proposed for flow physics research, flow modeling research, and the application of numerical predictions. The contributions of physical experiments, analytical solutions, and other numerical solutions are discussed, showing that each should be designed to achieve a distinctively separate purpose in building confidence in accuracy of CFD predictions. A number of examples are given for each approach to suggest methods for obtaining the highest value for CFD code quality assurance.

  11. IllinoisGRMHD: an open-source, user-friendly GRMHD code for dynamical spacetimes

    Science.gov (United States)

    Etienne, Zachariah B.; Paschalidis, Vasileios; Haas, Roland; Mösta, Philipp; Shapiro, Stuart L.

    2015-09-01

    In the extreme violence of merger and mass accretion, compact objects like black holes and neutron stars are thought to launch some of the most luminous outbursts of electromagnetic and gravitational wave energy in the Universe. Modeling these systems realistically is a central problem in theoretical astrophysics, but has proven extremely challenging, requiring the development of numerical relativity codes that solve Einstein's equations for the spacetime, coupled to the equations of general relativistic (ideal) magnetohydrodynamics (GRMHD) for the magnetized fluids. Over the past decade, the Illinois numerical relativity (ILNR) group's dynamical spacetime GRMHD code has proven itself as a robust and reliable tool for theoretical modeling of such GRMHD phenomena. However, the code was written ‘by experts and for experts’ of the code, with a steep learning curve that would severely hinder community adoption if it were open-sourced. Here we present IllinoisGRMHD, which is an open-source, highly extensible rewrite of the original closed-source GRMHD code of the ILNR group. Reducing the learning curve was the primary focus of this rewrite, with the goal of facilitating community involvement in the code's use and development, as well as the minimization of human effort in generating new science. IllinoisGRMHD also saves computer time, generating roundoff-precision identical output to the original code on adaptive-mesh grids, but nearly twice as fast at scales of hundreds to thousands of cores.

  12. Comparative study of Thermal Hydraulic Analysis Codes for Pressurized Water Reactors

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Yang Hoon; Jang, Mi Suk; Han, Kee Soo [Nuclear Engineering Service and Solution Co. Ltd., Daejeon (Korea, Republic of)

    2015-05-15

    Various codes are used for the thermal hydraulic analysis of nuclear reactors. The use of some codes among these is limited by user and some codes are not even open to general person. Thus, the use of alternative code is considered for some analysis. In this study, simple thermal hydraulic behaviors are analyzed using three codes to show that alternative codes are possible for the analysis of nuclear reactors. We established three models of the simple u-tube manometer using three different codes. RELAP5 (Reactor Excursion and Leak Analysis Program), SPACE (Safety and Performance Analysis CodE for nuclear power Plants), GOTHIC (Generation of Thermal Hydraulic Information for Containments) are selected for this analysis. RELAP5 is widely used codes for the analysis of system behavior of PWRs. SPACE has been developed based on RELAP5 for the analysis of system behavior of PWRs and licensing of the code is in progress. And GOTHIC code also has been widely used for the analysis of thermal hydraulic behavior in the containment system. The internal behavior of u-tube manometer was analyzed by RELAP5, SPACE and GOTHIC codes. The general transient behavior was similar among 3 codes. However, the stabilized status of the transient period analyzed by REPAP5 was different from the other codes. It would be resulted from the different physical models used in the other codes, which is specialized for the multi-phase thermal hydraulic behavior analysis.

  13. Development of the SCHAMBETA code for scoping analysis of HCDA

    Energy Technology Data Exchange (ETDEWEB)

    Suk, Soo Dong; Hahn, D. H

    2000-06-01

    A computer code, SCHAMBETA(Scoping Code for HCDA Analysis using Modified Bethe-Tait Method), is developed to investigate the core disassembly process following a meltdown accident in the framework of a mofified Bethe-Tait method as part of the scoping analysis work to demonstrate the inherent safety of conceptual designs of Korea Advanced Liquid Metal Reactor(KALIMER), A 150 Mwe pool-type sodium cooled prototype fast reactor that uses U-Pu-Zr metallic fuel. The methodologies adopted in the code ared particularly useful to perform various parametric studies for better understanding of core disassembly process of liquid metal fast reactors as well as to estimate upper-limit values of the energy release resulting from a power excursion. In the SCHAMBETA code, the core kinetics and hydraulic behavior of the KALIMER is followed over the period of the super-prompt critical power excursion induced by the ramp reactivity insertion, starting at the time that the sodium-voided core reaches the melting temperature of the metallic fuels. For this purpose, the equations of state of pressure-energy density relationship are derived for the saturated-vapor as well as the solid liquid of metallic uranium fuel, and implemenmted into the formulations of the disassembly reactivity. Mathematical formulations are then developed, in the framework of Modified Bethe-Tait method, in a form relevant to utilize the improved equations of state as well as to consider Doppler effects, for scoping analysis of the super-prompt-critical power excursions driven by a specified rate of reactivity insertion.

  14. Single stock dynamics on high-frequency data: from a compressed coding perspective.

    Directory of Open Access Journals (Sweden)

    Hsieh Fushing

    Full Text Available High-frequency return, trading volume and transaction number are digitally coded via a nonparametric computing algorithm, called hierarchical factor segmentation (HFS, and then are coupled together to reveal a single stock dynamics without global state-space structural assumptions. The base-8 digital coding sequence, which is capable of revealing contrasting aggregation against sparsity of extreme events, is further compressed into a shortened sequence of state transitions. This compressed digital code sequence vividly demonstrates that the aggregation of large absolute returns is the primary driving force for stimulating both the aggregations of large trading volumes and transaction numbers. The state of system-wise synchrony is manifested with very frequent recurrence in the stock dynamics. And this data-driven dynamic mechanism is seen to correspondingly vary as the global market transiting in and out of contraction-expansion cycles. These results not only elaborate the stock dynamics of interest to a fuller extent, but also contradict some classical theories in finance. Overall this version of stock dynamics is potentially more coherent and realistic, especially when the current financial market is increasingly powered by high-frequency trading via computer algorithms, rather than by individual investors.

  15. Single stock dynamics on high-frequency data: from a compressed coding perspective.

    Science.gov (United States)

    Fushing, Hsieh; Chen, Shu-Chun; Hwang, Chii-Ruey

    2014-01-01

    High-frequency return, trading volume and transaction number are digitally coded via a nonparametric computing algorithm, called hierarchical factor segmentation (HFS), and then are coupled together to reveal a single stock dynamics without global state-space structural assumptions. The base-8 digital coding sequence, which is capable of revealing contrasting aggregation against sparsity of extreme events, is further compressed into a shortened sequence of state transitions. This compressed digital code sequence vividly demonstrates that the aggregation of large absolute returns is the primary driving force for stimulating both the aggregations of large trading volumes and transaction numbers. The state of system-wise synchrony is manifested with very frequent recurrence in the stock dynamics. And this data-driven dynamic mechanism is seen to correspondingly vary as the global market transiting in and out of contraction-expansion cycles. These results not only elaborate the stock dynamics of interest to a fuller extent, but also contradict some classical theories in finance. Overall this version of stock dynamics is potentially more coherent and realistic, especially when the current financial market is increasingly powered by high-frequency trading via computer algorithms, rather than by individual investors.

  16. Application of the coupled code RELAP5-QUABOX/CUBBOX in the system analysis of nuclear power plants

    Energy Technology Data Exchange (ETDEWEB)

    Bencik, V.; Feretic, D.; Debrecin, N. [Faculty of Electrical Engineering and Computing, Zagreb (Croatia)

    2002-11-01

    Best estimate codes and methods for the realistic simulation of operational transients and accidents are being developed in two directions. First, computer codes with models of the interaction between multidimensional neutron kinetic and NPP dynamic behavior enable realistic simulation of transients characterized by strong coupling between neutronics and thermal-hydraulics as well as of transients that result in asymmetrical spatial core power distribution. Coupled codes consisting of a system thermal-hydraulic code and a multidimensional neutronic code are being developed worldwide in order to accomplish that task. Secondly, development of the qualified plant nodalization and of the models of plant protection and control systems is important for the realistic system analysis of operational transients and accidents. Comparison of the coupled code and point kinetic results is important for the validation of the coupled code and to gain more experience in the use of the coupled code in realistic analyses. In this paper the results of two transients for NPP Krsko using the coupled code RELAP5-QUABOX/CUBBOX (R5QC) and RELAP5 stand alone code are discussed. (orig.)

  17. Forming Teams for Teaching Programming based on Static Code Analysis

    CERN Document Server

    Arosemena-Trejos, Davis; Clunie, Clifton

    2012-01-01

    The use of team for teaching programming can be effective in the classroom because it helps students to generate and acquire new knowledge in less time, but these groups to be formed without taking into account some respects, may cause an adverse effect on the teaching-learning process. This paper proposes a tool for the formation of team based on the semantics of source code (SOFORG). This semantics is based on metrics extracted from the preferences, styles and good programming practices. All this is achieved through a static analysis of code that each student develops. In this way, you will have a record of students with the information extracted; it evaluates the best formation of teams in a given course. The team's formations are based on programming styles, skills, pair programming or with leader.

  18. Sensitivity Analysis of Hardwired Parameters in GALE Codes

    Energy Technology Data Exchange (ETDEWEB)

    Geelhood, Kenneth J.; Mitchell, Mark R.; Droppo, James G.

    2008-12-01

    The U.S. Nuclear Regulatory Commission asked Pacific Northwest National Laboratory to provide a data-gathering plan for updating the hardwired data tables and parameters of the Gaseous and Liquid Effluents (GALE) codes to reflect current nuclear reactor performance. This would enable the GALE codes to make more accurate predictions about the normal radioactive release source term applicable to currently operating reactors and to the cohort of reactors planned for construction in the next few years. A sensitivity analysis was conducted to define the importance of hardwired parameters in terms of each parameter’s effect on the emission rate of the nuclides that are most important in computing potential exposures. The results of this study were used to compile a list of parameters that should be updated based on the sensitivity of these parameters to outputs of interest.

  19. Forming Teams for Teaching Programming based on Static Code Analysis

    Directory of Open Access Journals (Sweden)

    Davis Arosemena-Trejos

    2012-03-01

    Full Text Available The use of team for teaching programming can be effective in the classroom because it helps students to generate and acquire new knowledge in less time, but these groups to be formed without taking into account some respects, may cause an adverse effect on the teaching-learning process. This paper proposes a tool for the formation of team based on the semantics of source code (SOFORG. This semantics is based on metrics extracted from the preferences, styles and good programming practices. All this is achieved through a static analysis of code that each student develops. In this way, you will have a record of students with the information extracted; it evaluates the best formation of teams in a given course. The team€™s formations are based on programming styles, skills, pair programming or with leader.

  20. MOLOCH computer code for molecular-dynamics simulation of processes in condensed matter

    Directory of Open Access Journals (Sweden)

    Derbenev I.V.

    2011-01-01

    Full Text Available Theoretical and experimental investigation into properties of condensed matter is one of the mainstreams in RFNC-VNIITF scientific activity. The method of molecular dynamics (MD is an innovative method of theoretical materials science. Modern supercomputers allow the direct simulation of collective effects in multibillion atom sample, making it possible to model physical processes on the atomistic level, including material response to dynamic load, radiation damage, influence of defects and alloying additions upon material mechanical properties, or aging of actinides. During past ten years, the computer code MOLOCH has been developed at RFNC-VNIITF. It is a parallel code suitable for massive parallel computing. Modern programming techniques were used to make the code almost 100% efficient. Practically all instruments required for modelling were implemented in the code: a potential builder for different materials, simulation of physical processes in arbitrary 3D geometry, and calculated data processing. A set of tests was developed to analyse algorithms efficiency. It can be used to compare codes with different MD implementation between each other.

  1. The Analysis of SBWR Critical Power Bundle Using Cobrag Code

    Directory of Open Access Journals (Sweden)

    Yohannes Sardjono

    2013-03-01

    Full Text Available The coolant mechanism of SBWR is similar with the Dodewaard Nuclear Power Plant (NPP in the Netherlands that first went critical in 1968. The similarity of both NPP is cooled by natural convection system. These coolant concept is very related with same parameters on fuel bundle design especially fuel bundle length, core pressure drop and core flow rate as well as critical power bundle. The analysis was carried out by using COBRAG computer code. COBRAG computer code is GE Company proprietary. Basically COBRAG computer code is a tool to solve compressible three-dimensional, two fluid, three field equations for two phase flow. The three fields are the vapor field, the continuous liquid field, and the liquid drop field. This code has been applied to analyses model flow and heat transfer within the reactor core. This volume describes the finitevolume equations and the numerical solution methods used to solve these equations. This analysis of same parameters has been done i.e.; inlet sub cooling 20 BTU/lbm and 40 BTU/lbm, 1000 psi pressure and R-factor is 1.038, mass flux are 0.5 Mlb/hr.ft2, 0.75 Mlb/hr.ft2, 1.00 Mlb/hr.ft2 and 1.25 Mlb/hr.ft2. Those conditions based on history operation of some type of the cell fuel bundle line at GE Nuclear Energy. According to the results, it can be concluded that SBWR critical power bundle is 10.5 % less than current BWR critical power bundle with length reduction of 12 ft to 9 ft.

  2. SPILADY: A parallel CPU and GPU code for spin-lattice magnetic molecular dynamics simulations

    Science.gov (United States)

    Ma, Pui-Wai; Dudarev, S. L.; Woo, C. H.

    2016-10-01

    Spin-lattice dynamics generalizes molecular dynamics to magnetic materials, where dynamic variables describing an evolving atomic system include not only coordinates and velocities of atoms but also directions and magnitudes of atomic magnetic moments (spins). Spin-lattice dynamics simulates the collective time evolution of spins and atoms, taking into account the effect of non-collinear magnetism on interatomic forces. Applications of the method include atomistic models for defects, dislocations and surfaces in magnetic materials, thermally activated diffusion of defects, magnetic phase transitions, and various magnetic and lattice relaxation phenomena. Spin-lattice dynamics retains all the capabilities of molecular dynamics, adding to them the treatment of non-collinear magnetic degrees of freedom. The spin-lattice dynamics time integration algorithm uses symplectic Suzuki-Trotter decomposition of atomic coordinate, velocity and spin evolution operators, and delivers highly accurate numerical solutions of dynamic evolution equations over extended intervals of time. The code is parallelized in coordinate and spin spaces, and is written in OpenMP C/C++ for CPU and in CUDA C/C++ for Nvidia GPU implementations. Temperatures of atoms and spins are controlled by Langevin thermostats. Conduction electrons are treated by coupling the discrete spin-lattice dynamics equations for atoms and spins to the heat transfer equation for the electrons. Worked examples include simulations of thermalization of ferromagnetic bcc iron, the dynamics of laser pulse demagnetization, and collision cascades.

  3. Business Ethics: International Analysis of Codes of Ethics and Conduct

    National Research Council Canada - National Science Library

    Josmar Andrade; Kavita Miadaira Hamza; Duarte Miguek Xara-Brasil

    2017-01-01

    ...">Codes of ethics and code of conduct formalize an ideal of expected behavior patterns to managers and employees of organizations, providing standards and orientation that states companies interactions...

  4. Business Ethics: International Analysis of Codes of Ethics and Conduct

    National Research Council Canada - National Science Library

    Josmar Andrade; Kavita Miadaira Hamza; Duarte Miguek Xara-Brasil

    2017-01-01

    Codes of ethics and code of conduct formalize an ideal of expected behavior patterns to managers and employees of organizations, providing standards and orientation that states companies interactions...

  5. On the development of LWR fuel analysis code (1). Analysis of the FEMAXI code and proposal of a new model

    Energy Technology Data Exchange (ETDEWEB)

    Lemehov, Sergei; Suzuki, Motoe [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment

    2000-01-01

    This report summarizes the review on the modeling features of FEMAXI code and proposal of a new theoretical equation model of clad creep on the basis of irradiation-induced microstructure change. It was pointed out that plutonium build-up in fuel matrix and non-uniform radial power profile at high burn-up affect significantly fuel behavior through the interconnected effects with such phenomena as clad irradiation-induced creep, fission gas release, fuel thermal conductivity degradation, rim porous band formation and associated fuel swelling. Therefore, these combined effects should be properly incorporated into the models of the FEMAXI code so that the code can carry out numerical analysis at the level of accuracy and elaboration that modern experimental data obtained in test reactors have. Also, the proposed new mechanistic clad creep model has a general formalism which allows the model to be flexibly applied for clad behavior analysis under normal operation conditions and power transients as well for Zr-based clad materials by the use of established out-of-pile mechanical properties. The model has been tested against experimental data, while further verification is needed with specific emphasis on power ramps and transients. (author)

  6. GASFLOW: A Computational Fluid Dynamics Code for Gases, Aerosols, and Combustion, Volume 2: User's Manual

    Energy Technology Data Exchange (ETDEWEB)

    Nichols, B. D.; Mueller, C.; Necker, G. A.; Travis, J. R.; Spore, J. W.; Lam, K. L.; Royl, P.; Wilson, T. L.

    1998-10-01

    Los Alamos National Laboratory (LANL) and Forschungszentrum Karlsruhe (FzK) are developing GASFLOW, a three-dimensional (3D) fluid dynamics field code as a best-estimate tool to characterize local phenomena within a flow field. Examples of 3D phenomena include circulation patterns; flow stratification; hydrogen distribution mixing and stratification; combustion and flame propagation; effects of noncondensable gas distribution on local condensation and evaporation; and aerosol entrainment, transport, and deposition. An analysis with GASFLOW will result in a prediction of the gas composition and discrete particle distribution in space and time throughout the facility and the resulting pressure and temperature loadings on the walls and internal structures with or without combustion. A major application of GASFLOW is for predicting the transport, mixing, and combustion of hydrogen and other gases in nuclear reactor containment and other facilities. It has been applied to situations involving transporting and distributing combustible gas mixtures. It has been used to study gas dynamic behavior in low-speed, buoyancy-driven flows, as well as sonic flows or diffusion dominated flows; and during chemically reacting flows, including deflagrations. The effects of controlling such mixtures by safety systems can be analyzed. The code version described in this manual is designated GASFLOW 2.1, which combines previous versions of the United States Nuclear Regulatory Commission code HMS (for Hydrogen Mixing Studies) and the Department of Energy and FzK versions of GASFLOW. The code was written in standard Fortran 90. This manual comprises three volumes. Volume I describes the governing physical equations and computational model. Volume II describes how to use the code to set up a model geometry, specify gas species and material properties, define initial and boundary conditions, and specify different outputs, especially graphical displays. Sample problems are included. Volume III

  7. Superimposed Code Theoretic Analysis of Deoxyribonucleic Acid (DNA) Codes and DNA Computing

    Science.gov (United States)

    2010-01-01

    DNA Codes Based on Fibonacci Ensembles of DNA Sequences ”, 2008 IEEE Proceedings of International Symposium on Information Theory, pp. 2292 – 2296...2008, pp. 525-34. 28. A. Macula, et al., “Random Coding Bounds for DNA Codes Based on Fibonacci Ensembles of DNA Sequences ”, 2008 IEEE...component of this innovation is the combinatorial method of bio-memory design and detection that encodes item or process information as numerical sequences

  8. Parallel TREE code for two-component ultracold plasma analysis

    Science.gov (United States)

    Jeon, Byoungseon; Kress, Joel D.; Collins, Lee A.; Grønbech-Jensen, Niels

    2008-02-01

    The TREE method has been widely used for long-range interaction N-body problems. We have developed a parallel TREE code for two-component classical plasmas with open boundary conditions and highly non-uniform charge distributions. The program efficiently handles millions of particles evolved over long relaxation times requiring millions of time steps. Appropriate domain decomposition and dynamic data management were employed, and large-scale parallel processing was achieved using an intermediate level of granularity of domain decomposition and ghost TREE communication. Even though the computational load is not fully distributed in fine grains, high parallel efficiency was achieved for ultracold plasma systems of charged particles. As an application, we performed simulations of an ultracold neutral plasma with a half million particles and a half million time steps. For the long temporal trajectories of relaxation between heavy ions and light electrons, large configurations of ultracold plasmas can now be investigated, which was not possible in past studies.

  9. Code Development and Analysis Program: developmental checkout of the BEACON/MOD2A code. [PWR

    Energy Technology Data Exchange (ETDEWEB)

    Ramsthaler, J. A.; Lime, J. F.; Sahota, M. S.

    1978-12-01

    A best-estimate transient containment code, BEACON, is being developed by EG and G Idaho, Inc. for the Nuclear Regulatory Commission's reactor safety research program. This is an advanced, two-dimensional fluid flow code designed to predict temperatures and pressures in a dry PWR containment during a hypothetical loss-of-coolant accident. The most recent version of the code, MOD2A, is presently in the final stages of production prior to being released to the National Energy Software Center. As part of the final code checkout, seven sample problems were selected to be run with BEACON/MOD2A.

  10. Viriato: A Fourier-Hermite spectral code for strongly magnetized fluid-kinetic plasma dynamics

    Science.gov (United States)

    Loureiro, N. F.; Dorland, W.; Fazendeiro, L.; Kanekar, A.; Mallet, A.; Vilelas, M. S.; Zocco, A.

    2016-09-01

    We report on the algorithms and numerical methods used in Viriato, a novel fluid-kinetic code that solves two distinct sets of equations: (i) the Kinetic Reduced Electron Heating Model (KREHM) equations (Zocco and Schekochihin, 2011) (which reduce to the standard Reduced-MHD equations in the appropriate limit) and (ii) the kinetic reduced MHD (KRMHD) equations (Schekochihin et al., 2009). Two main applications of these equations are magnetized (Alfvénic) plasma turbulence and magnetic reconnection. Viriato uses operator splitting (Strang or Godunov) to separate the dynamics parallel and perpendicular to the ambient magnetic field (assumed strong). Along the magnetic field, Viriato allows for either a second-order accurate MacCormack method or, for higher accuracy, a spectral-like scheme composed of the combination of a total variation diminishing (TVD) third order Runge-Kutta method for the time derivative with a 7th order upwind scheme for the fluxes. Perpendicular to the field Viriato is pseudo-spectral, and the time integration is performed by means of an iterative predictor-corrector scheme. In addition, a distinctive feature of Viriato is its spectral representation of the parallel velocity-space dependence, achieved by means of a Hermite representation of the perturbed distribution function. A series of linear and nonlinear benchmarks and tests are presented, including a detailed analysis of 2D and 3D Orszag-Tang-type decaying turbulence, both in fluid and kinetic regimes.

  11. Coding and Decoding for the Dynamic Decode and Forward Relay Protocol

    CERN Document Server

    Kumar, K Raj

    2008-01-01

    We study the Dynamic Decode and Forward (DDF) protocol for a single half-duplex relay, single-antenna channel with quasi-static fading. The DDF protocol is well-known and has been analyzed in terms of the Diversity-Multiplexing Tradeoff (DMT) in the infinite block length limit. We characterize the finite block length DMT and give new explicit code constructions. The finite block length analysis illuminates a few key aspects that have been neglected in the previous literature: 1) we show that one dominating cause of degradation with respect to the infinite block length regime is the event of decoding error at the relay; 2) we explicitly take into account the fact that the destination does not generally know a priori the relay decision time at which the relay switches from listening to transmit mode. Both the above problems can be tackled by a careful design of the decoding algorithm. In particular, we introduce a decision rejection criterion at the relay based on Forney's decision rule (a variant of the Neyman...

  12. Improving the Salammbo code modelling and using it to better predict radiation belts dynamics

    Science.gov (United States)

    Maget, Vincent; Sicard-Piet, Angelica; Grimald, Sandrine Rochel; Boscher, Daniel

    2016-07-01

    In the framework of the FP7-SPACESTORM project, one objective is to improve the reliability of the model-based predictions performed of the radiation belt dynamics (first developed during the FP7-SPACECAST project). In this purpose we have analyzed and improved the way the simulations using the ONERA Salammbô code are performed, especially in : - Better controlling the driving parameters of the simulation; - Improving the initialization of the simulation in order to be more accurate at most energies for L values between 4 to 6; - Improving the physics of the model. For first point a statistical analysis of the accuracy of the Kp index has been conducted. For point two we have based our method on a long duration simulation in order to extract typical radiation belt states depending on the solar wind stress and geomagnetic activity. For last point we have first improved separately the modelling of different processes acting in the radiation belts and then, we have analyzed the global improvements obtained when simulating them together. We'll discuss here on all these points and on the balance that has to be taken into account between modeled processes to globally improve the radiation belt modelling.

  13. Meanline Analysis of Turbines with Choked Flow in the Object-Oriented Turbomachinery Analysis Code

    Science.gov (United States)

    Hendricks, Eric S.

    2016-01-01

    The Object-Oriented Turbomachinery Analysis Code (OTAC) is a new meanline/streamline turbomachinery modeling tool being developed at NASA GRC. During the development process, a limitation of the code was discovered in relation to the analysis of choked flow in axial turbines. This paper describes the relevant physics for choked flow as well as the changes made to OTAC to enable analysis in this flow regime.

  14. Optimization and analysis of code-division multiplexed TES microcalorimeters

    CERN Document Server

    Fowler, J W; Hilton, G C; Irwin, K D; Schmidt, D R; Stiehl, G M; Swetz, D S; Ullom, J N; Vale., L R

    2011-01-01

    We are developing code-division multiplexing (CDM) systems for transition-edge sensor arrays with the goal of reaching multiplexing factors in the hundreds. We report on x-ray measurements made with a four-channel prototype CDM system that employs a flux-summing architecture, emphasizing data-analysis issues. We describe an empirical method to determine the demodulation matrix that minimizes cross-talk. This CDM system achieves energy resolutions of between 2.3 eV and 3.0 eV FWHM at 5.9 keV.

  15. Development of dynamic simulation code for fuel cycle of fusion reactor. 1. Single pulse operation simulation

    Energy Technology Data Exchange (ETDEWEB)

    Aoki, Isao; Seki, Yasushi [Japan Atomic Energy Research Inst., Naka, Ibaraki (Japan). Naka Fusion Research Establishment; Sasaki, Makoto; Shintani, Kiyonori; Kim, Yeong-Chan

    1997-11-01

    A dynamic simulation code for the fuel cycle of a fusion experimental reactor has been developed. The code follows the fuel inventory change with time in the plasma chamber and the fuel cycle system during a single pulse operation. The time dependence of the fuel inventory distribution is evaluated considering the fuel burn and exhaust in the plasma chamber, purification and supply functions. For each subsystem of the plasma chamber and the fuel cycle system, the fuel inventory equation is written based on the equation of state considering the function of fuel burn, exhaust, purification, and supply. The processing constants of subsystem for the steady states were taken from the values in the ITER Conceptual Design Activity (CDA) report. Using the code, the time dependence of the fuel supply and inventory depending on the burn state and subsystem processing functions are shown. (author)

  16. Research on verification and validation strategy of detonation fluid dynamics code of LAD2D

    Science.gov (United States)

    Wang, R. L.; Liang, X.; Liu, X. Z.

    2017-07-01

    The verification and validation (V&V) is an important approach in the software quality assurance of code in complex engineering application. Reasonable and efficient V&V strategy can achieve twice the result with half the effort. This article introduces the software-Lagrangian adaptive hydrodynamics code in 2D space (LAD2D), which is self-developed software in detonation CFD with plastic-elastic structure. The V&V strategy of this detonation CFD code is presented based on the foundation of V&V methodology for scientific software. The basic framework of the module verification and the function validation is proposed, composing the detonation fluid dynamics model V&V strategy of LAD2D.

  17. Enabling dynamics in face analysis

    NARCIS (Netherlands)

    Dibeklioğlu, H.

    2014-01-01

    Most of the approaches in automatic face analysis rely solely on static appearance. However, temporal analysis of expressions reveals interesting patterns. For a better understanding of the human face, this thesis focuses on temporal changes in the face, and dynamic patterns of expressions. In addit

  18. Dynamic analysis of process reactors

    Energy Technology Data Exchange (ETDEWEB)

    Shadle, L.J.; Lawson, L.O.; Noel, S.D.

    1995-06-01

    The approach and methodology of conducting a dynamic analysis is presented in this poster session in order to describe how this type of analysis can be used to evaluate the operation and control of process reactors. Dynamic analysis of the PyGas{trademark} gasification process is used to illustrate the utility of this approach. PyGas{trademark} is the gasifier being developed for the Gasification Product Improvement Facility (GPIF) by Jacobs-Siffine Engineering and Riley Stoker. In the first step of the analysis, process models are used to calculate the steady-state conditions and associated sensitivities for the process. For the PyGas{trademark} gasifier, the process models are non-linear mechanistic models of the jetting fluidized-bed pyrolyzer and the fixed-bed gasifier. These process sensitivities are key input, in the form of gain parameters or transfer functions, to the dynamic engineering models.

  19. Analysis of Non-binary Hybrid LDPC Codes

    CERN Document Server

    Sassatelli, Lucile

    2008-01-01

    In this paper, we analyse asymptotically a new class of LDPC codes called Non-binary Hybrid LDPC codes, which has been recently introduced. We use density evolution techniques to derive a stability condition for hybrid LDPC codes, and prove their threshold behavior. We study this stability condition to conclude on asymptotic advantages of hybrid LDPC codes compared to their non-hybrid counterparts.

  20. Sandia National Laboratories environmental fluid dynamics code : pH effects user manual.

    Energy Technology Data Exchange (ETDEWEB)

    Janardhanam, Vijay (University of New Mexico, Albuquerque, NM); James, Scott Carlton

    2012-02-01

    This document describes the implementation level changes in the source code and input files of Sandia National Laboratories Environmental Fluid Dynamics Code (SNL-EFDC) that are necessary for including pH effects into algae-growth dynamics. The document also gives a brief introduction to how pH effects are modeled into the algae-growth model. The document assumes that the reader is aware of the existing algae-growth model in SNL-EFDC. The existing model is described by James, Jarardhanam and more theoretical considerations behind modeling pH effects are presented therein. This document should be used in conjunction with the original EFDC manual and the original water-quality manual.

  1. A dynamic code for economic object valuation in prefrontal cortex neurons.

    Science.gov (United States)

    Tsutsui, Ken-Ichiro; Grabenhorst, Fabian; Kobayashi, Shunsuke; Schultz, Wolfram

    2016-09-13

    Neuronal reward valuations provide the physiological basis for economic behaviour. Yet, how such valuations are converted to economic decisions remains unclear. Here we show that the dorsolateral prefrontal cortex (DLPFC) implements a flexible value code based on object-specific valuations by single neurons. As monkeys perform a reward-based foraging task, individual DLPFC neurons signal the value of specific choice objects derived from recent experience. These neuronal object values satisfy principles of competitive choice mechanisms, track performance fluctuations and follow predictions of a classical behavioural model (Herrnstein's matching law). Individual neurons dynamically encode both, the updating of object values from recently experienced rewards, and their subsequent conversion to object choices during decision-making. Decoding from unselected populations enables a read-out of motivational and decision variables not emphasized by individual neurons. These findings suggest a dynamic single-neuron and population value code in DLPFC that advances from reward experiences to economic object values and future choices.

  2. Laminar and Temporal Expression Dynamics of Coding and Noncoding RNAs in the Mouse Neocortex

    Directory of Open Access Journals (Sweden)

    Sofia Fertuzinhos

    2014-03-01

    Full Text Available The hallmark of the cerebral neocortex is its organization into six layers, each containing a characteristic set of cell types and synaptic connections. The transcriptional events involved in laminar development and function still remain elusive. Here, we employed deep sequencing of mRNA and small RNA species to gain insights into transcriptional differences among layers and their temporal dynamics during postnatal development of the mouse primary somatosensory neocortex. We identify a number of coding and noncoding transcripts with specific spatiotemporal expression and splicing patterns. We also identify signature trajectories and gene coexpression networks associated with distinct biological processes and transcriptional overlap between these processes. Finally, we provide data that allow the study of potential miRNA and mRNA interactions. Overall, this study provides an integrated view of the laminar and temporal expression dynamics of coding and noncoding transcripts in the mouse neocortex and a resource for studies of neurodevelopment and transcriptome.

  3. Laminar and temporal expression dynamics of coding and noncoding RNAs in the mouse neocortex.

    Science.gov (United States)

    Fertuzinhos, Sofia; Li, Mingfeng; Kawasawa, Yuka Imamura; Ivic, Vedrana; Franjic, Daniel; Singh, Darshani; Crair, Michael; Sestan, Nenad

    2014-03-13

    The hallmark of the cerebral neocortex is its organization into six layers, each containing a characteristic set of cell types and synaptic connections. The transcriptional events involved in laminar development and function still remain elusive. Here, we employed deep sequencing of mRNA and small RNA species to gain insights into transcriptional differences among layers and their temporal dynamics during postnatal development of the mouse primary somatosensory neocortex. We identify a number of coding and noncoding transcripts with specific spatiotemporal expression and splicing patterns. We also identify signature trajectories and gene coexpression networks associated with distinct biological processes and transcriptional overlap between these processes. Finally, we provide data that allow the study of potential miRNA and mRNA interactions. Overall, this study provides an integrated view of the laminar and temporal expression dynamics of coding and noncoding transcripts in the mouse neocortex and a resource for studies of neurodevelopment and transcriptome.

  4. Users manual for CAFE-3D : a computational fluid dynamics fire code.

    Energy Technology Data Exchange (ETDEWEB)

    Khalil, Imane; Lopez, Carlos; Suo-Anttila, Ahti Jorma (Alion Science and Technology, Albuquerque, NM)

    2005-03-01

    The Container Analysis Fire Environment (CAFE) computer code has been developed to model all relevant fire physics for predicting the thermal response of massive objects engulfed in large fires. It provides realistic fire thermal boundary conditions for use in design of radioactive material packages and in risk-based transportation studies. The CAFE code can be coupled to commercial finite-element codes such as MSC PATRAN/THERMAL and ANSYS. This coupled system of codes can be used to determine the internal thermal response of finite element models of packages to a range of fire environments. This document is a user manual describing how to use the three-dimensional version of CAFE, as well as a description of CAFE input and output parameters. Since this is a user manual, only a brief theoretical description of the equations and physical models is included.

  5. N-MODY: A Code for Collisionless N-body Simulations in Modified Newtonian Dynamics

    Science.gov (United States)

    Londrillo, Pasquale; Nipoti, Carlo

    2011-02-01

    N-MODY is a parallel particle-mesh code for collisionless N-body simulations in modified Newtonian dynamics (MOND). N-MODY is based on a numerical potential solver in spherical coordinates that solves the non-linear MOND field equation, and is ideally suited to simulate isolated stellar systems. N-MODY can be used also to compute the MOND potential of arbitrary static density distributions. A few applications of N-MODY indicate that some astrophysically relevant dynamical processes are profoundly different in MOND and in Newtonian gravity with dark matter.

  6. DISCRETE DYNAMIC MODEL OF BEVEL GEAR – VERIFICATION THE PROGRAM SOURCE CODE FOR NUMERICAL SIMULATION

    Directory of Open Access Journals (Sweden)

    Krzysztof TWARDOCH

    2014-06-01

    Full Text Available In the article presented a new model of physical and mathematical bevel gear to study the influence of design parameters and operating factors on the dynamic state of the gear transmission. Discusses the process of verifying proper operation of copyright calculation program used to determine the solutions of the dynamic model of bevel gear. Presents the block diagram of a computing algorithm that was used to create a program for the numerical simulation. The program source code is written in an interactive environment to perform scientific and engineering calculations, MATLAB

  7. A QR Code Based Processing For Dynamic and Transparent Seat Allocation in Indian Railway

    Directory of Open Access Journals (Sweden)

    Man Mohan Swarup

    2012-05-01

    Full Text Available Indian Railway is worlds largest human transport system, is currently dealing with a lot of problems, such as availability of confirming seat to waiting list passengers after its departure. However there is no such system that presently in Indian railway gives accommodation to waiting list passengers during their journey. This paper proposes the Dynamic Seat Allocation (DSA system considering the advantage of QR code processing along with one of the standards of wireless communication. Moreover, dynamic authentication to the wireless device is incorporate which is passenger specific. Through this research paper our approach is to make fair processing in seat reservation or allocation in Indian Railway.

  8. Static Analysis for Dynamic XML

    DEFF Research Database (Denmark)

    Christensen, Aske Simon; Møller, Anders; Schwartzbach, Michael Ignatieff

    2002-01-01

    We describe the summary graph lattice for dataflow analysis of programs that dynamically construct XML documents. Summary graphs have successfully been used to provide static guarantees in the JWIG language for programming interactive Web services. In particular, the JWIG compiler is able to check...... validity of dynamically generated XHTML documents and to type check dynamic form data. In this paper we present summary graphs and indicate their applicability for various scenarios. We also show that the expressive power of summary graphs is similar to that of the regular expression types from XDuce......, but that the extra structure in summary graphs makes them more suitable for certain program analyses....

  9. Static Analysis for Dynamic XML

    DEFF Research Database (Denmark)

    Christensen, Aske Simon; Møller, Anders; Schwartzbach, Michael Ignatieff

    2002-01-01

    We describe the summary graph lattice for dataflow analysis of programs that dynamically construct XML documents. Summary graphs have successfully been used to provide static guarantees in the JWIG language for programming interactive Web services. In particular, the JWIG compiler is able to check...... validity of dynamically generated XHTML documents and to type check dynamic form data. In this paper we present summary graphs and indicate their applicability for various scenarios. We also show that the expressive power of summary graphs is similar to that of the regular expression types from XDuce...

  10. Neural dynamics of reward probability coding: a Magnetoencephalographic study in humans

    Directory of Open Access Journals (Sweden)

    Julie eThomas

    2013-11-01

    Full Text Available Prediction of future rewards and discrepancy between actual and expected outcomes (prediction error are crucial signals for adaptive behavior. In humans, a number of fMRI studies demonstrated that reward probability modulates these two signals in a large brain network. Yet, the spatio-temporal dynamics underlying the neural coding of reward probability remains unknown. Here, using magnetoencephalography, we investigated the neural dynamics of prediction and reward prediction error computations while subjects learned to associate cues of slot machines with monetary rewards with different probabilities. We showed that event-related magnetic fields (ERFs arising from the visual cortex coded the expected reward value 155 ms after the cue, demonstrating that reward value signals emerge early in the visual stream. Moreover, a prediction error was reflected in ERF peaking 300 ms after the rewarded outcome and showing decreasing amplitude with higher reward probability. This prediction error signal was generated in a network including the anterior and posterior cingulate cortex. These findings pinpoint the spatio-temporal characteristics underlying reward probability coding. Together, our results provide insights into the neural dynamics underlying the ability to learn probabilistic stimuli-reward contingencies.

  11. Dynamic Model for the Z Accelerator Vacuum Section Based on Transmission Line Code%Dynamic Model for the Z Accelerator Vacuum Section Based on Transmission Line Code

    Institute of Scientific and Technical Information of China (English)

    呼义翔; 雷天时; 吴撼宇; 郭宁; 韩娟娟; 邱爱慈; 王亮平; 黄涛; 丛培天; 张信军; 李岩; 曾正中; 孙铁平

    2011-01-01

    The transmission-line-circuit model of the Z accelerator, developed originally by W. A. STYGAR, P. A. CORCORAN, et al., is revised. The revised model uses different calculations for the electron loss and flow impedance in the magnetically insulated transmission line system of the Z accelerator before and after magnetic insulation is established. By including electron pressure and zero electric field at the cathode, a closed set of equations is obtained at each time step, and dynamic shunt resistance (used to represent any electron loss to the anode) and flow impedance are solved, which have been incorporated into the transmission line code for simulations of the vacuum section in the Z accelerator. Finally, the results are discussed in comparison with earlier findings to show the effectiveness and limitations of the model.

  12. A general-purpose contact detection algorithm for nonlinear structural analysis codes

    Energy Technology Data Exchange (ETDEWEB)

    Heinstein, M.W.; Attaway, S.W.; Swegle, J.W.; Mello, F.J.

    1993-05-01

    A new contact detection algorithm has been developed to address difficulties associated with the numerical simulation of contact in nonlinear finite element structural analysis codes. Problems including accurate and efficient detection of contact for self-contacting surfaces, tearing and eroding surfaces, and multi-body impact are addressed. The proposed algorithm is portable between dynamic and quasi-static codes and can efficiently model contact between a variety of finite element types including shells, bricks, beams and particles. The algorithm is composed of (1) a location strategy that uses a global search to decide which slave nodes are in proximity to a master surface and (2) an accurate detailed contact check that uses the projected motions of both master surface and slave node. In this report, currently used contact detection algorithms and their associated difficulties are discussed. Then the proposed algorithm and how it addresses these problems is described. Finally, the capability of the new algorithm is illustrated with several example problems.

  13. Analysis of selected Halden overpressure tests using the FALCON code

    Energy Technology Data Exchange (ETDEWEB)

    Khvostov, G., E-mail: grigori.khvostov@psi.ch [Paul Scherrer Institut, CH 5232 Villigen PSI (Switzerland); Wiesenack, W. [Institute for Energy Technology – OECD Halden Reactor Project, P.O. Box 173, N-1751 Halden (Norway)

    2016-12-15

    Highlights: • We analyse four Halden overpressure tests. • We determine a critical overpressure value for lift-off in a BWR fuel sample. • We show the role of bonding in over-pressurized rod behaviour. • We analytically quantify the degree of bonding via its impact on cladding elongation. • We hypothesize on an effect of circumferential cracks on thermal fuel response to overpressure. • We estimate a thermal effect of circumferential cracks based on interpretation of the data. - Abstract: Four Halden overpressure (lift-off) tests using samples with uranium dioxide fuel pre-irradiated in power reactors to a burnup of 60 MWd/kgU are analyzed. The FALCON code coupled to a mechanistic model, GRSW-A for fission gas release and gaseous-bubble swelling is used for the calculation. The advanced version of the FALCON code is shown to be applicable to best-estimate predictive analysis of overpressure tests using rods without, or weak pellet-cladding bonding, as well as scoping analysis of tests with fuels where stronger pellet-cladding bonding occurs. Significant effects of bonding and fuel cracking/relocation on the thermal and mechanical behaviour of highly over-pressurized rods are shown. The effect of bonding is particularly pronounced in the tests with the PWR samples. The present findings are basically consistent with an earlier analysis based on a direct interpretation of the experimental data. Additionally, in this paper, the specific effects are quantified based on the comparison of the data with the results of calculation. It is concluded that the identified effects are largely beyond the current traditional fuel-rod licensing analysis methods.

  14. Interface design of VSOP'94 computer code for safety analysis

    Science.gov (United States)

    Natsir, Khairina; Yazid, Putranto Ilham; Andiwijayakusuma, D.; Wahanani, Nursinta Adi

    2014-09-01

    Today, most software applications, also in the nuclear field, come with a graphical user interface. VSOP'94 (Very Superior Old Program), was designed to simplify the process of performing reactor simulation. VSOP is a integrated code system to simulate the life history of a nuclear reactor that is devoted in education and research. One advantage of VSOP program is its ability to calculate the neutron spectrum estimation, fuel cycle, 2-D diffusion, resonance integral, estimation of reactors fuel costs, and integrated thermal hydraulics. VSOP also can be used to comparative studies and simulation of reactor safety. However, existing VSOP is a conventional program, which was developed using Fortran 65 and have several problems in using it, for example, it is only operated on Dec Alpha mainframe platforms and provide text-based output, difficult to use, especially in data preparation and interpretation of results. We develop a GUI-VSOP, which is an interface program to facilitate the preparation of data, run the VSOP code and read the results in a more user friendly way and useable on the Personal 'Computer (PC). Modifications include the development of interfaces on preprocessing, processing and postprocessing. GUI-based interface for preprocessing aims to provide a convenience way in preparing data. Processing interface is intended to provide convenience in configuring input files and libraries and do compiling VSOP code. Postprocessing interface designed to visualized the VSOP output in table and graphic forms. GUI-VSOP expected to be useful to simplify and speed up the process and analysis of safety aspects.

  15. Preliminary Coupling of MATRA Code for Multi-physics Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Seongjin; Choi, Jinyoung; Yang, Yongsik; Kwon, Hyouk; Hwang, Daehyun [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)

    2014-05-15

    The boundary conditions such as the inlet temperature, mass flux, averaged heat flux, power distributions of the rods, and core geometry is given by constant values or functions of time. These conditions are separately calculated and provided by other codes, such as a neutronics or a system codes, into the MATRA code. In addition, the coupling of several codes in the different physics field is focused and embodied. In this study, multiphysics coupling methods were developed for a subchannel code (MATRA) with neutronics codes (MASTER, DeCART) and a fuel performance code (FRAPCON-3). Preliminary evaluation results for representative sample cases are presented. The MASTER and DeCART codes provide the power distribution of the rods in the core to the MATRA code. In case of the FRAPCON-3 code, the variation of the rod diameter induced by the thermal expansion is yielded and provided. The MATRA code transfers the thermal-hydraulic conditions that each code needs. Moreover, the coupling method with each code is described.

  16. Code Based Analysis for Object-Oriented Systems

    Institute of Scientific and Technical Information of China (English)

    Swapan Bhattacharya; Ananya Kanjilal

    2006-01-01

    The basic features of object-oriented software makes it difficult to apply traditional testing methods in objectoriented systems. Control Flow Graph (CFG) is a well-known model used for identification of independent paths in procedural software. This paper highlights the problem of constructing CFG in object-oriented systems and proposes a new model named Extended Control Flow Graph (ECFG) for code based analysis of Object-Oriented (OO) software. ECFG is a layered CFG where nodes refer to methods rather than statements. A new metrics - Extended Cyclomatic Complexity (E-CC) is developed which is analogous to McCabe's Cyclomatic Complexity (CC) and refers to the number of independent execution paths within the OO software. The different ways in which CFG's of individual methods are connected in an ECFG are presented and formulas for E-CC for these different cases are proposed. Finally we have considered an example in Java and based on its ECFG, applied these cases to arrive at the E-CC of the total system as well as proposed a methodology for calculating the basis set, i.e., the set of independent paths for the OO system that will help in creation of test cases for code testing.

  17. Outage probability analysis for superposition coded symmetric relaying

    Institute of Scientific and Technical Information of China (English)

    WU Yi; ZHENG Meng; FEI ZeSong; LARSSON Erik G.; KUANG JingMing

    2013-01-01

    Superposition coded symmetric relaying is a bandwidth-efficient cooperative scheme where each source node simultaneously transmits both its own "local" packet and "relay" packet that originated at its partner by adding the modulated local and relay signals in Euclidean space. This paper investigates the power allocation and outage probability of a superposition coded symmetric relaying system with finite-constellation signaling. We first derive the mutual information (MI) metrics for the system. The derived MI metrics consist of two parts: one represents the MI conveyed by the modulated signal corresponding to its own data, and the other represents the MI conveyed by the modulated signal corresponding to its partner's data. Using MI based effective signal-to-noise ratio mapping technique, we attain expressions for the outage probability. Furthermore, we discuss power allocation policies that minimize the outage probability. Simulation results are presented to verify the correctness of the outage probability analysis and the benefits of the power allocation.

  18. Static Analysis of Dynamic Languages

    DEFF Research Database (Denmark)

    Madsen, Magnus

    on the behaviour of these languages. A common theme is the reliance on static program analysis to over-approximate the behaviour of programs written in these languages. Specifically, the use of whole-program dataflow analysis. The research challenge of this line of work is the adaption of existing- and invention......Dynamic programming languages are highly popular and widely used. Java- Script is often called the lingua franca of the web and it is the de facto standard for client-side web programming. On the server-side the PHP, Python and Ruby languages are prevalent. What these languages have in common...... is an expressive power which is not easily captured by any static type system. These, and similar dynamic languages, are often praised for their ease-of-use and flexibility. Unfortunately, this dynamism comes at a great cost: The lack of a type system implies that most errors are not discovered until run...

  19. Coding and data analysis during qualitative empirical research in Practical Theology : original research

    National Research Council Canada - National Science Library

    Theron, Petria M

    2015-01-01

    .... In this article, it is argued for a more systematic approach during the coding and data analysis phase of qualitative research and the article concludes with a proposed model for coding and data...

  20. Analysis of truck platoon dynamics

    NARCIS (Netherlands)

    Verhoeff, L.; Zuurbier, J.; Lupker, H.A.

    1999-01-01

    This paper presents the analysis of truck platoon dynamics using a simulation environment. In this case the platoon consists of a number of trucks following each other using driver algorithms. MADYMO (a multibody and FE software package) and MATLAB/SIMULINK are coupled to allow for these driver algo

  1. Modeling compositional dynamics based on GC and purine contents of protein-coding sequences

    KAUST Repository

    Zhang, Zhang

    2010-11-08

    Background: Understanding the compositional dynamics of genomes and their coding sequences is of great significance in gaining clues into molecular evolution and a large number of publically-available genome sequences have allowed us to quantitatively predict deviations of empirical data from their theoretical counterparts. However, the quantification of theoretical compositional variations for a wide diversity of genomes remains a major challenge.Results: To model the compositional dynamics of protein-coding sequences, we propose two simple models that take into account both mutation and selection effects, which act differently at the three codon positions, and use both GC and purine contents as compositional parameters. The two models concern the theoretical composition of nucleotides, codons, and amino acids, with no prerequisite of homologous sequences or their alignments. We evaluated the two models by quantifying theoretical compositions of a large collection of protein-coding sequences (including 46 of Archaea, 686 of Bacteria, and 826 of Eukarya), yielding consistent theoretical compositions across all the collected sequences.Conclusions: We show that the compositions of nucleotides, codons, and amino acids are largely determined by both GC and purine contents and suggest that deviations of the observed from the expected compositions may reflect compositional signatures that arise from a complex interplay between mutation and selection via DNA replication and repair mechanisms.Reviewers: This article was reviewed by Zhaolei Zhang (nominated by Mark Gerstein), Guruprasad Ananda (nominated by Kateryna Makova), and Daniel Haft. 2010 Zhang and Yu; licensee BioMed Central Ltd.

  2. LSENS - GENERAL CHEMICAL KINETICS AND SENSITIVITY ANALYSIS CODE

    Science.gov (United States)

    Bittker, D. A.

    1994-01-01

    which provides the relationships between the predictions of a kinetics model and the input parameters of the problem. LSENS provides for efficient and accurate chemical kinetics computations and includes sensitivity analysis for a variety of problems, including nonisothermal conditions. LSENS replaces the previous NASA general chemical kinetics codes GCKP and GCKP84. LSENS is designed for flexibility, convenience and computational efficiency. A variety of chemical reaction models can be considered. The models include static system, steady one-dimensional inviscid flow, reaction behind an incident shock wave including boundary layer correction, and the perfectly stirred (highly backmixed) reactor. In addition, computations of equilibrium properties can be performed for the following assigned states, enthalpy and pressure, temperature and pressure, internal energy and volume, and temperature and volume. For static problems LSENS computes sensitivity coefficients with respect to the initial values of the dependent variables and/or the three rates coefficient parameters of each chemical reaction. To integrate the ODEs describing chemical kinetics problems, LSENS uses the packaged code LSODE, the Livermore Solver for Ordinary Differential Equations, because it has been shown to be the most efficient and accurate code for solving such problems. The sensitivity analysis computations use the decoupled direct method, as implemented by Dunker and modified by Radhakrishnan. This method has shown greater efficiency and stability with equal or better accuracy than other methods of sensitivity analysis. LSENS is written in FORTRAN 77 with the exception of the NAMELIST extensions used for input. While this makes the code fairly machine independent, execution times on IBM PC compatibles would be unacceptable to most users. LSENS has been successfully implemented on a Sun4 running SunOS and a DEC VAX running VMS. With minor modifications, it should also be easily implemented on other

  3. Relative-Residual-Based Dynamic Schedule for Belief Propagation Decoding of LDPC Codes

    Institute of Scientific and Technical Information of China (English)

    Huang Jie; Zhang Lijun

    2011-01-01

    Two Relative-Residual-based Dynamic Schedules (RRDS) for Belief Propagation (BP) decoding of Low-Density Parity-Check (LDPC) codes are proposed,in which the Variable code-RRDS (VN-RRDS) is a greediness-reduced version of the Check code-RRDS (CN-RRDS).The RRDS only processes the variable (or check) node,which has the maximum relative residual among all the variable (or check) nodes in each decoding iteration,thus keeping less greediness and decreased complexity in comparison with the edge-based Variable-to-Check Residual Belief Propagation (VC-RBP) algorithm.Moreover,VN-RRDS propagates first the message which has the largest residual based on all check equations.For different types of LDPC codes,simulation results show that the convergence rate of RRDS is higher than that of VC-RBP while keeping very low computational complexity.Furthermore,VN-RRDS achieves faster convergence as well as better performance than CN-RRDS.

  4. Analysis of the optimality of the standard genetic code.

    Science.gov (United States)

    Kumar, Balaji; Saini, Supreet

    2016-07-19

    Many theories have been proposed attempting to explain the origin of the genetic code. While strong reasons remain to believe that the genetic code evolved as a frozen accident, at least for the first few amino acids, other theories remain viable. In this work, we test the optimality of the standard genetic code against approximately 17 million genetic codes, and locate 29 which outperform the standard genetic code at the following three criteria: (a) robustness to point mutation; (b) robustness to frameshift mutation; and (c) ability to encode additional information in the coding region. We use a genetic algorithm to generate and score codes from different parts of the associated landscape, which are, as a result, presumably more representative of the entire landscape. Our results show that while the genetic code is sub-optimal for robustness to frameshift mutation and the ability to encode additional information in the coding region, it is very strongly selected for robustness to point mutation. This coupled with the observation that the different performance indicator scores for a particular genetic code are negatively correlated makes the standard genetic code nearly optimal for the three criteria tested in this work.

  5. Dynamic quality of service differentiation using fixed code weight in optical CDMA networks

    Science.gov (United States)

    Kakaee, Majid H.; Essa, Shawnim I.; Abd, Thanaa H.; Seyedzadeh, Saleh

    2015-11-01

    The emergence of network-driven applications, such as internet, video conferencing, and online gaming, brings in the need for a network the environments with capability of providing diverse Quality of Services (QoS). In this paper, a new code family of novel spreading sequences, called a Multi-Service (MS) code, has been constructed to support multiple services in Optical- Code Division Multiple Access (CDMA) system. The proposed method uses fixed weight for all services, however reducing the interfering codewords for the users requiring higher QoS. The performance of the proposed code is demonstrated using mathematical analysis. It shown that the total number of served users with satisfactory BER of 10-9 using NB=2 is 82, while they are only 36 and 10 when NB=3 and 4 respectively. The developed MS code is compared with variable-weight codes such as Variable Weight-Khazani Syed (VW-KS) and Multi-Weight-Random Diagonal (MW-RD). Different numbers of basic users (NB) are used to support triple-play services (audio, data and video) with different QoS requirements. Furthermore, reference to the BER of 10-12, 10-9, and 10-3 for video, data and audio, respectively, the system can support up to 45 total users. Hence, results show that the technique can clearly provide a relative QoS differentiation with lower value of basic users can support larger number of subscribers as well as better performance in terms of acceptable BER of 10-9 at fixed code weight.

  6. Fast dynamics of odor rate coding in the insect antennal lobe

    CERN Document Server

    Nawrot, Martin Paul; Farkhooi, Farzad; Menzel, Randolf

    2011-01-01

    Insects identify and evaluate behaviorally relevant odorants in complex natural scenes where odor concentrations and mixture composition can change rapidly. In the honeybee, a combinatorial code of activated and inactivated projection neurons (PNs) develops rapidly within tens of milliseconds at the first level of neural integration, the antennal lobe (AL). The phasic-tonic stimulus-response dynamics observed in the neural population code and in the firing rate profiles of single neurons is faithfully captured by two alternative models which rely either on short-term synaptic depression, or on spike frequency adaptation. Both mechanisms work independently and possibly in parallel to lateral inhibition. Short response latencies in local interneurons indicate that local processing within the AL network relies on fast lateral inhibition that can suppress effectively and specifically odor responses in single PNs. Reviewing recent findings obtained in different insect species, we conclude that the insect olfactory...

  7. Efficient Data Compression Scheme using Dynamic Huffman Code Applied on Arabic Language

    Directory of Open Access Journals (Sweden)

    Sameh Ghwanmeh

    2006-01-01

    Full Text Available The development of an efficient compression scheme to process the Arabic language represents a difficult task. This paper employs the dynamic Huffman coding on data compression with variable length bit coding, on the Arabic language. Experimental tests have been performed on both Arabic and English text. A comparison was made to measure the efficiency of compressing data results on both Arabic and English text. Also a comparison was made between the compression rate and the size of the file to be compressed. It has been found that as the file size increases, the compression ratio decreases for both Arabic and English text. The experimental results show that the average message length and the efficiency of compression on Arabic text was better than the compression on English text. Also, results show that the main factor which significantly affects compression ratio and average message length was the frequency of the symbols on the text.

  8. Structural Dynamics Verification of Rotorcraft Comprehensive Analysis System (RCAS)

    Energy Technology Data Exchange (ETDEWEB)

    Bir, G. S.

    2005-02-01

    The Rotorcraft Comprehensive Analysis System (RCAS) was acquired and evaluated as part of an ongoing effort by the U.S Department of Energy (DOE) and the National Renewable Energy Laboratory (NREL) to provide state-of-the-art wind turbine modeling and analysis technology for Government and industry. RCAS is an interdisciplinary tool offering aeroelastic modeling and analysis options not supported by current codes. RCAS was developed during a 4-year joint effort among the U.S. Army's Aeroflightdynamics Directorate, Advanced Rotorcraft Technology Inc., and the helicopter industry. The code draws heavily from its predecessor 2GCHAS (Second Generation Comprehensive Helicopter Analysis System), which required an additional 14 years to develop. Though developed for the rotorcraft industry, its general-purpose features allow it to model or analyze a general dynamic system. Its key feature is a specialized finite element that can model spinning flexible parts. The code, therefore, appears particularly suited for wind turbines whose dynamics is dominated by massive flexible spinning rotors. In addition to the simulation capability of the existing codes, RCAS [1-3] offers a range of unique capabilities, including aeroelastic stability analysis, trim, state-space modeling, operating modes, modal reduction, multi-blade coordinate transformation, periodic-system-specific analysis, choice of aerodynamic models, and a controls design/implementation graphical interface.

  9. Essentials of applied dynamic analysis

    CERN Document Server

    Jia, Junbo

    2014-01-01

    This book presents up-to-date knowledge of dynamic analysis in engineering world. To facilitate the understanding of the topics by readers with various backgrounds, general principles are linked to their applications from different angles. Special interesting topics such as statistics of motions and loading, damping modeling and measurement, nonlinear dynamics, fatigue assessment, vibration and buckling under axial loading, structural health monitoring, human body vibrations, and vehicle-structure interactions etc., are also presented. The target readers include industry professionals in civil, marine and mechanical engineering, as well as researchers and students in this area.

  10. Radial transport dynamics studies of SMBI with a newly developed TPSMBI code

    Science.gov (United States)

    Wang, Ya-Hui; Guo, Wen-Feng; Wang, Zhan-Hui; Ren, Qi-Long; Sun, Ai-Ping; Xu, Min; Wang, Ai-Ke; Xiang, Nong

    2016-10-01

    In tokamak plasma fueling, supersonic molecule beam injection (SMBI) with a higher fueling efficiency and a deeper penetration depth than the traditional gas puffing method has been developed and widely applied to many tokamak devices. It is crucial to study the transport dynamics of SMBI to improve its fueling efficiency, especially in the high confinement regime. A new one-dimensional (1D) code of TPSMBI has also been developed recently based on a six-field SMBI model in cylindrical coordinate. It couples plasma density and heat radial transport equations together with neutral density transport equations for both molecules and atoms and momentum radial transport equations for molecules. The dominant particle collisional interactions between plasmas and neutrals, such as molecule dissociation, atom ionization and charge-exchange effects, are included in the model. The code is verified to be correct with analytical solutions and also benchmarked well with the trans-neut module of BOUT++ code. Time-dependent radial transport dynamics and mean profile evolution are studied during SMBI with the TPSMBI code in both slab and cylindrical coordinates. Along the SMBI path, plasma density increases due to particle fuelling, while plasma temperature decreases due to heat cooling. Being different from slab coordinate, the curvature effect leads to larger front densities of molecule and atom during SMBI in cylindrical coordinate simulation. Project supported by the National Natural Science Foundation of China (Grant Nos. 11575055, 11375053, and 11475219) and the National Magnetic Confinement Fusion Science Program of China (Grant Nos. 2013GB111005, 2014GB108004, and 2015GB110001).

  11. Dynamic optical aberration correction with adaptive coded apertures techniques in conformal imaging

    Science.gov (United States)

    Li, Yan; Hu, Bin; Zhang, Pengbin; Zhang, Binglong

    2015-02-01

    Conformal imaging systems are confronted with dynamic aberration in optical design processing. In classical optical designs, for combination high requirements of field of view, optical speed, environmental adaption and imaging quality, further enhancements can be achieved only by the introduction of increased complexity of aberration corrector. In recent years of computational imaging, the adaptive coded apertures techniques which has several potential advantages over more traditional optical systems is particularly suitable for military infrared imaging systems. The merits of this new concept include low mass, volume and moments of inertia, potentially lower costs, graceful failure modes, steerable fields of regard with no macroscopic moving parts. Example application for conformal imaging system design where the elements of a set of binary coded aperture masks are applied are optimization designed is presented in this paper, simulation results show that the optical performance is closely related to the mask design and the reconstruction algorithm optimization. As a dynamic aberration corrector, a binary-amplitude mask located at the aperture stop is optimized to mitigate dynamic optical aberrations when the field of regard changes and allow sufficient information to be recorded by the detector for the recovery of a sharp image using digital image restoration in conformal optical system.

  12. Rigid body motion analysis system for off-line processing of time-coded video sequences

    Science.gov (United States)

    Snow, Walter L.; Shortis, Mark R.

    1995-09-01

    Photogrammetry affords the only noncontact means of providing unambiguous six-degree-of- freedom estimates for rigid body motion analysis. Video technology enables convenient off- the-shelf capability for obtaining and storing image data at frame (30 Hz) or field (60 Hz) rates. Videometry combines these technologies with frame capture capability accessible to PCs to allow unavailable measurements critical to the study of rigid body dynamics. To effectively utilize this capability, however, some means of editing, post processing, and sorting substantial amounts of time coded video data is required. This paper discusses a prototype motion analysis system built around PC and video disk technology, which is proving useful in exploring applications of these concepts to rigid body tracking and deformation analysis. Calibration issues and user interactive software development associated with this project will be discussed, as will examples of measurement projects and data reduction.

  13. Variable weight Khazani-Syed code using hybrid fixed-dynamic technique for optical code division multiple access system

    Science.gov (United States)

    Anas, Siti Barirah Ahmad; Seyedzadeh, Saleh; Mokhtar, Makhfudzah; Sahbudin, Ratna Kalos Zakiah

    2016-10-01

    Future Internet consists of a wide spectrum of applications with different bit rates and quality of service (QoS) requirements. Prioritizing the services is essential to ensure that the delivery of information is at its best. Existing technologies have demonstrated how service differentiation techniques can be implemented in optical networks using data link and network layer operations. However, a physical layer approach can further improve system performance at a prescribed received signal quality by applying control at the bit level. This paper proposes a coding algorithm to support optical domain service differentiation using spectral amplitude coding techniques within an optical code division multiple access (OCDMA) scenario. A particular user or service has a varying weight applied to obtain the desired signal quality. The properties of the new code are compared with other OCDMA codes proposed for service differentiation. In addition, a mathematical model is developed for performance evaluation of the proposed code using two different detection techniques, namely direct decoding and complementary subtraction.

  14. Sensitivity analysis on various parameters for lattice analysis of DUPIC fuel with WIMS-AECL code

    Energy Technology Data Exchange (ETDEWEB)

    Roh, Gyu Hong; Choi, Hang Bok; Park, Jee Won [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    1997-12-31

    The code WIMS-AECL has been used for the lattice analysis of DUPIC fuel. The lattice parameters calculated by the code is sensitive to the choice of number of parameters, such as the number of tracking lines, number of condensed groups, mesh spacing in the moderator region, other parameters vital to the calculation of probabilities and burnup analysis. We have studied this sensitivity with respect to these parameters and recommend their proper values which are necessary for carrying out the lattice analysis of DUPIC fuel.

  15. On the application of computational fluid dynamics codes for liquefied natural gas dispersion.

    Science.gov (United States)

    Luketa-Hanlin, Anay; Koopman, Ronald P; Ermak, Donald L

    2007-02-20

    Computational fluid dynamics (CFD) codes are increasingly being used in the liquefied natural gas (LNG) industry to predict natural gas dispersion distances. This paper addresses several issues regarding the use of CFD for LNG dispersion such as specification of the domain, grid, boundary and initial conditions. A description of the k-epsilon model is presented, along with modifications required for atmospheric flows. Validation issues pertaining to the experimental data from the Burro, Coyote, and Falcon series of LNG dispersion experiments are also discussed. A description of the atmosphere is provided as well as discussion on the inclusion of the Coriolis force to model very large LNG spills.

  16. Sodium fast reactor gaps analysis of computer codes and models for accident analysis and reactor safety.

    Energy Technology Data Exchange (ETDEWEB)

    Carbajo, Juan (Oak Ridge National Laboratory, Oak Ridge, TN); Jeong, Hae-Yong (Korea Atomic Energy Research Institute, Daejeon, Korea); Wigeland, Roald (Idaho National Laboratory, Idaho Falls, ID); Corradini, Michael (University of Wisconsin, Madison, WI); Schmidt, Rodney Cannon; Thomas, Justin (Argonne National Laboratory, Argonne, IL); Wei, Tom (Argonne National Laboratory, Argonne, IL); Sofu, Tanju (Argonne National Laboratory, Argonne, IL); Ludewig, Hans (Brookhaven National Laboratory, Upton, NY); Tobita, Yoshiharu (Japan Atomic Energy Agency, Ibaraki-ken, Japan); Ohshima, Hiroyuki (Japan Atomic Energy Agency, Ibaraki-ken, Japan); Serre, Frederic (Centre d' %C3%94etudes nucl%C3%94eaires de Cadarache %3CU%2B2013%3E CEA, France)

    2011-06-01

    This report summarizes the results of an expert-opinion elicitation activity designed to qualitatively assess the status and capabilities of currently available computer codes and models for accident analysis and reactor safety calculations of advanced sodium fast reactors, and identify important gaps. The twelve-member panel consisted of representatives from five U.S. National Laboratories (SNL, ANL, INL, ORNL, and BNL), the University of Wisconsin, the KAERI, the JAEA, and the CEA. The major portion of this elicitation activity occurred during a two-day meeting held on Aug. 10-11, 2010 at Argonne National Laboratory. There were two primary objectives of this work: (1) Identify computer codes currently available for SFR accident analysis and reactor safety calculations; and (2) Assess the status and capability of current US computer codes to adequately model the required accident scenarios and associated phenomena, and identify important gaps. During the review, panel members identified over 60 computer codes that are currently available in the international community to perform different aspects of SFR safety analysis for various event scenarios and accident categories. A brief description of each of these codes together with references (when available) is provided. An adaptation of the Predictive Capability Maturity Model (PCMM) for computational modeling and simulation is described for use in this work. The panel's assessment of the available US codes is presented in the form of nine tables, organized into groups of three for each of three risk categories considered: anticipated operational occurrences (AOOs), design basis accidents (DBA), and beyond design basis accidents (BDBA). A set of summary conclusions are drawn from the results obtained. At the highest level, the panel judged that current US code capabilities are adequate for licensing given reasonable margins, but expressed concern that US code development activities had stagnated and that the

  17. Dynamic Analysis of a Pendulum Dynamic Automatic Balancer

    Directory of Open Access Journals (Sweden)

    Jin-Seung Sohn

    2007-01-01

    Full Text Available The automatic dynamic balancer is a device to reduce the vibration from unbalanced mass of rotors. Instead of considering prevailing ball automatic dynamic balancer, pendulum automatic dynamic balancer is analyzed. For the analysis of dynamic stability and behavior, the nonlinear equations of motion for a system are derived with respect to polar coordinates by the Lagrange's equations. The perturbation method is applied to investigate the dynamic behavior of the system around the equilibrium position. Based on the linearized equations, the dynamic stability of the system around the equilibrium positions is investigated by the eigenvalue analysis.

  18. Analysis of Memory Codes and Cumulative Rehearsal in Observational Learning

    Science.gov (United States)

    Bandura, Albert; And Others

    1974-01-01

    The present study examined the influence of memory codes varying in meaningfulness and retrievability and cumulative rehearsal on retention of observationally learned responses over increasing temporal intervals. (Editor)

  19. Variable Dynamic Testbed Vehicle: Dynamics Analysis

    Science.gov (United States)

    Lee, A. Y.; Le, N. T.; Marriott, A. T.

    1997-01-01

    The Variable Dynamic Testbed Vehicle (VDTV) concept has been proposed as a tool to evaluate collision avoidance systems and to perform driving-related human factors research. The goal of this study is to analytically investigate to what extent a VDTV with adjustable front and rear anti-roll bar stiffnesses, programmable damping rates, and four-wheel-steering can emulate the lateral dynamics of a broad range of passenger vehicles.

  20. Feasibility Analysis on Simulation of PLCS Malfunction Event using SPACE Code

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Ung Soo; Lee, Cheol Shin; Sohn, Jong Joo [KEPCO-E and C, Daejeon (Korea, Republic of)

    2011-10-15

    A computer code named 'Safety and Performance Analysis Code (SPACE)' has been being developed in order to replace several existing computer codes used in designing nuclear power plant (NPP) in Korea. This SPACE code is a system code and should be able to simulate various plant events, needed for safety analysis of pressurized water reactors (PWRs), such as loss of coolant accident (LOCA), steam line break (SLB), feedwater line break (FLB), steam generator tube rupture (SGTR), and several anticipated operational occurrences (AOOs). Therefore, respective simulations of above events with the SPACE code should be verified and validated to utilize this code in the safety analysis. In this work, a feasibility analysis is performed for the simulation of pressurizer level control system (PLCS) malfunction event for the Shin-Kori units 3 and 4 (SKN 3 and 4)

  1. Development of a dynamical model of a nuclear processes simulator for analysis and training in classroom based in the RELAP/SCDAP codes; Desarrollo del modulo dinamico del simulador de procesos nucleares para analisis y entrenamiento en aula basado en los codigos RELAP/SCDAP

    Energy Technology Data Exchange (ETDEWEB)

    Salazar C, J.H.; Ramos P, J.C.; Salazar S, E.; Chavez M, C. [UNAM, Fac. de Ingenieria, Laboratorio de Analisis en Ingenieria de Reactores Nucleares, DEPFI Campus Morelos, Cuernavaca (Mexico)]. e-mail: cchavez2@cableonline.com.mx

    2003-07-01

    The present work illustrates the application of the concept of a simulator for analysis, design, instruction and training in a classroom environment associated to a nuclear power station. Emphasis is made on the methodology used to incorporate the best estimate codes RELAP/SCDAP to a prototype under development at the Nuclear Reactor Engineering Analysis Laboratory (NREAL). This methodology is based on a modular structure where multiple processes can be executed in an independent way and where the generated information is stored in shared memory segments and distributed by means of communication routines developed in the C programming language. The utility of the system is demonstrated using highly interactive graphics (mimic diagrams, pictorials and tendency graphs) for the simultaneous dynamic visualization of the most significant variables of a typical transient event (feed water controller failure in a BWR). A fundamental part of the system is its advanced graphic interface. This interface, of the type of direct manipulation, reproduces instruments and controls whose functionality is similar to those found in the current replica simulator for the Laguna Verde Nuclear Power Station. Finally the evaluation process is described. The general behavior of the main variables for the selected transitory event is interpreted, corroborating that they follow the same tendency that those reported for a BWR. The obtained results allow to conclude that the developed system works satisfactorily and that the use of al 1 x 1 real time visualization tools offers important advantages regarding other traditional methods of analysis. (Author)

  2. A preliminary uncertainty analysis of phenomenological inputs in TEXAS-V code

    Energy Technology Data Exchange (ETDEWEB)

    Park, S. H.; Kim, H. D.; Ahn, K. I. [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)

    2010-10-15

    Uncertainty analysis is important step in safety analysis of nuclear power plants. The better estimate for the computer codes is on the increase instead of conservative codes. These efforts aim to get more precise evaluation of safety margins, and aim at determining the rate of change in the prediction of codes with one or more input parameters varies within its range of interest. From this point of view, a severe accident uncertainty analysis system, SAUNA, has been improved for TEXAS-V FCI uncertainty analysis. The main objective of this paper is to present the TEXAS FCI uncertainty analysis results implemented through the SAUNA code

  3. Recent developments in seismic analysis in the code Aster; Les developpements recents en analyse sismique dans le code aster

    Energy Technology Data Exchange (ETDEWEB)

    Guihot, P.; Devesa, G.; Dumond, A.; Panet, M.; Waeckel, F.

    1996-12-31

    Progress in the field of seismic qualification and design methods made these last few years allows physical phenomena actually in play to be better considered, while cutting down the conservatism associated with some simplified design methods. So following the change in methods and developing the most advantageous ones among them contributes to the process of the seismic margins assessment and the preparation of new design tools for future series. In this paper, the main developments and improvements in methods which have been made these last two years in the Code Aster, in order to improve seismic calculation methods and seismic margin assessment are presented. The first development relates to making the MISS3D soil structure interaction code available, thanks to an interface made with the Code Aster. The second relates to the possibility of making modal basis time calculations on multi-supported structures by considering local non linearities like impact, friction or squeeze fluid forces. Recent developments in random dynamics and postprocessing devoted to earthquake designs are then mentioned. Three applications of these developments are then ut forward. The first application relates to a test case for soil structure interaction design using MISS3D-Aster coupling. The second is a test case for a multi-supported structure. The last application, more for manufacturing, refers to seismic qualification of Main Live Steam stop valves. First results of the independent validation of the Code Aster seismic design functionalities, which provide and improve the quality of software, are also recalled. (authors). 11 refs.

  4. A dynamical systems proof of Kraft-McMillan inequality and its converse for prefix-free codes

    Science.gov (United States)

    Nagaraj, Nithin

    2009-03-01

    Uniquely decodable codes are central to lossless data compression in both classical and quantum communication systems. The Kraft-McMillan inequality is a basic result in information theory which gives a necessary and sufficient condition for a code to be uniquely decodable and also has a quantum analogue. In this letter, we provide a novel dynamical systems proof of this inequality and its converse for prefix-free codes (no codeword is a prefix of another—the popular Huffman codes are an example). For constrained sources, the problem is still open.

  5. Design of an Object-Oriented Turbomachinery Analysis Code: Initial Results

    Science.gov (United States)

    Jones, Scott M.

    2015-01-01

    Performance prediction of turbomachines is a significant part of aircraft propulsion design. In the conceptual design stage, there is an important need to quantify compressor and turbine aerodynamic performance and develop initial geometry parameters at the 2-D level prior to more extensive Computational Fluid Dynamics (CFD) analyses. The Object-oriented Turbomachinery Analysis Code (OTAC) is being developed to perform 2-D meridional flowthrough analysis of turbomachines using an implicit formulation of the governing equations to solve for the conditions at the exit of each blade row. OTAC is designed to perform meanline or streamline calculations; for streamline analyses simple radial equilibrium is used as a governing equation to solve for spanwise property variations. While the goal for OTAC is to allow simulation of physical effects and architectural features unavailable in other existing codes, it must first prove capable of performing calculations for conventional turbomachines. OTAC is being developed using the interpreted language features available in the Numerical Propulsion System Simulation (NPSS) code described by Claus et al (1991). Using the NPSS framework came with several distinct advantages, including access to the pre-existing NPSS thermodynamic property packages and the NPSS Newton-Raphson solver. The remaining objects necessary for OTAC were written in the NPSS framework interpreted language. These new objects form the core of OTAC and are the BladeRow, BladeSegment, TransitionSection, Expander, Reducer, and OTACstart Elements. The BladeRow and BladeSegment consumed the initial bulk of the development effort and required determining the equations applicable to flow through turbomachinery blade rows given specific assumptions about the nature of that flow. Once these objects were completed, OTAC was tested and found to agree with existing solutions from other codes; these tests included various meanline and streamline comparisons of axial

  6. Solution of the neutronics code dynamic benchmark by finite element method

    Science.gov (United States)

    Avvakumov, A. V.; Vabishchevich, P. N.; Vasilev, A. O.; Strizhov, V. F.

    2016-10-01

    The objective is to analyze the dynamic benchmark developed by Atomic Energy Research for the verification of best-estimate neutronics codes. The benchmark scenario includes asymmetrical ejection of a control rod in a water-type hexagonal reactor at hot zero power. A simple Doppler feedback mechanism assuming adiabatic fuel temperature heating is proposed. The finite element method on triangular calculation grids is used to solve the three-dimensional neutron kinetics problem. The software has been developed using the engineering and scientific calculation library FEniCS. The matrix spectral problem is solved using the scalable and flexible toolkit SLEPc. The solution accuracy of the dynamic benchmark is analyzed by condensing calculation grid and varying degree of finite elements.

  7. GANDALF: Graphical Astrophysics code for N-body Dynamics And Lagrangian Fluids

    Science.gov (United States)

    Hubber, David; Rosotti, Giovanni

    2016-02-01

    GANDALF, a successor to SEREN (ascl:1102.010), is a hybrid self-gravitating fluid dynamics and collisional N-body code primarily designed for investigating star formation and planet formation problems. GANDALF uses various implementations of Smoothed Particle Hydrodynamics (SPH) to perform hydrodynamical simulations of gas clouds undergoing gravitational collapse to form new stars (or other objects), and can perform simulations of pure N-body dynamics using high accuracy N-body integrators, model the intermediate phase of cluster evolution, and provide visualizations via its python interface as well as interactive simulations. Although based on many of the SEREN routines, GANDALF has been largely re-written from scratch in C++ using more optimal algorithms and data structures.

  8. Adaptive coded spreading OFDM signal for dynamic-λ optical access network

    Science.gov (United States)

    Liu, Bo; Zhang, Lijia; Xin, Xiangjun

    2015-12-01

    This paper proposes and experimentally demonstrates a novel adaptive coded spreading (ACS) orthogonal frequency division multiplexing (OFDM) signal for dynamic distributed optical ring-based access network. The wavelength can be assigned to different remote nodes (RNs) according to the traffic demand of optical network unit (ONU). The ACS can provide dynamic spreading gain to different signals according to the split ratio or transmission length, which offers flexible power budget for the network. A 10×13.12 Gb/s OFDM access with ACS is successfully demonstrated over two RNs and 120 km transmission in the experiment. The demonstrated method may be viewed as one promising for future optical metro access network.

  9. PERFORMANCE ANALYSIS OF CHANNEL ESTIMATION FOR LDPC-CODED OFDM SYSTEM IN MULTIPATH FADING CHANNEL

    Institute of Scientific and Technical Information of China (English)

    Zhu Qi; Li Hao; Feng Guangzeng

    2006-01-01

    In this paper, the channel estimation techniques for Orthogonal Frequency Division Multiplexing (OFDM) systems based on pilot arrangement are studied and we apply Low Density Parity Check (LDPC) codes to the system of IEEE 802.16a with OFDM modulation. First investigated is the influence of channel estimation schemes on LDPC-code based OFDM system in static and multipath fading channels. According to the different propagation environments in 802.16a system, a dynamic channel estimation scheme is proposed.A good irregular LDPC code is designed with code rate of 1/2 and code length of 1200. Simulation results show that the performance of LDPC coded OFDM system proposed in this paper is better than that of the convolution Turbo coded OFDM system proposed in IEEE standard 802.16a.

  10. GASFLOW: A Computational Fluid Dynamics Code for Gases, Aerosols, and Combustion, Volume 1: Theory and Computational Model

    Energy Technology Data Exchange (ETDEWEB)

    Nichols, B.D.; Mueller, C.; Necker, G.A.; Travis, J.R.; Spore, J.W.; Lam, K.L.; Royl, P.; Redlinger, R.; Wilson, T.L.

    1998-10-01

    Los Alamos National Laboratory (LANL) and Forschungszentrum Karlsruhe (FzK) are developing GASFLOW, a three-dimensional (3D) fluid dynamics field code as a best-estimate tool to characterize local phenomena within a flow field. Examples of 3D phenomena include circulation patterns; flow stratification; hydrogen distribution mixing and stratification; combustion and flame propagation; effects of noncondensable gas distribution on local condensation and evaporation; and aerosol entrainment, transport, and deposition. An analysis with GASFLOW will result in a prediction of the gas composition and discrete particle distribution in space and time throughout the facility and the resulting pressure and temperature loadings on the walls and internal structures with or without combustion. A major application of GASFLOW is for predicting the transport, mixing, and combustion of hydrogen and other gases in nuclear reactor containments and other facilities. It has been applied to situations involving transporting and distributing combustible gas mixtures. It has been used to study gas dynamic behavior (1) in low-speed, buoyancy-driven flows, as well as sonic flows or diffusion dominated flows; and (2) during chemically reacting flows, including deflagrations. The effects of controlling such mixtures by safety systems can be analyzed. The code version described in this manual is designated GASFLOW 2.1, which combines previous versions of the United States Nuclear Regulatory Commission code HMS (for Hydrogen Mixing Studies) and the Department of Energy and FzK versions of GASFLOW. The code was written in standard Fortran 90. This manual comprises three volumes. Volume I describes the governing physical equations and computational model. Volume II describes how to use the code to set up a model geometry, specify gas species and material properties, define initial and boundary conditions, and specify different outputs, especially graphical displays. Sample problems are included

  11. Introducing ZEUS-MP A 3D, Parallel, Multiphysics Code for Astrophysical Fluid Dynamics

    CERN Document Server

    Norman, M L

    2000-01-01

    We describe ZEUS-MP: a Multi-Physics, Massively-Parallel, Message-Passing code for astrophysical fluid dynamics simulations in 3 dimensions. ZEUS-MP is a follow-on to the sequential ZEUS-2D and ZEUS-3D codes developed and disseminated by the Laboratory for Computational Astrophysics (lca.ncsa.uiuc.edu) at NCSA. V1.0 released 1/1/2000 includes the following physics modules: ideal hydrodynamics, ideal MHD, and self-gravity. Future releases will include flux-limited radiation diffusion, thermal heat conduction, two-temperature plasma, and heating and cooling functions. The covariant equations are cast on a moving Eulerian grid with Cartesian, cylindrical, and spherical polar coordinates currently supported. Parallelization is done by domain decomposition and implemented in F77 and MPI. The code is portable across a wide range of platforms from networks of workstations to massively parallel processors. Some parallel performance results are presented as well as an application to turbulent star formation.

  12. Coding of Markov dynamics for multiscale community detection in complex networks

    CERN Document Server

    Schaub, Michael T; Barahona, Mauricio

    2011-01-01

    The detection of community structure in complex networks is intimately related to the problem of finding a concise description of the network in terms of its modules. This notion has been recently exploited by the Map equation formalism (M. Rosvall and C. T. Bergstrom, PNAS, vol. 105, no. 4, pp. 1118-1123, 2008) through an information-theoretic characterization of the process of coding the transitions of a random walker inside and between communities at stationarity. However, a thorough consideration of the relationship between a time-evolving Markov dynamics and the coding mechanism is still lacking. We show that the original one-step coding scheme used by the Map equation method neglects the internal structure of the communities and introduces an upper scale, the 'field-of-view' limit, for the communities that it can detect. Although the Map equation method is known for its good performance on clique-like graphs, the field-of-view limit can result in undesirable overpartitioning when communities are far fro...

  13. On the implementation of error handling in dynamic interfaces to scientific codes

    Energy Technology Data Exchange (ETDEWEB)

    Solomon, Cynthia Jean [Univ. of California, Davis, CA (United States)

    1993-11-01

    With the advent of powerful workstations with windowing systems, the scientific community has become interested in user friendly interfaces as a means of promoting the distribution of scientific codes to colleagues. Distributing scientific codes to a wider audience can, however, be problematic because scientists, who are familiar with the problem being addressed but not aware of necessary operational details, are encouraged to use the codes. A more friendly environment that not only guides user inputs, but also helps catch errors is needed. This thesis presents a dynamic graphical user interface (GUI) creation system with user controlled support for error detection and handling. The system checks a series of constraints defining a valid input set whenever the state of the system changes and notifies the user when an error has occurred. A naive checking scheme was implemented that checks every constraint every time the system changes. However, this method examines many constraints whose values have not changed. Therefore, a minimum evaluation scheme that only checks those constraints that may have been violated was implemented. This system was implemented in a prototype and user testing was used to determine if it was a success. Users examined both the GUI creation system and the end-user environment. The users found both to be easy to use and efficient enough for practical use. Moreover, they concluded that the system would promote distribution.

  14. [Analysis of selected changes in project the penal code].

    Science.gov (United States)

    Berent, Jarosław; Jurczyk, Agnieszka P; Szram, Stefan

    2002-01-01

    In this paper the authors have analysed selected proposals of changes in the project of amendments in the penal code. Special attention has been placed on problem of the legality of the "comma" in art. 156 of the penal code. In this matter also a review of court jurisdiction has been made.

  15. Development of a multi-dimensional realistic thermal-hydraulic system analysis code, MARS 1.3 and its verification

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Won Jae; Chung, Bub Dong; Jeong, Jae Jun; Ha, Kwi Seok [Korea Atomic Energy Research Institute, Taejon (Korea)

    1998-06-01

    A multi-dimensional realistic thermal-hydraulic system analysis code, MARS version 1.3 has been developed. Main purpose of MARS 1.3 development is to have the realistic analysis capability of transient two-phase thermal-hydraulics of Pressurized Water Reactors (PWRs) especially during Large Break Loss of Coolant Accidents (LBLOCAs) where the multi-dimensional phenomena domain the transients. MARS code is a unified version of USNRC developed COBRA-TF, domain the transients. MARS code is a unified version of USNRC developed COBRA-TF, three-dimensional (3D) reactor vessel analysis code, and RELAP5/MOD3.2.1.2, one-dimensional (1D) reactor system analysis code., Developmental requirements for MARS are chosen not only to best utilize the existing capability of the codes but also to have the enhanced capability in code maintenance, user accessibility, user friendliness, code portability, code readability, and code flexibility. For the maintenance of existing codes capability and the enhancement of code maintenance capability, user accessibility and user friendliness, MARS has been unified to be a single code consisting of 1D module (RELAP5) and 3D module (COBRA-TF). This is realized by implicitly integrating the system pressure matrix equations of hydrodynamic models and solving them simultaneously, by modifying the 1D/3D calculation sequence operable under a single Central Processor Unit (CPU) and by unifying the input structure and the light water property routines of both modules. In addition, the code structure of 1D module is completely restructured using the modular data structure of standard FORTRAN 90, which greatly improves the code maintenance capability, readability and portability. For the code flexibility, a dynamic memory management scheme is applied in both modules. MARS 1.3 now runs on PC/Windows and HP/UNIX platforms having a single CPU, and users have the options to select the 3D module to model the 3D thermal-hydraulics in the reactor vessel or other

  16. Chemo -- dynamical, multi -- fragmented SPH code for evolution of star forming disk galaxies

    Science.gov (United States)

    Berczik, P.

    The problem of chemical and dynamical evolution of galaxies is one of the most attracting and complex problems of modern astrophysics. Within the framework of the given paper the standard dynamic Smoothed Particle Hydrodynamics (SPH) code (Monaghan J.J. 1992, ARAA, 30, 543) is noticeably expanded. Our work concernes with the changes and incorporation of new ideas into the algorithmic inclusion of Star Formation (SF) and Super Novae (SN) explosions in SPH (Berczik P. & Kravchuk S.G., 1996, ApSpSci, 245, 27). The proposed energy criterion for definition of a place and efficiency of SF results in the successfully explain Star Formation History (SFH) in isolated galaxies of different types. On the base of original ideas we expand a code in a more realistic way of the description of effects of return of a hot, chemical enriched gas in Interstellar Matter (ISM). In addition to the account of SNII, we offer the self-agreed account of SNIa and PN. This allows to describe not only the ISM content of O^16 but also the content of Fe^56 . This model will allow to investigate adequately also a well known G - dwarf problem.

  17. MMA, A Computer Code for Multi-Model Analysis

    Science.gov (United States)

    Poeter, Eileen P.; Hill, Mary C.

    2007-01-01

    This report documents the Multi-Model Analysis (MMA) computer code. MMA can be used to evaluate results from alternative models of a single system using the same set of observations for all models. As long as the observations, the observation weighting, and system being represented are the same, the models can differ in nearly any way imaginable. For example, they may include different processes, different simulation software, different temporal definitions (for example, steady-state and transient models could be considered), and so on. The multiple models need to be calibrated by nonlinear regression. Calibration of the individual models needs to be completed before application of MMA. MMA can be used to rank models and calculate posterior model probabilities. These can be used to (1) determine the relative importance of the characteristics embodied in the alternative models, (2) calculate model-averaged parameter estimates and predictions, and (3) quantify the uncertainty of parameter estimates and predictions in a way that integrates the variations represented by the alternative models. There is a lack of consensus on what model analysis methods are best, so MMA provides four default methods. Two are based on Kullback-Leibler information, and use the AIC (Akaike Information Criterion) or AICc (second-order-bias-corrected AIC) model discrimination criteria. The other two default methods are the BIC (Bayesian Information Criterion) and the KIC (Kashyap Information Criterion) model discrimination criteria. Use of the KIC criterion is equivalent to using the maximum-likelihood Bayesian model averaging (MLBMA) method. AIC, AICc, and BIC can be derived from Frequentist or Bayesian arguments. The default methods based on Kullback-Leibler information have a number of theoretical advantages, including that they tend to favor more complicated models as more data become available than do the other methods, which makes sense in many situations. Many applications of MMA will

  18. A study on the interlink of CANDU safety analysis codes with development of GUI system

    Energy Technology Data Exchange (ETDEWEB)

    Lee, J. J.; Jeo, Y. J.; Park, Q. C. [Seoul National Univ., Seoul (Korea, Republic of); Kim, H. T.; Min, B. J. [KAERI, Taejon (Korea, Republic of)

    2003-10-01

    In order to improve the CANDU safety analysis code system, the interlink of containment analysis code, PRESCON2 to the system thermal hydraulics analysis code, CATHENA, has been implemented with development of the GUI system. Before the GUI development, we partly corrected two codes to optimize on the PC environment. The interlink of two codes could be executed by introducing three interlinking variables, mass flux, mixture enthalpy, and mixture specific volume. To guarantee the robustness of the codes, two codes are extremely linked by using the GUI system. The GUI system provides much of user-friendly functions and will be improved step by step. This study is expected to improve the safety assessment system and technology for CANDU NPPs.

  19. A dynamical systems approach to characterizing the contribution of neurogenesis to neural coding

    Directory of Open Access Journals (Sweden)

    Merav Stern

    2014-03-01

    index reaches a maximum when young neurons having hyper-excitability ratio of ~4 comprised ~2% of the population. This agrees with experimental estimates (Cameron and McKay, 2001; Deng et al., 2010; Spalding et al., 2013; Tashiro et al., 2007 without any adjustable parameters in the model. Figure 2. Computational analysis of distributed coding in heterogeneous networks. Networks can be efficiently trained only in the regime where prior to training to reproduce target input patterns they exhibit chaotic dynamics (Sussillo and Abbott, 2009. For a model network based on one type of rate neurons (Sompolinsky et al., 1988, the transition to chaotic dynamics occurs when at least some modes in the network respond to perturbation with exponents (eigenvalues that have real parts > 1 (purple line. Imaginary parts indicate oscillatory dynamics along the respective modes, and are not relevant indicators of chaotic dynamics. Our analytic estimate for the limits of exponents (blue circle matches the numerical simulation (small open circles, each circle is a separate mode. In contrast, predictions based on average synaptic weights (red circle are not accurate. This example network is in the chaotic regime prior to training, because some modes have exponents with real parts > 1. The corresponding neural responses over time from different types of neurons (group 1 and group 2 are shown on the right.

  20. Implementation of discrete transfer radiation method into swift computational fluid dynamics code

    Directory of Open Access Journals (Sweden)

    Baburić Mario

    2004-01-01

    Full Text Available The Computational Fluid Dynamics (CFD has developed into a powerful tool widely used in science, technology and industrial design applications, when ever fluid flow, heat transfer, combustion, or other complicated physical processes, are involved. During decades of development of CFD codes scientists were writing their own codes, that had to include not only the model of processes that were of interest, but also a whole spectrum of necessary CFD procedures, numerical techniques, pre-processing and post-processing. That has arrested much of the scientist effort in work that has been copied many times over, and was not actually producing the added value. The arrival of commercial CFD codes brought relief to many engineers that could now use the user-function approach for mod el ling purposes, en trusting the application to do the rest of the work. This pa per shows the implementation of Discrete Transfer Radiation Method into AVL’s commercial CFD code SWIFT with the help of user defined functions. Few standard verification test cases were per formed first, and in order to check the implementation of the radiation method it self, where the comparisons with available analytic solution could be performed. After wards, the validation was done by simulating the combustion in the experimental furnace at IJmuiden (Netherlands, for which the experimental measurements were available. The importance of radiation prediction in such real-size furnaces is proved again to be substantial, where radiation itself takes the major fraction of over all heat transfer. The oil-combustion model used in simulations was the semi-empirical one that has been developed at the Power Engineering Department, and which is suit able for a wide range of typical oil flames.

  1. Significance-linked connected component analysis for wavelet image coding.

    Science.gov (United States)

    Chai, B B; Vass, J; Zhuang, X

    1999-01-01

    Recent success in wavelet image coding is mainly attributed to a recognition of the importance of data organization and representation. There have been several very competitive wavelet coders developed, namely, Shapiro's (1993) embedded zerotree wavelets (EZW), Servetto et al.'s (1995) morphological representation of wavelet data (MRWD), and Said and Pearlman's (see IEEE Trans. Circuits Syst. Video Technol., vol.6, p.245-50, 1996) set partitioning in hierarchical trees (SPIHT). We develop a novel wavelet image coder called significance-linked connected component analysis (SLCCA) of wavelet coefficients that extends MRWD by exploiting both within-subband clustering of significant coefficients and cross-subband dependency in significant fields. Extensive computer experiments on both natural and texture images show convincingly that the proposed SLCCA outperforms EZW, MRWD, and SPIHT. For example, for the Barbara image, at 0.25 b/pixel, SLCCA outperforms EZW, MRWD, and SPIHT by 1.41 dB, 0.32 dB, and 0.60 dB in PSNR, respectively. It is also observed that SLCCA works extremely well for images with a large portion of texture. For eight typical 256x256 grayscale texture images compressed at 0.40 b/pixel, SLCCA outperforms SPIHT by 0.16 dB-0.63 dB in PSNR. This performance is achieved without using any optimal bit allocation procedure. Thus both the encoding and decoding procedures are fast.

  2. Formal analysis of design process dynamics

    NARCIS (Netherlands)

    Bosse, T.; Jonker, C.M.; Treur, J.

    2010-01-01

    This paper presents a formal analysis of design process dynamics. Such a formal analysis is a prerequisite to come to a formal theory of design and for the development of automated support for the dynamics of design processes. The analysis was geared toward the identification of dynamic design

  3. Formal analysis of design process dynamics

    NARCIS (Netherlands)

    Bosse, T.; Jonker, C.M.; Treur, J.

    2010-01-01

    This paper presents a formal analysis of design process dynamics. Such a formal analysis is a prerequisite to come to a formal theory of design and for the development of automated support for the dynamics of design processes. The analysis was geared toward the identification of dynamic design prope

  4. MMA, A Computer Code for Multi-Model Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Eileen P. Poeter and Mary C. Hill

    2007-08-20

    This report documents the Multi-Model Analysis (MMA) computer code. MMA can be used to evaluate results from alternative models of a single system using the same set of observations for all models. As long as the observations, the observation weighting, and system being represented are the same, the models can differ in nearly any way imaginable. For example, they may include different processes, different simulation software, different temporal definitions (for example, steady-state and transient models could be considered), and so on. The multiple models need to be calibrated by nonlinear regression. Calibration of the individual models needs to be completed before application of MMA. MMA can be used to rank models and calculate posterior model probabilities. These can be used to (1) determine the relative importance of the characteristics embodied in the alternative models, (2) calculate model-averaged parameter estimates and predictions, and (3) quantify the uncertainty of parameter estimates and predictions in a way that integrates the variations represented by the alternative models. There is a lack of consensus on what model analysis methods are best, so MMA provides four default methods. Two are based on Kullback-Leibler information, and use the AIC (Akaike Information Criterion) or AICc (second-order-bias-corrected AIC) model discrimination criteria. The other two default methods are the BIC (Bayesian Information Criterion) and the KIC (Kashyap Information Criterion) model discrimination criteria. Use of the KIC criterion is equivalent to using the maximum-likelihood Bayesian model averaging (MLBMA) method. AIC, AICc, and BIC can be derived from Frequentist or Bayesian arguments. The default methods based on Kullback-Leibler information have a number of theoretical advantages, including that they tend to favor more complicated models as more data become available than do the other methods, which makes sense in many situations.

  5. Substitution dynamical systems spectral analysis

    CERN Document Server

    Queffélec, Martine

    2010-01-01

    This volume mainly deals with the dynamics of finitely valued sequences, and more specifically, of sequences generated by substitutions and automata. Those sequences demonstrate fairly simple combinatorical and arithmetical properties and naturally appear in various domains. As the title suggests, the aim of the initial version of this book was the spectral study of the associated dynamical systems: the first chapters consisted in a detailed introduction to the mathematical notions involved, and the description of the spectral invariants followed in the closing chapters. This approach, combined with new material added to the new edition, results in a nearly self-contained book on the subject. New tools - which have also proven helpful in other contexts - had to be developed for this study. Moreover, its findings can be concretely applied, the method providing an algorithm to exhibit the spectral measures and the spectral multiplicity, as is demonstrated in several examples. Beyond this advanced analysis, many...

  6. Correcting sequencing errors in DNA coding regions using a dynamic programming approach.

    Science.gov (United States)

    Xu, Y; Mural, R J; Uberbacher, E C

    1995-04-01

    This paper presents an algorithm for detecting and 'correcting' sequencing errors that occur in DNA coding regions. The types of sequencing errors addressed are insertions and deletions (indels) of DNA bases. The goal is to provide a capability which makes single-pass or low-redundancy sequence data more informative, reducing the need for high-redundancy sequencing for gene identification and characterization purposes. This would permit improved sequencing efficiency and reduce genome sequencing costs. The algorithm detects sequencing errors by discovering changes in the statistically preferred reading frame within a putative coding region and then inserts a number of 'neutral' bases at a perceived reading frame transition point to make the putative exon candidate frame consistent. We have implemented the algorithm as a front-end subsystem of the GRAIL DNA sequence analysis system to construct a version which is very error tolerant and also intend to use this as a testbed for further development of sequencing error-correction technology. Preliminary test results have shown the usefulness of this algorithm and also exhibited some of its weakness, providing possible directions for further improvement. On a test set consisting of 68 human DNA sequences with 1% randomly generated indels in coding regions, the algorithm detected and corrected 76% of the indels. The average distance between the position of an indel and the predicted one was 9.4 bases. With this subsystem in place, GRAIL correctly predicted 89% of the coding messages with 10% false message on the 'corrected' sequences, compared to 69% correctly predicted coding messages and 11% falsely predicted messages on the 'corrupted' sequences using standard GRAIL II method (version 1.2).(ABSTRACT TRUNCATED AT 250 WORDS)

  7. RAVEN, a New Software for Dynamic Risk Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Cristian Rabiti; Andrea Alfonsi; Joshua Cogliati; Diego Mandelli; Robert Kinoshita

    2014-06-01

    RAVEN is a generic software driver to perform parametric and probabilistic analysis of code simulating complex systems. Initially developed to provide dynamic risk analysis capabilities to the RELAP-7 code [1] is currently being generalized with the addition of Application Programming Interfaces (APIs). These interfaces are used to extend RAVEN capabilities to any software as long as all the parameters that need to be perturbed are accessible by inputs files or directly via python interfaces. RAVEN is capable to investigate the system response probing the input space using Monte Carlo, grid strategies, or Latin Hyper Cube schemes, but its strength is its focus toward system feature discovery like limit surfaces separating regions of the input space leading to system failure using dynamic supervised learning techniques. The paper will present an overview of the software capabilities and their implementation schemes followed by same application examples.

  8. Promoter Analysis Reveals Globally Differential Regulation of Human Long Non-Coding RNA and Protein-Coding Genes

    KAUST Repository

    Alam, Tanvir

    2014-10-02

    Transcriptional regulation of protein-coding genes is increasingly well-understood on a global scale, yet no comparable information exists for long non-coding RNA (lncRNA) genes, which were recently recognized to be as numerous as protein-coding genes in mammalian genomes. We performed a genome-wide comparative analysis of the promoters of human lncRNA and protein-coding genes, finding global differences in specific genetic and epigenetic features relevant to transcriptional regulation. These two groups of genes are hence subject to separate transcriptional regulatory programs, including distinct transcription factor (TF) proteins that significantly favor lncRNA, rather than coding-gene, promoters. We report a specific signature of promoter-proximal transcriptional regulation of lncRNA genes, including several distinct transcription factor binding sites (TFBS). Experimental DNase I hypersensitive site profiles are consistent with active configurations of these lncRNA TFBS sets in diverse human cell types. TFBS ChIP-seq datasets confirm the binding events that we predicted using computational approaches for a subset of factors. For several TFs known to be directly regulated by lncRNAs, we find that their putative TFBSs are enriched at lncRNA promoters, suggesting that the TFs and the lncRNAs may participate in a bidirectional feedback loop regulatory network. Accordingly, cells may be able to modulate lncRNA expression levels independently of mRNA levels via distinct regulatory pathways. Our results also raise the possibility that, given the historical reliance on protein-coding gene catalogs to define the chromatin states of active promoters, a revision of these chromatin signature profiles to incorporate expressed lncRNA genes is warranted in the future.

  9. Exposure calculation code module for reactor core analysis: BURNER

    Energy Technology Data Exchange (ETDEWEB)

    Vondy, D.R.; Cunningham, G.W.

    1979-02-01

    The code module BURNER for nuclear reactor exposure calculations is presented. The computer requirements are shown, as are the reference data and interface data file requirements, and the programmed equations and procedure of calculation are described. The operating history of a reactor is followed over the period between solutions of the space, energy neutronics problem. The end-of-period nuclide concentrations are determined given the necessary information. A steady state, continuous fueling model is treated in addition to the usual fixed fuel model. The control options provide flexibility to select among an unusually wide variety of programmed procedures. The code also provides user option to make a number of auxiliary calculations and print such information as the local gamma source, cumulative exposure, and a fine scale power density distribution in a selected zone. The code is used locally in a system for computation which contains the VENTURE diffusion theory neutronics code and other modules.

  10. Vehicle Codes and Standards: Overview and Gap Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Blake, C.; Buttner, W.; Rivkin, C.

    2010-02-01

    This report identifies gaps in vehicle codes and standards and recommends ways to fill the gaps, focusing on six alternative fuels: biodiesel, natural gas, electricity, ethanol, hydrogen, and propane.

  11. Exposure calculation code module for reactor core analysis: BURNER

    Energy Technology Data Exchange (ETDEWEB)

    Vondy, D.R.; Cunningham, G.W.

    1979-02-01

    The code module BURNER for nuclear reactor exposure calculations is presented. The computer requirements are shown, as are the reference data and interface data file requirements, and the programmed equations and procedure of calculation are described. The operating history of a reactor is followed over the period between solutions of the space, energy neutronics problem. The end-of-period nuclide concentrations are determined given the necessary information. A steady state, continuous fueling model is treated in addition to the usual fixed fuel model. The control options provide flexibility to select among an unusually wide variety of programmed procedures. The code also provides user option to make a number of auxiliary calculations and print such information as the local gamma source, cumulative exposure, and a fine scale power density distribution in a selected zone. The code is used locally in a system for computation which contains the VENTURE diffusion theory neutronics code and other modules.

  12. A Deformation Analysis Code of CANDU Fuel under the Postulated Accident: ELOCA

    Energy Technology Data Exchange (ETDEWEB)

    Park, Joo Hwan; Jung, Jong Yeob

    2006-11-15

    Deformations of the fuel element or fuel channel might be the main cause of the fuel failure. Therefore, the accurate prediction of the deformation and the analysis capabilities are closely related to the increase of the safety margin of the reactor. In this report, among the performance analysis or the transient behavior prediction computer codes, the analysis codes for deformation such as the ELOCA, HOTSPOT, CONTACT-1, and PTDFORM are briefly introduced and each code's objectives, applicability, and relations are explained. Especially, the user manual for ELOCA code which is the analysis code for the fuel deformation and the release of fission product during the transient period after the postulated accidents is provided so that it can be the guidance to the potential users of the code and save the time and economic loss by reducing the trial and err000.

  13. Performance Analysis of a Decoding Algorithm for Algebraic Geometry Codes

    DEFF Research Database (Denmark)

    Jensen, Helge Elbrønd; Nielsen, Rasmus Refslund; Høholdt, Tom

    1998-01-01

    We analyse the known decoding algorithms for algebraic geometry codes in the case where the number of errors is greater than or equal to [(dFR-1)/2]+1, where dFR is the Feng-Rao distance......We analyse the known decoding algorithms for algebraic geometry codes in the case where the number of errors is greater than or equal to [(dFR-1)/2]+1, where dFR is the Feng-Rao distance...

  14. User Manual and Source Code for a LAMMPS Implementation of Constant Energy Dissipative Particle Dynamics (DPD-E)

    Science.gov (United States)

    2014-06-01

    User Manual and Source Code for a LAMMPS Implementation of Constant Energy Dissipative Particle Dynamics ( DPD -E) by James P. Larentzos...Energy Dissipative Particle Dynamics ( DPD -E) James P. Larentzos Engility Corporation John K. Brennan, Joshua D. Moore, and William D. Mattson...Constant Energy Dissipative Particle Dynamics ( DPD -E) 5a. CONTRACT NUMBER 5b. GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER 6. AUTHOR(S) James P

  15. Analysis of the KUCA MEU experiments using the ANL code system

    Energy Technology Data Exchange (ETDEWEB)

    Shiroya, S.; Hayashi, M.; Kanda, K.; Shibata, T.; Woodruff, W.L.; Matos, J.E.

    1982-01-01

    This paper provides some preliminary results on the analysis of the KUCA critical experiments using the ANL code system. Since this system was employed in the earlier neutronics calculations for the KUHFR, it is important to assess its capabilities for the KUHFR. The KUHFR has a unique core configuration which is difficult to model precisely with current diffusion theory codes. This paper also provides some results from a finite-element diffusion code (2D-FEM-KUR), which was developed in a cooperative research program between KURRI and JAERI. This code provides the capability for mockup of a complex core configuration as the KUHFR. Using the same group constants generated by the EPRI-CELL code, the results of the 2D-FEM-KUR code are compared with the finite difference diffusion code (DIF3D(2D) which is mainly employed in this analysis.

  16. Estimation and analysis of Galileo differential code biases

    Science.gov (United States)

    Li, Min; Yuan, Yunbin; Wang, Ningbo; Li, Zishen; Li, Ying; Huo, Xingliang

    2017-03-01

    When sensing the Earth's ionosphere using dual-frequency pseudorange observations of global navigation satellite systems (GNSS), the satellite and receiver differential code biases (DCBs) account for one of the main sources of error. For the Galileo system, limited knowledge is available about the determination and characteristic analysis of the satellite and receiver DCBs. To better understand the characteristics of satellite and receiver DCBs of Galileo, the IGGDCB (IGG, Institute of Geodesy and Geophysics, Wuhan, China) method is extended to estimate the satellite and receiver DCBs of Galileo, with the combined use of GPS and Galileo observations. The experimental data were collected from the Multi-GNSS Experiment network, covering the period of 2013-2015. The stability of both Galileo satellite and receiver DCBs over a time period of 36 months was thereby analyzed for the current state of the Galileo system. Good agreement of Galileo satellite DCBs is found between the IGGDCB-based DCB estimates and those from the German Aerospace Center (DLR), at the level of 0.22 ns. Moreover, high-level stability of the Galileo satellite DCB estimates is obtained over the selected time span (less than 0.25 ns in terms of standard deviation) by both IGGDCB and DLR algorithms. The Galileo receiver DCB estimates are also relatively stable for the case in which the receiver hardware device stays unchanged. It can also be concluded that the receiver DCB estimates are rather sensitive to the change of the firmware version and that the receiver antenna type has no great impact on receiver DCBs.

  17. Analysis and design of Raptor codes using a multi-edge framework

    OpenAIRE

    Jayasooriya, Sachini; Shirvanimoghaddam, Mahyar; Ong, Lawrence; Johnson, Sarah J.

    2017-01-01

    The focus of this paper is on the analysis and design of Raptor codes using a multi-edge framework. In this regard, we first represent the Raptor code as a multi-edge type low-density parity-check (METLDPC) code. This MET representation gives a general framework to analyze and design Raptor codes over a binary input additive white Gaussian noise channel using MET density evolution (MET-DE). We consider a joint decoding scheme based on the belief propagation (BP) decoding for Raptor codes in t...

  18. Orbital Data Confirms Dynamic Fractal Firework Universe Having 3D-spiral Code

    CERN Document Server

    Savov, E P

    2005-01-01

    The understanding of the universe is confused by the unknown nature of about 95% of its matter, required to confine the motions of space objects in cosmic structures. The proposed self-similar transformations of one all-building basic matter suggest that the universe is dynamic fractal that has 3D-spiral code. Its dynamic fractal elements create unifying force described by equation of the unifying interaction, drawn in the new fundamental dynamic fractal framework. The equation of unifying interaction converges into the inverse square laws and the principle of uncertainty at laboratory scales. It is further confirmed with calculations based on the orbital data of bodies moving around the nucleus of the Milky Way Galaxy, the centers of the Earth and the Sun and also orbiting the 87 Sylvia asteroid. Some of the testable predictions say that 87 Sylvia will have density > 1.2 g/cm2 and variations in gravity force less or comparable to 5.4 x 10exp(-4) m/s2, having space scales less or comparable to 10exp(3) km, wi...

  19. Bit Error Rate Performance Analysis on Modulation Techniques of Wideband Code Division Multiple Access

    CERN Document Server

    Masud, M A; Rahman, M A

    2010-01-01

    In the beginning of 21st century there has been a dramatic shift in the market dynamics of telecommunication services. The transmission from base station to mobile or downlink transmission using M-ary Quadrature Amplitude modulation (QAM) and Quadrature phase shift keying (QPSK) modulation schemes are considered in Wideband-Code Division Multiple Access (W-CDMA) system. We have done the performance analysis of these modulation techniques when the system is subjected to Additive White Gaussian Noise (AWGN) and multipath Rayleigh fading are considered in the channel. The research has been performed by using MATLAB 7.6 for simulation and evaluation of Bit Error Rate (BER) and Signal-To-Noise Ratio (SNR) for W-CDMA system models. It is shows that the analysis of Quadrature phases shift key and 16-ary Quadrature Amplitude modulations which are being used in wideband code division multiple access system, Therefore, the system could go for more suitable modulation technique to suit the channel quality, thus we can d...

  20. OSSMETER D3.4 – Language-Specific Source Code Quality Analysis

    NARCIS (Netherlands)

    Vinju, J.J.; Shahi, A.; Basten, H.J.S.

    2014-01-01

    This deliverable is part of WP3: Source Code Quality and Activity Analysis. It provides descriptions and prototypes of the tools that are needed for source code quality analysis in open source software projects. It builds upon the results of: • Deliverable 3.1 where infra-structure and a domain anal

  1. Atemporal equilibria: pro- and retroactive coding in the dynamics of cognitive microstructures

    Directory of Open Access Journals (Sweden)

    Mark A Elliott

    2014-09-01

    Full Text Available Detection reaction times to a grouping target are expedited when the target is preceded by repeated presentation of a priming stimulus presented below detection thresholds in a premask matrix presented at specific frequencies between 27-68 Hz. Frequency-specific priming can be explained as a function of the return phase of the priming stimulus relative to premask matrix frequency, indicating one consequence of repeating stimulation is pre-activation of a prime. By varying the frequency and phase of priming-stimulus relative to target presentation it is also shown that given an interaction in phase between the target presentations relative to the return phase of the prime, target coding is expedited by a prime that achieves maximum activation at a phase that would precede priming-stimulus presentation by several tens of milliseconds. However, this cognitive response is flexible enough to be able to achieve an identical prime retroactively, that is, at a phase during or subsequent to priming-stimulus presentation. This occurs if there is no interaction between the phase of target presentation and priming-stimulus presentation. On this basis, it can be concluded that by virtue of the relationship between stimulus events and the dynamics of microstructural cognition, cognition may exist in a temporal context that can shift from past to future states. Consequently and at this low level of psychological function, the conventional, one-dimensional model of time flow - from future to past states does not fully explain cognitive function. Instead, depending upon the interaction in phase between different coding frequencies, the same form of cognition can anticipate or retroactively code events. In so far as our cognition provides a content structure for consciousness, our psychological lives may be based upon an ability to travel backwards and forwards across very short intervals of time.

  2. Automatic Generation of OpenMP Directives and Its Application to Computational Fluid Dynamics Codes

    Science.gov (United States)

    Yan, Jerry; Jin, Haoqiang; Frumkin, Michael; Yan, Jerry (Technical Monitor)

    2000-01-01

    The shared-memory programming model is a very effective way to achieve parallelism on shared memory parallel computers. As great progress was made in hardware and software technologies, performance of parallel programs with compiler directives has demonstrated large improvement. The introduction of OpenMP directives, the industrial standard for shared-memory programming, has minimized the issue of portability. In this study, we have extended CAPTools, a computer-aided parallelization toolkit, to automatically generate OpenMP-based parallel programs with nominal user assistance. We outline techniques used in the implementation of the tool and discuss the application of this tool on the NAS Parallel Benchmarks and several computational fluid dynamics codes. This work demonstrates the great potential of using the tool to quickly port parallel programs and also achieve good performance that exceeds some of the commercial tools.

  3. The neural dynamics of reward value and risk coding in the human orbitofrontal cortex.

    Science.gov (United States)

    Li, Yansong; Vanni-Mercier, Giovanna; Isnard, Jean; Mauguière, François; Dreher, Jean-Claude

    2016-04-01

    The orbitofrontal cortex is known to carry information regarding expected reward, risk and experienced outcome. Yet, due to inherent limitations in lesion and neuroimaging methods, the neural dynamics of these computations has remained elusive in humans. Here, taking advantage of the high temporal definition of intracranial recordings, we characterize the neurophysiological signatures of the intact orbitofrontal cortex in processing information relevant for risky decisions. Local field potentials were recorded from the intact orbitofrontal cortex of patients suffering from drug-refractory partial epilepsy with implanted depth electrodes as they performed a probabilistic reward learning task that required them to associate visual cues with distinct reward probabilities. We observed three successive signals: (i) around 400 ms after cue presentation, the amplitudes of the local field potentials increased with reward probability; (ii) a risk signal emerged during the late phase of reward anticipation and during the outcome phase; and (iii) an experienced value signal appeared at the time of reward delivery. Both the medial and lateral orbitofrontal cortex encoded risk and reward probability while the lateral orbitofrontal cortex played a dominant role in coding experienced value. The present study provides the first evidence from intracranial recordings that the human orbitofrontal cortex codes reward risk both during late reward anticipation and during the outcome phase at a time scale of milliseconds. Our findings offer insights into the rapid mechanisms underlying the ability to learn structural relationships from the environment.

  4. Image Encryption Algorithm Based on Dynamic DNA Coding and Chen’s Hyperchaotic System

    Directory of Open Access Journals (Sweden)

    Jian Zhang

    2016-01-01

    Full Text Available With the development of national information processes, specific image information from secret departments or individuals is often required to be confidentially transmitted. Numerous image encryption methods exist, especially since the initial value sensitivity and other characteristics of chaos theory and chaos theory-based encryption have become increasingly important in recent years. At present, DNA coding constitutes a new research direction of image encryption that uses the four base pairs of DNA code and image pixel values to establish a special correspondence, in order to achieve pixel diffusion. There are eight DNA encoding rules, and current methods of selecting the DNA encoding rules are largely fixed. Thus, the security of encoded data is not high. In this paper, we use the Lorenz chaotic system, Chen’s hyperchaotic system, and the DNA encoding combination and present a new image encryption algorithm that can dynamically select eight types of DNA encoding rules and eight types of DNA addition and subtraction rules, with significant improvements in security. Through simulation experiments and histograms, correlations, and NPCR analyses, we have determined that the algorithm possesses numerous desirable features, including good encryption effects and antishear and antinoise performances.

  5. Representation and coding of large-scale 3D dynamic maps

    Science.gov (United States)

    Cohen, Robert A.; Tian, Dong; Krivokuća, Maja; Sugimoto, Kazuo; Vetro, Anthony; Wakimoto, Koji; Sekiguchi, Shunichi

    2016-09-01

    combined with depth and color measurements of the surrounding environment. Localization could be achieved with GPS, inertial measurement units (IMU), cameras, or combinations of these and other devices, while the depth measurements could be achieved with time-of-flight, radar or laser scanning systems. The resulting 3D maps, which are composed of 3D point clouds with various attributes, could be used for a variety of applications, including finding your way around indoor spaces, navigating vehicles around a city, space planning, topographical surveying or public surveying of infrastructure and roads, augmented reality, immersive online experiences, and much more. This paper discusses application requirements related to the representation and coding of large-scale 3D dynamic maps. In particular, we address requirements related to different types of acquisition environments, scalability in terms of progressive transmission and efficiently rendering different levels of details, as well as key attributes to be included in the representation. Additionally, an overview of recently developed coding techniques is presented, including an assessment of current performance. Finally, technical challenges and needs for future standardization are discussed.

  6. REBOUND: An open-source multi-purpose N-body code for collisional dynamics

    CERN Document Server

    Rein, Hanno

    2011-01-01

    REBOUND is a new multi-purpose N-body code which is freely available under an open-source license. It was designed for collisional dynamics such as planetary rings but can also solve the classical N-body problem. It is highly modular and can be customized easily to work on a wide variety of different problems in astrophysics and beyond. REBOUND comes with three symplectic integrators: leap-frog, the symplectic epicycle integrator (SEI) and a Wisdom-Holman mapping (WH). It supports open, periodic and shearing-sheet boundary conditions. REBOUND can use a Barnes-Hut tree to calculate both self-gravity and collisions. These modules are fully parallelized with MPI as well as OpenMP. The former makes use of a static domain decomposition and a distributed essential tree. Two new collision detection modules based on a plane-sweep algorithm are also implemented. The performance of the plane-sweep algorithm is superior to a tree code for simulations in which one dimension is much longer than the other two and in simula...

  7. Evaluation of the analysis models in the ASTRA nuclear design code system

    Energy Technology Data Exchange (ETDEWEB)

    Cho, Nam Jin; Park, Chang Jea; Kim, Do Sam; Lee, Kyeong Taek; Kim, Jong Woon [Korea Advanced Institute of Science and Technology, Taejon (Korea, Republic of)

    2000-11-15

    In the field of nuclear reactor design, main practice was the application of the improved design code systems. During the process, a lot of basis and knowledge were accumulated in processing input data, nuclear fuel reload design, production and analysis of design data, et al. However less efforts were done in the analysis of the methodology and in the development or improvement of those code systems. Recently, KEPO Nuclear Fuel Company (KNFC) developed the ASTRA (Advanced Static and Transient Reactor Analyzer) code system for the purpose of nuclear reactor design and analysis. In the code system, two group constants were generated from the CASMO-3 code system. The objective of this research is to analyze the analysis models used in the ASTRA/CASMO-3 code system. This evaluation requires indepth comprehension of the models, which is important so much as the development of the code system itself. Currently, most of the code systems used in domestic Nuclear Power Plant were imported, so it is very difficult to maintain and treat the change of the situation in the system. Therefore, the evaluation of analysis models in the ASTRA nuclear reactor design code system in very important.

  8. Analysis of error performance on Turbo coded FDPIM

    Institute of Scientific and Technical Information of China (English)

    ZHU Yin-bing; WANG Hong-Xing; ZHANG Tie-Ying

    2008-01-01

    Due to variable symbol length of digital pulse interval modulation(DPIM), it is difficult to analyze the error performances of Turbo ceded DPIM. To solve this problem, a fixed-length digital pulse interval modulation(FDPIM) method is provided.The FDPIM modulation structure is introduced. The packet error rates of uncoded FDPIM are analyzed and compared with that of DPIM. Bit error rates of Turbo coded FDPIM are simulated based on three kinds of analytical models under weak turbulence channel. The results show that packet error rate of uncoded FDPIM is inferior to that of uncoded DPIM.However, FDPIM is easy to be implemented and easy to be combined, with Turbo code for soft-decision because of its fixed length. Besides, the introduction of Turbo code in this modulation can decrease the average power about 10 dBm,which means that it can improve the error performance of the system effectively.

  9. Analysis of extensive air showers with the hybrid code SENECA

    CERN Document Server

    Ortiz, J A; Medina-Tanco, G; Ortiz, Jeferson A.; Souza, Vitor de; Medina-Tanco, Gustavo

    2005-01-01

    The ultrahigh energy tail of the cosmic ray spectrum has been explored with unprecedented detail. For this reason, new experiments are exerting a severe pressure on extensive air shower modeling. Detailed fast codes are in need in order to extract and understand the richness of information now available. In this sense we explore the potential of SENECA, an efficient hybrid tridimensional simulation code, as a valid practical alternative to full Monte Carlo simulations of extensive air showers generated by ultrahigh energy cosmic rays. We discuss the influence of this approach on the main longitudinal characteristics of proton, iron nucleus and gamma induced air showers for different hadronic interaction models. We also show the comparisons of our predictions with those of CORSIKA code.

  10. Analysis of extensive air showers with the hybrid code SENECA

    Science.gov (United States)

    Ortiz, Jeferson A.; de Souza, Vitor; Medina-Tanco, Gustavo

    The ultrahigh energy tail of the cosmic ray spectrum has been explored with unprecedented detail. For this reason, new experiments are exerting a severe pressure on extensive air shower modeling. Detailed fast codes are in need in order to extract and understand the richness of information now available. In this sense we explore the potential of SENECA, an efficient hybrid tridimensional simulation code, as a valid practical alternative to full Monte Carlo simulations of extensive air showers generated by ultrahigh energy cosmic rays. We discuss the influence of this approach on the main longitudinal characteristics of proton, iron nucleus and gamma induced air showers for different hadronic interaction models. We also show the comparisons of our predictions with those of CORSIKA code.

  11. ATR PDQ and MCWO Fuel Burnup Analysis Codes Evaluation

    Energy Technology Data Exchange (ETDEWEB)

    G.S. Chang; P. A. Roth; M. A. Lillo

    2009-11-01

    The Advanced Test Reactor (ATR) at the Idaho National Laboratory (INL) is being studied to determine the feasibility of converting it from the highly enriched Uranium (HEU) fuel that is currently uses to low enriched Uranium (LEU) fuel. In order to achieve this goal, it would be best to qualify some different computational methods than those that have been used at ATR for the past 40 years. This paper discusses two methods of calculating the burnup of ATR fuel elements. The existing method, that uses the PDQ code, is compared to a modern method that uses A General Monte Carlo N-Particle Transport Code (MCNP) combined with the Origen2.2 code. This modern method, MCNP with ORIGEN2.2 (MCWO), is found to give excellent agreement with the existing method (PDQ). Both of MCWO and PDQ are also in a very good agreement to the 235U burnup data generated by an analytical method.

  12. Code Blue Emergencies: A Team Task Analysis and Educational Initiative

    Directory of Open Access Journals (Sweden)

    James W. Price

    2012-04-01

    Full Text Available Introduction: The objective of this study was to identify factors that have a positive or negative influence on resuscitation team performance during emergencies in the operating room (OR and post-operative recovery unit (PAR at a major Canadian teaching hospital. This information was then used to implement a team training program for code blue emergencies. Methods: In 2009/10, all OR and PAR nurses and 19 anesthesiologists at Vancouver General Hospital (VGH were invited to complete an anonymous, 10 minute written questionnaire regarding their code blue experience. Survey questions were devised by 10 recovery room and operation room nurses as well as 5 anesthesiologists representing 4 different hospitals in British Columbia. Three iterations of the survey were reviewed by a pilot group of nurses and anesthesiologists and their feedback was integrated into the final version of the survey. Results: Both nursing staff (n = 49 and anesthesiologists (n = 19 supported code blue training and believed that team training would improve patient outcome. Nurses noted that it was often difficult to identify the leader of the resuscitation team. Both nursing staff and anesthesiologists strongly agreed that too many people attending the code blue with no assigned role hindered team performance. Conclusion: Identifiable leadership and clear communication of roles were identified as keys to resuscitation team functioning. Decreasing the number of people attending code blue emergencies with no specific role, increased access to mock code blue training, and debriefing after crises were all identified as areas requiring improvement. Initial team training exercises have been well received by staff.

  13. Analysis of planetary evolution with emphasis on differentiation and dynamics

    Science.gov (United States)

    Kaula, William M.; Newman, William I.

    1987-01-01

    In order to address the early stages of nebula evolution, a three-dimensional collapse code which includes not only hydrodynamics and radiative transfer, but also the effects of ionization and, possibly, magnetic fields is being addressed. As part of the examination of solar system evolution, an N-body code was developed which describes the latter stages of planet formation from the accretion of planetesimals. To test the code for accuracy and run-time efficiency, and to develop a stronger theoretical foundation, problems were studied in orbital dynamics. A regional analysis of the correlation in the gravity and topography fields of Venus was performed in order to determine the small and intermediate scale subsurface structure.

  14. Design and Analysis of LT Codes with Decreasing Ripple Size

    DEFF Research Database (Denmark)

    Sørensen, Jesper Hemming; Popovski, Petar; Østergaard, Jan

    2012-01-01

    In this paper we propose a new design of LT codes, which decreases the amount of necessary overhead in comparison to existing designs. The design focuses on a parameter of the LT decoding process called the ripple size. This parameter was also a key element in the design proposed in the original...... work by Luby. Specifically, Luby argued that an LT code should provide a constant ripple size during decoding. In this work we show that the ripple size should decrease during decoding, in order to reduce the necessary overhead. Initially we motivate this claim by analytical results related...

  15. Sandia Engineering Analysis Code Access System v. 2.0.1

    Energy Technology Data Exchange (ETDEWEB)

    2017-10-30

    The Sandia Engineering Analysis Code Access System (SEACAS) is a suite of preprocessing, post processing, translation, visualization, and utility applications supporting finite element analysis software using the Exodus database file format.

  16. Application of CFD Codes in Nuclear Reactor Safety Analysis

    Directory of Open Access Journals (Sweden)

    T. Höhne

    2010-01-01

    Full Text Available Computational Fluid Dynamics (CFD is increasingly being used in nuclear reactor safety (NRS analyses as a tool that enables safety relevant phenomena occurring in the reactor coolant system to be described in more detail. Numerical investigations on single phase coolant mixing in Pressurised Water Reactors (PWR have been performed at the FZD for almost a decade. The work is aimed at describing the mixing phenomena relevant for both safety analysis, particularly in steam line break and boron dilution scenarios, and mixing phenomena of interest for economical operation and the structural integrity. For the experimental investigation of horizontal two phase flows, different non pressurized channels and the TOPFLOW Hot Leg model in a pressure chamber was build and simulated with ANSYS CFX. In a common project between the University of Applied Sciences Zittau/Görlitz and FZD the behaviour of insulation material released by a LOCA released into the containment and might compromise the long term emergency cooling systems is investigated. Moreover, the actual capability of CFD is shown to contribute to fuel rod bundle design with a good CHF performance.

  17. LDPC Codes--Structural Analysis and Decoding Techniques

    Science.gov (United States)

    Zhang, Xiaojie

    2012-01-01

    Low-density parity-check (LDPC) codes have been the focus of much research over the past decade thanks to their near Shannon limit performance and to their efficient message-passing (MP) decoding algorithms. However, the error floor phenomenon observed in MP decoding, which manifests itself as an abrupt change in the slope of the error-rate curve,…

  18. Quantitative Analysis of Standardized Dress Code and Minority Academic Achievement

    Science.gov (United States)

    Proctor, J. R.

    2013-01-01

    This study was designed to investigate if a statistically significant variance exists in African American and Hispanic students' attendance and Texas Assessment of Knowledge and Skills test scores in mathematics before and after the implementation of a standardized dress code. For almost two decades supporters and opponents of public school…

  19. Broadcast Coded Slotted ALOHA: A Finite Frame Length Analysis

    DEFF Research Database (Denmark)

    Ivanov, Mikhail; Brännström, Frederik; Graell i Amat, Alexandre;

    2016-01-01

    We propose an uncoordinated medium access control (MAC) protocol, called all-to-all broadcast coded slotted ALOHA (B-CSA) for reliable all-to-all broadcast with strict latency constraints. In B-CSA, each user acts as both transmitter and receiver in a half-duplex mode. The half-duplex mode gives...

  20. Validation of numerical codes for the analysis of plasma discharges

    Energy Technology Data Exchange (ETDEWEB)

    Albanese, R. (Univ. di Salerno, Dipt. di Ingegneria Elettronica, Fisciano (Italy)); Bottura, L. (NET Team, Garching (Germany)); Chiocchio, S. (NET Team, Garching (Germany)); Coccorese, E. (Univ. di Reggio Calabria, Ist. di Ingegneria Elettronica (Italy)); Gernhardt, J. (Max Planck IPP, Garching (Germany)); Gruber, O. (Max Planck IPP, Garching (Germany)); Fresa, R. (Univ. di Salerno, Dipt. di Ingegneria Elettronica, Fisciano (Italy)); Martone, R. (Univ. di Salerno, Dipt. di Ingegneria Elettronica, Fisciano (Italy)); Portone, A. (NET Team, Garching (Germany)); Seidel, U. (Max Planck IPP, Garching (Germany))

    1994-01-01

    Electromagnetic aspects in the design of ITER-like reactors call for an extensive use of complex and advanced numerical codes. For this reason a strong attention has been paid within the NET-Team to the code development. In particular, through a cooperation with some Italian universities, during the last years a number of numerical procedures were developed and integrated. In order to assess the code reliability and to gain confidence on their predictions for next generation ITER-like reactors, the validation of the codes against experiments has to be considered as a strict requirement. Aim of this paper is to give a comprehensive presentation of this problem in the light of the results of a campaign of validation runs. The main outcome of this work is that the computational procedures, which have been developed for the NET project and then extensively used also for ITER studies, can be considered as experimentally validated in a sufficiently wide range of cases of interest. In particular, computed values are compared with experimental measurements made during some typical ASDEX-Upgrade discharges. From the electromagnetic point of view, many features of this machine are common to the ITER concept, so that the results of the validation can reasonably be extended to the ITER case. (orig.)

  1. Comparison of Activation Analysis Codes between CINDER'90 and ORIGEN-S

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Jeong Dong; Choi, Hong Yeop; Lee, Yong Deok; Kim, Hodong [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)

    2014-05-15

    A Slowing Down Time Spectrometer (SDTS) system is the most feasible technology among the non-destructive techniques to directly analyze the content of isotopic fissile material. SDTS is necessary to a source neutron for inducing isotopic fissile fission. The source neutron is produced between the electron beam and a metal target by an (e,γ)(γ,n) reaction in the target. The target is required to have a high intensity neutron source through a proper target design. The status of activation on the designed target is analyzed through the activation code. Also, an activation evaluation of the material of the shielding facilities for SDTS system is required. The radioactivity intensity and kind of nuclides are measured through an activation analysis. ORIGEN-S and CINDER'90 codes are an activation code and are used in combination with the MCNPX code. ORIGEN-S code interprets a problem as one point about target. It cannot describe the geometry. CINDER'90 code can describe a 3D-geometry, and the result of CINDER'90 has high reliability when using a multi-group library. In this research, CINDER'90 was introduced as an activation analysis code and compared with the ORIGEN-S code. An activation analysis was conducted on the materials of the designed target. The ORIGEN-S and CINDER'90 code simulation results are provided for a selection of the activation analysis code. A Slowing Down Time Spectrometer (SDTS) system is a highly efficient technique in a nuclear material analysis. An activation analysis on the shielding and target material was required for the SDTS system. The activation results of CINDER'90 and ORIGEN-S codes were similar or different according to the nuclides because the cross section library of the codes is different. In utilizing the code, CINDER'90 code is more convenient than ORIGEN-S. It can describe the 3D-geometry, and therefore the activation information can be obtained by one simulation. The results of the activation

  2. An establishment of MELCOR code to generate source terms for off site consequence analysis

    Energy Technology Data Exchange (ETDEWEB)

    Park, S. H.; Han, S.; Ahn, K. I. [KAERI, Daejeon (Korea, Republic of)

    2012-10-15

    Since the Fukushima accident, an effective approach to a source term analysis for off site consequence analyses has been needed. The MELCOR code has the capability to assess the source term characteristics for this kind of demand. A comprehensive effort is required to use the MELCOR code for a source term analysis effectively. For this purpose, the following works are required: - Review and assess the MELCOR model relevant to source term characterization - Generate input files for source term analysis - Utilize the source term parameters This paper shows an effort to establish the MELCOR code to generate source terms for an off site consequence analysis.

  3. Development and Application of Subchannel Analysis Code Technology for Advanced Reactor Systems

    Energy Technology Data Exchange (ETDEWEB)

    Hwang, Dae Hyun; Seo, K. W

    2006-01-15

    A study has been performed for the development and assessment of a subchannel analysis code which is purposed to be used for the analysis of advanced reactor conditions with various configurations of reactor core and several kinds of reactor coolant fluids. The subchannel analysis code was developed on the basis of MATRA code which is being developed at KAERI. A GUI (Graphic User Interface) system was adopted in order to reduce input error and to enhance user convenience. The subchannel code was complemented in the property calculation modules by including various fluids such as heavy liquid metal, gas, refrigerant,and supercritical water. The subchannel code was applied to calculate the local thermal hydraulic conditions inside the non-square test bundles which was employed for the analysis of CHF. The applicability of the subchannel code was evaluated for a high temperature gas cooled reactor condition and supercritical pressure conditions with water and Freon. A subchannel analysis has been conducted for European ADS(Accelerator-Driven subcritical System) with Pb-Bi coolant through the international cooperation work between KAERI and FZK, Germany. In addition, the prediction capability of the subchannel code was evaluated for the subchannel void distribution data by participating an international code benchmark program which was organized by OECD/NRC.

  4. Dynamic analysis of spur gears using computer program DANST

    Science.gov (United States)

    Oswald, Fred B.; Lin, Hsiang H.; Liou, Chuen-Huei; Valco, Mark J.

    1993-06-01

    DANST is a computer program for static and dynamic analysis of spur gear systems. The program can be used for parametric studies to predict the effect on dynamic load and tooth bending stress of spur gears due to operating speed, torque, stiffness, damping, inertia, and tooth profile. DANST performs geometric modeling and dynamic analysis for low- or high-contact-ratio spur gears. DANST can simulate gear systems with contact ratio ranging from one to three. It was designed to be easy to use, and it is extensively documented by comments in the source code. This report describes the installation and use of DANST. It covers input data requirements and presents examples. The report also compares DANST predictions for gear tooth loads and bending stress to experimental and finite element results.

  5. Dynamical analysis of highly excited molecular spectra

    Energy Technology Data Exchange (ETDEWEB)

    Kellman, M.E. [Univ. of Oregon, Eugene (United States)

    1993-12-01

    The goal of this program is new methods for analysis of spectra and dynamics of highly excited vibrational states of molecules. In these systems, strong mode coupling and anharmonicity give rise to complicated classical dynamics, and make the simple normal modes analysis unsatisfactory. New methods of spectral analysis, pattern recognition, and assignment are sought using techniques of nonlinear dynamics including bifurcation theory, phase space classification, and quantization of phase space structures. The emphasis is chaotic systems and systems with many degrees of freedom.

  6. Aerodynamic analysis of flapping foils using volume grid deformation code

    Energy Technology Data Exchange (ETDEWEB)

    Ko, Jin Hwan [Seoul National University, Seoul (Korea, Republic of); Kim, Jee Woong; Park, Soo Hyung; Byun, Do Young [Konkuk University, Seoul (Korea, Republic of)

    2009-06-15

    Nature-inspired flapping foils have attracted interest for their high thrust efficiency, but the large motions of their boundaries need to be considered. It is challenging to develop robust, efficient grid deformation algorithms appropriate for the large motions in three dimensions. In this paper, a volume grid deformation code is developed based on finite macro-element and transfinite interpolation, which successfully interfaces to a structured multi-block Navier-Stokes code. A suitable condition that generates the macro-elements with efficiency and improves the robustness of grid regularity is presented as well. As demonstrated by an airfoil with various motions related to flapping, the numerical results of aerodynamic forces by the developed method are shown to be in good agreement with those of an experimental data or a previous numerical solution

  7. Parallelization of Finite Element Analysis Codes Using Heterogeneous Distributed Computing

    Science.gov (United States)

    Ozguner, Fusun

    1996-01-01

    Performance gains in computer design are quickly consumed as users seek to analyze larger problems to a higher degree of accuracy. Innovative computational methods, such as parallel and distributed computing, seek to multiply the power of existing hardware technology to satisfy the computational demands of large applications. In the early stages of this project, experiments were performed using two large, coarse-grained applications, CSTEM and METCAN. These applications were parallelized on an Intel iPSC/860 hypercube. It was found that the overall speedup was very low, due to large, inherently sequential code segments present in the applications. The overall execution time T(sub par), of the application is dependent on these sequential segments. If these segments make up a significant fraction of the overall code, the application will have a poor speedup measure.

  8. Load Flow Analysis Using Real Coded Genetic Algorithm

    Directory of Open Access Journals (Sweden)

    Himakar Udatha

    2014-02-01

    Full Text Available This paper presents a Real Coded Genetic Algorithm (RCGA for finding the load flow solution of electrical power systems. The proposed method is based on the minimization of the real and reactive power mismatches at various buses. The traditional methods such as Gauss-Seidel method and Newton-Raphson (NR method have certain drawbacks under abnormal operating condition. In order to overcome these problems, the load flow solution based on Real Coded Genetic Algorithm (RCGA is presented in this paper. Two cross over techniques, Arithmetic crossover and heuristic crossover are used to solve the power flow problem. The proposed method is applied for 3-bus, 5-bus and 6-bus systems and the results are presented.

  9. Ambiguity and nonidentifiability in the statistical analysis of neural codes

    Science.gov (United States)

    Amarasingham, Asohan; Geman, Stuart; Harrison, Matthew T.

    2015-01-01

    Many experimental studies of neural coding rely on a statistical interpretation of the theoretical notion of the rate at which a neuron fires spikes. For example, neuroscientists often ask, “Does a population of neurons exhibit more synchronous spiking than one would expect from the covariability of their instantaneous firing rates?” For another example, “How much of a neuron’s observed spiking variability is caused by the variability of its instantaneous firing rate, and how much is caused by spike timing variability?” However, a neuron’s theoretical firing rate is not necessarily well-defined. Consequently, neuroscientific questions involving the theoretical firing rate do not have a meaning in isolation but can only be interpreted in light of additional statistical modeling choices. Ignoring this ambiguity can lead to inconsistent reasoning or wayward conclusions. We illustrate these issues with examples drawn from the neural-coding literature. PMID:25934918

  10. Counteracting Byzantine Adversaries with Network Coding: An Overhead Analysis

    CERN Document Server

    Kim, MinJi; Barros, Joao

    2008-01-01

    Network coding increases throughput and is robust against failures and erasures. However, since it allows mixing of information within the network, a single corrupted packet generated by a Byzantine attacker can easily contaminate the information to multiple destinations. In this paper, we study the transmission overhead associated with detecting Byzantine adversaries at a trusted node using network coding. We consider three different schemes: end-to-end error correction, packet-based Byzantine detection scheme, and generation-based Byzantine detection scheme. In end-to-end error correction, it is known that we can correct up to the min-cut between the source and destinations. However, if we use Byzantine detection schemes, we can detect polluted data, drop them, and therefore, only transmit valid data. For the dropped data, the destinations perform erasure correction, which is computationally lighter than error correction. We show that, with enough attackers present in the network, Byzantine detection scheme...

  11. Statistical Mechanics Analysis of LDPC Coding in MIMO Gaussian Channels

    OpenAIRE

    Alamino, Roberto C.; Saad, David

    2007-01-01

    Using analytical methods of statistical mechanics, we analyse the typical behaviour of a multiple-input multiple-output (MIMO) Gaussian channel with binary inputs under LDPC network coding and joint decoding. The saddle point equations for the replica symmetric solution are found in particular realizations of this channel, including a small and large number of transmitters and receivers. In particular, we examine the cases of a single transmitter, a single receiver and the symmetric and asymm...

  12. Performance analysis of LDPC codes on OOK terahertz wireless channels

    Science.gov (United States)

    Chun, Liu; Chang, Wang; Jun-Cheng, Cao

    2016-02-01

    Atmospheric absorption, scattering, and scintillation are the major causes to deteriorate the transmission quality of terahertz (THz) wireless communications. An error control coding scheme based on low density parity check (LDPC) codes with soft decision decoding algorithm is proposed to improve the bit-error-rate (BER) performance of an on-off keying (OOK) modulated THz signal through atmospheric channel. The THz wave propagation characteristics and channel model in atmosphere is set up. Numerical simulations validate the great performance of LDPC codes against the atmospheric fading and demonstrate the huge potential in future ultra-high speed beyond Gbps THz communications. Project supported by the National Key Basic Research Program of China (Grant No. 2014CB339803), the National High Technology Research and Development Program of China (Grant No. 2011AA010205), the National Natural Science Foundation of China (Grant Nos. 61131006, 61321492, and 61204135), the Major National Development Project of Scientific Instrument and Equipment (Grant No. 2011YQ150021), the National Science and Technology Major Project (Grant No. 2011ZX02707), the International Collaboration and Innovation Program on High Mobility Materials Engineering of the Chinese Academy of Sciences, and the Shanghai Municipal Commission of Science and Technology (Grant No. 14530711300).

  13. Severe accident analysis using dynamic accident progression event trees

    Science.gov (United States)

    Hakobyan, Aram P.

    In present, the development and analysis of Accident Progression Event Trees (APETs) are performed in a manner that is computationally time consuming, difficult to reproduce and also can be phenomenologically inconsistent. One of the principal deficiencies lies in the static nature of conventional APETs. In the conventional event tree techniques, the sequence of events is pre-determined in a fixed order based on the expert judgments. The main objective of this PhD dissertation was to develop a software tool (ADAPT) for automated APET generation using the concept of dynamic event trees. As implied by the name, in dynamic event trees the order and timing of events are determined by the progression of the accident. The tool determines the branching times from a severe accident analysis code based on user specified criteria for branching. It assigns user specified probabilities to every branch, tracks the total branch probability, and truncates branches based on the given pruning/truncation rules to avoid an unmanageable number of scenarios. The function of a dynamic APET developed includes prediction of the conditions, timing, and location of containment failure or bypass leading to the release of radioactive material, and calculation of probabilities of those failures. Thus, scenarios that can potentially lead to early containment failure or bypass, such as through accident induced failure of steam generator tubes, are of particular interest. Also, the work is focused on treatment of uncertainties in severe accident phenomena such as creep rupture of major RCS components, hydrogen burn, containment failure, timing of power recovery, etc. Although the ADAPT methodology (Analysis of Dynamic Accident Progression Trees) could be applied to any severe accident analysis code, in this dissertation the approach is demonstrated by applying it to the MELCOR code [1]. A case study is presented involving station blackout with the loss of auxiliary feedwater system for a

  14. The SCEC-USGS Dynamic Earthquake Rupture Code Comparison Exercise - Simulations of Large Earthquakes and Strong Ground Motions

    Science.gov (United States)

    Harris, R.

    2015-12-01

    I summarize the progress by the Southern California Earthquake Center (SCEC) and U.S. Geological Survey (USGS) Dynamic Rupture Code Comparison Group, that examines if the results produced by multiple researchers' earthquake simulation codes agree with each other when computing benchmark scenarios of dynamically propagating earthquake ruptures. These types of computer simulations have no analytical solutions with which to compare, so we use qualitative and quantitative inter-code comparisons to check if they are operating satisfactorily. To date we have tested the codes against benchmark exercises that incorporate a range of features, including single and multiple planar faults, single rough faults, slip-weakening, rate-state, and thermal pressurization friction, elastic and visco-plastic off-fault behavior, complete stress drops that lead to extreme ground motion, heterogeneous initial stresses, and heterogeneous material (rock) structure. Our goal is reproducibility, and we focus on the types of earthquake-simulation assumptions that have been or will be used in basic studies of earthquake physics, or in direct applications to specific earthquake hazard problems. Our group's goals are to make sure that when our earthquake-simulation codes simulate these types of earthquake scenarios along with the resulting simulated strong ground shaking, that the codes are operating as expected. For more introductory information about our group and our work, please see our group's overview papers, Harris et al., Seismological Research Letters, 2009, and Harris et al., Seismological Research Letters, 2011, along with our website, scecdata.usc.edu/cvws.

  15. Atmospheric Cluster Dynamics Code: a flexible method for solution of the birth-death equations

    Directory of Open Access Journals (Sweden)

    M. J. McGrath

    2012-03-01

    Full Text Available The Atmospheric Cluster Dynamics Code (ACDC is presented and explored. This program was created to study the first steps of atmospheric new particle formation by examining the formation of molecular clusters from atmospherically relevant molecules. The program models the cluster kinetics by explicit solution of the birth–death equations, using an efficient computer script for their generation and the MATLAB ode15s routine for their solution. Through the use of evaporation rate coefficients derived from formation free energies calculated by quantum chemical methods for clusters containing dimethylamine or ammonia and sulphuric acid, we have explored the effect of changing various parameters at atmospherically relevant monomer concentrations. We have included in our model clusters with 0–4 base molecules and 0–4 sulfuric acid molecules for which we have commensurable quantum chemical data. The tests demonstrate that large effects can be seen for even small changes in different parameters, due to the non-linearity of the system. In particular, changing the temperature had a significant impact on the steady-state concentrations of all clusters, while the boundary effects (allowing clusters to grow to sizes beyond the largest cluster that the code keeps track of, or forbidding such processes, coagulation sink terms, non-monomer collisions, sticking probabilities and monomer concentrations did not show as large effects under the conditions studied. Removal of coagulation sink terms prevented the system from reaching the steady state when all the initial cluster concentrations were set to the default value of 1 m−3, which is probably an effect caused by studying only relatively small cluster sizes.

  16. The Use of Legal Terminology in Student Conduct Codes: A Content Analysis

    Science.gov (United States)

    Martin, Janice E.; Janosik, Steven M.

    2004-01-01

    A content analysis of 20 student conduct codes identified through stratified random sampling was performed to determine the extent to which legal terminology had been eliminated, as suggested by judicial affairs experts. The results showed that 80% of the codes selected in this study still contained some legal terms. These words and phrases are…

  17. Analysis of Energy Saving Impacts of New Residential Energy Codes for the Gulf Coast

    Energy Technology Data Exchange (ETDEWEB)

    Lucas, Robert G.

    2007-01-01

    At the request of the Government Accountability Office (GAO), DOE’s Building Energy Codes Program performed an analysis of the energy savings and cost impacts associated with the use of newer and more efficient residential building energy codes in the states of Louisiana and Mississippi.

  18. Code Analysis and Refactoring with Clang Tools, Version 0.1

    Energy Technology Data Exchange (ETDEWEB)

    2016-12-23

    Code Analysis and Refactoring with Clang Tools is a small set of example code that demonstrates techniques for applying tools distributed with the open source Clang compiler. Examples include analyzing where variables are used and replacing old data structures with standard structures.

  19. Mathematical Description of Wafer-1, a Three-Dimensional Code for LWR Fuel Performance Analysis

    DEFF Research Database (Denmark)

    Kjær-Pedersen, Niels

    1975-01-01

    This article describes in detail the mathematical formulation used in the WAFER-1 code, which is presently used for three-dimensional analysis of LWR fuel pin performance. The code aims at a prediction of the local stress-strain history in the cladding, especially with regard to the ridging pheno...

  20. Analysis and design of raptor codes for joint decoding using Information Content evolution

    CERN Document Server

    Venkiah, Auguste; Declercq, David

    2007-01-01

    In this paper, we present an analytical analysis of the convergence of raptor codes under joint decoding over the binary input additive white noise channel (BIAWGNC), and derive an optimization method. We use Information Content evolution under Gaussian approximation, and focus on a new decoding scheme that proves to be more efficient: the joint decoding of the two code components of the raptor code. In our general model, the classical tandem decoding scheme appears to be a subcase, and thus, the design of LT codes is also possible.

  1. Dynamics Analysis of Wind Energy Production Development

    Science.gov (United States)

    Berg, V. I.; Zakirzakov, A. G.; Gordievskaya, E. F.

    2017-01-01

    The paper presents the analysis of the introduction experience and dynamics development of the world wind energy production. Calculated the amount of wind energy sources investments and the production capacity growth dynamics of the wind turbines. The studies have shown that the introduction dynamics of new wind energy sources is higher than any other energy source.

  2. Reusability of coded data in the primary care electronic medical record : A dynamic cohort study concerning cancer diagnoses

    NARCIS (Netherlands)

    Sollie, Annet; Sijmons, Rolf H.; Helsper, Charles W.; Numans, Mattijs E.

    2017-01-01

    Objectives: To assess quality and reusability of coded cancer diagnoses in routine primary care data. To identify factors that influence data quality and areas for improvement. Methods: A dynamic cohort study in a Dutch network database containing 250,000 anonymized electronic medical records (EMRs)

  3. Reusability of coded data in the primary care electronic medical record : A dynamic cohort study concerning cancer diagnoses

    NARCIS (Netherlands)

    Sollie, Annet; Sijmons, Rolf H; Helsper, Charles; Numans, Mattijs E

    2017-01-01

    OBJECTIVES: To assess quality and reusability of coded cancer diagnoses in routine primary care data. To identify factors that influence data quality and areas for improvement. METHODS: A dynamic cohort study in a Dutch network database containing 250,000 anonymized electronic medical records (EMRs)

  4. Development of heat and mass balance analysis code in out-of-pile hydrogen production system for HTTR heat utilization system (contract research)

    Energy Technology Data Exchange (ETDEWEB)

    Inaba, Yoshitomo; Inagaki, Yoshiyuki; Hayashi, Koji; Suyama, Kazumasa [Japan Atomic Energy Research Inst., Oarai, Ibaraki (Japan). Oarai Research Establishment

    1999-03-01

    A heat and mass balance analysis code has been developed to examine test conditions, to investigate transient behavior etc. in the out-of-pile hydrogen production system for the HTTR heat utilization system. The code can analyze temperature, mass and pressure profiles of helium and process gases and behavior of the control system under both static state (case of steady operation) and dynamic state (case of transient operation). This report describes analytical methods, basic equations and constitution of the code, and how to make of the input data, estimate of the analytical results and so on. (author)

  5. Development of heat and mass balance analysis code in out-of-pile hydrogen production system for HTTR heat utilization system (contract research)

    Energy Technology Data Exchange (ETDEWEB)

    Inaba, Yoshitomo; Inagaki, Yoshiyuki; Hayashi, Koji; Suyama, Kazumasa [Japan Atomic Energy Research Inst., Oarai, Ibaraki (Japan). Oarai Research Establishment

    1999-03-01

    A heat and mass balance analysis code has been developed to examine test conditions, to investigate transient behavior etc. in the out-of-pile hydrogen production system for the HTTR heat utilization system. The code can analyze temperature, mass and pressure profiles of helium and process gases and behavior of the control system under both static state (case of steady operation) and dynamic state (case of transient operation). This report describes analytical methods, basic equations and constitution of the code, and how to make of the input data, estimate of the analytical results and so on. (author)

  6. An Introduction to Thermodynamic Performance Analysis of Aircraft Gas Turbine Engine Cycles Using the Numerical Propulsion System Simulation Code

    Science.gov (United States)

    Jones, Scott M.

    2007-01-01

    This document is intended as an introduction to the analysis of gas turbine engine cycles using the Numerical Propulsion System Simulation (NPSS) code. It is assumed that the analyst has a firm understanding of fluid flow, gas dynamics, thermodynamics, and turbomachinery theory. The purpose of this paper is to provide for the novice the information necessary to begin cycle analysis using NPSS. This paper and the annotated example serve as a starting point and by no means cover the entire range of information and experience necessary for engine performance simulation. NPSS syntax is presented but for a more detailed explanation of the code the user is referred to the NPSS User Guide and Reference document (ref. 1).

  7. Precise capacity analysis in binary networks with multiple coding level inputs.

    Science.gov (United States)

    Amit, Yali; Huang, Yibi

    2010-03-01

    We compute retrieval probabilities as a function of pattern age for networks with binary neurons and synapses updated with the simple Hebbian learning model studied in Amit and Fusi ( 1994 ). The analysis depends on choosing a neural threshold that enables patterns to stabilize in the neural dynamics. In contrast to most earlier work, where selective neurons for each pattern are drawn independently with fixed probability f, here we analyze the situation where f is drawn from some distribution on a range of coding levels. In order to set a workable threshold in this setting, it is necessary to introduce a simple inhibition in the neural dynamics whose magnitude depends on the total activity of the network. Proper choice of the threshold depends on the value of the covariances between the synapses for which we provide an explicit formula. Retrieval probabilities depend on the distribution of the fields induced by a learned pattern. We show that the field induced by the first learned pattern evolves as a Markov chain during subsequent learning epochs, leading to a recursive formula for the distribution. Alternatively, the distribution can be computed using a normal approximation, which involves the value of the synaptic covariances. Capacity is computed as the sum of the retrieval probabilities over all ages. We show through simulation that the chosen threshold enables retrieval with asynchronous dynamics even in the presence of significant noise in the initial state of the pattern. The computed probabilities with both methods are shown to be very close to probabilities estimated from simulation. The analysis is extended to randomly connected networks.

  8. A computer code for analysis of severe accidents in LWRs

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    2001-07-01

    The ICARE2 computer code, developed and validated since 1988 at IPSN (nuclear safety and protection institute), calculates in a mechanistic way the physical and chemical phenomena involved in the core degradation process during possible severe accidents in LWR's. The coupling between ICARE2 and the best-estimate thermal-hydraulics code CATHARE2 was completed at IPSN and led to the release of a first ICARE/CATHARE V1 version in 1999, followed by 2 successive revisions in 2000 and 2001. This documents gathers all the contributions presented at the first international ICARE/CATHARE users'club seminar that took place in November 2001. This seminar was characterized by a high quality and variety of the presentations, showing an increase of reactor applications and user needs in this area (2D/3D aspects, reflooding, corium slumping into the lower head,...). 2 sessions were organized. The first one was dedicated to the applications of ICARE2 V3mod1 against small-scale experiments such as PHEBUS FPT2 and FPT3 tests, PHEBUS AIC, QUENCH experiments, NRU-FLHT-5 test, ACRR-MP1 and DC1 experiments, CORA-PWR tests, and PBF-SFD1.4 test. The second session involved ICARE/CATHARE V1mod1 reactor applications and users'guidelines. Among reactor applications we found: code applicability to high burn-up fuel rods, simulation of the TMI-2 transient, simulation of a PWR-900 high pressure severe accident sequence, and the simulation of a VVER-1000 large break LOCA scenario. (A.C.)

  9. User's manual for seismic analysis code 'SONATINA-2V'

    Energy Technology Data Exchange (ETDEWEB)

    Hanawa, Satoshi; Iyoku, Tatsuo [Japan Atomic Energy Research Inst., Oarai, Ibaraki (Japan). Oarai Research Establishment

    2001-08-01

    The seismic analysis code, SONATINA-2V, has been developed to analyze the behavior of the HTTR core graphite components under seismic excitation. The SONATINA-2V code is a two-dimensional computer program capable of analyzing the vertical arrangement of the HTTR graphite components, such as fuel blocks, replaceable reflector blocks, permanent reflector blocks, as well as their restraint structures. In the analytical model, each block is treated as rigid body and is restrained by dowel pins which restrict relative horizontal movement but allow vertical and rocking motions between upper and lower blocks. Moreover, the SONATINA-2V code is capable of analyzing the core vibration behavior under both simultaneous excitations of vertical and horizontal directions. The SONATINA-2V code is composed of the main program, pri-processor for making the input data to SONATINA-2V and post-processor for data processing and making the graphics from analytical results. Though the SONATINA-2V code was developed in order to work in the MSP computer system of Japan Atomic Energy Research Institute (JAERI), the computer system was abolished with the technical progress of computer. Therefore, improvement of this analysis code was carried out in order to operate the code under the UNIX machine, SR8000 computer system, of the JAERI. The users manual for seismic analysis code, SONATINA-2V, including pri- and post-processor is given in the present report. (author)

  10. High-Fidelity Buckling Analysis of Composite Cylinders Using the STAGS Finite Element Code

    Science.gov (United States)

    Hilburger, Mark W.

    2014-01-01

    Results from previous shell buckling studies are presented that illustrate some of the unique and powerful capabilities in the STAGS finite element analysis code that have made it an indispensable tool in structures research at NASA over the past few decades. In particular, prototypical results from the development and validation of high-fidelity buckling simulations are presented for several unstiffened thin-walled compression-loaded graphite-epoxy cylindrical shells along with a discussion on the specific methods and user-defined subroutines in STAGS that are used to carry out the high-fidelity simulations. These simulations accurately account for the effects of geometric shell-wall imperfections, shell-wall thickness variations, local shell-wall ply-gaps associated with the fabrication process, shell-end geometric imperfections, nonuniform applied end loads, and elastic boundary conditions. The analysis procedure uses a combination of nonlinear quasi-static and transient dynamic solution algorithms to predict the prebuckling and unstable collapse response characteristics of the cylinders. Finally, the use of high-fidelity models in the development of analysis-based shell-buckling knockdown (design) factors is demonstrated.

  11. Unraveling the distributed neural code of facial identity through spatiotemporal pattern analysis.

    Science.gov (United States)

    Nestor, Adrian; Plaut, David C; Behrmann, Marlene

    2011-06-14

    Face individuation is one of the most impressive achievements of our visual system, and yet uncovering the neural mechanisms subserving this feat appears to elude traditional approaches to functional brain data analysis. The present study investigates the neural code of facial identity perception with the aim of ascertaining its distributed nature and informational basis. To this end, we use a sequence of multivariate pattern analyses applied to functional magnetic resonance imaging (fMRI) data. First, we combine information-based brain mapping and dynamic discrimination analysis to locate spatiotemporal patterns that support face classification at the individual level. This analysis reveals a network of fusiform and anterior temporal areas that carry information about facial identity and provides evidence that the fusiform face area responds with distinct patterns of activation to different face identities. Second, we assess the information structure of the network using recursive feature elimination. We find that diagnostic information is distributed evenly among anterior regions of the mapped network and that a right anterior region of the fusiform gyrus plays a central role within the information network mediating face individuation. These findings serve to map out and characterize a cortical system responsible for individuation. More generally, in the context of functionally defined networks, they provide an account of distributed processing grounded in information-based architectures.

  12. Jointly Decoded Raptor Codes: Analysis and Design for the BIAWGN Channel

    Directory of Open Access Journals (Sweden)

    Venkiah Auguste

    2009-01-01

    Full Text Available Abstract We are interested in the analysis and optimization of Raptor codes under a joint decoding framework, that is, when the precode and the fountain code exchange soft information iteratively. We develop an analytical asymptotic convergence analysis of the joint decoder, derive an optimization method for the design of efficient output degree distributions, and show that the new optimized distributions outperform the existing ones, both at long and moderate lengths. We also show that jointly decoded Raptor codes are robust to channel variation: they perform reasonably well over a wide range of channel capacities. This robustness property was already known for the erasure channel but not for the Gaussian channel. Finally, we discuss some finite length code design issues. Contrary to what is commonly believed, we show by simulations that using a relatively low rate for the precode , we can improve greatly the error floor performance of the Raptor code.

  13. Analysis Code for High Gradient Dielectric Insulator Surface Breakdown

    Energy Technology Data Exchange (ETDEWEB)

    Ives, Robert Lawrence [Calabazas Creek Research, Inc.; Verboncoeur, John [University of California - Berkeley; Aldan, Manuel [University of California, Berkeley

    2010-05-30

    High voltage (HV) insulators are critical components in high-energy, accelerator and pulsed power systems that drive diverse applications in the national security, nuclear weapons science, defense and industrial arenas. In these systems, the insulator may separate vacuum/non-vacuum regions or conductors with high electrical field gradients. These insulators will often fail at electric fields over an order of magnitude lower than their intrinsic dielectric strength due to flashover at the dielectric interface. Decades of studies have produced a wealth of information on fundamental processes and mechanisms important for flashover initiation, but only for relatively simple insulator configurations in controlled environments. Accelerator and pulsed power system designers are faced with applying the fundamental knowledge to complex, operational devices with escalating HV requirements. Designers are forced to rely on “best practices” and expensive prototype testing, providing boundaries for successful operation. However, the safety margin is difficult to estimate, and system design must be very conservative for situations where testing is not practicable, or replacement of failed parts is disruptive or expensive. The Phase I program demonstrated the feasibility of developing an advanced code for modeling insulator breakdown. Such a code would be of great interest for a number of applications, including high energy physics, microwave source development, fusion sciences, and other research and industrial applications using high voltage devices.

  14. Generator dynamics in aeroelastic analysis and simulations

    DEFF Research Database (Denmark)

    Larsen, Torben J.; Hansen, Morten Hartvig; Iov, F.

    2003-01-01

    This report contains a description of a dynamic model for a doubly-fed induction generator. The model has physical input parameters (voltage, resistance, reactance etc.) and can be used to calculate rotor and stator currents, hence active and reactivepower. A perturbation method has been used...... to reduce the original generator model equations to a set of equations which can be solved with the same time steps as a typical aeroelastic code. The method is used to separate the fast transients of the modelfrom the slow variations and deduce a reduced order expression for the slow part. Dynamic effects...

  15. Dynamic Hurricane Data Analysis Tool

    Science.gov (United States)

    Knosp, Brian W.; Li, Peggy; Vu, Quoc A.

    2009-01-01

    A dynamic hurricane data analysis tool allows users of the JPL Tropical Cyclone Information System (TCIS) to analyze data over a Web medium. The TCIS software is described in the previous article, Tropical Cyclone Information System (TCIS) (NPO-45748). This tool interfaces with the TCIS database to pull in data from several different atmospheric and oceanic data sets, both observed by instruments. Users can use this information to generate histograms, maps, and profile plots for specific storms. The tool also displays statistical values for the user-selected parameter for the mean, standard deviation, median, minimum, and maximum values. There is little wait time, allowing for fast data plots over date and spatial ranges. Users may also zoom-in for a closer look at a particular spatial range. This is version 1 of the software. Researchers will use the data and tools on the TCIS to understand hurricane processes, improve hurricane forecast models and identify what types of measurements the next generation of instruments will need to collect.

  16. Viriato: a Fourier-Hermite spectral code for strongly magnetised fluid-kinetic plasma dynamics

    Science.gov (United States)

    Loureiro, Nuno; Dorland, William; Fazendeiro, Luis; Kanekar, Anjor; Mallet, Alfred; Zocco, Alessandro

    2015-11-01

    We report on the algorithms and numerical methods used in Viriato, a novel fluid-kinetic code that solves two distinct sets of equations: (i) the Kinetic Reduced Electron Heating Model equations [Zocco & Schekochihin, 2011] and (ii) the kinetic reduced MHD (KRMHD) equations [Schekochihin et al., 2009]. Two main applications of these equations are magnetised (Alfvnénic) plasma turbulence and magnetic reconnection. Viriato uses operator splitting to separate the dynamics parallel and perpendicular to the ambient magnetic field (assumed strong). Along the magnetic field, Viriato allows for either a second-order accurate MacCormack method or, for higher accuracy, a spectral-like scheme. Perpendicular to the field Viriato is pseudo-spectral, and the time integration is performed by means of an iterative predictor-corrector scheme. In addition, a distinctive feature of Viriato is its spectral representation of the parallel velocity-space dependence, achieved by means of a Hermite representation of the perturbed distribution function. A series of linear and nonlinear benchmarks and tests are presented, with focus on 3D decaying kinetic turbulence. Work partially supported by Fundação para a Ciência e Tecnologia via Grants UID/FIS/50010/2013 and IF/00530/2013.

  17. Spontaneous activity in the piriform cortex extends the dynamic range of cortical odor coding.

    Science.gov (United States)

    Tantirigama, Malinda L S; Huang, Helena H-Y; Bekkers, John M

    2017-02-28

    Neurons in the neocortex exhibit spontaneous spiking activity in the absence of external stimuli, but the origin and functions of this activity remain uncertain. Here, we show that spontaneous spiking is also prominent in a sensory paleocortex, the primary olfactory (piriform) cortex of mice. In the absence of applied odors, piriform neurons exhibit spontaneous firing at mean rates that vary systematically among neuronal classes. This activity requires the participation of NMDA receptors and is entirely driven by bottom-up spontaneous input from the olfactory bulb. Odor stimulation produces two types of spatially dispersed, odor-distinctive patterns of responses in piriform cortex layer 2 principal cells: Approximately 15% of cells are excited by odor, and another approximately 15% have their spontaneous activity suppressed. Our results show that, by allowing odor-evoked suppression as well as excitation, the responsiveness of piriform neurons is at least twofold less sparse than currently believed. Hence, by enabling bidirectional changes in spiking around an elevated baseline, spontaneous activity in the piriform cortex extends the dynamic range of odor representation and enriches the coding space for the representation of complex olfactory stimuli.

  18. Analysis of the Length of Braille Texts in English Braille American Edition, the Nemeth Code, and Computer Braille Code versus the Unified English Braille Code

    Science.gov (United States)

    Knowlton, Marie; Wetzel, Robin

    2006-01-01

    This study compared the length of text in English Braille American Edition, the Nemeth code, and the computer braille code with the Unified English Braille Code (UEBC)--also known as Unified English Braille (UEB). The findings indicate that differences in the length of text are dependent on the type of material that is transcribed and the grade…

  19. First results for fluid dynamics, neutronics and fission product behavior in HTR applying the HTR code package (HCP) prototype

    Energy Technology Data Exchange (ETDEWEB)

    Allelein, H.-J., E-mail: h.j.allelein@fz-juelich.de [Forschungszentrum Jülich, 52425 Jülich (Germany); Institute for Reactor Safety and Reactor Technology, RWTH Aachen University, 52064 Aachen (Germany); Kasselmann, S.; Xhonneux, A.; Tantillo, F.; Trabadela, A.; Lambertz, D. [Forschungszentrum Jülich, 52425 Jülich (Germany)

    2016-09-15

    To simulate the different aspects of High Temperature Reactor (HTR) cores, a variety of specialized computer codes have been developed at Forschungszentrum Jülich (IEK-6) and Aachen University (LRST) in the last decades. In order to preserve knowledge, to overcome present limitations and to make these codes applicable to modern computer clusters, these individual programs are being integrated into a consistent code package. The so-called HTR code package (HCP) couples the related and recently applied physics models in a highly integrated manner and therefore allows to simulate phenomena with higher precision in space and time while at the same time applying state-of-the-art programming techniques and standards. This paper provides an overview of the status of the HCP and reports about first benchmark results for an HCP prototype which couples the fluid dynamics and time dependent neutronics code MGT-3D, the burn up code TNT and the fission product release code STACY. Due to the coupling of MGT-3D and TNT, a first step towards a new reactor operation and accident simulation code was made, where nuclide concentrations calculated by TNT lead to new cross sections, which are fed back into MGT-3D. Selected operation scenarios of the HTR-Module 200 concept plant and the HTTR were chosen to be simulated with the HCP prototype. The fission product release during normal operation conditions will be calculated with STACY based on a core status derived from SERPENT and MGT-3D. Comparisons will be shown against data generated by SERPENT and the legacy codes VSOP99/11, NAKURE and FRESCO-II.

  20. De-randomizing Shannon: The Design and Analysis of a Capacity-Achieving Rateless Code

    CERN Document Server

    Balakrishnan, Hari; Perry, Jonathan; Shah, Devavrat

    2012-01-01

    This paper presents an analysis of spinal codes, a class of rateless codes proposed recently. We prove that spinal codes achieve Shannon capacity for the binary symmetric channel (BSC) and the additive white Gaussian noise (AWGN) channel with an efficient polynomial-time encoder and decoder. They are the first rateless codes with proofs of these properties for BSC and AWGN. The key idea in the spinal code is the sequential application of a hash function over the message bits. The sequential structure of the code turns out to be crucial for efficient decoding. Moreover, counter to the wisdom of having an expander structure in good codes, we show that the spinal code, despite its sequential structure, achieves capacity. The pseudo-randomness provided by a hash function suffices for this purpose. Our proof introduces a variant of Gallager's result characterizing the error exponent of random codes for any memoryless channel. We present a novel application of these error-exponent results within the framework of an...

  1. Measuring and Detecting Errors in Occupational Coding: an Analysis of SHARE Data

    Directory of Open Access Journals (Sweden)

    Belloni Michele

    2016-12-01

    Full Text Available This article studies coding errors in occupational data, as the quality of this data is important but often neglected. In particular, we recoded open-ended questions on occupation for last and current job in the Dutch sample of the “Survey of Health, Ageing and Retirement in Europe” (SHARE using a high-quality software program for ex-post coding (CASCOT software. Taking CASCOT coding as our benchmark, our results suggest that the incidence of coding errors in SHARE is high, even when the comparison is made at the level of one-digit occupational codes (28% for last job and 30% for current job. This finding highlights the complexity of occupational coding and suggests that processing errors due to miscoding should be taken into account when undertaking statistical analyses or writing econometric models. Our analysis suggests strategies to alleviate such coding errors, and we propose a set of equations that can predict error. These equations may complement coding software and improve the quality of occupational coding.

  2. Sensitivity analysis codes fire simulation; Analisis de sensibilidad en codigos de simulacion contra incendios

    Energy Technology Data Exchange (ETDEWEB)

    Cordero Garcia, S.; Peco Espinosa, J.

    2010-07-01

    The fire simulation and analytical validation procedures have been gaining importance in the context of safety analysis and probabilistic deterministic Fire at Nuclear Power Plants. Meanwhile, the fire simulation codes have been accepted as valuable tools for risk analysis in fire protection. To evaluate the quality and predictability of the codes used in fire protection, they are validated with different experiments. This work is performed to find the dependency of the FDS output parameters against a certain input.

  3. Clarifying Analysis and Interpretation in Grounded Theory: Using a Conditional Relationship Guide and Reflective Coding Matrix

    Directory of Open Access Journals (Sweden)

    Karen Wilson Scott PhD

    2008-06-01

    Full Text Available Although qualitative methods, grounded theory included, cannot be reduced to formulaic procedures, research tools can clarify the process. The authors discuss two instruments supporting grounded theory analysis and interpretation using two examples from doctoral students. The conditional relationship guide contextualizes the central phenomenon and relates categories linking structure with process. The reflective coding matrix serves as a bridge to the final phase of grounded theory analysis, selective coding and interpretation, and, ultimately, to substantive theory generation.

  4. A new thermal hydraulics code coupled to agent for light water reactor analysis

    Science.gov (United States)

    Eklund, Matthew Deric

    A new numerical model for coupling a thermal hydraulics method based on the Drift Flux and Homogeneous Equilibrium Mixture (HEM) models, with a deterministic neutronics code system AGENT (Arbitrary Geometry Neutron Transport), is developed. Named the TH thermal hydraulics code, it is based on the mass continuity, momentum, and energy equations integrated with appropriate relations for liquid and vapor phasic velocities. The modified conservation equations are then evaluated in one-dimensional (1D) steady-state conditions for LWR coolant subchannel in the axial direction. This permits faster computation times without sacrificing significant accuracy, as compared to other three-dimensional (3D) codes such as RELAP5/TRACE. AGENT is a deterministic neutronics code system based on the Method of Characteristics to solve the 2D/3D neutron transport equation in current and future reactor systems. The coupling scheme between the TH and AGENT codes is accomplished by computing the normalized fission rate profile in the LWR fuel elements by AGENT. The normalized fission rate profile is then transferred to the TH thermal hydraulics code for computing the reactor coolant properties. In conjunction with the 1D axial TH code, a separate 1D radial heat transfer model within the TH code is used to determine the average fuel temperature at each node where coolant properties are calculated. These properties then are entered into Scale 6.1, a criticality analysis code, to recalculate fuel pin neutron interaction cross sections based on thermal feedback. With updated fuel neutron interaction cross sections, the fission rate profile is recalculated in AGENT, and the cycle continues until convergence is reached. The TH code and coupled AGENT-TH code are benchmarked against the TRACE reactor analysis software, showing required agreement in evaluating the basic reactor parameters.

  5. Assessment on the characteristics of the analysis code for KALIMER PSDRS

    Energy Technology Data Exchange (ETDEWEB)

    Eoh, Jae Hyuk; Sim, Yoon Sub; Kim, Seong O.; Kim, Yeon Sik; Kim, Eui Kwang; Wi, Myung Hwan [Korea Atomic Energy Research Institute, Taejeon (Korea)

    2002-04-01

    The PARS2 code was developed to analyze the RHR(Residual Heat Removal) system, especially PSDRS(Passive Safety Decay Heat Removal System), of KALIMER. In this report, preliminary verification and sensitivity analyses for PARS2 code were performed. From the results of the analyses, the PARS2 code has a good agreement with the experimental data of CRIEPI in the range of turbulent airside flow, and also the radiation heat transfer mode was well predicted. In this verification work, it was founded that the code calculation stopped in a very low air flowrate, and the numerical scheme related to the convergence of PARS2 code was adjusted to solve this problem. Through the sensitivity analysis on the PARS2 calculation results from the change of the input parameters, the pool-mixing coefficient related to the heat capacity of the structure in the system was improved such that the physical phenomenon can be well predicted. Also the initial conditions for the code calculation such as the hot and cold pool temperatures at the PSDRS commencing time were set up by using the transient analysis of the COMMIX code, and the surface emissivity of PSDRS was investigated and its permitted variation rage was set up. From this study, overall sensitivity characteristics of the PARS2 code were investigated and the results of the sensitivity analyses can be used in the design of the RHR system of KALIMER. 14 refs., 28 figs., 2 tabs. (Author)

  6. Plutonium and uranium isotopic analysis: recent developments of the MGA++ code suite

    Energy Technology Data Exchange (ETDEWEB)

    Buckley, W; Clark, D; Parker, W E; Romine, W; Ruhter, W; Wang, T F

    1999-09-17

    The Lawrence Livermore National Laboratory develops sophisticated gamma-ray analysis codes for isotopic determinations of nuclear materials based on the principles of the MultiGroup Analysis (MGA). MGA methodology has been upgraded and expanded and is now comprised of a suite of codes known as MGA++. A graphical user interface has also been developed for viewing the data and the fitting procedure. The code suite provides plutonium and uranium isotopic analysis for data collected with high-purity germanium planar and/or coaxial detector systems. The most recent addition to the MGA++ code suite, MGAHI, analyzes Pu data using higher-energy gamma rays (200 keV and higher) and is particularly useful for Pu samples that are enclosed in thick-walled containers. Additionally, the code suite can perform isotopic analysis of uranium spectra collected with cadmium-zinc-telluride (CZT) detectors. We are currently developing new codes with will integrate into the MGA++ suite. These will include Pu isotopic analysis capabilities for data collected with CZT detectors, and U isotopic analysis with high-purity germanium detectors, which utilizes only higher energy gamma rays. Future development of MGA++ will include a capability for isotopic analyses on mixtures of Pu and U.

  7. Computer code system for the R and D of nuclear fuel cycle with fast reactor. 5. Development and application of reactor analysis code system

    Energy Technology Data Exchange (ETDEWEB)

    Yokoyama, Kenji; Hazama, Taira; Chiba, Go; Ohki, Shigeo; Ishikawa, Makoto [Japan Nuclear Cycle Development Inst., Oarai, Ibaraki (Japan). Oarai Engineering Center

    2002-12-01

    In the core design of fast reactors (FRs), it is very important to improve the prediction accuracy of the nuclear characteristics for both reducing cost and ensuring reliability of FR plants. A nuclear reactor analysis code system for FRs has been developed by the Japan Nuclear Cycle Development Institute (JNC). This paper describes the outline of the calculation models and methods in the system consisting of several analysis codes, such as the cell calculation code CASUP, the core calculation code TRITAC and the sensitivity analysis code SAGEP. Some examples of verification results and improvement of the design accuracy are also introduced based on the measurement data from critical assemblies, e.g, the JUPITER experiment (USA/Japan), FCA (Japan), MASURCA (France), and BFS (Russia). Furthermore, application fields and future plans, such as the development of new generation nuclear constants and applications to MA{center_dot}FP transmutation, are described. (author)

  8. Hierarchical Symbolic Analysis of Large Analog Circuits with Totally Coded Method

    Institute of Scientific and Technical Information of China (English)

    XU Jing-bo

    2006-01-01

    Symbolic analysis has many applications in the design of analog circuits. Existing approaches rely on two forms of symbolic-expression representation: expanded sum-ofproduct form and arbitrarily nested form. Expanded form suffers the problem that the number of product terms grows exponentially with the size of a circuit. Nested form is neither canonical nor amenable to symbolic manipulation. In this paper, we present a new approach to exact and canonical symbolic analysis by exploiting the sparsity and sharing of product terms. This algorithm, called totally coded method (TCM), consists of representing the symbolic determinant of a circuit matrix by code series and performing symbolic analysis by code manipulation. We describe an efficient code-ordering heuristic and prove that it is optimum for ladder-structured circuits. For practical analog circuits, TCM not only covers all advantages of the algorithm via determinant decision diagrams (DDD) but is more simple and efficient than DDD method.

  9. On the Analysis and Evaluation of Direct Containement Heating with the Multidimensional Multiphase Flow Code MC3D

    Directory of Open Access Journals (Sweden)

    Tanguy Janin

    2010-01-01

    Full Text Available In the course of a postulated severe accident in an NPP, Direct Containment Heating (DCH may occur after an eventual failure of the vessel. DCH is related to dynamical, thermal, and chemical phenomena involved by the eventual fine fragmentation and dispersal of the corium melt out of the vessel pit. It may threaten the integrity of the containment by pressurization of its atmosphere. Several simplified modellings have been proposed in the past but they require a very strong fitting which renders any extrapolation regarding geometry, material, and scales rather doubtful. With the development of multidimensional multiphase flow computer codes, it is now possible to investigate the phenomenon numerically with more details. We present an analysis of the potential of the MC3D code to support the analysis of this phenomenon, restricting our discussion to the dynamical processes. The analysis is applied to the case of French 1300 MWe PWR reactors for which we derive a correlation for the corium dispersal rate for application in a Probabilistic Safety Analysis (PSA level 2 study.

  10. Finite Dynamic Elements and Modal Analysis

    Directory of Open Access Journals (Sweden)

    N.J. Fergusson

    1993-01-01

    Full Text Available A general modal analysis scheme is derived for forced response that makes use of high accuracy modes computed by the dynamic element method. The new procedure differs from the usual modal analysis in that the modes are obtained from a power series expansion for the dynamic stiffness matrix that includes an extra dynamic correction term in addition to the static stiffness matrix and the consistent mass matrix based on static displacement. A cantilevered beam example is used to demonstrate the relative accuracies of the dynamic element and the traditional finite element methods.

  11. Identification and Analysis of Critical Gaps in Nuclear Fuel Cycle Codes Required by the SINEMA Program

    Energy Technology Data Exchange (ETDEWEB)

    Adrian Miron; Joshua Valentine; John Christenson; Majd Hawwari; Santosh Bhatt; Mary Lou Dunzik-Gougar: Michael Lineberry

    2009-10-01

    The current state of the art in nuclear fuel cycle (NFC) modeling is an eclectic mixture of codes with various levels of applicability, flexibility, and availability. In support of the advanced fuel cycle systems analyses, especially those by the Advanced Fuel Cycle Initiative (AFCI), Unviery of Cincinnati in collaboration with Idaho State University carried out a detailed review of the existing codes describing various aspects of the nuclear fuel cycle and identified the research and development needs required for a comprehensive model of the global nuclear energy infrastructure and the associated nuclear fuel cycles. Relevant information obtained on the NFC codes was compiled into a relational database that allows easy access to various codes' properties. Additionally, the research analyzed the gaps in the NFC computer codes with respect to their potential integration into programs that perform comprehensive NFC analysis.

  12. Non binary LDPC codes over the binary erasure channel: density evolution analysis

    CERN Document Server

    Savin, Valentin

    2008-01-01

    In this paper we present a thorough analysis of non binary LDPC codes over the binary erasure channel. First, the decoding of non binary LDPC codes is investigated. The proposed algorithm performs on-the-fly decoding, i.e. it starts decoding as soon as the first symbols are received, which generalizes the erasure decoding of binary LDPC codes. Next, we evaluate the asymptotical performance of ensembles of non binary LDPC codes, by using the density evolution method. Density evolution equations are derived by taking into consideration both the irregularity of the bipartite graph and the probability distribution of the graph edge labels. Finally, infinite-length performance of some ensembles of non binary LDPC codes for different edge label distributions are shown.

  13. Application of a Two-dimensional Unsteady Viscous Analysis Code to a Supersonic Throughflow Fan Stage

    Science.gov (United States)

    Steinke, Ronald J.

    1989-01-01

    The Rai ROTOR1 code for two-dimensional, unsteady viscous flow analysis was applied to a supersonic throughflow fan stage design. The axial Mach number for this fan design increases from 2.0 at the inlet to 2.9 at the outlet. The Rai code uses overlapped O- and H-grids that are appropriately packed. The Rai code was run on a Cray XMP computer; then data postprocessing and graphics were performed to obtain detailed insight into the stage flow. The large rotor wakes uniformly traversed the rotor-stator interface and dispersed as they passed through the stator passage. Only weak blade shock losses were computerd, which supports the design goals. High viscous effects caused large blade wakes and a low fan efficiency. Rai code flow predictions were essentially steady for the rotor, and they compared well with Chima rotor viscous code predictions based on a C-grid of similar density.

  14. Development of statistical analysis code for meteorological data (W-View)

    Energy Technology Data Exchange (ETDEWEB)

    Tachibana, Haruo; Sekita, Tsutomu; Yamaguchi, Takenori [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment

    2003-03-01

    A computer code (W-View: Weather View) was developed to analyze the meteorological data statistically based on 'the guideline of meteorological statistics for the safety analysis of nuclear power reactor' (Nuclear Safety Commission on January 28, 1982; revised on March 29, 2001). The code gives statistical meteorological data to assess the public dose in case of normal operation and severe accident to get the license of nuclear reactor operation. This code was revised from the original code used in a large office computer code to enable a personal computer user to analyze the meteorological data simply and conveniently and to make the statistical data tables and figures of meteorology. (author)

  15. Comparative Criticality Analysis of Two Monte Carlo Codes on Centrifugal Atomizer: MCNPS and SCALE

    Energy Technology Data Exchange (ETDEWEB)

    Kang, H-S; Jang, M-S; Kim, S-R [NESS, Daejeon (Korea, Republic of); Park, J-M; Kim, K-N [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)

    2015-10-15

    There are two well-known Monte Carlo codes for criticality analysis, MCNP5 and SCALE. MCNP5 is a general-purpose Monte Carlo N-Particle code that can be used for neutron, photon, electron or coupled neutron / photon / electron transport, including the capability to calculate eigenvalues for critical system as a main analysis code. SCALE provides a comprehensive, verified and validated, user-friendly tool set for criticality safety, reactor physics, radiation shielding, radioactive source term characterization, and sensitivity and uncertainty analysis. SCALE was conceived and funded by US NRC to perform standardized computer analysis for licensing evaluation and is used widely in the world. We performed a validation test of MCNP5 and a comparative analysis of Monte Carlo codes, MCNP5 and SCALE, in terms of the critical analysis of centrifugal atomizer. In the criticality analysis using MCNP5 code, we obtained the statistically reliable results by using a large number of source histories per cycle and performing of uncertainty analysis.

  16. Code Pulse: Software Assurance (SWA) Visual Analytics for Dynamic Analysis of Code

    Science.gov (United States)

    2014-09-01

    ZAP ) DAST tool to measure its testing coverage, and several test...source world, OWASP ZAP . Distinctive within the penetration testing tool space however is the scarci- ty of visualization tools, particularly as they...Transmission Control Protocol UI User Interface ZAP Zed Attack Proxy Approved for Public Release; Distribution Unlimited. 37

  17. A HERMENEUTIC ANALYSIS OF THE NEW CIVIL PROCEDURE CODE ADVANCES

    Directory of Open Access Journals (Sweden)

    Lenio Luiz Streck

    2016-07-01

    Full Text Available I've never been unwillingly with CPC/15. Everything I wrote to criticize the procedural instrumentalism and its side effects, present until the Rapporteur, Deputy Paulo Teixeira, assumed courageously the thesis that there was something more to be treated in the Project. This plus concerned the philosophical paradigms and the need to control the judicial decisions. Anyway, I believe that some guiding principles of the new code can be drawn from the project and its complexity, such as the need to maintain the consistency and integrity of the case law (including the precedents, the prohibition of the free convincing, which implies minor role and the need to adopt the intersubjectivism paradigm, that is, the subjectivity of the judge should be suspended and controlled by structuring intersubjectivity of law. This is the holding of the new "system". Without understanding it, we run the risk of making a reverse revolution. Small-gnosiological reasoning still seated in objectivist and subjectivist paradigm (or its voluntarist vulgatas can quickly cause the downfall of a good idea.

  18. Analysis the Performance of Coded WSK-DWDM Transmission System

    Directory of Open Access Journals (Sweden)

    Bobby Barua

    2012-12-01

    Full Text Available Dense Wavelength Division Multiplexing (DWDM is the system with more than eight active wavelengths per fiber. Again high data rates as well as long spans between amplifiers in a chain require high optical power per channel to satisfy the signal to noise ratio (SNR requirements. So the DWDM systems with long repeater-less spans, the simultaneous requirements of high launched power and low dispersion fibers lead to the generation of new waves by four-wave mixing (FWM, which degrades the performance of a multi-channel transmission system. Several methods have been proposed to mitigate the effect of FWM crosstalk. All these works are performed considering only binary WSK scheme. Although M-ary WSK (M>2 schemes have higher spectral efficiency than binary WSK system. Again, the BER performances for M-ary WDM system are not satisfactory with the effect of FWM. Therefore, in this paper we include the effect of FWM on the performance of an M-ary WDM system and try to mitigate the effect by employing the energy efficient convolution code in a normal dispersive fiber as well as in a dispersion shifted fiber (DSF.

  19. Transient analysis techniques in performing impact and crash dynamic studies

    Science.gov (United States)

    Pifko, A. B.; Winter, R.

    1989-01-01

    Because of the emphasis being placed on crashworthiness as a design requirement, increasing demands are being made by various organizations to analyze a wide range of complex structures that must perform safely when subjected to severe impact loads, such as those generated in a crash event. The ultimate goal of crashworthiness design and analysis is to produce vehicles with the ability to reduce the dynamic forces experienced by the occupants to specified levels, while maintaining a survivable envelope around them during a specified crash event. DYCAST is a nonlinear structural dynamic finite element computer code that started from the plans systems of a finite element program for static nonlinear structural analysis. The essential features of DYCAST are outlined.

  20. Passivhaus: indoor comfort and energy dynamic analysis.

    Science.gov (United States)

    Guida, Antonella; Pagliuca, Antonello; Cardinale, Nicola; Rospi, Gianluca

    2013-04-01

    The research aims to verify the energy performance as well as the indoor comfort of an energy class A+ building, built so that the sum of the heat passive contributions of solar radiation, transmitted through the windows, and the heat generated inside the building, are adeguate to compensate for the envelope loss during the cold season. The building, located in Emilia Romagna (Italy), was built using a wooden structure, an envelope realized using a pinewood sandwich panels (transmittance U = 0.250 W/m2K) and, inside, a wool flax insulation layer and thermal window frame with low-emissivity glass (U = 0524 W/m2K). The building design and construction process has followed the guidelines set by "CasaClima". The building has been modeled in the code of dynamic calculation "Energy Plus" by the Design Builder application and divided it into homogenous thermal zones, characterized by winter indoor temperature set at 20 ° (+ / - 1 °) and summer indoor temperature set at 26 ° (+ / - 1 °). It has modeled: the envelope, as described above, the "free" heat contributions, the air conditioning system, the Mechanical Ventilation system as well as home automation solutions. The air conditioning system is an heat pump, able to guarantee an optimization of energy consumption (in fact, it uses the "free" heat offered by the external environment for conditioning indoor environment). As regards the air recirculation system, it has been used a mechanical ventilation system with internal heat cross-flow exchanger, with an efficiency equal to 50%. The domotic solutions, instead, regard a system for the control of windows external screening using reeds, adjustable as a function of incident solar radiation and a lighting management system adjusted automatically using a dimmer. A so realized building meets the requirement imposed from Italian standard UNI/TS 11300 1, UNI/TS 11300 2 and UNI/TS 11300 3. The analysis was performed according to two different configurations: in "spontaneous

  1. Python for hydrological modeling: interfacing C code with ctypes, dynamic typing and introspection

    Science.gov (United States)

    Bogaart, P. W.

    2008-12-01

    The Python programming language has several features that make it an ideal front-end language for user-friendly numerical modelling of hydrological systems. In this presentation we will demonstrate this with a comprehensive hillslope hydrological modelling framework, where the following features are highlighted: •[Extensibility:] Often a low-level langage like C is better suited for the core functionality of numerical models, for instance because of the inherent higher computational speed, or the availability of specialized algorithms, like the well-known Numerical Recipes or the open source GNU Scientific Library. The now standard Python "ctypes" enables easy manipulation of the C functions and datastructures, provided the C code is compiled into a shared library. High-level Python wrapper functions or classes are easily constructed. •[Flexibility and introspection:] Python is strong but dynamically typed, meaning that variables can change type. This enables the construction of highly flexible functions that operate on a single model parameter, or a list of parameters, or a function that yields parameter values. Related to this, is the capacity of Python to inspect its own types and variables during runtime. So, depending on the actual type of a function argument (scalar, list, function) different actions are taken. Python classes are highly flexibly in the sense that member fields can be added to them during runtime. Python's introspection capacities enables finding out which member fields are actually present. This enables the construction of 'smart' functions that probe an object for the presence or absence of specified members (using their name), and then taking action. On example that will be worked out is a flexible parameter optimizer that takes a reference to a model object, and a list of a parameter names that are to be optimized.

  2. Fuel burnup analysis for IRIS reactor using MCNPX and WIMS-D5 codes

    Science.gov (United States)

    Amin, E. A.; Bashter, I. I.; Hassan, Nabil M.; Mustafa, S. S.

    2017-02-01

    International Reactor Innovative and Secure (IRIS) reactor is a compact power reactor designed with especial features. It contains Integral Fuel Burnable Absorber (IFBA). The core is heterogeneous both axially and radially. This work provides the full core burn up analysis for IRIS reactor using MCNPX and WIMDS-D5 codes. Criticality calculations, radial and axial power distributions and nuclear peaking factor at the different stages of burnup were studied. Effective multiplication factor values for the core were estimated by coupling MCNPX code with WIMS-D5 code and compared with SAS2H/KENO-V code values at different stages of burnup. The two calculation codes show good agreement and correlation. The values of radial and axial powers for the full core were also compared with published results given by SAS2H/KENO-V code (at the beginning and end of reactor operation). The behavior of both radial and axial power distribution is quiet similar to the other data published by SAS2H/KENO-V code. The peaking factor values estimated in the present work are close to its values calculated by SAS2H/KENO-V code.

  3. A study of transonic aerodynamic analysis methods for use with a hypersonic aircraft synthesis code

    Science.gov (United States)

    Sandlin, Doral R.; Davis, Paul Christopher

    1992-01-01

    A means of performing routine transonic lift, drag, and moment analyses on hypersonic all-body and wing-body configurations were studied. The analysis method is to be used in conjunction with the Hypersonic Vehicle Optimization Code (HAVOC). A review of existing techniques is presented, after which three methods, chosen to represent a spectrum of capabilities, are tested and the results are compared with experimental data. The three methods consist of a wave drag code, a full potential code, and a Navier-Stokes code. The wave drag code, representing the empirical approach, has very fast CPU times, but very limited and sporadic results. The full potential code provides results which compare favorably to the wind tunnel data, but with a dramatic increase in computational time. Even more extreme is the Navier-Stokes code, which provides the most favorable and complete results, but with a very large turnaround time. The full potential code, TRANAIR, is used for additional analyses, because of the superior results it can provide over empirical and semi-empirical methods, and because of its automated grid generation. TRANAIR analyses include an all body hypersonic cruise configuration and an oblique flying wing supersonic transport.

  4. Verification of Dinamika-5 code on experimental data of water level behaviour in PGV-440 under dynamic conditions

    Energy Technology Data Exchange (ETDEWEB)

    Beljaev, Y.V.; Zaitsev, S.I.; Tarankov, G.A. [OKB Gidropress (Russian Federation)

    1995-12-31

    Comparison of the results of calculational analysis with experimental data on water level behaviour in horizontal steam generator (PGV-440) under the conditions with cessation of feedwater supply is presented in the report. Calculational analysis is performed using DIMANIKA-5 code, experimental data are obtained at Kola NPP-4. (orig.). 2 refs.

  5. Semi-implicit scheme for treating radiation under M1 closure in general relativistic conservative fluid dynamics codes

    CERN Document Server

    Sdowski, Aleksander; Tchekhovskoy, Alexander; Zhu, Yucong

    2012-01-01

    A numerical scheme is described for including radiation in multi-dimensional general-relativistic conservative fluid dynamics codes. In this method, a covariant form of the M1 closure scheme is used to close the radiation moments, and the radiative source terms are treated semi-implicitly in order to handle both optically thin and optically thick regimes. The scheme has been implemented in a conservative general relativistic radiation hydrodynamics code KORAL. The robustness of the code is demonstrated on a number of test problems, including radiative relativistic shock tubes, static radiation pressure supported atmosphere, shadows, beams of light in curved spacetime, and radiative Bondi accretion. The advantages of M1 closure relative to other approaches such as Eddington closure and flux-limited diffusion are discussed, and its limitations are also highlighted.

  6. Dynamic Blowout Risk Analysis Using Loss Functions.

    Science.gov (United States)

    Abimbola, Majeed; Khan, Faisal

    2017-08-11

    Most risk analysis approaches are static; failing to capture evolving conditions. Blowout, the most feared accident during a drilling operation, is a complex and dynamic event. The traditional risk analysis methods are useful in the early design stage of drilling operation while falling short during evolving operational decision making. A new dynamic risk analysis approach is presented to capture evolving situations through dynamic probability and consequence models. The dynamic consequence models, the focus of this study, are developed in terms of loss functions. These models are subsequently integrated with the probability to estimate operational risk, providing a real-time risk analysis. The real-time evolving situation is considered dependent on the changing bottom-hole pressure as drilling progresses. The application of the methodology and models are demonstrated with a case study of an offshore drilling operation evolving to a blowout. © 2017 Society for Risk Analysis.

  7. NASCRAC - A computer code for fracture mechanics analysis of crack growth

    Science.gov (United States)

    Harris, D. O.; Eason, E. D.; Thomas, J. M.; Bianca, C. J.; Salter, L. D.

    1987-01-01

    NASCRAC - a computer code for fracture mechanics analysis of crack growth - is described in this paper. The need for such a code is increasing as requirements grow for high reliability and low weight in aerospace components. The code is comprehensive and versatile, as well as user friendly. The major purpose of the code is calculation of fatigue, corrosion fatigue, or stress corrosion crack growth, and a variety of crack growth relations can be selected by the user. Additionally, crack retardation models are included. A very wide variety of stress intensity factor solutions are contained in the code, and extensive use is made of influence functions. This allows complex stress gradients in three-dimensional crack problems to be treated easily and economically. In cases where previous stress intensity factor solutions are not adequate, new influence functions can be calculated by the code. Additional features include incorporation of J-integral solutions from the literature and a capability for estimating elastic-plastic stress redistribution from the results of a corresponding elastic analysis. An example problem is presented which shows typical outputs from the code.

  8. A Multi-Code Analysis Toolkit for Astrophysical Simulation Data

    OpenAIRE

    Turk, Matthew J.; Smith., Britton D.; Oishi, Jeffrey S.; Skory, Stephen; Skillman, Samuel W.; Abel, Tom; Norman, Michael L.

    2010-01-01

    The analysis of complex multiphysics astrophysical simulations presents a unique and rapidly growing set of challenges: reproducibility, parallelization, and vast increases in data size and complexity chief among them. In order to meet these challenges, and in order to open up new avenues for collaboration between users of multiple simulation platforms, we present yt (available at http://yt.enzotools.org/), an open source, community-developed astrophysical analysis and visualization toolkit. ...

  9. TPASS: a gamma-ray spectrum analysis and isotope identification computer code

    Energy Technology Data Exchange (ETDEWEB)

    Dickens, J.K.

    1981-03-01

    The gamma-ray spectral data-reduction and analysis computer code TPASS is described. This computer code is used to analyze complex Ge(Li) gamma-ray spectra to obtain peak areas corrected for detector efficiencies, from which are determined gamma-ray yields. These yields are compared with an isotope gamma-ray data file to determine the contributions to the observed spectrum from decay of specific radionuclides. A complete FORTRAN listing of the code and a complex test case are given.

  10. Dynamics of genetic variation at gliadin-coding loci in bread wheat cultivars developed in small grains research center (Kragujevac during last 35 years

    Directory of Open Access Journals (Sweden)

    Novosljska-Dragovič Aleksandra

    2005-01-01

    Full Text Available Multiple alleles of gliadin-coding loci are well-known genetic markers of common wheat genotypes. Based on analysis of gliadin patterns in common wheat cultivars developed at the Small Grains Research Center in Kragujevac dynamics of genetic variability at gliadin-coding loci has been surveyed for the period of 35 years. It was shown that long-term breeding of the wheat cultivars involved gradual replacement of ancient alleles for those widely spread in some regions in the world, which belong to well-known cultivars-donor of some important traits. Developing cultivars whose pedigree involved much new foreign genetic material has increased genetic diversity as well as has changed frequency of alleles of gliadin-coding loci. So we can conclude that the genetic profile of modern Serbian cultivars has changed considerably. Genetic formula of gliadin was made for each the cultivar studied. The most frequent alleles of gliadin-coding loci among modern cultivars should be of great interest of breeders because these alleles are probably linked with genes that confer advantage to their carriers at present.

  11. RSA Power Analysis Obfuscation: A Dynamic FPGA Architecture

    Science.gov (United States)

    2012-03-01

    research provides a VHDL coded dynamic architecture for synthesization on a Xilinx Virtex-5 FPGA. This architecture provides two-way communication...Component Under Test (CUT) is the dynamic RSA implementation. This dynamic hardware is synthesized from VHDL onto a Xilinx Virtex-5 FPGA. The built in...The hardware platform used for this research is a the Xil- inx Virtex-5 FX FPGA. VHDL code is synthesized using the Xilinx design suite and downloaded

  12. Development of 3-D Flow Analysis Code for Fuel Assembly using Unstructured Grid System

    Energy Technology Data Exchange (ETDEWEB)

    Myong, Hyon Kook; Kim, Jong Eun; Ahn, Jong Ki; Yang, Seung Yong [Kookmin Univ., Seoul (Korea, Republic of)

    2007-03-15

    The flow through a nuclear rod bundle with mixing vanes are very complex and required a suitable turbulence model to be predicted accurately. Final objective of this study is to develop a CFD code for fluid flow and heat transfer analysis in a nuclear fuel assembly using unstructured grid system. In order to develop a CFD code for fluid flow and heat transfer analysis in a nuclear fuel assembly using unstructured grid system, the following researches are made: - Development of numerical algorithm for CFD code's solver - Grid and geometric connectivity data - Development of software(PowerCFD code) for fluid flow and heat transfer analysis in a nuclear fuel assembly using unstructured grid system - Modulation of software(PowerCFD code) - Development of turbulence model - Development of analysis module of RANS/LES hybrid models - Analysis of turbulent flow and heat transfer - Basic study on LES analysis - Development of main frame on pre/post processors based on GUI - Algorithm for fully-developed flow.

  13. Introducing fluid dynamics using dimensional analysis

    DEFF Research Database (Denmark)

    Jensen, Jens Højgaard

    2013-01-01

    Many aspects of fluid dynamics can be introduced using dimensional analysis, combined with some basic physical principles. This approach is concise and allows exploration of both the laminar and turbulent limits—including important phenomena that are not normally dealt with when fluid dynamics...

  14. Nonlinear coupled dynamics analysis of a truss spar platform

    Science.gov (United States)

    Li, Cheng-xi; Zhang, Jun

    2016-12-01

    Accurate prediction of the offshore structure motion response and associate mooring line tension is important in both technical applications and scientific research. In our study, a truss spar platform, operated in Gulf of Mexico, is numerically simulated and analyzed by an in-house numerical code `COUPLE'. Both the platform motion responses and associated mooring line tension are calculated and investigated through a time domain nonlinear coupled dynamic analysis. Satisfactory agreement between the simulation and corresponding field measurements is in general reached, indicating that the numerical code can be used to conduct the time-domain analysis of a truss spar interacting with its mooring and riser system. Based on the comparison between linear and nonlinear results, the relative importance of nonlinearity in predicting the platform motion response and mooring line tensions is assessed and presented. Through the coupled and quasi-static analysis, the importance of the dynamic coupling effect between the platform hull and the mooring/riser system in predicting the mooring line tension and platform motions is quantified. These results may provide essential information pertaining to facilitate the numerical simulation and design of the large scale offshore structures.

  15. Code interoperability and standard data formats in quantum chemistry and quantum dynamics: The Q5/D5Cost data model.

    Science.gov (United States)

    Rossi, Elda; Evangelisti, Stefano; Laganà, Antonio; Monari, Antonio; Rampino, Sergio; Verdicchio, Marco; Baldridge, Kim K; Bendazzoli, Gian Luigi; Borini, Stefano; Cimiraglia, Renzo; Angeli, Celestino; Kallay, Peter; Lüthi, Hans P; Ruud, Kenneth; Sanchez-Marin, José; Scemama, Anthony; Szalay, Peter G; Tajti, Attila

    2014-03-30

    Code interoperability and the search for domain-specific standard data formats represent critical issues in many areas of computational science. The advent of novel computing infrastructures such as computational grids and clouds make these issues even more urgent. The design and implementation of a common data format for quantum chemistry (QC) and quantum dynamics (QD) computer programs is discussed with reference to the research performed in the course of two Collaboration in Science and Technology Actions. The specific data models adopted, Q5Cost and D5Cost, are shown to work for a number of interoperating codes, regardless of the type and amount of information (small or large datasets) to be exchanged. The codes are either interfaced directly, or transfer data by means of wrappers; both types of data exchange are supported by the Q5/D5Cost library. Further, the exchange of data between QC and QD codes is addressed. As a proof of concept, the H + H2 reaction is discussed. The proposed scheme is shown to provide an excellent basis for cooperative code development, even across domain boundaries. Moreover, the scheme presented is found to be useful also as a production tool in the grid distributed computing environment.

  16. Automated face analysis by feature point tracking has high concurrent validity with manual FACS coding.

    Science.gov (United States)

    Cohn, J F; Zlochower, A J; Lien, J; Kanade, T

    1999-01-01

    The face is a rich source of information about human behavior. Available methods for coding facial displays, however, are human-observer dependent, labor intensive, and difficult to standardize. To enable rigorous and efficient quantitative measurement of facial displays, we have developed an automated method of facial display analysis. In this report, we compare the results with this automated system with those of manual FACS (Facial Action Coding System, Ekman & Friesen, 1978a) coding. One hundred university students were videotaped while performing a series of facial displays. The image sequences were coded from videotape by certified FACS coders. Fifteen action units and action unit combinations that occurred a minimum of 25 times were selected for automated analysis. Facial features were automatically tracked in digitized image sequences using a hierarchical algorithm for estimating optical flow. The measurements were normalized for variation in position, orientation, and scale. The image sequences were randomly divided into a training set and a cross-validation set, and discriminant function analyses were conducted on the feature point measurements. In the training set, average agreement with manual FACS coding was 92% or higher for action units in the brow, eye, and mouth regions. In the cross-validation set, average agreement was 91%, 88%, and 81% for action units in the brow, eye, and mouth regions, respectively. Automated face analysis by feature point tracking demonstrated high concurrent validity with manual FACS coding.

  17. Studies on the liquid fluoride thorium reactor: Comparative neutronics analysis of MCNP6 code with SRAC95 reactor analysis code based on FUJI-U3-(0)

    Energy Technology Data Exchange (ETDEWEB)

    Jaradat, S.Q., E-mail: sqjxv3@mst.edu; Alajo, A.B., E-mail: alajoa@mst.edu

    2017-04-01

    Highlights: • The verification for FUJI-U3-(0)—a molten salt reactor—was performed. • The MCNP6 was used to study the reactor physics characteristics for FUJI-U3 type. • The results from the MCNP6 were comparable with the ones obtained from literature. - Abstract: The verification for FUJI-U3-(0)—a molten salt reactor—was performed. The reactor used LiF-BeF2-ThF4-UF4 as the mixed liquid fuel salt, and the core was graphite moderated. The MCNP6 code was used to study the reactor physics characteristics for the FUJI-U3-(0) reactor. Results for reactor physics characteristic of the FUJI-U3-(0) exist in literature, which were used as reference. The reference results were obtained using SRAC95 (a reactor analysis code) coupled with ORIGEN2 (a depletion code). Some modifications were made in the reconstruction of the FUJI-U3-(0) reactor in MCNP due to unavailability of more detailed description of the reactor core. The assumptions resulted in two representative models of the reactor. The results from the MCNP6 models were compared with the reference results obtained from literature. The results were comparable with each other, but with some notable differences. The differences are because of the approximations that were done on the SRAC95 model of the FUJI-U3 to simplify the simulation. Based on the results, it is concluded that MCNP6 code predicts well the overall simulation of neutronics analysis to the previous simulation works using SRAC95 code.

  18. Stability Analysis of MEMS Gyroscope Dynamic Systems

    OpenAIRE

    M. Naser-Moghadasi; S. A. Olamaei; F. Setoudeh

    2013-01-01

    In this paper, the existence of a common quadratic Lyapunov function for stability analysis of MEMS Gyroscope dynamic systems has been studied then a new method based on stochastic stability of MEMS Gyroscope system has been proposed.

  19. PARAMETRIC ANALYSIS OF THE DYNAMIC PROPERTIES OF ...

    African Journals Online (AJOL)

    static and dynamic analysis of structures [2, 3,4]. ... than by the expected complexity of their behavior. This fact has been .... The computational cost of extracting the vibration modes can be reduced by applying one of the condensation.

  20. Dynamic Wireless Power Transfer - Grid Impacts Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Markel, Tony; Meintz, Andrew; Gonder, Jeff

    2015-12-04

    This presentation discusses the current status of analysis of the electricity grid impacts of a dynamic wireless power transfer system deployed to the Atlanta region on select high traffic roadway segments.

  1. Development of the Verification and Validation Matrix for Safety Analysis Code SPACE

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Yo Han; Ha, Sang Jun; Yang, Chang Keun [Korea Electric Power Research Institute, Daejeon (Korea, Republic of)

    2009-10-15

    Korea Electric Power Research Institute (KEPRI) has been developed the safety analysis code, called as SPACE (Safety and Performance Analysis CodE for Nuclear Power Plant), for typical pressurized water reactors (PWR). Current safety analysis codes were conducted from foreign vendors, such as Westinghouse Electric Corp., ABB Combustion Engineering Inc., Kraftwerk Union, etc. Considering the conservatism and inflexibility of the foreign code systems, it is difficult to expand the application areas and analysis scopes. To overcome the mentioned problems KEPRI has launched the project to develop the native safety analysis code with Korea Power Engineering Co.(KOPEC), Korea Atomic Energy Research Inst.(KAERI), Korea Nuclear Fuel(KNF), and Korea Hydro and Nuclear Power Co.(KHNP) under the funding of Ministry of Knowledge Economy (MKE). As a result of the project, the demo-version of SPACE has been released in July 2009. As an advance preparation of the next step, KEPRI and colleagues have developed the verification and validation (V and V) matrix for SPACE. To develop the matrix, the preceding studies and experiments were reviewed. After mature consideration, the V and V matrix has been developed and the experiment plans were designed for the next step to compensate the lack of data.

  2. On parsing the neural code in the prefrontal cortex of primates using principal dynamic modes.

    Science.gov (United States)

    Marmarelis, V Z; Shin, D C; Song, D; Hampson, R E; Deadwyler, S A; Berger, T W

    2014-06-01

    Nonlinear modeling of multi-input multi-output (MIMO) neuronal systems using Principal Dynamic Modes (PDMs) provides a novel method for analyzing the functional connectivity between neuronal groups. This paper presents the PDM-based modeling methodology and initial results from actual multi-unit recordings in the prefrontal cortex of non-human primates. We used the PDMs to analyze the dynamic transformations of spike train activity from Layer 2 (input) to Layer 5 (output) of the prefrontal cortex in primates performing a Delayed-Match-to-Sample task. The PDM-based models reduce the complexity of representing large-scale neural MIMO systems that involve large numbers of neurons, and also offer the prospect of improved biological/physiological interpretation of the obtained models. PDM analysis of neuronal connectivity in this system revealed "input-output channels of communication" corresponding to specific bands of neural rhythms that quantify the relative importance of these frequency-specific PDMs across a variety of different tasks. We found that behavioral performance during the Delayed-Match-to-Sample task (correct vs. incorrect outcome) was associated with differential activation of frequency-specific PDMs in the prefrontal cortex.

  3. Dynamic Response Analysis of Motorized Spindle System

    Institute of Scientific and Technical Information of China (English)

    ZHANG Li; LUO Yi-chao; XU Juan; XIAO Ru-feng; LI Xian-hui

    2013-01-01

    As to motorized spindle system, this paper builds a simplified 3D model of spindle and bearing, performs structure modal analysis, reveals its dynamic characteristics under the free model;furthermore, modifies bearing radial stiffness and number of model, and studies the change of modal parameters. On this basis, through the harmonic response analysis of the finite element model, dy-namic response characteristic caused by imbalance of monitored spindle system and law of vibration response to different amount of unbalance is analyzed.

  4. The Analysis of Dimensionality Reduction Techniques in Cryptographic Object Code Classification

    Energy Technology Data Exchange (ETDEWEB)

    Jason L. Wright; Milos Manic

    2010-05-01

    This paper compares the application of three different dimension reduction techniques to the problem of locating cryptography in compiled object code. A simple classi?er is used to compare dimension reduction via sorted covariance, principal component analysis, and correlation-based feature subset selection. The analysis concentrates on the classi?cation accuracy as the number of dimensions is increased.

  5. Improvement in the Plutonium Parameter Files of the FRAM Isotopic Analysis Code

    Energy Technology Data Exchange (ETDEWEB)

    D. T. Vo; T. E. Sampson

    2000-09-01

    The isotopic analysis code Fixed-energy Response-function Analysis with Multiple efficiency (FRAM) employs user-editable parameter sets to analyze a broad range of sample types. This report presents new parameter files, based upon a new set of plutonium branding ratios, which give more accurate isotope results than the current parameter files that use FRAM.

  6. Installation of VENUS-2 code for the analysis of transition phase of LMR HCDA

    Energy Technology Data Exchange (ETDEWEB)

    Choi, Yong Joon; Suk, S. D.; Hahn, D. H

    2000-06-01

    VENUS-2 is a two-dimensional coupled neutronics-hydrodynamics program that calculate the dynamic behavior of an LMFBR during a prompt-critical disassembly excursion. The code, was developed by the Argonne National Laboratory(ANL) in 1972, was originally written in FORTRAN-4 computer language to run on the CDC7600 or IBM3033 computer systems. A number of modifications were done on the code to make it work in the DIGITAL Fortran system of the MS-Windows. In particular, an effort has been made to revise I/O status system and graphical plotting routines using Excel and Origin 6.0. Test runs with the sample problem show that the modified VENUS-2 code provides the results in general agreement to sample outputs described in the manual.

  7. Random Matrix Theory in molecular dynamics analysis.

    Science.gov (United States)

    Palese, Luigi Leonardo

    2015-01-01

    It is well known that, in some situations, principal component analysis (PCA) carried out on molecular dynamics data results in the appearance of cosine-shaped low index projections. Because this is reminiscent of the results obtained by performing PCA on a multidimensional Brownian dynamics, it has been suggested that short-time protein dynamics is essentially nothing more than a noisy signal. Here we use Random Matrix Theory to analyze a series of short-time molecular dynamics experiments which are specifically designed to be simulations with high cosine content. We use as a model system the protein apoCox17, a mitochondrial copper chaperone. Spectral analysis on correlation matrices allows to easily differentiate random correlations, simply deriving from the finite length of the process, from non-random signals reflecting the intrinsic system properties. Our results clearly show that protein dynamics is not really Brownian also in presence of the cosine-shaped low index projections on principal axes.

  8. Dynamic Analysis of a Military- Tracked Vehicle

    Directory of Open Access Journals (Sweden)

    V. Balamurugan

    2000-04-01

    Full Text Available 'The ride dynamic characteristics of a typical medium weight, high speed military-tracked vehicle for negotiating rough cross-counlry terrain have been studied. The vehicle is modelled using finiteelement simulation method with beam and shell elements. An eigenvalue analysis has been done to estimate natural modes ofNibration of the vehicle. The dynamic response of certain salient locations is obtained by carrying out a transient dynamic analysis using implicit Newmark beta method. A constant forwar vehicle speed and non-deformable sinusoidal terrain profile are assumed.

  9. String Analysis for Dynamic Field Access

    DEFF Research Database (Denmark)

    Madsen, Magnus; Andreasen, Esben

    2014-01-01

    In JavaScript, and scripting languages in general, dynamic field access is a commonly used feature. Unfortunately, current static analysis tools either completely ignore dynamic field access or use overly conservative approximations that lead to poor precision and scalability. We present new string...... domains to reason about dynamic field access in a static analysis tool. A key feature of the domains is that the equal, concatenate and join operations take Ο(1) time. Experimental evaluation on four common JavaScript libraries, including jQuery and Prototype, shows that traditional string domains...

  10. Sensitivity Analysis of Fire Dynamics Simulation

    DEFF Research Database (Denmark)

    Brohus, Henrik; Nielsen, Peter V.; Petersen, Arnkell J.

    2007-01-01

    equations require solution of the issues of combustion and gas radiation to mention a few. This paper performs a sensitivity analysis of a fire dynamics simulation on a benchmark case where measurement results are available for comparison. The analysis is performed using the method of Elementary Effects......In case of fire dynamics simulation requirements to reliable results are most often very high due to the severe consequences of erroneous results. At the same time it is a well known fact that fire dynamics simulation constitutes rather complex physical phenomena which apart from flow and energy...

  11. Numerical study of subcooled boiling phenomena using a component analysis code, CUPID

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Ba-Ro; Lee, Yeon-Gun [Jeju National University, Jeju (Korea, Republic of)

    2015-10-15

    In this study, a couple of subcooled boiling experiments at high- (> 10 bar) and low-pressure (near atmospheric pressure) conditions are analyzed using a three-dimensional thermal-hydraulic component code, CUPID. And then the analysis results compared with the results using MARS-KS code. Subcooled boiling experiments at high- and low pressure conditions are analyzed using a three dimensional thermal-hydraulic component code, CUPID. The predictions of the CUPID code shows good agreement with Christenses's data and Bartolomey's data obtained at high pressure conditions. Subcooled boiling is encountered in many industrial applications in the power and process industry. In nuclear reactors, under certain conditions, subcooled boiling may be encountered in the core. The movement of bubbles generated by subcooled boiling affect the heat transfer characteristics and the pressure drop of the system. Thus some experimental and analysis using safety codes works have been already performed by previous investigators. It has been reported that the existing safety analysis codes have some weaknesses in predicting subcooled boiling phenomena at low pressure conditions. Thus, it is required to improve the predictive capability of thermal-hydraulic analysis codes on subcooled boiling phenomenon at low-pressure conditions. At low pressure condition, the CUPID code generally is overestimated prediction of the void fraction. Thus, we did selected submodels in the heat partitioning model by sensitivity analysis. Selected submodels of M{sub c}ase 4 are Kocamustafaogullari and Ishii correlation model of active nucleate site density, N' and Fritz correlation model of bubble departure diameter, d{sub Bd} . And then, case 5 - 8 are reanalysis using submodels of M{sub c}ase 4. The calculated void fraction is compared the default CUPID code model to the modified CUPID code model. As a result, average void fraction error was reduced from 0.081 to 0.011 and 0.128 to 0.024, 0

  12. Secure and Practical Defense Against Code-Injection Attacks using Software Dynamic Translation

    Science.gov (United States)

    2006-06-16

    Cache inst1 inst2 … instx inst3 inst4 cmpl %eax,%ecx trampoline Code Fragment1 inst7 inst8 … trampoline Code Fragment2 Context Switch Fetch Decode...inst4 cmpl %eax,%ecx bne L4 inst5 inst6 … jmp L8 L4: inst7 inst8 … Application Text CFn CFn+1 CFn+2 CFn+3 CFn+4 CFn+5 CFn+x inst5 inst6 … trampoline

  13. Demonstration of Emulator-Based Bayesian Calibration of Safety Analysis Codes: Theory and Formulation

    Directory of Open Access Journals (Sweden)

    Joseph P. Yurko

    2015-01-01

    Full Text Available System codes for simulation of safety performance of nuclear plants may contain parameters whose values are not known very accurately. New information from tests or operating experience is incorporated into safety codes by a process known as calibration, which reduces uncertainty in the output of the code and thereby improves its support for decision-making. The work reported here implements several improvements on classic calibration techniques afforded by modern analysis techniques. The key innovation has come from development of code surrogate model (or code emulator construction and prediction algorithms. Use of a fast emulator makes the calibration processes used here with Markov Chain Monte Carlo (MCMC sampling feasible. This work uses Gaussian Process (GP based emulators, which have been used previously to emulate computer codes in the nuclear field. The present work describes the formulation of an emulator that incorporates GPs into a factor analysis-type or pattern recognition-type model. This “function factorization” Gaussian Process (FFGP model allows overcoming limitations present in standard GP emulators, thereby improving both accuracy and speed of the emulator-based calibration process. Calibration of a friction-factor example using a Method of Manufactured Solution is performed to illustrate key properties of the FFGP based process.

  14. Comparative Analysis of the Corporate Governance Codes of the Five BRICS Countries

    Directory of Open Access Journals (Sweden)

    Marcelle Colares Oliveira

    2014-12-01

    Full Text Available The objective of the study was to compare the codes of good corporate governance (CG practices adopted by the five BRICS countries (Brazil, Russia, India, China and South Africa using the set of practices recommended by the United Nations (UN as categories of analysis. The present study was exploratory (our comparison of the CG codes of the five BRICS countries in light of the UN recommended practices has no precedent in the literature. Its nature is qualitative because we performed a multiple case study focused on a specific group of countries – BRICS – and used the content analysis technic to analyze and cross data with the literature reviewed. The research is documental because it is based on a review of documents and the literature. Among the five BRICS countries, the Corporate Governance codes of Brazil, Russia and South Africa are those most closely aligned with the recommendations set forth by the UN. It was observed that although these countries are developing their codes of good corporate governance practices focused on the requirements of international investors, corporate governance and codes in these countries are at very different stages. In comparison, the low level of convergence of the Chinese and Indian codes with the UN recommendations suggests a greater concern in those countries with the local context than with international standards.

  15. Impact Analysis of Baseband Quantizer on Coding Efficiency for HDR Video

    Science.gov (United States)

    Wong, Chau-Wai; Su, Guan-Ming; Wu, Min

    2016-10-01

    Digitally acquired high dynamic range (HDR) video baseband signal can take 10 to 12 bits per color channel. It is economically important to be able to reuse the legacy 8 or 10-bit video codecs to efficiently compress the HDR video. Linear or nonlinear mapping on the intensity can be applied to the baseband signal to reduce the dynamic range before the signal is sent to the codec, and we refer to this range reduction step as a baseband quantization. We show analytically and verify using test sequences that the use of the baseband quantizer lowers the coding efficiency. Experiments show that as the baseband quantizer is strengthened by 1.6 bits, the drop of PSNR at a high bitrate is up to 1.60dB. Our result suggests that in order to achieve high coding efficiency, information reduction of videos in terms of quantization error should be introduced in the video codec instead of on the baseband signal.

  16. Development of Advanced Suite of Deterministic Codes for VHTR Physics Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Kang Seog; Cho, J. Y.; Lee, K. H. (and others)

    2007-07-15

    Advanced Suites of deterministic codes for VHTR physics analysis has been developed for detailed analysis of current and advanced reactor designs as part of a US-ROK collaborative I-NERI project. These code suites include the conventional 2-step procedure in which a few group constants are generated by a transport lattice calculation, and the reactor physics analysis is performed by a 3-dimensional diffusion calculation, and a whole core transport code that can model local heterogeneities directly at the core level. Particular modeling issues in physics analysis of the gas-cooled VHTRs were resolved, which include a double heterogeneity of the coated fuel particles, a neutron streaming in the coolant channels, a strong core-reflector interaction, and large spectrum shifts due to changes of the surrounding environment, temperature and burnup. And the geometry handling capability of the DeCART code were extended to deal with the hexagonal fuel elements of the VHTR core. The developed code suites were validated and verified by comparing the computational results with those of the Monte Carlo calculations for the benchmark problems.

  17. Decoding of responses to mixed frequency and phase coded visual stimuli using multiset canonical correlation analysis.

    Science.gov (United States)

    Suefusa, Kaori; Tanaka, Toshihisa

    2016-08-01

    Brain-computer interfacing (BCI) based on steady-state visual evoked potentials (SSVEPs) is one of the most practical BCIs because of its high recognition accuracies and little training of a user. Mixed frequency and phase coding which can implement a number of commands and achieve a high information transfer rate (ITR) has recently been gaining much attention. In order to implement mixed-coded SSVEP-BCI as a reliable interface, it is important to detect commands fast and accurately. This paper presents a novel method to recognize mixed-coded SSVEPs which achieves high performance. The method employs multiset canonical correlation analysis to obtain spatial filters which enhance SSVEP components. An experiment with a mixed-coded SSVEP-BCI was conducted to evaluate performance of the proposed method compared with the previous work. The experimental results showed that the proposed method achieved significantly higher command recognition accuracy and ITR than the state-of-the-art.

  18. Statistical mechanical analysis of a hierarchical random code ensemble in signal processing

    Energy Technology Data Exchange (ETDEWEB)

    Obuchi, Tomoyuki [Department of Earth and Space Science, Faculty of Science, Osaka University, Toyonaka 560-0043 (Japan); Takahashi, Kazutaka [Department of Physics, Tokyo Institute of Technology, Tokyo 152-8551 (Japan); Takeda, Koujin, E-mail: takeda@sp.dis.titech.ac.jp [Department of Computational Intelligence and Systems Science, Tokyo Institute of Technology, Yokohama 226-8502 (Japan)

    2011-02-25

    We study a random code ensemble with a hierarchical structure, which is closely related to the generalized random energy model with discrete energy values. Based on this correspondence, we analyze the hierarchical random code ensemble by using the replica method in two situations: lossy data compression and channel coding. For both the situations, the exponents of large deviation analysis characterizing the performance of the ensemble, the distortion rate of lossy data compression and the error exponent of channel coding in Gallager's formalism, are accessible by a generating function of the generalized random energy model. We discuss that the transitions of those exponents observed in the preceding work can be interpreted as phase transitions with respect to the replica number. We also show that the replica symmetry breaking plays an essential role in these transitions.

  19. A New Phenomenon in Saudi Females’ Code-switching: A Morphemic Analysis

    Directory of Open Access Journals (Sweden)

    Mona O. Turjoman

    2016-12-01

    Full Text Available This sociolinguistics study investigates a new phenomenon that has recently surfaced in the field of code-switching among Saudi females residing in the Western region of Saudi Arabia. This phenomenon basically combines bound Arabic pronouns, tense markers or definite article to English free morphemes or the combination of bound English affixes to Arabic morphemes. Moreover, the study examines the factors that affect this type of code-switching. The results of the study indicate that this phenomenon provides data that invalidates Poplack’s (1980 universality of the ‘Free Morpheme Constraint’. It is also concluded that the main factors that influence this type of code-switching is solidarity and group identity among other factors. Keywords: Code-switching, Saudi females, sociolinguistics, CS factors, morphemic analysis

  20. Network coding based joint signaling and dynamic bandwidth allocation scheme for inter optical network unit communication in passive optical networks

    Science.gov (United States)

    Wei, Pei; Gu, Rentao; Ji, Yuefeng

    2014-06-01

    As an innovative and promising technology, network coding has been introduced to passive optical networks (PON) in recent years to support inter optical network unit (ONU) communication, yet the signaling process and dynamic bandwidth allocation (DBA) in PON with network coding (NC-PON) still need further study. Thus, we propose a joint signaling and DBA scheme for efficiently supporting differentiated services of inter ONU communication in NC-PON. In the proposed joint scheme, the signaling process lays the foundation to fulfill network coding in PON, and it can not only avoid the potential threat to downstream security in previous schemes but also be suitable for the proposed hybrid dynamic bandwidth allocation (HDBA) scheme. In HDBA, a DBA cycle is divided into two sub-cycles for applying different coding, scheduling and bandwidth allocation strategies to differentiated classes of services. Besides, as network traffic load varies, the entire upstream transmission window for all REPORT messages slides accordingly, leaving the transmission time of one or two sub-cycles to overlap with the bandwidth allocation calculation time at the optical line terminal (the OLT), so that the upstream idle time can be efficiently eliminated. Performance evaluation results validate that compared with the existing two DBA algorithms deployed in NC-PON, HDBA demonstrates the best quality of service (QoS) support in terms of delay for all classes of services, especially guarantees the end-to-end delay bound of high class services. Specifically, HDBA can eliminate queuing delay and scheduling delay of high class services, reduce those of lower class services by at least 20%, and reduce the average end-to-end delay of all services over 50%. Moreover, HDBA also achieves the maximum delay fairness between coded and uncoded lower class services, and medium delay fairness for high class services.

  1. Development and application of a deflagration pressure analysis code for high level waste processing

    Energy Technology Data Exchange (ETDEWEB)

    Hensel, S.J.; Thomas, J.K.

    1994-06-01

    The Deflagration Pressure Analysis Code (DPAC) was developed primarily to evaluate peak pressures for deflagrations in radioactive waste storage and process facilities at the Savannah River Site (SRS). Deflagrations in these facilities are generally considered to be incredible events, but it was judged prudent to develop modeling capabilities in order to facilitate risk estimates. DPAC is essentially an engineering analysis tool, as opposed to a detailed thermal hydraulics code. It accounts for mass loss via venting, energy dissipation by radiative heat transfer, and gas PdV work. Volume increases due to vessel deformation can also be included using pressure-volume data from a structural analysis of the enclosure. This paper presents an overview of the code, benchmarking, and applications at SRS.

  2. Development of a code system DEURACS for theoretical analysis and prediction of deuteron-induced reactions

    Directory of Open Access Journals (Sweden)

    Nakayama Shinsuke

    2017-01-01

    Full Text Available We have developed an integrated code system dedicated for theoretical analysis and prediction of deuteron-induced reactions, which is called DEUteron-induced Reaction Analysis Code System (DEURACS. DEURACS consists of several calculation codes based on theoretical models to describe respective reaction mechanisms and it was successfully applied to (d,xp and (d,xn reactions. In the present work, the analysis of (d,xn reactions is extended to higher incident energy up to nearly 100 MeV and also DEURACS is applied to (d,xd reactions at 80 and 100 MeV. The DEURACS calculations reproduce the experimental double-differential cross sections for the (d,xn and (d,xd reactions well.

  3. GRABGAM: A Gamma Analysis Code for Ultra-Low-Level HPGe SPECTRA

    Energy Technology Data Exchange (ETDEWEB)

    Winn, W.G.

    1999-07-28

    The GRABGAM code has been developed for analysis of ultra-low-level HPGe gamma spectra. The code employs three different size filters for the peak search, where the largest filter provides best sensitivity for identifying low-level peaks and the smallest filter has the best resolution for distinguishing peaks within a multiplet. GRABGAM basically generates an integral probability F-function for each singlet or multiplet peak analysis, bypassing the usual peak fitting analysis for a differential f-function probability model. Because F is defined by the peak data, statistical limitations for peak fitting are avoided; however, the F-function does provide generic values for peak centroid, full width at half maximum, and tail that are consistent with a Gaussian formalism. GRABGAM has successfully analyzed over 10,000 customer samples, and it interfaces with a variety of supplementary codes for deriving detector efficiencies, backgrounds, and quality checks.

  4. Development of a code system DEURACS for theoretical analysis and prediction of deuteron-induced reactions

    Science.gov (United States)

    Nakayama, Shinsuke; Kouno, Hiroshi; Watanabe, Yukinobu; Iwamoto, Osamu; Ye, Tao; Ogata, Kazuyuki

    2017-09-01

    We have developed an integrated code system dedicated for theoretical analysis and prediction of deuteron-induced reactions, which is called DEUteron-induced Reaction Analysis Code System (DEURACS). DEURACS consists of several calculation codes based on theoretical models to describe respective reaction mechanisms and it was successfully applied to (d,xp) and (d,xn) reactions. In the present work, the analysis of (d,xn) reactions is extended to higher incident energy up to nearly 100 MeV and also DEURACS is applied to (d,xd) reactions at 80 and 100 MeV. The DEURACS calculations reproduce the experimental double-differential cross sections for the (d,xn) and (d,xd) reactions well.

  5. Regulation of Coding and Non-coding Genes : New insights obtained through analysis of high-throughput sequencing data

    NARCIS (Netherlands)

    K. Rooijers (Koos)

    2016-01-01

    markdownabstractThe genetic code of a cell is kept in its DNA. However, a vast number of functions of a cell are carried out by proteins. Through gene expression the genetic code can be expressed and give rise to proteins. The expression of genes into proteins follows two steps: transcription of DNA

  6. Regulation of Coding and Non-coding Genes : New insights obtained through analysis of high-throughput sequencing data

    NARCIS (Netherlands)

    K. Rooijers (Koos)

    2016-01-01

    markdownabstractThe genetic code of a cell is kept in its DNA. However, a vast number of functions of a cell are carried out by proteins. Through gene expression the genetic code can be expressed and give rise to proteins. The expression of genes into proteins follows two steps: transcription of

  7. Dynamic Analysis of Structures Using Neural Networks

    Directory of Open Access Journals (Sweden)

    N. Ahmadi

    2008-01-01

    Full Text Available In the recent years, neural networks are considered as the best candidate for fast approximation with arbitrary accuracy in the time consuming problems. Dynamic analysis of structures against earthquake has the time consuming process. We employed two kinds of neural networks: Generalized Regression neural network (GR and Back-Propagation Wavenet neural network (BPW, for approximating of dynamic time history response of frame structures. GR is a traditional radial basis function neural network while BPW categorized as a wavelet neural network. In BPW, sigmoid activation functions of hidden layer neurons are substituted with wavelets and weights training are achieved using Scaled Conjugate Gradient (SCG algorithm. Comparison the results of BPW with those of GR in the dynamic analysis of eight story steel frame indicates that accuracy of the properly trained BPW was better than that of GR and therefore, BPW can be efficiently used for approximate dynamic analysis of structures.

  8. Nuclear Energy Advanced Modeling and Simulation (NEAMS) waste Integrated Performance and Safety Codes (IPSC) : gap analysis for high fidelity and performance assessment code development.

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Joon H.; Siegel, Malcolm Dean; Arguello, Jose Guadalupe, Jr.; Webb, Stephen Walter; Dewers, Thomas A.; Mariner, Paul E.; Edwards, Harold Carter; Fuller, Timothy J.; Freeze, Geoffrey A.; Jove-Colon, Carlos F.; Wang, Yifeng

    2011-03-01

    This report describes a gap analysis performed in the process of developing the Waste Integrated Performance and Safety Codes (IPSC) in support of the U.S. Department of Energy (DOE) Office of Nuclear Energy Advanced Modeling and Simulation (NEAMS) Campaign. The goal of the Waste IPSC is to develop an integrated suite of computational modeling and simulation capabilities to quantitatively assess the long-term performance of waste forms in the engineered and geologic environments of a radioactive waste storage or disposal system. The Waste IPSC will provide this simulation capability (1) for a range of disposal concepts, waste form types, engineered repository designs, and geologic settings, (2) for a range of time scales and distances, (3) with appropriate consideration of the inherent uncertainties, and (4) in accordance with rigorous verification, validation, and software quality requirements. The gap analyses documented in this report were are performed during an initial gap analysis to identify candidate codes and tools to support the development and integration of the Waste IPSC, and during follow-on activities that delved into more detailed assessments of the various codes that were acquired, studied, and tested. The current Waste IPSC strategy is to acquire and integrate the necessary Waste IPSC capabilities wherever feasible, and develop only those capabilities that cannot be acquired or suitably integrated, verified, or validated. The gap analysis indicates that significant capabilities may already exist in the existing THC codes although there is no single code able to fully account for all physical and chemical processes involved in a waste disposal system. Large gaps exist in modeling chemical processes and their couplings with other processes. The coupling of chemical processes with flow transport and mechanical deformation remains challenging. The data for extreme environments (e.g., for elevated temperature and high ionic strength media) that are

  9. Employment of CB models for non-linear dynamic analysis

    Science.gov (United States)

    Klein, M. R. M.; Deloo, P.; Fournier-Sicre, A.

    1990-01-01

    The non-linear dynamic analysis of large structures is always very time, effort and CPU consuming. Whenever possible the reduction of the size of the mathematical model involved is of main importance to speed up the computational procedures. Such reduction can be performed for the part of the structure which perform linearly. Most of the time, the classical Guyan reduction process is used. For non-linear dynamic process where the non-linearity is present at interfaces between different structures, Craig-Bampton models can provide a very rich information, and allow easy selection of the relevant modes with respect to the phenomenon driving the non-linearity. The paper presents the employment of Craig-Bampton models combined with Newmark direct integration for solving non-linear friction problems appearing at the interface between the Hubble Space Telescope and its solar arrays during in-orbit maneuvers. Theory, implementation in the FEM code ASKA, and practical results are shown.

  10. A restructuring proposal based on MELCOR for severe accident analysis code development

    Energy Technology Data Exchange (ETDEWEB)

    Park, Sun Hee; Song, Y. M.; Kim, D. H. [Korea Atomic Energy Research Institute, Taejeon (Korea)

    2000-03-01

    In order to develop a template based on existing MELCOR code, current data saving and transferring methods used in MELCOR are addressed first. Then a naming convention for the constructed module is suggested and an automatic program to convert old variables into new derived type variables has been developed. Finally, a restructured module for the SPR package has been developed to be applied to MELCOR. The current MELCOR code ensures a fixed-size storage for four different data types, and manages the variable-sized data within the storage limit by storing the data on the stacked packages. It uses pointer to identify the variables between the packages. This technique causes a difficult grasping of the meaning of the variables as well as memory waste. New features of FORTRAN90, however, make it possible to allocate the storage dynamically, and to use the user-defined data type which lead to a restructured module development for the SPR package. An efficient memory treatment and as easy understanding of the code are allowed in this developed module. The validation of the template has been done by comparing the results of the modified code with those from the existing code, and it is confirmed that the results are the same. The template for the SPR package suggested in this report hints the extension of the template to the entire code. It is expected that the template will accelerate the code domestication thanks to direct understanding of each variable and easy implementation of modified or newly developed models. 3 refs., 15 figs., 16 tabs. (Author)

  11. Experimental assessment of computer codes used for safety analysis of integral reactors

    Energy Technology Data Exchange (ETDEWEB)

    Falkov, A.A.; Kuul, V.S.; Samoilov, O.B. [OKB Mechanical Engineering, Nizhny Novgorod (Russian Federation)

    1995-09-01

    Peculiarities of integral reactor thermohydraulics in accidents are associated with presence of noncondensable gas in built-in pressurizer, absence of pumped ECCS, use of guard vessel for LOCAs localisation and passive RHRS through in-reactor HX`s. These features defined the main trends in experimental investigations and verification efforts for computer codes applied. The paper reviews briefly the performed experimental investigation of thermohydraulics of AST-500, VPBER600-type integral reactors. The characteristic of UROVEN/MB-3 code for LOCAs analysis in integral reactors and results of its verification are given. The assessment of RELAP5/mod3 applicability for accident analysis in integral reactor is presented.

  12. Analysis of isoplanatic high resolution stellar fields by the StarFinder code

    Science.gov (United States)

    Diolaiti, E.; Bendinelli, O.; Bonaccini, D.; Close, L.; Currie, D.; Parmeggiani, G.

    2000-12-01

    We describe a new code for the deep analysis of stellar fields, designed for Adaptive Optics (AO) Nyquist-sampled images with high and low Strehl ratio. The Point Spread Function (PSF) is extracted directly from the image frame, to take into account the actual structure of the instrumental response and the atmospheric effects. The code is written in IDL language and organized in the form of a self-contained widget-based application, provided with a series of tools for data visualization and analysis. A description of the method and some applications to AO data are presented.

  13. Performance Analysis of Iterative Decoding Algorithms for PEG LDPC Codes in Nakagami Fading Channels

    Directory of Open Access Journals (Sweden)

    O. Al Rasheed

    2013-11-01

    Full Text Available In this paper we give a comparative analysis of decoding algorithms of Low Density Parity Check (LDPC codes in a channel with the Nakagami distribution of the fading envelope. We consider the Progressive Edge-Growth (PEG method and Improved PEG method for the parity check matrix construction, which can be used to avoid short girths, small trapping sets and a high level of error floor. A comparative analysis of several classes of LDPC codes in various propagation conditions and decoded using different decoding algorithms is also presented.

  14. Analysis of detailed aerodynamic field measurements using results from an aeroelastic code

    Energy Technology Data Exchange (ETDEWEB)

    Schepers, J.G. [Energy Research Centre, Petten (Netherlands); Feigl, L. [Ecotecnia S. coop.c.l. (Spain); Rooij, R. van; Bruining, A. [Delft Univ. of Technology (Netherlands)

    2004-07-01

    In this article an analysis is given of aerodynamic field measurements on wind turbine blades. The analysis starts with a consistency check on the measurements, by relating the measured local aerodynamic segment forces to the overall rotor loads. It is found that the results are very consistent. Moreover, a comparison is made between measured results and results calculated from an aeroelastic code. On the basis of this comparison, the aerodynamic modelling in the aeroelastic code could be improved. This holds in particular for the modelling of 3D stall effects, not only on the lift but also on the drag, and for the modelling of tip effects (author)

  15. HSI colour-coded analysis of scattered light of single plasmonic nanoparticles

    Science.gov (United States)

    Zhou, Jun; Lei, Gang; Zheng, Lin Ling; Gao, Peng Fei; Huang, Cheng Zhi

    2016-06-01

    Single plasmonic nanoparticles (PNPs) analysis with dark-field microscopic imaging (iDFM) has attracted much attention in recent years. The ability for quantitative analysis of iDFM is critical, but cumbersome, for characterizing and analyzing the scattered light of single PNPs. Here, a simple automatic HSI colour coding method is established for coding dark-field microscopic (DFM) images of single PNPs with localized surface plasmon resonance (LSPR) scattered light, showing that hue value in the HSI system can realize accurate quantitative analysis of iDFM and providing a novel approach for quantitative chemical and biochemical imaging at the single nanoparticle level.Single plasmonic nanoparticles (PNPs) analysis with dark-field microscopic imaging (iDFM) has attracted much attention in recent years. The ability for quantitative analysis of iDFM is critical, but cumbersome, for characterizing and analyzing the scattered light of single PNPs. Here, a simple automatic HSI colour coding method is established for coding dark-field microscopic (DFM) images of single PNPs with localized surface plasmon resonance (LSPR) scattered light, showing that hue value in the HSI system can realize accurate quantitative analysis of iDFM and providing a novel approach for quantitative chemical and biochemical imaging at the single nanoparticle level. Electronic supplementary information (ESI) available: Experimental section and additional figures. See DOI: 10.1039/c6nr01089j

  16. Validation of a Subchannel Analysis Code MATRA Version 1.0

    Energy Technology Data Exchange (ETDEWEB)

    Hwang, Dae Hyun; Seo, Kyung Won; Kwon, Hyouk

    2008-10-15

    A subchannel analysis code MATRA has been developed for the thermal hydraulic analysis of SMART core. The governing equations and important models were established, and validation calculations have been performed for subchannel flow and enthalpy distributions in rod bundles under steady-state conditions. The governing equations of the MATRA were on the basis of integral balance equation of the two-phase mixture. The effects of non-homogeneous and non-equilibrium states were considered by employing the subcooled boiling model and the phasic slip model. Solution scheme and main structure of the MATRA code, as well as the difference of MATRA and COBRA-IV-I codes, were summarized. Eight different test data sets were employed for the validation of the MATRA code. The collected data consisted of single-phase subchannel flow and temperature distribution data, single-phase inlet flow maldistribution data, single-phase partial flow blockage data, and two-phase subchannel flow and enthalpy distribution data. The prediction accuracy as well as the limitation of the MATRA code was evaluated from this analysis.

  17. Sensitivity Analysis of FEAST-Metal Fuel Performance Code: Initial Results

    Energy Technology Data Exchange (ETDEWEB)

    Edelmann, Paul Guy [Los Alamos National Laboratory; Williams, Brian J. [Los Alamos National Laboratory; Unal, Cetin [Los Alamos National Laboratory; Yacout, Abdellatif [Argonne National Laboratories

    2012-06-27

    This memo documents the completion of the LANL milestone, M3FT-12LA0202041, describing methodologies and initial results using FEAST-Metal. The FEAST-Metal code calculations for this work are being conducted at LANL in support of on-going activities related to sensitivity analysis of fuel performance codes. The objective is to identify important macroscopic parameters of interest to modeling and simulation of metallic fuel performance. This report summarizes our preliminary results for the sensitivity analysis using 6 calibration datasets for metallic fuel developed at ANL for EBR-II experiments. Sensitivity ranking methodology was deployed to narrow down the selected parameters for the current study. There are approximately 84 calibration parameters in the FEAST-Metal code, of which 32 were ultimately used in Phase II of this study. Preliminary results of this sensitivity analysis led to the following ranking of FEAST models for future calibration and improvements: fuel conductivity, fission gas transport/release, fuel creep, and precipitation kinetics. More validation data is needed to validate calibrated parameter distributions for future uncertainty quantification studies with FEAST-Metal. Results of this study also served to point out some code deficiencies and possible errors, and these are being investigated in order to determine root causes and to improve upon the existing code models.

  18. Nonlinear Dynamical Analysis of Fibrillation

    Science.gov (United States)

    Kerin, John A.; Sporrer, Justin M.; Egolf, David A.

    2013-03-01

    The development of spatiotemporal chaotic behavior in heart tissue, termed fibrillation, is a devastating, life-threatening condition. The chaotic behavior of electrochemical signals, in the form of spiral waves, causes the muscles of the heart to contract in an incoherent manner, hindering the heart's ability to pump blood. We have applied the mathematical tools of nonlinear dynamics to large-scale simulations of a model of fibrillating heart tissue to uncover the dynamical modes driving this chaos. By studying the evolution of Lyapunov vectors and exponents over short times, we have found that the fibrillating tissue is sensitive to electrical perturbations only in narrow regions immediately in front of the leading edges of spiral waves, especially when these waves collide, break apart, or hit the edges of the tissue sample. Using this knowledge, we have applied small stimuli to areas of varying sensitivity. By studying the evolution of the effects of these perturbations, we have made progress toward controlling the electrochemical patterns associated with heart fibrillation. This work was supported by the U.S. National Science Foundation (DMR-0094178) and Research Corporation.

  19. Analysis of a gas stratification break-up by a vertical jet using the GOTHIC code

    Energy Technology Data Exchange (ETDEWEB)

    Fernández-Cosials, Mikel Kevin; Jimenez, Gonzalo, E-mail: gonzalo.jimenez@upm.es; Lopez-Alonso, Emma

    2016-02-15

    Highlights: • Study of a light gas distribution with the GOTHIC code based on the OECD/NEA IBE-3. • Sensitivity analysis on turbulence model, discretization scheme and heat transfer. • The jet erosion phenomena is captured properly with a relatively coarse mesh. • Development of a tool to evaluate the influence of each parameter on the simulation. • Several recommendation on modeling a stratification break-up are included. - Abstract: During a severe accident in light water reactor (LWR), hydrogen concentration can overpass the flammability limits locally, so the correct simulation of its behavior during a release is critical. The capability assessment of computational fluid dynamics tools to calculate the hydrogen distribution under different conditions has been the focus of intense research worldwide. In this context, the OECD/NEA conducted an international benchmark exercise (IBE-3), which was focused on the break-up of a stratified layer of a light gas by a vertical jet. The participants performed their simulations before the experiment data was released. When the data was released, it was noticed that a combination of several parameters like the mesh, turbulence model or solver controls were responsible for the broad differences between the participants’ results. To obtain information about how each parameter affects the simulation, a post-test sensitivity analysis has been done by the UPM. In this paper, the IBE-3 experiment simulation with GOTHIC 8.0 is presented along with extensive sensitivity analyses of the relevant parameters. The first objective of the work is to test the capability of GOTHIC 8.0 to simulate properly a gas stratification break-up by a vertical jet with a relatively coarse mesh. The second objective of the paper is to relate each sensitivity parameter with each other and with the experiment through the Parameter Influence Chart, a helpful tool specially designed for this purpose.

  20. Recent development and application of a new safety analysis code for fusion reactors

    Energy Technology Data Exchange (ETDEWEB)

    Merrill, Brad J., E-mail: Brad.Merrill@inl.gov; Humrickhouse, Paul W.; Shimada, Masashi

    2016-11-01

    Highlights: • This paper presents recent code development activities for the MELCOR for fusion and Tritium Migration Analysis Program computer codes at the Idaho National Engineering Laboratory. • The capabilities of these computer codes are being merged into a single safety analysis tool for fusion reactor accidents. • The result of benchmarking these codes against previous code versions is presented by the authors of this paper. • This new capability is applied to study the tritium inventory and permeation rate for a water cold tungsten divertor that has neutron damage at 0.3 dpa. - Abstract: This paper describes the recent progress made in the development of two codes for fusion reactor safety assessments at the Idaho National Laboratory (INL): MELCOR for fusion and the Tritium Migration Analysis Program (TMAP). During the ITER engineering design activity (EDA), the INL Fusion Safety Program (FSP) modified the MELCOR 1.8.2 code for fusion applications to perform ITER thermal hydraulic safety analyses. Because MELCOR has undergone many improvements at SNL-NM since version 1.8.2 was released, the INL FSP recently imported these same fusion modifications into the MELCOR 1.8.6 code, along with the multiple fluids modifications of MELCOR 1.8.5 for fusion used in US advanced fusion reactor design studies. TMAP has also been under development for several decades at the INL by the FSP. TMAP treats multi-specie surface absorption and diffusion in composite materials with dislocation traps, plus the movement of these species from room to room by fluid flow within a given facility. Recently, TMAP was updated to consider multiple trap site types to allow the simulation of experimental data from neutron irradiated tungsten. The natural development path for both of these codes is to merge their capabilities into one computer code to provide a more comprehensive safety tool for analyzing accidents in fusion reactors. In this paper we detail recent developments in this

  1. Development of the ANL plant dynamics code and control strategies for the supercritical carbon dioxide Brayton cycle and code validation with data from the Sandia small-scale supercritical carbon dioxide Brayton cycle test loop.

    Energy Technology Data Exchange (ETDEWEB)

    Moisseytsev, A.; Sienicki, J. J. (Nuclear Engineering Division)

    2011-11-07

    Significant progress has been made in the ongoing development of the Argonne National Laboratory (ANL) Plant Dynamics Code (PDC), the ongoing investigation and development of control strategies, and the analysis of system transient behavior for supercritical carbon dioxide (S-CO{sub 2}) Brayton cycles. Several code modifications have been introduced during FY2011 to extend the range of applicability of the PDC and to improve its calculational stability and speed. A new and innovative approach was developed to couple the Plant Dynamics Code for S-CO{sub 2} cycle calculations with SAS4A/SASSYS-1 Liquid Metal Reactor Code System calculations for the transient system level behavior on the reactor side of a Sodium-Cooled Fast Reactor (SFR) or Lead-Cooled Fast Reactor (LFR). The new code system allows use of the full capabilities of both codes such that whole-plant transients can now be simulated without additional user interaction. Several other code modifications, including the introduction of compressor surge control, a new approach for determining the solution time step for efficient computational speed, an updated treatment of S-CO{sub 2} cycle flow mergers and splits, a modified enthalpy equation to improve the treatment of negative flow, and a revised solution of the reactor heat exchanger (RHX) equations coupling the S-CO{sub 2} cycle to the reactor, were introduced to the PDC in FY2011. All of these modifications have improved the code computational stability and computational speed, while not significantly affecting the results of transient calculations. The improved PDC was used to continue the investigation of S-CO{sub 2} cycle control and transient behavior. The coupled PDC-SAS4A/SASSYS-1 code capability was used to study the dynamic characteristics of a S-CO{sub 2} cycle coupled to a SFR plant. Cycle control was investigated in terms of the ability of the cycle to respond to a linear reduction in the electrical grid demand from 100% to 0% at a rate of 5

  2. A multi-platform linking code for fuel burnup and radiotoxicity analysis

    Science.gov (United States)

    Cunha, R.; Pereira, C.; Veloso, M. A. F.; Cardoso, F.; Costa, A. L.

    2014-02-01

    A linking code between ORIGEN2.1 and MCNP has been developed at the Departamento de Engenharia Nuclear/UFMG to calculate coupled neutronic/isotopic results for nuclear systems and to produce a large number of criticality, burnup and radiotoxicity results. In its previous version, it evaluated the isotopic composition evolution in a Heat Pipe Power System model as well as the radiotoxicity and radioactivity during lifetime cycles. In the new version, the code presents features such as multi-platform execution and automatic results analysis. Improvements made in the code allow it to perform simulations in a simpler and faster way without compromising accuracy. Initially, the code generates a new input for MCNP based on the decisions of the user. After that, MCNP is run and data, such as recoverable energy per prompt fission neutron, reaction rates and keff, are automatically extracted from the output and used to calculate neutron flux and cross sections. These data are then used to construct new ORIGEN inputs, one for each cell in the core. Each new input is run on ORIGEN and generates outputs that represent the complete isotopic composition of the core on that time step. The results show good agreement between GB (Coupled Neutronic/Isotopic code) and Monteburns (Automated, Multi-Step Monte Carlo Burnup Code System), developed by the Los Alamos National Laboratory.

  3. On the Analysis of Weighted Nonbinary Repeat Multiple-Accumulate Codes

    CERN Document Server

    Rosnes, Eirik

    2011-01-01

    In this paper, we consider weighted nonbinary repeat multiple-accumulate (WNRMA) code ensembles obtained from the serial concatenation of a nonbinary rate-1/n repeat code and the cascade of L>= 1 accumulators, where each encoder is followed by a nonbinary random weighter. The WNRMA codes are assumed to be iteratively decoded using the turbo principle with maximum a posteriori constituent decoders. We derive the exact weight enumerator of nonbinary accumulators and subsequently give the weight enumerators for WNRMA code ensembles. We formally prove that the symbol-wise minimum distance of WNRMA code ensembles asymptotically grows linearly with the block length when L >= 3 and n >= 2, and L=2 and n >= 3, for all powers of primes q >= 3 considered, where q is the field size. Thus, WNRMA code ensembles are asymptotically good for these parameters. We also give iterative decoding thresholds, computed by an extrinsic information transfer chart analysis, on the q-ary symmetric channel to show the convergence propert...

  4. An Overview of the XGAM Code and Related Software for Gamma-ray Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Younes, W. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2014-11-13

    The XGAM spectrum-fitting code and associated software were developed specifically to analyze the complex gamma-ray spectra that can result from neutron-induced reactions. The XGAM code is designed to fit a spectrum over the entire available gamma-ray energy range as a single entity, in contrast to the more traditional piecewise approaches. This global-fit philosophy enforces background continuity as well as consistency between local and global behavior throughout the spectrum, and in a natural way. This report presents XGAM and the suite of programs built around it with an emphasis on how they fit into an overall analysis methodology for complex gamma-ray data. An application to the analysis of time-dependent delayed gamma-ray yields from 235U fission is shown in order to showcase the codes and how they interact.

  5. Artificial viscosity method for the design of supercritical airfoils. [Analysis code H

    Energy Technology Data Exchange (ETDEWEB)

    McFadden, G.B.

    1979-07-01

    The need for increased efficiency in the use of our energy resources has stimulated applied research in many areas. Recently progress has been made in the field of aerodynamics, where the development of the supercritical wing promises significant savings in the fuel consumption of aircraft operating near the speed of sound. Computational transonic aerodynamics has proved to be a useful tool in the design and evaluation of these wings. A numerical technique for the design of two-dimensional supercritical wing sections with low wave drag is presented. The method is actually a design mode of the analysis code H developed by Bauer, Garabedian, and Korn. This analysis code gives excellent agreement with experimental results and is used widely by the aircraft industry. The addition of a conceptually simple design version should make this code even more useful to the engineering public.

  6. Adaptive Mesh Computations with the PLUTO Code for Astrophysical Fluid Dynamics

    Science.gov (United States)

    Mignone, A.; Zanni, C.

    2012-07-01

    We present an overview of the current version of the PLUTO code for numerical simulations of astrophysical fluid flows over block-structured adaptive grids. The code preserves its modular framework for the solution of the classical or relativistic magnetohydrodynamics (MHD) equations while exploiting the distributed infrastructure of the Chombo library for multidimensional adaptive mesh refinement (AMR) parallel computations. Equations are evolved in time using an explicit second-order, dimensionally unsplit time stepping scheme based on a cell-centered discretization of the flow variables. Efficiency and robustness are shown through multidimensional benchmarks and applications to problems of astrophysical relevance.

  7. Dynamics and vibrations progress in nonlinear analysis

    CERN Document Server

    Kachapi, Seyed Habibollah Hashemi

    2014-01-01

    Dynamical and vibratory systems are basically an application of mathematics and applied sciences to the solution of real world problems. Before being able to solve real world problems, it is necessary to carefully study dynamical and vibratory systems and solve all available problems in case of linear and nonlinear equations using analytical and numerical methods. It is of great importance to study nonlinearity in dynamics and vibration; because almost all applied processes act nonlinearly, and on the other hand, nonlinear analysis of complex systems is one of the most important and complicated tasks, especially in engineering and applied sciences problems. There are probably a handful of books on nonlinear dynamics and vibrations analysis. Some of these books are written at a fundamental level that may not meet ambitious engineering program requirements. Others are specialized in certain fields of oscillatory systems, including modeling and simulations. In this book, we attempt to strike a balance between th...

  8. Spectral analysis of bedform dynamics

    DEFF Research Database (Denmark)

    Winter, Christian; Ernstsen, Verner Brandbyge; Noormets, Riko

    . An assessment of bedform migration was achieved, as the growth and displacement of every single constituent can be distinguished. It can be shown that the changes in amplitude remain small for all harmonic constituents, whereas the phase shifts differ significantly. Thus the harmonics can be classified....... The proposed method overcomes the above mentioned problems of common descriptive analysis as it is an objective and straightforward mathematical process. The spectral decomposition of superimposed dunes allows a detailed description and analysis of dune patterns and migration....

  9. Development and Verification of Smoothed Particle Hydrodynamics Code for Analysis of Tsunami near NPP

    Energy Technology Data Exchange (ETDEWEB)

    Jo, Young Beom; Kim, Eung Soo [Seoul National Univ., Seoul (Korea, Republic of)

    2014-10-15

    It becomes more complicated when considering the shape and phase of the ground below the seawater. Therefore, some different attempts are required to precisely analyze the behavior of tsunami. This paper introduces an on-going activities on code development in SNU based on an unconventional mesh-free fluid analysis method called Smoothed Particle Hydrodynamics (SPH) and its verification work with some practice simulations. This paper summarizes the on-going development and verification activities on Lagrangian mesh-free SPH code in SNU. The newly developed code can cover equation of motions and heat conduction equation so far, and verification of each models is completed. In addition, parallel computation using GPU is now possible, and GUI is also prepared. If users change input geometry or input values, they can simulate for various conditions geometries. A SPH method has large advantages and potential in modeling of free surface, highly deformable geometry and multi-phase problems that traditional grid-based code has difficulties in analysis. Therefore, by incorporating more complex physical models such as turbulent flow, phase change, two-phase flow, and even solid mechanics, application of the current SPH code is expected to be much more extended including molten fuel behaviors in the sever accident.

  10. Analysis of Cooperative Networks Based on WiMAX LDPC Code

    Directory of Open Access Journals (Sweden)

    M.B. Khan

    2014-11-01

    Full Text Available This study focus on the performance analysis of Cooperative communication networks based on WiMAX Low Density Parity Check (LDPC codes. The channel capacity approaching coding technique LDPC having coding gain method Bit Interleave Coded Modulation with Iterative Decoding (BICM-ID is used. The different fading environment is analyze to counter the challenges in wireless communication and provides solutions for the drawbacks in the multiple input multiple output MIMO technology. The relays are used in Cooperative communications networks to increases the range and link reliability at a lower transmit power because once the signal transmit power loses its strength it’s amplify on the relay node and when it suffers from noise it is also decoded at the relay node which increases the link reliability. LDPC with iterative decoding are used to gain BER performance only a small amount of decibel to attain Shannon limit. This performance analysis open the way for WiMAX technology can be used with Cooperative networks by using LDPC codes. The above mention communication system will provides rate, range and reliability at a lower cost, less complexity and lower transmit power.

  11. NACOM code for analysis of postulated sodium spray fires in LMFBRs

    Energy Technology Data Exchange (ETDEWEB)

    Tsai, S.S.

    1980-03-01

    An analysis of potential sodium spills and fires in liquid metal fast breeder reactors has been made to assess the maximum equipment cell loading conditions. A computer code called NACOM (sodium combustion) has been developed at Brookhaven National Laobratory (BNL) to analyze sodium spray fires. This report contains a detailed description of physical models used in this code as well as programming aspects. The single droplet combustion model and the model describing the droplets' motion are verified. Comparisons between NACOM predictions and SPRAY-3A predictions of the Atomics International (AI) LTV Jet Tests are made. Good agreement is found between the NACOM predictions and the experimental data. NACOM predictions of the pressure rise are more accurate than SPRAY-3A predictions for most of the cases studied. The code has been verified for oxygen concentrations ranging from 0 to 21%. NACOM utilizes more realistic single droplet and spray combustion models than SPRAY-3A. Moreover, NACOM does not utilize adjustable parameters for the burning rate equations, contrary to the approach taken with SPRAY-3A. Thus, the NACOM code is a more reliable code for use in the analysis of large-scale sodium spray fires in LMFBR containment cells. 24 refs., 32 figs.

  12. Development of MATRA-LMR code {alpha}-version for LMR subchannel analysis

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Won Seok; Kim, Young Gyun; Kim, Young Gin

    1998-05-01

    Since the sodium boiling point is very high, maximum cladding and pin temperature are used for design limit condition in sodium cooled liquid metal reactor. It is necessary to predict accurately the core temperature distribution to increase the sodium coolant efficiency. Based on the MATRA code, which is developed for PWR analysis, MATRA-LMR is being developed for LMR. The major modification are as follows : A) The sodium properties table is implemented as subprogram in the code. B) Heat transfer coefficients are changed for LMR C) The pressure drop correlations are changed for more accurate calculations, which are Novendstern, Chiu-Rohsenow-Todreas, and Cheng-Todreas correlations. To assess the development status of MATRA-LMR code, calculations have been performed for ORNL 19 pin and EBR-II 61 pin tests. MATRA-LMR calculation results are also compared with the results obtained by the ALTHEN code, which uses more simplied thermal hydraulic model. The MATRA-LMR predictions are found to agree well to the measured values. The differences in results between MATRA-LMR and SLTHEN have occurred because SLTHEN code uses the very simplied thermal-hydraulic model to reduce computing time. MATRA-LMR can be used only for single assembly analysis, but it is planned to extend for multi-assembly calculation. (author). 18 refs., 8 tabs., 14 figs.

  13. GRADSPH: A parallel smoothed particle hydrodynamics code for self-gravitating astrophysical fluid dynamics

    NARCIS (Netherlands)

    Vanaverbeke, S.; Keppens, R.; Poedts, S.; Boffin, H.

    2009-01-01

    We describe the algorithms implemented in the first version of GRADSPH, a parallel, tree-based, smoothed particle hydrodynamics code for simulating self-gravitating astrophysical systems written in FORTRAN 90. The paper presents details on the implementation of the Smoothed Particle Hydro (SPH) desc

  14. The Proteus Navier-Stokes code. [two and three dimensional computational fluid dynamics

    Science.gov (United States)

    Towne, Charles E.; Schwab, John R.

    1992-01-01

    An effort is currently underway at NASA Lewis to develop two and three dimensional Navier-Stokes codes, called Proteus, for aerospace propulsion applications. Proteus solves the Reynolds-averaged, unsteady, compressible Navier-Stokes equations in strong conservation law form. Turbulence is modeled using a Baldwin-Lomax based algebraic eddy viscosity model. In addition, options are available to solve thin layer or Euler equations, and to eliminate the energy equation by assuming constant stagnation enthalpy. An extensive series of validation cases have been run, primarily using the two dimensional planar/axisymmetric version of the code. Several flows were computed that have exact solution such as: fully developed channel and pipe flow; Couette flow with and without pressure gradients; unsteady Couette flow formation; flow near a suddenly accelerated flat plate; flow between concentric rotating cylinders; and flow near a rotating disk. The two dimensional version of the Proteus code has been released, and the three dimensional code is scheduled for release in late 1991.

  15. Light water reactor fuel analysis code FEMAXI-V (Ver.1)

    Energy Technology Data Exchange (ETDEWEB)

    Suzuki, Motoe [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment

    2000-09-01

    A light water fuel analysis code FEMAXI-V is an advanced version which has been produced by integrating FEMAXI-IV(Ver.2), high burn-up fuel code EXBURN-I, and a number of functional improvements and extensions, to predict fuel rod behavior in normal and transient (not accident) conditions. The present report describes in detail the basic theories and structure, models and numerical solutions applied, improvements and extensions, and the material properties adopted in FEMAXI-V(Ver.1). FEMAXI-V deals with a single fuel rod. It predicts thermal and mechanical response of fuel rod to irradiation, including FP gas release. The thermal analysis predicts rod temperature distribution on the basis of pellet heat generation, changes in pellet thermal conductivity and gap thermal conductance, (transient) change in surface heat transfer to coolant, using radial one-dimensional geometry. The heat generation density profile of pellet can be determined by adopting the calculated results of burning analysis code. The mechanical analysis performs elastic/plastic, creep and PCMI calculations by FEM. The FP gas release model calculates diffusion of FP gas atoms and accumulation in bubbles, release and increase in internal pressure of rod. In every analysis, it is possible to allow some materials properties and empirical equations to depend on the local burnup or heat flux, which enables particularly analysis of high burnup fuel behavior and boiling transient of BWR rod. In order to facilitate effective and wide-ranging application of the code, formats and methods of input/output of the code are also described, and a sample output in an actual form is included. (author)

  16. Analysis of BER Performance of the Spread Spectrum Communication System with Constrained Spreading Code

    OpenAIRE

    長谷川, 孝明; 羽渕, 裕真

    1996-01-01

    Copyright notice. c1996 IEICE All rights reserved. "Analysis of BER Performance of the Spread Spectrum Communication System with Constrained Spreading Code"Hiromasa HABUCHI, Toshio TAKEBAYASHI, Takaaki HASEGAWA. IEICE TRANSACTIONS on Fundamentals of Electronics, Communications and Computer Sciences ,1996 Vol.E79-A No.12 pp. 2078-2080 許諾No.07RB0055.

  17. Analytical signal analysis of strange nonchaotic dynamics.

    Science.gov (United States)

    Gupta, Kopal; Prasad, Awadhesh; Singh, Harinder P; Ramaswamy, Ramakrishna

    2008-04-01

    We apply an analytical signal analysis to strange nonchaotic dynamics. Through this technique it is possible to obtain the spectrum of instantaneous intrinsic mode frequencies that are present in a given signal. We find that the second-mode frequency and its variance are good order parameters for dynamical transitions from quasiperiodic tori to strange nonchaotic attractors (SNAs) and from SNAs to chaotic attractors. Phase fluctuation analysis shows that SNAs and chaotic attractors behave identically within short time windows as a consequence of local instabilities in the dynamics. In longer time windows, however, the globally stable character of SNAs becomes apparent. This methodology can be of great utility in the analysis of experimental time series, and representative applications are made to signals obtained from Rössler and Duffing oscillators.

  18. Development of human reliability analysis methodology and its computer code during low power/shutdown operation

    Energy Technology Data Exchange (ETDEWEB)

    Chung, Chang Hyun; You, Young Woo; Huh, Chang Wook; Kim, Ju Yeul; Kim Do Hyung; Kim, Yoon Ik; Yang, Hui Chang [Seoul National University, Seoul (Korea, Republic of); Jae, Moo Sung [Hansung University, Seoul (Korea, Republic of)

    1997-07-01

    The objective of this study is to develop the appropriate procedure that can evaluate the human error in LP/S(lower power/shutdown) and the computer code that calculate the human error probabilities(HEPs) using this framework. The assessment of applicability of the typical HRA methodologies to LP/S is conducted and a new HRA procedure, SEPLOT (Systematic Evaluation Procedure for LP/S Operation Tasks) which presents the characteristics of LP/S is developed by selection and categorization of human actions by reviewing present studies. This procedure is applied to evaluate the LOOP(Loss of Off-site Power) sequence and the HEPs obtained by using SEPLOT are used to quantitative evaluation of the core uncovery frequency. In this evaluation one of the dynamic reliability computer codes, DYLAM-3 which has the advantages against the ET/FT is used. The SEPLOT developed in this study can give the basis and arrangement as to the human error evaluation technique. And this procedure can make it possible to assess the dynamic aspects of accidents leading to core uncovery applying the HEPs obtained by using the SEPLOT as input data to DYLAM-3 code, Eventually, it is expected that the results of this study will contribute to improve safety in LP/S and reduce uncertainties in risk. 57 refs. 17 tabs., 33 figs. (author)

  19. Linear Stability Analysis of Dynamical Quadratic Gravity

    CERN Document Server

    Ayzenberg, Dimitry; Yunes, Nicolas

    2013-01-01

    We perform a linear stability analysis of dynamical, quadratic gravity in the high-frequency, geometric optics approximation. This analysis is based on a study of gravitational and scalar modes propagating on spherically-symmetric and axially-symmetric, vacuum solutions of the theory. We find dispersion relations that do no lead to exponential growth of the propagating modes, suggesting the theory is linearly stable on these backgrounds. The modes are found to propagate at subluminal and superluminal speeds, depending on the propagating modes' direction relative to the background geometry, just as in dynamical Chern-Simons gravity.

  20. Rotor dynamic analysis of main coolant pump

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Chong Won; Seo, Jeong Hwan; Kim, Choong Hwan; Shin, Jae Chul; Wang, Lei Tian [Korea Advanced Institute of Science and Technology, Taejon (Korea)

    1999-03-01

    A rotor dynamic analysis program DARBS/MCP, for the main coolant pump of the integral reactor, has been developed. The dynamic analysis model of the main coolant pump includes a vertical shaft, three grooved radial journal bearings and gaps that represent the structure-fluid interaction effects between the rotor and the lubricant fluid. The electromagnetic force from the motor and the hydro-dynamic force induced by impeller are the major sources of vibration that may affect the rotor system stability. DARBS/MCP is a software that is developed to effectively analyze the dynamics of MCP rotor systems effectively by applying powerful numerical algorithms such as FEM with modal truncation and {lambda}-matrix method for harmonic analysis. Main design control parameters, that have much influence to the dynamic stability, have been found by Taguchi's sensitivity analysis method. Design suggestions to improve the stability of MCP rotor system have been documented. The dynamic bearing parameters of the journal bearings used for main coolant pump have been determined by directly solving the Reynolds equation using FDM method. Fluid-structure interaction effect that occurs at the small gaps between the rotor and the stator were modeled as equivalent seals, the electromagnetic force effect was regarded as a linear negative radial spring and the impeller was modeled as a rigid disk with hydrodynamic and static radial force. Although there exist critical speeds in the range of operational speeds for type I and II rotor systems, the amplitude of vibration appears to be less than the vibration limit set by the API standards. Further more, it has been verified that the main design parameters such as the clearance and length of journal bearings, and the static radial force of impeller should be properly adjusted, in order to the improve dynamic stability of the rotor system. (author). 39 refs., 81 figs., 17 tabs.

  1. Shared and Distributed Memory Parallel Security Analysis of Large-Scale Source Code and Binary Applications

    Energy Technology Data Exchange (ETDEWEB)

    Quinlan, D; Barany, G; Panas, T

    2007-08-30

    Many forms of security analysis on large scale applications can be substantially automated but the size and complexity can exceed the time and memory available on conventional desktop computers. Most commercial tools are understandably focused on such conventional desktop resources. This paper presents research work on the parallelization of security analysis of both source code and binaries within our Compass tool, which is implemented using the ROSE source-to-source open compiler infrastructure. We have focused on both shared and distributed memory parallelization of the evaluation of rules implemented as checkers for a wide range of secure programming rules, applicable to desktop machines, networks of workstations and dedicated clusters. While Compass as a tool focuses on source code analysis and reports violations of an extensible set of rules, the binary analysis work uses the exact same infrastructure but is less well developed into an equivalent final tool.

  2. Context-based coding of bilevel images enhanced by digital straight line analysis

    DEFF Research Database (Denmark)

    Aghito, Shankar Manuel; Forchhammer, Søren

    2006-01-01

    , or segmentation maps are also encoded efficiently. The algorithm is not targeted at document images with text, which can be coded efficiently with dictionary-based techniques as in JBIG2. The scheme is based on a local analysis of the digital straightness of the causal part of the object boundary, which is used...... in the context definition for arithmetic encoding. Tested on individual images of standard TV resolution binary shapes and the binary layers of a digital map, the proposed algorithm outperforms PWC, JBIG, JBIG2, and MPEG-4 CAE. On the binary shapes, the code lengths are reduced by 21%, 27 %, 28 %, and 41...

  3. BETA. A code for {beta}{sub eff} measurement and analysis

    Energy Technology Data Exchange (ETDEWEB)

    Kato, Yuichi; Okajima, Shigeaki; Sakurai, Takeshi [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment

    1999-03-01

    The code BETA has been developed to calculate the following reactor physics parameters which are used for the {beta}{sub eff} measurement and analysis; Diven factor, Spatial correction factor (g-factor) for neutron correlation experiment, Adjoint weighted g-factor, Fission rate integrated in whole reactor and Adjoint weighted fission rate integrated in whole reactor: The code also calculates the effective delayed neutron fraction with different evaluated delayed neutron data. These parameters are calculated with using the forward and adjoint fluxes given by SLAROM, POPLARS or TWOTRAN-II. This report describes the input data and job control statements instructions, file requirements and sample input output data. (author)

  4. Structural Dynamics and Data Analysis

    Science.gov (United States)

    Luthman, Briana L.

    2013-01-01

    This project consists of two parts, the first will be the post-flight analysis of data from a Delta IV launch vehicle, and the second will be a Finite Element Analysis of a CubeSat. Shock and vibration data was collected on WGS-5 (Wideband Global SATCOM- 5) which was launched on a Delta IV launch vehicle. Using CAM (CAlculation with Matrices) software, the data is to be plotted into Time History, Shock Response Spectrum, and SPL (Sound Pressure Level) curves. In this format the data is to be reviewed and compared to flight instrumentation data from previous flights of the same launch vehicle. This is done to ensure the current mission environments, such as shock, random vibration, and acoustics, are not out of family with existing flight experience. In family means the peaks on the SRS curve for WGS-5 are similar to the peaks from the previous flights and there are no major outliers. The curves from the data will then be compiled into a useful format so that is can be peer reviewed then presented before an engineering review board if required. Also, the reviewed data will be uploaded to the Engineering Review Board Information System (ERBIS) to archive. The second part of this project is conducting Finite Element Analysis of a CubeSat. In 2010, Merritt Island High School partnered with NASA to design, build and launch a CubeSat. The team is now called StangSat in honor of their mascot, the mustang. Over the past few years, the StangSat team has built a satellite and has now been manifested for flight on a SpaceX Falcon 9 launch in 2014. To prepare for the final launch, a test flight was conducted in Mojave, California. StangSat was launched on a Prospector 18D, a high altitude rocket made by Garvey Spacecraft Corporation, along with their sister satellite CP9 built by California Polytechnic University. However, StangSat was damaged during an off nominal landing and this project will give beneficial insights into what loads the CubeSat experienced during the crash

  5. Development of best estimate auditing code for CANDU thermal-hydraulic safety analysis

    Energy Technology Data Exchange (ETDEWEB)

    Chung, Bub Dong; Lee, Won Jae; Hwang, Moon Kyu; Lim, Hong Sik [Korea Atomic Energy Research Institute, Taejeon (Korea)

    2002-04-01

    The main purpose of this study is to develop a thermal hydraulic auditing code for the CANDU reactor, modifying the model of existing PWR auditing tool, i.e. RELAP5/MOD3.The study was performed by reconsideration of the previous code assessment works and phenomena identification for essential accident scenario. Improvement areas of model development for auditing tool were identified based on the code comparison and PIRT results. Nine models have been improved significantly for the analysis of LOCA and Mon LOCA event. Conceptual problem or separate effect assessment have been performed to verify the model improvement. The linking calculation with CONTAIN 2.0 has been also enabled to establish the unified auditing code system. Analysis for the CANDU plant real transient and hypothetical LOCA bas been performed using the improved version. It has been concluded that the developed version can be utilized for the auditing analysis of LOCA and non-LOCA event for the CANDU reactor. 25 refs., 84 figs., 36 tabs. (Author)

  6. Research on the improvement of nuclear safety -The development of a severe accident analysis code-

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Heui Dong; Cho, Sung Won; Park, Jong Hwa; Hong, Sung Wan; Yoo, Dong Han; Hwang, Moon Kyoo; Noh, Kee Man; Song, Yong Man [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    1995-07-01

    For prevention and mitigation of the containment failure during severe accident, the study is focused on the severe accident phenomena, especially, the ones occurring inside the cavity and is intended to improve existing models and develop analytical tools for the assessment of severe accidents. A correlation equation of the flame velocity of pre mixture gas of H{sub 2}/air/steam has been suggested and combustion flame characteristic was analyzed using a developed computer code. For the analysis of the expansion phase of vapor explosion, the mechanical model has been developed. The development of a debris entrainment model in a reactor cavity with captured volume has been continued to review and examine the limitation and deficiencies of the existing models. Pre-test calculation was performed to support the severe accident experiment for molten corium concrete interaction study and the crust formation process and heat transfer characteristics of the crust have been carried out. A stress analysis code was developed using finite element method for the reactor vessel lower head failure analysis. Through international program of PHEBUS-FP and participation in the software development, the research on the core degradation process and fission products release and transportation are undergoing. CONTAIN and MELCOR codes were continuously updated under the cooperation with USNRC and French developed computer codes such as ICARE2, ESCADRE, SOPHAEROS were also installed into the SUN workstation. 204 figs, 61 tabs, 87 refs. (Author).

  7. Beam Dynamics in an Electron Lens with the Warp Particle-in-cell Code

    CERN Document Server

    Stancari, Giulio; Redaelli, Stefano

    2014-01-01

    Electron lenses are a mature technique for beam manipulation in colliders and storage rings. In an electron lens, a pulsed, magnetically confined electron beam with a given current-density profile interacts with the circulating beam to obtain the desired effect. Electron lenses were used in the Fermilab Tevatron collider for beam-beam compensation, for abort-gap clearing, and for halo scraping. They will be used in RHIC at BNL for head-on beam-beam compensation, and their application to the Large Hadron Collider for halo control is under development. At Fermilab, electron lenses will be implemented as lattice elements for nonlinear integrable optics. The design of electron lenses requires tools to calculate the kicks and wakefields experienced by the circulating beam. We use the Warp particle-in-cell code to study generation, transport, and evolution of the electron beam. For the first time, a fully 3-dimensional code is used for this purpose.

  8. A Coding Scheme Development Methodology Using Grounded Theory For Qualitative Analysis Of Pair Programming

    Directory of Open Access Journals (Sweden)

    Stephan Salinger

    2008-01-01

    Full Text Available A number of quantitative studies of pair programming (the practice of two programmers working together using just one computer have partially conflicting results. Qualitative studies are needed to explain what is really going on. We support such studies by taking a grounded theory (GT approach for deriving a coding scheme for the objective conceptual description of specific pair programming sessions independent of a particular research goal. The present article explains why our initial attempts at using GT failed and describes how to avoid these difficulties by a predetermined perspective on the data, concept naming rules, an analysis results metamodel, and pair coding. These practices may be helpful in all GT situations, particularly those involving very rich data such as video data. We illustrate the operation and usefulness of these practices by real examples derived from our coding work and present a few preliminary hypotheses regarding pair programming that have surfaced.

  9. Analysis of Doppler Effect on the Pulse Compression of Different Codes Emitted by an Ultrasonic LPS

    Directory of Open Access Journals (Sweden)

    Jorge Morera

    2011-11-01

    Full Text Available This work analyses the effect of the receiver movement on the detection by pulse compression of different families of codes characterizing the emissions of an Ultrasonic Local Positioning System. Three families of codes have been compared: Kasami, Complementary Sets of Sequences and Loosely Synchronous, considering in all cases three different lengths close to 64, 256 and 1,024 bits. This comparison is first carried out by using a system model in order to obtain a set of results that are then experimentally validated with the help of an electric slider that provides radial speeds up to 2 m/s. The performance of the codes under analysis has been characterized by means of the auto-correlation and cross-correlation bounds. The results derived from this study should be of interest to anyone performing matched filtering of ultrasonic signals with a moving emitter/receiver.

  10. Analysis of Doppler effect on the pulse compression of different codes emitted by an ultrasonic LPS.

    Science.gov (United States)

    Paredes, José A; Aguilera, Teodoro; Alvarez, Fernando J; Lozano, Jesús; Morera, Jorge

    2011-01-01

    This work analyses the effect of the receiver movement on the detection by pulse compression of different families of codes characterizing the emissions of an ultrasonic local positioning system. Three families of codes have been compared: Kasami, Complementary Sets of Sequences and Loosely Synchronous, considering in all cases three different lengths close to 64, 256 and 1,024 bits. This comparison is first carried out by using a system model in order to obtain a set of results that are then experimentally validated with the help of an electric slider that provides radial speeds up to 2 m/s. The performance of the codes under analysis has been characterized by means of the auto-correlation and cross-correlation bounds. The results derived from this study should be of interest to anyone performing matched filtering of ultrasonic signals with a moving emitter/receiver.

  11. Development of an Object-Oriented Turbomachinery Analysis Code within the NPSS Framework

    Science.gov (United States)

    Jones, Scott M.

    2014-01-01

    During the preliminary or conceptual design phase of an aircraft engine, the turbomachinery designer has a need to estimate the effects of a large number of design parameters such as flow size, stage count, blade count, radial position, etc. on the weight and efficiency of a turbomachine. Computer codes are invariably used to perform this task however, such codes are often very old, written in outdated languages with arcane input files, and rarely adaptable to new architectures or unconventional layouts. Given the need to perform these kinds of preliminary design trades, a modern 2-D turbomachinery design and analysis code has been written using the Numerical Propulsion System Simulation (NPSS) framework. This paper discusses the development of the governing equations and the structure of the primary objects used in OTAC.

  12. Architectural and Algorithmic Requirements for a Next-Generation System Analysis Code

    Energy Technology Data Exchange (ETDEWEB)

    V.A. Mousseau

    2010-05-01

    This document presents high-level architectural and system requirements for a next-generation system analysis code (NGSAC) to support reactor safety decision-making by plant operators and others, especially in the context of light water reactor plant life extension. The capabilities of NGSAC will be different from those of current-generation codes, not only because computers have evolved significantly in the generations since the current paradigm was first implemented, but because the decision-making processes that need the support of next-generation codes are very different from the decision-making processes that drove the licensing and design of the current fleet of commercial nuclear power reactors. The implications of these newer decision-making processes for NGSAC requirements are discussed, and resulting top-level goals for the NGSAC are formulated. From these goals, the general architectural and system requirements for the NGSAC are derived.

  13. Dynamic analysis and design of offshore structures

    CERN Document Server

    Chandrasekaran, Srinivasan

    2015-01-01

    This book  attempts to provide readers with an overall idea of various types of offshore platform geometries. It covers the various environmental loads encountered by these structures, a detailed description of the fundamentals of structural dynamics in a class-room style, estimate of damping in offshore structures and their applications in the preliminary analysis and design. Basic concepts of structural dynamics are emphasized through simple illustrative examples and exercises. Design methodologies and guidelines, which are FORM based concepts are explained through a few applied example structures. Each chapter also has tutorials and exercises for self-learning. A dedicated chapter on stochastic dynamics will help the students to extend the basic concepts of structural dynamics to this advanced domain of research. Hydrodynamic response of offshore structures with perforated members is one of the recent research applications, which is found to be one of the effective manner of retrofitting offshore structur...

  14. Calculation of the Thermal Radiation Benchmark Problems for a CANDU Fuel Channel Analysis Using the CFX-10 Code

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Hyoung Tae; Park, Joo Hwan; Rhee, Bo Wook

    2006-07-15

    To justify the use of a commercial Computational Fluid Dynamics (CFD) code for a CANDU fuel channel analysis, especially for the radiation heat transfer dominant conditions, the CFX-10 code is tested against three benchmark problems which were used for the validation of a radiation heat transfer in the CANDU analysis code, a CATHENA. These three benchmark problems are representative of the CANDU fuel channel configurations from a simple geometry to whole fuel channel geometry. With assumptions of a non-participating medium completely enclosed with the diffuse, gray and opaque surfaces, the solutions of the benchmark problems are obtained by the concept of surface resistance to radiation accounting for the view factors and the emissivities. The view factors are calculated by the program MATRIX version 1.0 avoiding the difficulty of hand calculation for the complex geometries. For the solutions of the benchmark problems, the temperature or the net radiation heat flux boundary conditions are prescribed for each radiating surface to determine the radiation heat transfer rate or the surface temperature, respectively by using the network method. The Discrete Transfer Model (DTM) is used for the CFX-10 radiation model and its calculation results are compared with the solutions of the benchmark problems. The CFX-10 results for the three benchmark problems are in close agreement with these solutions, so it is concluded that the CFX-10 with a DTM radiation model can be applied to the CANDU fuel channel analysis where a surface radiation heat transfer is a dominant mode of the heat transfer.

  15. Phonemic coding might be a result of sensory-motorcoupling dynamics

    OpenAIRE

    2003-01-01

    in the Proceedings of the 7th International Conference on the Simulation of Adaptive Behavior, pp. 406-416, eds. B. Hallam, D. Floreano, J. Hallam, G. Hayes, J-A. Meyer, MIT Press.; Human sound systems are invariably phonemically coded. Furthermore,phoneme inventories follow very particular tendancies. To explain these phenomena, there existed so far three kinds of approaches : "Chomskyan"/cognitive innatism, morpho-perceptual innatism and the more recent approach of "language as a complex cu...

  16. Clean Energy in City Codes: A Baseline Analysis of Municipal Codification across the United States

    Energy Technology Data Exchange (ETDEWEB)

    Cook, Jeffrey J. [National Renewable Energy Lab. (NREL), Golden, CO (United States); Aznar, Alexandra [National Renewable Energy Lab. (NREL), Golden, CO (United States); Dane, Alexander [National Renewable Energy Lab. (NREL), Golden, CO (United States); Day, Megan [National Renewable Energy Lab. (NREL), Golden, CO (United States); Mathur, Sivani [National Renewable Energy Lab. (NREL), Golden, CO (United States); Doris, Elizabeth [National Renewable Energy Lab. (NREL), Golden, CO (United States)

    2016-12-01

    Municipal governments in the United States are well positioned to influence clean energy (energy efficiency and alternative energy) and transportation technology and strategy implementation within their jurisdictions through planning, programs, and codification. Municipal governments are leveraging planning processes and programs to shape their energy futures. There is limited understanding in the literature related to codification, the primary way that municipal governments enact enforceable policies. The authors fill the gap in the literature by documenting the status of municipal codification of clean energy and transportation across the United States. More directly, we leverage online databases of municipal codes to develop national and state-specific representative samples of municipal governments by population size. Our analysis finds that municipal governments with the authority to set residential building energy codes within their jurisdictions frequently do so. In some cases, communities set codes higher than their respective state governments. Examination of codes across the nation indicates that municipal governments are employing their code as a policy mechanism to address clean energy and transportation.

  17. Thermal-hydraulic analysis of SWAMUP facility using ATHLET-SC code

    Directory of Open Access Journals (Sweden)

    Zidi eWang

    2015-03-01

    Full Text Available During the loss of coolant accident (LOCA of supercritical water cooled reactor (SCWR, the pressure in the reactor system will undergo a rapid decrease from the supercritical pressure to the subcritical condition. This process is called trans-critical transients, which is of crucial importance for the LOCA analysis of SCWR. In order to simulate the trans-critical transient, a number of system codes for SCWR have been developed up to date. However, the validation work for the trans-critical models in these codes is still missing. The test facility Supercritical WAter MUltiPurpose loop (SWAMUP with 2×2 rod bundle in Shanghai Jiao Tong University (SJTU will be applied to provide test data for code validation. Some pre-test calculations are important and necessary to show the feasibility of the experiment. In this study, trans-critical transient analysis is performed for the SWAMUP facility with the system code ATHLET-SC, which is modified in SJTU, for supercritical water system. This paper presents the system behavior e.g. system pressure, coolant mass flow, cladding temperature during the depressurization. The effects of some important parameters such as heating power, depressurization rate on the system characteristics are also investigated in this paper. Additionally, some sensitivities study of the code models, e.g. heat transfer coefficient, CHF correlation, are analyzed and discussed. The results indicate that the revised system code ATHLET-SC is capable of simulating thermal hydraulic behavior during the trans-critical transient. According to the results, the cladding temperature during the transient is kept at a low value. However, the pressure difference of the heat exchanger after depressurization could reach 6 MPa, which should be considered in the experiment.

  18. A Dynamic Model for Energy Structure Analysis

    Institute of Scientific and Technical Information of China (English)

    2006-01-01

    Energy structure is a complicated system concerning economic development, natural resources, technological innovation, ecological balance, social progress and many other elements. It is not easy to explain clearly the developmental mechanism of an energy system and the mutual relations between the energy system and its related environments by the traditional methods. It is necessary to develop a suitable dynamic model, which can reflect the dynamic characteristics and the mutual relations of the energy system and its related environments. In this paper, the historical development of China's energy structure was analyzed. A new quantitative analysis model was developed based on system dynamics principles through analysis of energy resources, and the production and consumption of energy in China and comparison with the world. Finally, this model was used to predict China's future energy structures under different conditions.

  19. Dynamic Analysis of The Intelligent Sprayer Boom

    DEFF Research Database (Denmark)

    Wiggers, Sine Leergaard; Maagaard, Jørgen; Terp, Christian Istjord

    As part of the 3 year project “The intelligent Sprayer Boom”, financed by The Danish National Advanced Technology Foundation, the dynamics of the sprayer boom is to be analysed. In order to minimize the amount of herbicides used to kill the weeds in agriculture a new sprayer boom is being developed...... called “The intelligent sprayer boom”. For the sprayer boom the primary challenge is to hit the weeds with precision from a movable platform. Since the sprayer boom is mounted on a tractor the system will react to bumps in the field. The intelligent sprayer boom has an integrated camera technology...... system to work properly. At the University of Southern Denmark (SDU) a patent for an active damping system of the sprayer boom has been obtained. The subject of this paper is analysis of the dynamics of the Sprayer boom. The analysis is based on a Multibody Dynamics model of the sprayer boom and is made...

  20. Comparative Neutronics Analysis of DIMPLE S06 Criticality Benchmark with Contemporary Reactor Core Analysis Computer Code Systems

    Directory of Open Access Journals (Sweden)

    Wonkyeong Kim

    2015-01-01

    Full Text Available A high-leakage core has been known to be a challenging problem not only for a two-step homogenization approach but also for a direct heterogeneous approach. In this paper the DIMPLE S06 core, which is a small high-leakage core, has been analyzed by a direct heterogeneous modeling approach and by a two-step homogenization modeling approach, using contemporary code systems developed for reactor core analysis. The focus of this work is a comprehensive comparative analysis of the conventional approaches and codes with a small core design, DIMPLE S06 critical experiment. The calculation procedure for the two approaches is explicitly presented in this paper. Comprehensive comparative analysis is performed by neutronics parameters: multiplication factor and assembly power distribution. Comparison of two-group homogenized cross sections from each lattice physics codes shows that the generated transport cross section has significant difference according to the transport approximation to treat anisotropic scattering effect. The necessity of the ADF to correct the discontinuity at the assembly interfaces is clearly presented by the flux distributions and the result of two-step approach. Finally, the two approaches show consistent results for all codes, while the comparison with the reference generated by MCNP shows significant error except for another Monte Carlo code, SERPENT2.

  1. WITS-HEX: A renovated code for analysis of PIXE X-ray spectra

    Science.gov (United States)

    Lipworth, A. D.; Annegarn, H. J.; Bauman, S.; Molokomme, T.; Walker, A. J.

    1990-04-01

    The code REX [H.C. Kaufmann and K.R. Akselsson, Adv. X-ray Anal. 18 (1975) 353] and its successor, HEX were developed at Florida State University and University of Lund in the early seventies for quantitative reduction of PIXE spectra. HEX modelled many of the physical processes, including target self-absorption. Lack of adequate documentation and several inherent software related problems hampered widespread adoption and further development. HEX was designed using a hierarchical functional design approach, coded in FORTRAN IV. Its element library and request list are tightly constrained. ANSI standard FORTRAN provides few structured constructs, nor does it support dynamic data structures. HEX has a batch mode of interaction which affords minimal (interim) user interactions. To overcome these limitations we have modernised the program to run on IBM PC compatible computers by using object-oriented design techniques, documented in Program Description Language (pseudo-code) and implemented in a modern programming language, Pascal. Turbo Pascal 5.5 provides an integrated program development environment, a high resolution graphics library, an operating system interface and many software development tools to improve programmer productivity. A menu-based highly interactive screen management library is used for the user interface. A demonstration version is available for user evaluation. Further rigorous testing and the development of user acquired spectra interfacing code will be developed before the first release. The renovated code, WITS-HEX, has extracted the intellectual value embodied in HEX. It will be inherently more stable, reliable, versatile, maintainable and easy to use to facilitate future process modifications.

  2. Mytoe: automatic analysis of mitochondrial dynamics.

    NARCIS (Netherlands)

    Lihavainen, E.; Makela, J.; Spelbrink, J.N.; Ribeiro, A.S.

    2012-01-01

    SUMMARY: We present Mytoe, a tool for analyzing mitochondrial morphology and dynamics from fluorescence microscope images. The tool provides automated quantitative analysis of mitochondrial motion by optical flow estimation and of morphology by segmentation of individual branches of the network-like

  3. Dynamical Analysis of DTNN with Impulsive Effect

    Directory of Open Access Journals (Sweden)

    Chao Chen

    2009-01-01

    Full Text Available We present dynamical analysis of discrete-time delayed neural networks with impulsive effect. Under impulsive effect, we derive some new criteria for the invariance and attractivity of discrete-time neural networks by using decomposition approach and delay difference inequalities. Our results improve or extend the existing ones.

  4. Performance analysis of ARQ schemes with code combining over Nakagami-m fading channel

    Institute of Scientific and Technical Information of China (English)

    XU Wen-bo; NIU Kai; LIN Jia-ru; HE Zhi-qiang

    2009-01-01

    This article investigates the performance of hybrid automatic repeat request (HARQ) with code combining over the ideally interleaved Nakagami-m fading channel. Two retransmission protocols with coherent equal gain code combining are adopted, where the entire frame and several selected portions of the frame are repeated in protocols Ⅰ and Ⅱ, respectively. Protocol Ⅱ could be viewed as a generalization of the recently proposed reliability-based HARQ. To facilitate performance analysis, an approximation of the product of two independent Nakagami-m distributed random variables is first developed. Then the approximate analysis is utilized to obtain exact frame error probability (FEP) for protocol Ⅰ, and the upper bound of the FEP for protocol Ⅱ. Furthermore, the throughput performance of both two protocols is presented. Simulation results show the reliability of the theoretical analysis, where protocol Ⅱ outperforms protocol Ⅰ in the throughput performance due to the reduced amount of transmitted information.

  5. Data-driven inference of network connectivity for modeling the dynamics of neural codes in the insect antennal lobe

    Directory of Open Access Journals (Sweden)

    Eli eShlizerman

    2014-08-01

    Full Text Available The antennal lobe (AL, olfactory processing center in insects, is able to process stimuli into distinct neural activity patterns, called olfactory neural codes. To model their dynamics we perform multichannel recordings from the projection neurons in the AL driven by different odorants. We then derive a dynamic neuronal network from the electrophysiological data. The network consists of lateral-inhibitory neurons and excitatory neurons (modeled as firing-rate units, and is capable of producing unique olfactory neural codes for the tested odorants. To construct the network, we (i design a projection, an odor space, for the neural recording from the AL, which discriminates between distinct odorants trajectories (ii characterize scent recognition, i.e., decision-making based on olfactory signals and (iii infer the wiring of the neural circuit, the connectome of the AL. We show that the constructed model is consistent with biological observations, such as contrast enhancement and robustness to noise. The study suggests a data-driven approach to answer a key biological question in identifying how lateral inhibitory neurons can be wired to excitatory neurons to permit robust activity patterns.

  6. Development of RBMK-1500 Model for BDBA Analysis Using RELAP/SCDAPSIM Code

    Science.gov (United States)

    Uspuras, Eugenijus; Kaliatka, Algirdas

    This article discusses the specificity of RBMK (channel type, boiling water, graphite moderated) reactors and problems of Reactor Cooling System modelling employing computer codes. The article presents, how the RELAP/SCDAPSIM code, which is originally designed for modelling of accidents in vessel type reactors, is fit to simulate the phenomena in the RBMK reactor core and RCS in case of Beyond Design Basis Accidents. For this reason, use of two RELAP/SCDAPSIM models is recommended. First model with described complete geometry of RCS is recommended for analysis of initial phase of accident. The calculations results, received using this model, are used as boundary conditions in simplified model for simulation of later phases of severe accidents. The simplified model was tested comparing results of simulation performed using RELAP5 and RELAP/SCDAPSIM codes. As the typical example of BDBA, large break LOCA in reactor cooling system with failure of emergency core cooling system was analyzed. Use of developed models allows to receive behaviour of thermal-hydraulic parameters, temperatures of core components, amount of generated hydrogen due to steam-zirconium reaction. These parameters will be used as input for other special codes, designed for analysis of processes in reactor containment.

  7. Development of the thermal behavior analysis code DIRAD and the fuel design procedure for LMFBR

    Science.gov (United States)

    Nakae, N.; Tanaka, K.; Nakajima, H.; Matsumoto, M.

    1992-06-01

    It is very important to increase the fuel linear heat rating for improvement of economy in LMFBR without any degradation in safety. A reduction of the design margin is helpful to achieve the high power operation. The development of a fuel design code and a design procedure is effective on the reduction of the design margin. The thermal behavior analysis code DIRAD has been developed with respect to fuel restructuring and gap conductance models. These models have been calibrated and revised using irradiation data of fresh fuel. It is, therefore, found that the code is applicable for the thermal analysis with fresh fuel. The uncertainties in fuel irradiation condition and fuel fabrication tolerance together with the uncertainty of the code prediction have major contributions to the design margin. In the current fuel design the first two uncertainties independently contribute to temperature increment. Another method which can rationally explain the effect of the uncertainties on the temperature increment is adopted here. Then, the design margin may be rationally reduced.

  8. BILAM: a composite laminate failure-analysis code using bilinear stress-strain approximations

    Energy Technology Data Exchange (ETDEWEB)

    McLaughlin, P.V. Jr.; Dasgupta, A.; Chun, Y.W.

    1980-10-01

    The BILAM code which uses constant strain laminate analysis to generate in-plane load/deformation or stress/strain history of composite laminates to the point of laminate failure is described. The program uses bilinear stress-strain curves to model layer stress-strain behavior. Composite laminates are used for flywheels. The use of this computer code will help to develop data on the behavior of fiber composite materials which can be used by flywheel designers. In this program the stress-strain curves are modelled by assuming linear response in axial tension while using bilinear approximations (2 linear segments) for stress-strain response to axial compressive, transverse tensile, transverse compressive and axial shear loadings. It should be noted that the program attempts to empirically simulate the effects of the phenomena which cause nonlinear stress-strain behavior, instead of mathematically modelling the micromechanics involved. This code, therefore, performs a bilinear laminate analysis, and, in conjunction with several user-defined failure interaction criteria, is designed to provide sequential information on all layer failures up to and including the first fiber failure. The modus operandi is described. Code BILAM can be used to: predict the load-deformation/stress-strain behavior of a composite laminate subjected to a given combination of in-plane loads, and make analytical predictions of laminate strength.

  9. GPU上计算流体力学的加速%Acceleration of Computational Fluid Dynamics Codes on GPU

    Institute of Scientific and Technical Information of China (English)

    董廷星; 李新亮; 李森; 迟学斌

    2011-01-01

    Computational Fluid Dynamic (CFD) codes based on incompressible Navier-Stokes, compressible Euler and compressible Navier-Stokes solvers are ported on NVIDIA GPU. As validation test, we have simulated a two-dimension cavity flow, Riemann problem and a transonic flow over a RAE2822 airfoil. Maximum 33.2x speedup is reported in our test. To maximum the GPU code performance, we also explore a number of GPU-specific optimization strategies. It demonstrates GPU code gives the expected results compared CPU code and experimental result and GPU computing has good compatibility and bright future.%本文将计算流体力学中的可压缩的纳维叶-斯托克斯(Navier-Stokes),不可压缩的Navier-Stokes和欧拉(Euler)方程移植到NVIDIA GPU上.模拟了3个测试例子,2维的黎曼问题,方腔流问题和RAE2822型的机翼绕流.相比于CPU,我们在GPU平台上最高得到了33.2倍的加速比.为了最大程度提高代码的性能,针对GPU平台上探索了几种优化策略.和CPU以及实验结果对比表明,利用计算流体力学在GPU平台上能够得到预想的结果,具有很好的应用前景.

  10. Systematic analysis of coding and noncoding DNA sequences using methods of statistical linguistics

    Science.gov (United States)

    Mantegna, R. N.; Buldyrev, S. V.; Goldberger, A. L.; Havlin, S.; Peng, C. K.; Simons, M.; Stanley, H. E.

    1995-01-01

    We compare the statistical properties of coding and noncoding regions in eukaryotic and viral DNA sequences by adapting two tests developed for the analysis of natural languages and symbolic sequences. The data set comprises all 30 sequences of length above 50 000 base pairs in GenBank Release No. 81.0, as well as the recently published sequences of C. elegans chromosome III (2.2 Mbp) and yeast chromosome XI (661 Kbp). We find that for the three chromosomes we studied the statistical properties of noncoding regions appear to be closer to those observed in natural languages than those of coding regions. In particular, (i) a n-tuple Zipf analysis of noncoding regions reveals a regime close to power-law behavior while the coding regions show logarithmic behavior over a wide interval, while (ii) an n-gram entropy measurement shows that the noncoding regions have a lower n-gram entropy (and hence a larger "n-gram redundancy") than the coding regions. In contrast to the three chromosomes, we find that for vertebrates such as primates and rodents and for viral DNA, the difference between the statistical properties of coding and noncoding regions is not pronounced and therefore the results of the analyses of the investigated sequences are less conclusive. After noting the intrinsic limitations of the n-gram redundancy analysis, we also briefly discuss the failure of the zeroth- and first-order Markovian models or simple nucleotide repeats to account fully for these "linguistic" features of DNA. Finally, we emphasize that our results by no means prove the existence of a "language" in noncoding DNA.

  11. Systematic analysis of coding and noncoding DNA sequences using methods of statistical linguistics

    Science.gov (United States)

    Mantegna, R. N.; Buldyrev, S. V.; Goldberger, A. L.; Havlin, S.; Peng, C. K.; Simons, M.; Stanley, H. E.

    1995-01-01

    We compare the statistical properties of coding and noncoding regions in eukaryotic and viral DNA sequences by adapting two tests developed for the analysis of natural languages and symbolic sequences. The data set comprises all 30 sequences of length above 50 000 base pairs in GenBank Release No. 81.0, as well as the recently published sequences of C. elegans chromosome III (2.2 Mbp) and yeast chromosome XI (661 Kbp). We find that for the three chromosomes we studied the statistical properties of noncoding regions appear to be closer to those observed in natural languages than those of coding regions. In particular, (i) a n-tuple Zipf analysis of noncoding regions reveals a regime close to power-law behavior while the coding regions show logarithmic behavior over a wide interval, while (ii) an n-gram entropy measurement shows that the noncoding regions have a lower n-gram entropy (and hence a larger "n-gram redundancy") than the coding regions. In contrast to the three chromosomes, we find that for vertebrates such as primates and rodents and for viral DNA, the difference between the statistical properties of coding and noncoding regions is not pronounced and therefore the results of the analyses of the investigated sequences are less conclusive. After noting the intrinsic limitations of the n-gram redundancy analysis, we also briefly discuss the failure of the zeroth- and first-order Markovian models or simple nucleotide repeats to account fully for these "linguistic" features of DNA. Finally, we emphasize that our results by no means prove the existence of a "language" in noncoding DNA.

  12. Dynamic Allocation and Efficient Distribution of Data Among Multiple Clouds Using Network Coding

    DEFF Research Database (Denmark)

    Sipos, Marton A.; Fitzek, Frank; Roetter, Daniel Enrique Lucani

    2014-01-01

    Distributed storage has attracted large interest lately from both industry and researchers as a flexible, cost-efficient, high performance, and potentially secure solution for geographically distributed data centers, edge caching or sharing storage among users. This paper studies the benefits...... of random linear network coding to exploit multiple commercially available cloud storage providers simultaneously with the possibility to constantly adapt to changing cloud performance in order to optimize data retrieval times. The main contribution of this paper is a new data distribution mechanisms...

  13. BROMOCEA Code: An Improved Grand Canonical Monte Carlo/Brownian Dynamics Algorithm Including Explicit Atoms.

    Science.gov (United States)

    Solano, Carlos J F; Pothula, Karunakar R; Prajapati, Jigneshkumar D; De Biase, Pablo M; Noskov, Sergei Yu; Kleinekathöfer, Ulrich

    2016-05-10

    All-atom molecular dynamics simulations have a long history of applications studying ion and substrate permeation across biological and artificial pores. While offering unprecedented insights into the underpinning transport processes, MD simulations are limited in time-scales and ability to simulate physiological membrane potentials or asymmetric salt solutions and require substantial computational power. While several approaches to circumvent all of these limitations were developed, Brownian dynamics simulations remain an attractive option to the field. The main limitation, however, is an apparent lack of protein flexibility important for the accurate description of permeation events. In the present contribution, we report an extension of the Brownian dynamics scheme which includes conformational dynamics. To achieve this goal, the dynamics of amino-acid residues was incorporated into the many-body potential of mean force and into the Langevin equations of motion. The developed software solution, called BROMOCEA, was applied to ion transport through OmpC as a test case. Compared to fully atomistic simulations, the results show a clear improvement in the ratio of permeating anions and cations. The present tests strongly indicate that pore flexibility can enhance permeation properties which will become even more important in future applications to substrate translocation.

  14. Motion-compensated coding and frame rate up-conversion: models and analysis.

    Science.gov (United States)

    Dar, Yehuda; Bruckstein, Alfred M

    2015-07-01

    Block-based motion estimation (ME) and motion compensation (MC) techniques are widely used in modern video processing algorithms and compression systems. The great variety of video applications and devices results in diverse compression specifications, such as frame rates and bit rates. In this paper, we study the effect of frame rate and compression bit rate on block-based ME and MC as commonly utilized in inter-frame coding and frame rate up-conversion (FRUC). This joint examination yields a theoretical foundation for comparing MC procedures in coding and FRUC. First, the video signal is locally modeled as a noisy translational motion of an image. Then, we theoretically model the motion-compensated prediction of available and absent frames as in coding and FRUC applications, respectively. The theoretic MC-prediction error is studied further and its autocorrelation function is calculated, yielding useful separable-simplifications for the coding application. We argue that a linear relation exists between the variance of the MC-prediction error and temporal distance. While the relevant distance in MC coding is between the predicted and reference frames, MC-FRUC is affected by the distance between the frames available for interpolation. We compare our estimates with experimental results and show that the theory explains qualitatively the empirical behavior. Then, we use the models proposed to analyze a system for improving of video coding at low bit rates, using a spatio-temporal scaling. Although this concept is practically employed in various forms, so far it lacked a theoretical justification. We here harness the proposed MC models and present a comprehensive analysis of the system, to qualitatively predict the experimental results.

  15. HYDRA-II: A hydrothermal analysis computer code: Volume 3, Verification/validation assessments

    Energy Technology Data Exchange (ETDEWEB)

    McCann, R.A.; Lowery, P.S.

    1987-10-01

    HYDRA-II is a hydrothermal computer code capable of three-dimensional analysis of coupled conduction, convection, and thermal radiation problems. This code is especially appropriate for simulating the steady-state performance of spent fuel storage systems. The code has been evaluated for this application for the US Department of Energy's Commercial Spent Fuel Management Program. HYDRA-II provides a finite difference solution in cartesian coordinates to the equations governing the conservation of mass, momentum, and energy. A cylindrical coordinate system may also be used to enclose the cartesian coordinate system. This exterior coordinate system is useful for modeling cylindrical cask bodies. The difference equations for conservation of momentum are enhanced by the incorporation of directional porosities and permeabilities that aid in modeling solid structures whose dimensions may be smaller than the computational mesh. The equation for conservation of energy permits modeling of orthotropic physical properties and film resistances. Several automated procedures are available to model radiation transfer within enclosures and from fuel rod to fuel rod. The documentation of HYDRA-II is presented in three separate volumes. Volume I - Equations and Numerics describes the basic differential equations, illustrates how the difference equations are formulated, and gives the solution procedures employed. Volume II - User's Manual contains code flow charts, discusses the code structure, provides detailed instructions for preparing an input file, and illustrates the operation of the code by means of a model problem. This volume, Volume III - Verification/Validation Assessments, provides a comparison between the analytical solution and the numerical simulation for problems with a known solution. This volume also documents comparisons between the results of simulations of single- and multiassembly storage systems and actual experimental data. 11 refs., 55 figs., 13 tabs.

  16. HYDRA-II: A hydrothermal analysis computer code: Volume 2, User's manual

    Energy Technology Data Exchange (ETDEWEB)

    McCann, R.A.; Lowery, P.S.; Lessor, D.L.

    1987-09-01

    HYDRA-II is a hydrothermal computer code capable of three-dimensional analysis of coupled conduction, convection, and thermal radiation problems. This code is especially appropriate for simulating the steady-state performance of spent fuel storage systems. The code has been evaluated for this application for the US Department of Energy's Commercial Spent Fuel Management Program. HYDRA-II provides a finite-difference solution in cartesian coordinates to the equations governing the conservation of mass, momentum, and energy. A cylindrical coordinate system may also be used to enclose the cartesian coordinate system. This exterior coordinate system is useful for modeling cylindrical cask bodies. The difference equations for conservation of momentum incorporate directional porosities and permeabilities that are available to model solid structures whose dimensions may be smaller than the computational mesh. The equation for conservation of energy permits modeling of orthotropic physical properties and film resistances. Several automated methods are available to model radiation transfer within enclosures and from fuel rod to fuel rod. The documentation of HYDRA-II is presented in three separate volumes. Volume 1 - Equations and Numerics describes the basic differential equations, illustrates how the difference equations are formulated, and gives the solution procedures employed. This volume, Volume 2 - User's Manual, contains code flow charts, discusses the code structure, provides detailed instructions for preparing an input file, and illustrates the operation of the code by means of a sample problem. The final volume, Volume 3 - Verification/Validation Assessments, provides a comparison between the analytical solution and the numerical simulation for problems with a known solution. 6 refs.

  17. SIMA: Python software for analysis of dynamic fluorescence imaging data

    Directory of Open Access Journals (Sweden)

    Patrick eKaifosh

    2014-09-01

    Full Text Available Fluorescence imaging is a powerful method for monitoring dynamic signals in the nervous system. However, analysis of dynamic fluorescence imaging data remains burdensome, in part due to the shortage of available software tools. To address this need, we have developed SIMA, an open source Python package that facilitates common analysis tasks related to fluorescence imaging. Functionality of this package includes correction of motion artifacts occurring during in vivo imaging with laser-scanning microscopy, segmentation of imaged fields into regions of interest (ROIs, and extraction of signals from the segmented ROIs. We have also developed a graphical user interface (GUI for manual editing of the automatically segmented ROIs and automated registration of ROIs across multiple imaging datasets. This software has been designed with flexibility in mind to allow for future extension with different analysis methods and potential integration with other packages. Software, documentation, and source code for the SIMA package and ROI Buddy GUI are freely available at http://www.losonczylab.org/sima/.

  18. Nonlinear dynamics and quantitative EEG analysis.

    Science.gov (United States)

    Jansen, B H

    1996-01-01

    Quantitative, computerized electroencephalogram (EEG) analysis appears to be based on a phenomenological approach to EEG interpretation, and is primarily rooted in linear systems theory. A fundamentally different approach to computerized EEG analysis, however, is making its way into the laboratories. The basic idea, inspired by recent advances in the area of nonlinear dynamics and chaos theory, is to view an EEG as the output of a deterministic system of relatively simple complexity, but containing nonlinearities. This suggests that studying the geometrical dynamics of EEGs, and the development of neurophysiologically realistic models of EEG generation may produce more successful automated EEG analysis techniques than the classical, stochastic methods. A review of the fundamentals of chaos theory is provided. Evidence supporting the nonlinear dynamics paradigm to EEG interpretation is presented, and the kind of new information that can be extracted from the EEG is discussed. A case is made that a nonlinear dynamic systems viewpoint to EEG generation will profoundly affect the way EEG interpretation is currently done.

  19. A mono-dimensional nuclear fuel performance analysis code, PUMA, development from a coupled approach

    Energy Technology Data Exchange (ETDEWEB)

    Cheon, J. S.; Lee, B. O.; Lee, C. B. [Korea Atomic Energy Research Institute, 989-111 Daedeok-daero, Yuseong, Daejeon, 305-353 (Korea, Republic of); Yacout, A. M. [Argonne National Laboratory, 9700 South Cass Avenue, Argonne, IL 60439 (United States)

    2013-07-01

    Multidimensional-multi-physical phenomena in nuclear fuels are treated as a set of mono-dimensional-coupled problems which encompass heat, displacement, fuel constituent redistribution, and fission gas release. Rather than uncoupling these coupled equations as in conventional fuel performance analysis codes, efforts are put into to obtain fully coupled solutions by relying on the recent advances of numerical analysis. Through this approach, a new SFR metal fuel performance analysis code, called PUMA (Performance of Uranium Metal fuel rod Analysis code) is under development. Although coupling between temperature and fuel constituent was made easily, the coupling between the mechanical equilibrium equation and a set of stiff kinetics equations for fission gas release is accomplished by introducing one-level Newton scheme through backward differentiation formula. Displacement equations from 1D finite element formulation of the mechanical equilibrium equation are solved simultaneously with stress equation, creep equation, swelling equation, and FGR equations. Calculations was made successfully such that the swelling and the hydrostatic pressure are interrelated each other. (authors)

  20. Error Correction Coding Meets Cyber-Physical Systems: Message-Passing Analysis of Self-Healing Interdependent Networks

    CERN Document Server

    Behfarnia, Ali

    2016-01-01

    Coupling cyber and physical systems gives rise to numerous engineering challenges and opportunities. An important challenge is the contagion of failure from one system to another, which can lead to large-scale cascading failures. However, the \\textit{self-healing} ability emerges as a valuable opportunity where the overlaying cyber network can cure failures in the underlying physical network. To capture both self-healing and contagion, this paper considers a graphical model representation of an interdependent cyber-physical system, in which nodes represent various cyber or physical functionalities, and edges capture the interactions between the nodes. A message-passing algorithm used in low-density parity-check codes is extended to this representation to study the dynamics of failure propagation and healing. By applying a density evolution analysis to this algorithm, network reaction to initial disruptions is investigated. It is proved that as the number of message-passing iterations increases, the network re...