WorldWideScience

Sample records for benchmark experiments

  1. Aeroelastic Benchmark Experiments Project

    Data.gov (United States)

    National Aeronautics and Space Administration — M4 Engineering proposes to conduct canonical aeroelastic benchmark experiments. These experiments will augment existing sources for aeroelastic data in the...

  2. Benchmark experiments for nuclear data

    International Nuclear Information System (INIS)

    Benchmark experiments offer the most direct method for validation of nuclear data. Benchmark experiments for several areas of application of nuclear data were specified by CSEWG. These experiments are surveyed and tests of recent versions of ENDF/B are presented. (U.S.)

  3. TRIGA Mark II benchmark experiment

    International Nuclear Information System (INIS)

    The experimental results of startup tests after reconstruction and modification of the TRIGA Mark II reactor in Ljubljana are presented. The experiments were performed with a completely fresh, compact, and uniform core. The operating conditions were well defined and controlled, so that the results can be used as a benchmark test case for TRIGA reactor calculations. Both steady-state and pulse mode operation were tested. In this paper, the following steady-state experiments are treated: critical core and excess reactivity, control rod worths, fuel element reactivity worth distribution, fuel temperature distribution, and fuel temperature reactivity coefficient

  4. Analysis of VENUS-3 benchmark experiment

    International Nuclear Information System (INIS)

    The paper presents the revision and the analysis of VENUS-3 benchmark experiment performed at CEN/SCK, Mol (Belgium). This benchmark was found to be particularly suitable for validation of current calculation tools like 3-D neutron transport codes, and in particular of the 3D sensitivity and uncertainty analysis code developed within the EFF project. The compilation of the integral experiment was integrated into the SINBAD electronic data base for storing and retrieving information about the shielding experiments for nuclear systems. SINBAD now includes 33 reviewed benchmark descriptions and several compilations waiting for the review, among them many benchmarks relevant for pressure vessel dosimetry system validation.(author)

  5. The design and analysis of benchmark experiments

    OpenAIRE

    Hothorn, Torsten; Leisch, Friedrich; Zeileis, Achim; Hornik, Kurt

    2003-01-01

    The assessment of the performance of learners by means of benchmark experiments is established exercise. In practice, benchmark studies are a tool to compare the performance of several competing algorithms for a certain learning problem. Cross-validation or resampling techniques are commonly used to derive point estimates of the performances which are compared to identify algorithms with good properties. For several benchmarking problems, test procedures taking the variability of those point ...

  6. TRIGA Mark II benchmark experiment

    International Nuclear Information System (INIS)

    Experimental results of pulse parameters and control rod worth measurements at TRIGA Mark 2 reactor in Ljubljana are presented. The measurements were performed with a completely fresh, uniform, and compact core. Only standard fuel elements with 12 wt% uranium were used. Special efforts were made to get reliable and accurate results at well-defined experimental conditions, and it is proposed to use the results as a benchmark test case for TRIGA reactors

  7. Direct Simulation of a Solidification Benchmark Experiment

    OpenAIRE

    Carozzani, Tommy; Gandin, Charles-André; Digonnet, Hugues; Bellet, Michel; Zaidat, Kader; Fautrelle, Yves

    2013-01-01

    International audience A solidification benchmark experiment is simulated using a three-dimensional cellular automaton-finite element solidification model. The experiment consists of a rectangular cavity containing a Sn-3 wt pct Pb alloy. The alloy is first melted and then solidified in the cavity. A dense array of thermocouples permits monitoring of temperatures in the cavity and in the heat exchangers surrounding the cavity. After solidification, the grain structure is revealed by metall...

  8. Benchmark calculations of sodium fast critical experiments

    International Nuclear Information System (INIS)

    The high expectations from fast critical experiments impose the additional requirements on reliability of final reconstructed values, obtained in experiments at critical facility. Benchmark calculations of critical experiments are characterized by impossibility of complete experiment reconstruction, the large amounts of input data (dependent and independent) with very different reliability. It should also take into account different sensitivity of the measured and appropriate calculated characteristics to the identical changes of geometry parameters, temperature, and isotopic composition of individual materials. The calculations of critical facility experiments are produced for the benchmark models, generated by the specific reconstructing codes with its features when adjusting model parameters, and using the nuclear data library. The generated benchmark model, providing the agreed calculated and experimental values for one or more neutronic characteristics can lead to considerable differences for other key characteristics. The sensitivity of key neutronic characteristics to the extra steel allocation in the core, and ENDF/B nuclear data sources is performed using a few calculated models of BFS-62-3A and BFS1-97 critical assemblies. The comparative analysis of the calculated effective multiplication factor, spectral indices, sodium void reactivity, and radial fission-rate distributions leads to quite different models, providing the best agreement the calculated and experimental neutronic characteristics. This fact should be considered during the refinement of computational models and code-verification purpose. (author)

  9. Uncertainty analysis of benchmark experiments using MCBEND

    International Nuclear Information System (INIS)

    Differences between measurement and calculation for shielding benchmark experiments can arise from uncertainties in a number of areas including nuclear data, radiation transport modelling, source specification, geometry modelling, measurement, and calculation statistics. In order to understand the significance of these differences, detailed sensitivity analysis of these various uncertainties is required. This is of particular importance when considering the requirements for nuclear data improvements aimed at providing better agreement between calculation and measurement. As part of a programme of validation activity associated with the international JEFF data project, the Monte Carlo code MCBEND has been used to analyse a range of benchmark experiments using JEF-2.2 based nuclear data together with modern dosimetry data. This paper describes detailed uncertainty analyses that have been performed for the following Winfrith material benchmark experiments: graphite, water, iron, graphite/steel and steel/water. Conclusions are reported and compared with calculations using other nuclear data libraries. In addition, the effect that nuclear data uncertainties have on the calculated results is discussed by making use of the data adjustment code DATAK. Requirements for further nuclear data evaluation arising from this work are identified. (author)

  10. Adiabatic microcalorimetry in shielding benchmark experiments

    International Nuclear Information System (INIS)

    The application of a newly developed microcalorimeter is described : (1) for measuring energy-deposition rates in the mixed radiation fields of zero-energy reactors and shielding benchmark experiments. Methods of calculation for energy-deposition (n + γ) are usually validated by measuring the neutron component with the aid of activation detectors and the gamma-ray component using TLDs or ion-chambers. The major limitation with the use of calorimeters in low-power radiation fields has been lack of sensitivity. The aim of the present work has been to investigate the performance of a calorimeter which can measure heating-rates in the range down to 10 μW/g by comparison with conventional dosimetry techniques in a graphite benchmark experiment conducted in the NESSUS reference field in the NESTOR reactor at Winfrith. Major problems have been encountered with the neutron sensitivity of both TLDs and gamma-ray ion-chambers. When appropriate corrections are made good agreement can be achieved between all the dosimetry techniques and the results provide a benchmark test of calculational methods for energy-deposition in graphite. In power reactors, steel-walled calorimeters are used for the dosimetry of materials such as graphite and the net effect of electron migration between the sample and steel walls significantly increases the heating rate in the specimen. In the NESSUS experiments, an increase of 18% was observed in the graphite heating rate above that expected from the enhanced gamma source, when the calorimeter wall was changed from graphite to iron. (author)

  11. Consistent utilization of shielding benchmark experiments

    Energy Technology Data Exchange (ETDEWEB)

    D' Angelo, A. (Univ. of Calabria, Cosenza, Italy); Oliva, A.; Palmiotti, G.; Salvatores, M.; Zero, S.

    1978-03-01

    Benchmark experiments of neutron propagation in iron and iron--sodium mixtures were used to generate an ''adjusted'' ENDF/B data file for iron, Mat = 1192. In particular, the secondary neutron energy distribution in the continuous level energy range was adjusted by use of such high-energy responses as the /sup 32/S(n,p)/sup 32/P reaction, which are significantly sensitive to changes in that probability distribution. The experimental analysis used carefully checked two-dimensional transport methods to avoid bias in the adjustment procedure due to inadequate calculational methods. 10 figures, 13 tables.

  12. Benchmark analysis of KRITZ-2 critical experiments

    International Nuclear Information System (INIS)

    In the KRITZ-2 critical experiments, criticality and pin power distributions were measured at room temperature and high temperature (about 245 degC) for three different cores (KRITZ-2:1, KRITZ-2:13, KRITZ-2:19) loading slightly enriched UO2 or MOX fuels. Recently, international benchmark problems were provided by ORNL and OECD/NEA based on the KRITZ-2 experimental data. The published experimental data for the system with slightly enriched fuels at high temperature are rare in the world and they are valuable for nuclear data testing. Thus, the benchmark analysis was carried out with a continuous-energy Monte Carlo code MVP and its four nuclear data libraries based on JENDL-3.2, JENDL-3.3, JEF-2.2 and ENDF/B-VI.8. As a result, fairly good agreements with the experimental data were obtained with any libraries for the pin power distributions. However, the JENDL-3.3 and ENDF/B-VI.8 give under-prediction of criticality and too negative isothermal temperature coefficients for slightly enriched UO2 cores, although the older nuclear data JENDL-3.2 and JEF-2.2 give rather good agreements with the experimental data. From the detailed study with an infinite unit cell model, it was found that the differences among the results with different libraries are mainly due to the different fission cross section of U-235 in the energy range below 1.0 eV. (author)

  13. Karma1.1 benchmark calculations for the numerical benchmark problems and the critical experiments

    International Nuclear Information System (INIS)

    The transport lattice code KARMA 1.1 has been developed at KAERI for the reactor physics analysis of the pressurized water reactor. This program includes the multi-group library processed from ENDF/B-VI R8 and also utilizes the macroscopic cross sections for the benchmark problems. Benchmark calculations were performed for the C5G7 and the KAERI benchmark problems given with seven group cross sections, for various fuels loaded in the operating pressurized water reactors in South Korea, and for the critical experiments including CE, B and W and KRITZ. Benchmark results show that KARMA 1.1 is working reasonably. (author)

  14. Quantitative consistency testing of thermal benchmark lattice experiments

    International Nuclear Information System (INIS)

    The paper sets forth a general method to demonstrate the quantitative consistency (or inconsistency) of results of thermal reactor lattice experiments. The method is of particular importance in selecting standard ''benchmark'' experiments for comparison testing of lattice analysis codes and neutron cross sections. ''Benchmark'' thermal lattice experiments are currently selected by consensus, which usually means the experiment is geometrically simple, well-documented, reasonably complete, and qualitatively consistent. A literature search has not revealed any general quantitative test that has been applied to experimental results to demonstrate consistency, although some experiments must have been subjected to some form or other of quantitative test. The consistency method is based on a two-group neutron balance condition that is capable of revealing the quantitative consistency (or inconsistency) of reported thermal benchmark lattice integral parameters. This equation is used in conjunction with a second equation in the following discussion to assess the consistency (or inconsistency) of: (1) several Cross Section Evaluation Working Group (CSEWG) defined thermal benchmark lattices, (2) SRL experiments on the Mark 5R and Mark 15 lattices, and (3) several D2O lattices encountered as proposed thermal benchmark lattices. Nineteen thermal benchmark lattice experiments were subjected to a quantitative test of consistency between the reported experimental integral parameters. Results of this testing showed only two lattice experiments to be generally useful as ''benchmarks,'' three lattice experiments to be of limited usefulness, three lattice experiments to be potentially useful, and 11 lattice experiments to be not useful. These results are tabulated with the lattices identified

  15. Introduction to 'International Handbook of Criticality Safety Benchmark Experiments'

    International Nuclear Information System (INIS)

    The Criticality Safety Benchmark Evaluation Project (CSBEP) was initiated in 1992 by the United States Department of Energy. The project quickly became an international effort as scientists from other interested countries became involved. The International Criticality Safety Benchmark Evaluation Project (ICSBEP) is now an official activity of the Organization for Economic Cooperation and Development-Nuclear Energy Agency (OECD-NEA). 'International Handbook of Criticality Safety Benchmark Experiments' was prepared and is updated year by year by the working group of the project. This handbook contains criticality safety benchmark specifications that have been derived from experiments that were performed at various nuclear critical facilities around the world. The benchmark specifications are intended for use by criticality safety engineers to validate calculation techniques used. The author briefly introduces the informative handbook and would like to encourage Japanese engineers who are in charge of nuclear criticality safety to use the handbook. (author)

  16. HPC in Java: Experiences in Implementing the NAS Parallel Benchmarks

    OpenAIRE

    Amedro, Brian; Caromel, Denis; Huet, Fabrice; Bodnartchouk, Vladimir; Delbé, Christian; L. Taboada, Guillermo

    2010-01-01

    This paper reports on the design, implementation and benchmarking of a Java version of the Nas Parallel Benchmarks. We first briefly describe the implementation and the performance pitfalls. We then compare the overall performance of the Fortran MPI (PGI) version with a Java implementation using the ProActive middleware for distribution. All Java experiments were conducted on virtual machines with different vendors and versions. We show that the performance varies with the type of computation...

  17. Benchmark exercise on SBLOCA experiment of PWR PACTEL facility

    International Nuclear Information System (INIS)

    Highlights: • PWR PACTEL, the facility with EPR type steam generators, is introduced. • The focus of the benchmark was on the analyses of the SBLOCA test with PWR PACTEL. • System codes with several modeling approaches were utilized to analyze the test. • Proper consideration of heat and pressure losses improves simulation remarkably. - Abstract: The PWR PACTEL benchmark exercise was organized in Lappeenranta, Finland by Lappeenranta University of Technology. The benchmark consisted of two phases, i.e. a blind and an open calculation task. Seven organizations from the Czech Republic, Germany, Italy, Sweden and Finland participated in the benchmark exercise, and four system codes were utilized in the benchmark simulation tasks. Two workshops were organized for launching and concluding the benchmark, the latter of which involved presentations of the calculation results as well as discussions on the related modeling issues. The chosen experiment for the benchmark was a small break loss of coolant accident experiment which was performed to study the natural circulation behavior over a continuous range of primary side coolant inventories. For the blind calculation task, the detailed facility descriptions, the measured pressure and heat losses as well as the results of a short characterizing transient were provided. For the open calculation task part, the experiment results were released. According to the simulation results, the benchmark experiment was quite challenging to model. Several improvements were found and utilized especially for the open calculation case. The issues concerned model construction, heat and pressure losses impact, interpreting measured and calculated data, non-condensable gas effect, testing several condensation and CCFL correlations, sensitivity studies, as well as break modeling. There is a clear need for user guidelines or for a collection of best practices in modeling for every code. The benchmark offered a unique opportunity to test

  18. Pre-evaluation of fusion shielding benchmark experiment

    Energy Technology Data Exchange (ETDEWEB)

    Hayashi, K.; Handa, H. [Hitachi Engineering Company, Ltd., Ibaraki (Japan); Konno, C. [Japan Atomic Energy Research Inst. Ibaraki (Japan)] [and others

    1994-12-31

    Shielding benchmark experiment is very useful to test the design code and nuclear data for fusion devices. There are many types of benchmark experiments that should be done in fusion shielding problems, but time and budget are limited. Therefore it will be important to select and determine the effective experimental configurations by precalculation before the experiment. The authors did three types of pre-evaluation to determine the experimental assembly configurations of shielding benchmark experiments planned in FNS, JAERI. (1) Void Effect Experiment - The purpose of this experiment is to measure the local increase of dose and nuclear heating behind small void(s) in shield material. Dimension of the voids and its arrangements were decided as follows. Dose and nuclear heating were calculated both for with and without void(s). Minimum size of the void was determined so that the ratio of these two results may be larger than error of the measurement system. (2) Auxiliary Shield Experiment - The purpose of this experiment is to measure shielding properties of B{sub 4}C, Pb, W, and dose around superconducting magnet (SCM). Thickness of B{sub 4}C, Pb, W and their arrangement including multilayer configuration were determined. (3) SCM Nuclear Heating Experiment - The purpose of this experiment is to measure nuclear heating and dose distribution in SCM material. Because it is difficult to use liquid helium as a part of SCM mock up material, material composition of SCM mock up are surveyed to have similar nuclear heating property of real SCM composition.

  19. Providing Nuclear Criticality Safety Analysis Education through Benchmark Experiment Evaluation

    International Nuclear Information System (INIS)

    One of the challenges that today's new workforce of nuclear criticality safety engineers face is the opportunity to provide assessment of nuclear systems and establish safety guidelines without having received significant experience or hands-on training prior to graduation. Participation in the International Criticality Safety Benchmark Evaluation Project (ICSBEP) and/or the International Reactor Physics Experiment Evaluation Project (IRPhEP) provides students and young professionals the opportunity to gain experience and enhance critical engineering skills.

  20. Analyses of shielding benchmark experiments in thick iron

    Energy Technology Data Exchange (ETDEWEB)

    Oka, Y.; Ohashi, S.; Hashikura, H.; An, S. (Tokyo Univ. (Japan). Nuclear Engineering Research Lab.)

    1981-03-01

    The results of the iron related shielding benchmark experiments that have been performed in the world are summarized and analysed by using discrete ordinates transport calculations. Analyses of three one-dimensional experiments show good agreement above 10/sup 5/ eV and disagreement below 10/sup 5/ eV in neutron spectra. Analyses of four two-dimensional experiments reveal overestimation of threshold-type reaction rates up to 40 cm iron thickness and underestimation above 50 cm.

  1. International handbook of evaluated criticality safety benchmark experiments

    International Nuclear Information System (INIS)

    The Criticality Safety Benchmark Evaluation Project (CSBEP) was initiated in October of 1992 by the United States Department of Energy. The project quickly became an international effort as scientists from other interested countries became involved. The International Criticality Safety Benchmark Evaluation Project (ICSBEP) became an official activity of the Organization for Economic Cooperation and Development - Nuclear Energy Agency (OECD-NEA) in 1995. This handbook contains criticality safety benchmark specifications that have been derived from experiments performed at various nuclear critical facilities around the world. The benchmark specifications are intended for use by criticality safety engineers to validate calculational techniques used to establish minimum subcritical margins for operations with fissile material and to determine criticality alarm requirement and placement. Many of the specifications are also useful for nuclear data testing. Example calculations are presented; however, these calculations do not constitute a validation of the codes or cross section data. The evaluated criticality safety benchmark data are given in nine volumes. These volumes span over 55,000 pages and contain 516 evaluations with benchmark specifications for 4,405 critical, near critical, or subcritical configurations, 24 criticality alarm placement / shielding configurations with multiple dose points for each, and 200 configurations that have been categorized as fundamental physics measurements that are relevant to criticality safety applications. Experiments that are found unacceptable for use as criticality safety benchmark experiments are discussed in these evaluations; however, benchmark specifications are not derived for such experiments (in some cases models are provided in an appendix). Approximately 770 experimental configurations are categorized as unacceptable for use as criticality safety benchmark experiments. Additional evaluations are in progress and will be

  2. Pre-evaluation of fusion shielding benchmark experiment

    Energy Technology Data Exchange (ETDEWEB)

    Hayashi, K. [Hitachi Engineering Co. Ltd., Ibaraki (Japan); Handa, H. [Hitachi Engineering Co. Ltd., Ibaraki (Japan); Konno, C. [Japan Atomic Energy Research Institute, Tokai-mura, Ibaraki, 319-11 (Japan); Maekawa, F. [Japan Atomic Energy Research Institute, Tokai-mura, Ibaraki, 319-11 (Japan); Maekawa, H. [Japan Atomic Energy Research Institute, Tokai-mura, Ibaraki, 319-11 (Japan); Maki, K. [Energy Research Laboratory, Hitachi Ltd., Omika-cho, Hitachi, Ibaraki, 316 (Japan); Yamada, K. [Business Automation Co. Ltd., Toranomon, 1-24-10, Minato-ku, Tokyo, 105 (Japan); Abe, T. [Business Automation Co. Ltd., Toranomon, 1-24-10, Minato-ku, Tokyo, 105 (Japan)

    1995-03-01

    A shielding benchmark experiment is very useful to test the design code and nuclear data for fusion devices. There are many types of benchmark experiment that should be done in fusion shielding problems, but time and budget are limited. Therefore it will be important to select and determine the effective experimental configurations by precalculation before the experiment. We did pre-evaluations of three types of shielding benchmark experiment to determine the experimental assembly configurations. The types of experiment discussed are the void effect experiment, the auxiliary shield experiment, and the SCM (superconductive magnet) nuclear heating experiment. These calculations were performed by using two-dimensional discrete ordinate transport code DOT3.5 with first collision source prepared by GRTUNCL code. Group constants used was FUSION-40 (neutron 42 group, photon 21 group, P5 Legendre expansion) processed from Japanese Evaluated Nuclear Data Library JENDL-3. All three types of configuration were finally determined with consideration of detector efficiencies and measurement time. (orig.).

  3. International handbook of evaluated criticality safety benchmark experiments

    International Nuclear Information System (INIS)

    The primary purpose of the International Criticality Safety Benchmark Evaluation Project (ICSBEP) Working Group is to compile critical and subcritical benchmark experiment data into a standardised format that allows criticality safety analysts to easily use the data to validate calculation tools and cross-section libraries. ICSBEP work includes: - Identifying a comprehensive set of critical benchmark data and, to the extent possible, verify the data by reviewing original and subsequently revised documentation, and by talking with the experimenters or individuals who are familiar with the experimenters or the experimental facility; - Evaluating the data and quantify overall uncertainties through various types of sensitivity analysis; - Compiling the data into a standardised format; - Performing calculations of each experiment with standard criticality safety codes; - Formally documenting the work into a single source of verified benchmark critical data. The work of the ICSBEP is documented as an International Handbook of Evaluated Criticality Safety Benchmark Experiments (ICSBEP Handbook). Currently, the handbook spans nearly 67,000 pages and contains 561 evaluations representing 4839 critical, near-critical, or subcritical configurations, 24 criticality alarm placement/shielding configurations with multiple dose points for each, and 207 configurations that have been categorised as fundamental physics measurements that are relevant to criticality safety applications. The handbook is intended for use by criticality safety analysts to perform necessary validations of their calculational techniques and is expected to be a valuable tool for decades to come. The ICSBEP Handbook is produced in electronic format (pdf files) where the experiments are grouped into evaluations and categorised by: fissile media (plutonium, highly enriched uranium, intermediate and mixed enrichment uranium, low enriched uranium, uranium-233, mixed plutonium-uranium and special isotope systems

  4. Bulk shielding benchmark experiment at Frascati neutron generator (FNG)

    Energy Technology Data Exchange (ETDEWEB)

    Batistoni, P.; Angelone, M.; Martone, M.; Pillon, M.; Rado, V. [Associazione EURATOM-ENEA sulla Fusione, Frascati (Italy); Santamarina, A.; Abidi, I.; Gastaldi, B.; Martini, M.; Marquette, J.P. [CEA Centre d`Etudes de Cadarache, 13 - Saint-Paul-lez-Durance (France)

    1994-11-01

    In the framework of the European Fusion Technology Program, ENEA (Italian Agency for New Technologies, Energy and the Environment) - Frascati and CEA (Commissariat a` l`Energie Atomique) - Cadarache, in collaboration performed a bulk shielding benchmark experiment, using the 14-MeV Frascati neutron generator (FNG), aimed at obtaining accurate experimental data for improving the nuclear data base and methods used in shielding designs. The experiment consisted of the irradiation of a stainless steel block by 14-MeV neutrons. The experimental results have been compared with numerical results calculated using both Sn and Monte Carlo transport codes and the cross section library EFF.1 (european fusion file).

  5. TRIGA Mark II Criticality Benchmark Experiment with Burned Fuel

    International Nuclear Information System (INIS)

    The experimental results of criticality benchmark experiments performed at the Jozef Stefan Institute TRIGA Mark II reactor are presented. The experiments were performed with partly burned fuel in two compact and uniform core configurations in the same arrangements as were used in the fresh fuel criticality benchmark experiment performed in 1991. In the experiments, both core configurations contained only 12 wt% U-ZrH fuel with 20% enriched uranium. The first experimental core contained 43 fuel elements with average burnup of 1.22 MWd or 2.8% 235U burned. The last experimental core configuration was composed of 48 fuel elements with average burnup of 1.15 MWd or 2.6% 235U burned. The experimental determination of keff for both core configurations, one subcritical and one critical, are presented. Burnup for all fuel elements was calculated in two-dimensional four-group diffusion approximation using the TRIGLAV code. The burnup of several fuel elements was measured also by the reactivity method

  6. Benchmark enclosure fire suppression experiments - phase 1 test report.

    Energy Technology Data Exchange (ETDEWEB)

    Figueroa, Victor G.; Nichols, Robert Thomas; Blanchat, Thomas K.

    2007-06-01

    A series of fire benchmark water suppression tests were performed that may provide guidance for dispersal systems for the protection of high value assets. The test results provide boundary and temporal data necessary for water spray suppression model development and validation. A review of fire suppression in presented for both gaseous suppression and water mist fire suppression. The experimental setup and procedure for gathering water suppression performance data are shown. Characteristics of the nozzles used in the testing are presented. Results of the experiments are discussed.

  7. Rushover analysis of concrete shear walls: benchmarking of CAMUS experiment

    International Nuclear Information System (INIS)

    A number of shake table tests have been conducted on the scaled down model of a concrete shear wall as part of CAMUS experiment in France. The experiments were conducted between 1996 and 1998 in the CEA facilities in Saclay. Benchmarking of CAMUS experiments was undertaken as a part of the coordinated research program initiated by IAEA. Technique of deflection based method was adopted for benchmarking exercise. Pushover analysis, which is an important component of deflection based method is an efficient analytical tool to look into the additional capacities to withstand seismic loading effect that a structural system can offer considering the effect of redundancies and inelastic deformation. Pushover analysis tries to determine the response of the structures against various values of base shears in order to plot the force displacement (base shear - roof displacement) curve of the structure. This is done by step-by-step static non-linear analysis of the structure with increasing value of load. For calculating the performance point of the structure corresponding to a specified input motion, established procedures like FEMA-273, ATC-40 are already available. (authors)

  8. Benchmarking Benchmarks

    NARCIS (Netherlands)

    D.C. Blitz (David)

    2011-01-01

    textabstractBenchmarking benchmarks is a bundle of six studies that are inspired by the prevalence of benchmarking in academic finance research as well as in investment practice. Three studies examine if current benchmark asset pricing models adequately describe the cross-section of stock returns. W

  9. Benchmark-experiments for Pb and Bi neutron data testing

    International Nuclear Information System (INIS)

    The expedience of accurate estimation of neutron data for Pb and Bi has increased recently in connection with the Accelerator-driven system (ADS) projects and the new generation fast reactors under development, which shall use lead or lead-bismuth coolant. Still the significant difference (10%) in the energy range of 100 keV - 500 keV, for the σtot from various data sets has been observed. The differences found are associated with the energy range, for which experimental information is lacking. The situation with Bi data is not better. In this connection, several benchmarks were assembled at BFS with uranium and plutonium fuel and lead or lead-bismuth coolant. The scope of the investigations included the measurements of the spectral indexes, distributions of the fission rates of the main isotopes, small samples worths and coolant voiding. The special program was connected with minor actinides. The influence of the plutonium isotope composition was investigated at the assemblies with reactor and weapon grade Pu. Calculations of the measured parameters were carried out using the most modern versions of nuclear data libraries. All the results of these experiments and their analysis have prepared for the construction of the benchmarks and planed as the candidates for the International data base IRPhEP. (authors)

  10. Enthalpy benchmark experiments for numerical ice sheet models

    Directory of Open Access Journals (Sweden)

    T. Kleiner

    2014-06-01

    Full Text Available We present benchmark experiments to test the implementation of enthalpy and the corresponding boundary conditions in numerical ice sheet models. The first experiment tests particularly the functionality of the boundary condition scheme and the basal melt rate calculation during transient simulations. The second experiment addresses the steady-state enthalpy profile and the resulting position of the cold–temperate transition surface (CTS. For both experiments we assume ice flow in a parallel-sided slab decoupled from the thermal regime. Since we impose several assumptions on the experiment design, analytical solutions can be formulated for the proposed numerical experiments. We compare simulation results achieved by three different ice flow-models with these analytical solutions. The models agree well to the analytical solutions, if the change in conductivity between cold and temperate ice is properly considered in the model. In particular, the enthalpy gradient at the cold side of the CTS vanishes in the limit of vanishing conductivity in the temperate ice part as required from the physical jump conditions at the CTS.

  11. Monte Carlo uncertainty analysis for an iron shielding benchmark experiment

    Energy Technology Data Exchange (ETDEWEB)

    Fischer, U.; Tsige-Tamirat, H. [Association Euratom-FZK Forschungszentrum Karlsruhe (Germany); Perel, R.L. [Hebrew Univ., Jerusalem (Israel); Wu, Y. [Institute of Plasma Physics, Heifi (China)

    1998-07-01

    This work is devoted to the computational uncertainty analysis of an iron benchmark experiment having been performed previously at the Technical University of Dresden (TUD). The analysis is based on the use of a novel Monte Carlo approach for calculating sensitivities of point detectors and focuses on the new {sup 56}Fe evaluation of the European Fusion File EFF-3. The calculated uncertainties of the neutron leakage fluxes are shown to be significantly smaller than with previous data. Above 5 MeV the calculated uncertainties are larger than the experimental ones. As the measured neutron leakage fluxes are underestimated by about 10 - 20 % in that energy range, it is concluded that the {sup 56}Fe cross-section data have to be further improved. (authors)

  12. Benchmark Data Through The International Reactor Physics Experiment Evaluation Project (IRPHEP)

    Energy Technology Data Exchange (ETDEWEB)

    J. Blair Briggs; Dr. Enrico Sartori

    2005-09-01

    The International Reactor Physics Experiments Evaluation Project (IRPhEP) was initiated by the Organization for Economic Cooperation and Development (OECD) Nuclear Energy Agency’s (NEA) Nuclear Science Committee (NSC) in June of 2002. The IRPhEP focus is on the derivation of internationally peer reviewed benchmark models for several types of integral measurements, in addition to the critical configuration. While the benchmarks produced by the IRPhEP are of primary interest to the Reactor Physics Community, many of the benchmarks can be of significant value to the Criticality Safety and Nuclear Data Communities. Benchmarks that support the Next Generation Nuclear Plant (NGNP), for example, also support fuel manufacture, handling, transportation, and storage activities and could challenge current analytical methods. The IRPhEP is patterned after the International Criticality Safety Benchmark Evaluation Project (ICSBEP) and is closely coordinated with the ICSBEP. This paper highlights the benchmarks that are currently being prepared by the IRPhEP that are also of interest to the Criticality Safety Community. The different types of measurements and associated benchmarks that can be expected in the first publication and beyond are described. The protocol for inclusion of IRPhEP benchmarks as ICSBEP benchmarks and for inclusion of ICSBEP benchmarks as IRPhEP benchmarks is detailed. The format for IRPhEP benchmark evaluations is described as an extension of the ICSBEP format. Benchmarks produced by the IRPhEP add new dimension to criticality safety benchmarking efforts and expand the collection of available integral benchmarks for nuclear data testing. The first publication of the "International Handbook of Evaluated Reactor Physics Benchmark Experiments" is scheduled for January of 2006.

  13. Benchmark physics experiment of metallic-fueled LMFBR at FCA

    International Nuclear Information System (INIS)

    A benchmark physics experiment of a metallic-fueled LMFBR was performed at Japan Atomic Energy Research Institute's Fast Critical Assembly (FCA) in order to examine availability of data and method for a design of metallic-fueled core. The nuclear data and the calculation methods used for a LMFBR core design have been improved based on the oxide fuel core experiments. A metallic-fueled core has a harder neutron spectrum than an oxide-fueled core and has typical nuclear characteristics affected by the neutron spectrum. In this study, availability of the conventional calculation method for the design of the metallic-fueled core was examined by comparing the calculation values of the nuclear characteristics with the measured values. The experimental core (FCA assembly XVI-1) was selected by referring to the conceptual design of Central Research Institute of Electric Power Industry. The calculated-to-experiment (C/E) value for keff of assembly XVI-1 was 1.001. From this, as far as the criticality the prediction accuracy of the conventional calculation for the metallic-fueled core was concluded to be similar to that of an oxide-fueled core. (author)

  14. Nuclear data benchmark through pulsed sphere experiment conducted at OKTAVIAN

    International Nuclear Information System (INIS)

    In order to make benchmark validation of the nuclear data for the existing evaluated nuclear data libraries, neutron leakage spectra from spherical piles were measured with a time-of-flight technique using the intense 14 MeV neutron source, OKTAVIAN in the energy range from 0.1 to 15 MeV. The measured spectra were compared with those by theoretical calculation using a Monte Carlo neutron transport code, MCNP with several libraries. Comparison has been made with the spectrum shape, the C/E values of neutron numbers integrated in 4 energy regions and the calculated spectra unfolded by the number of collisions, especially those after a single collision. The new libraries predicted the experiment fairly well in general. The detailed discussion has been given for Cr, Mn and Cu samples. Other than the problem found in the calculation with older nuclear data, 24% overestimation of the Cr experiment with EFF-2 has been found between 1 and 5-MeV neutron energy region, presumably because of small overestimation of inelastic cross section and 52Cr(n,2n) cross section and the problem in energy and angular distribution of secondary neutrons in EFF-2. For Cu, ENDF/B-VI and EFF-2 overestimated the experiment by about 20 to 30-% in the energy range between 5 and 12-MeV, presumably from the problem in inelastic scattering cross section. (author)

  15. Benchmarking ~(232)Th Evaluations With KBR and Thor Experiments

    Institute of Scientific and Technical Information of China (English)

    2011-01-01

    The n+232Th evaluations from CENDL-3.1, ENDF/B-Ⅶ.0, JENDL-3.3 and JENDL-4.0 were tested with KBR series and THOR benchmark from ICSBEP Handbook. THOR is Plutonium-Metal-Fast (PMF) criticality benchmark reflected with metal thorium.

  16. INTEGRAL BENCHMARKS AVAILABLE THROUGH THE INTERNATIONAL REACTOR PHYSICS EXPERIMENT EVALUATION PROJECT AND THE INTERNATIONAL CRITICALITY SAFETY BENCHMARK EVALUATION PROJECT

    Energy Technology Data Exchange (ETDEWEB)

    J. Blair Briggs; Lori Scott; Enrico Sartori; Yolanda Rugama

    2008-09-01

    Interest in high-quality integral benchmark data is increasing as efforts to quantify and reduce calculational uncertainties accelerate to meet the demands of next generation reactor and advanced fuel cycle concepts. The International Reactor Physics Experiment Evaluation Project (IRPhEP) and the International Criticality Safety Benchmark Evaluation Project (ICSBEP) continue to expand their efforts and broaden their scope to identify, evaluate, and provide integral benchmark data for method and data validation. Benchmark model specifications provided by these two projects are used heavily by the international reactor physics, nuclear data, and criticality safety communities. Thus far, 14 countries have contributed to the IRPhEP, and 20 have contributed to the ICSBEP. The status of the IRPhEP and ICSBEP is discussed in this paper, and the future of the two projects is outlined and discussed. Selected benchmarks that have been added to the IRPhEP and ICSBEP handbooks since PHYSOR’06 are highlighted, and the future of the two projects is discussed.

  17. Benchmark validation by means of pulsed sphere experiment at OKTAVIAN

    Energy Technology Data Exchange (ETDEWEB)

    Ichihara, Chihiro [Kyoto Univ., Kumatori, Osaka (Japan). Research Reactor Inst.; Hayashi, Shu A.; Kimura, Itsuro; Yamamoto, Junji; Takahashi, Akito

    1998-03-01

    In order to make benchmark validation of the existing evaluated nuclear data for fusion related material, neutron leakage spectra from spherical piles were measured with a time-of-flight technique using the intense 14 MeV neutron source, OKTAVIAN in the energy range from 0.1 to 15 MeV. The neutron energy spectra were obtained as the absolute value normalized per the source neutron. The measured spectra were compared with those by theoretical calculation using a Monte Carlo neutron transport code, MCNP with several libraries processed from the evaluated nuclear data files. Comparison has been made with the spectrum shape, the C/E values of neutron numbers integrated in 4 energy regions and the calculated spectra unfolded by the number of collisions, especially those after a single collision. The new libraries predicted the experiment fairly well for Li, Cr, Mn, Cu and Mo. For Al, Si, Zr, Nb and W, new data files could give fair prediction. However, C/E differed more than 20% for several regions. For LiF, CF{sub 2}, Ti and Co, no calculation could predict the experiment. The detailed discussion has been given for Cr, Mn and Cu samples. EFF-2 calculation overestimated by 24% for the Cr experiment between 1 and 5-MeV neutron energy region, presumably because of overestimation of inelastic cross section and {sup 52}Cr(n,2n) cross section and the problem in energy and angular distribution of secondary neutrons in EFF-2. For Cu, ENDF/B-VI and EFF-2 overestimated the experiment by about 20 to 30-% in the energy range between 5 and 12-MeV, presumably from the problem in inelastic scattering cross section. (author)

  18. Performance analysis for neutronics benchmark experiments with partial adjoint contribution estimated by forward Monte Carlo calculation

    International Nuclear Information System (INIS)

    Highlights: • Performance estimation of nuclear-data benchmark was investigated. • Point detector contribution played a benchmark role not only to the neutron producing the detector contribution but also equally to all the upstream transport neutrons. • New functions were defined to give how well the contribution could be interpreted for benchmarking. • Benchmark performance could be evaluated only by a forward Monte Carlo calculation. -- Abstract: The author's group has been investigating how the performance estimation of nuclear-data benchmark using experiment and its analysis by Monte Carlo code should be carried out especially at 14 MeV. We have recently found that a detector contribution played a benchmark role not only to the neutron producing the detector contribution but also equally to all the upstream neutrons during the neutron history. This result would propose that the benchmark performance could be evaluated only by a forward Monte Carlo calculation. In this study, we thus defined new functions to give how well the contribution could be utilized for benchmarking using the point detector, and described that it was deeply related to the newly introduced “partial adjoint contribution”. By preparing these functions before benchmark experiments, one could know beforehand how well and for which nuclear data the experiment results could do benchmarking in forward Monte Carlo calculations

  19. Review of recent benchmark experiments on integral test for high energy nuclear data evaluation

    Energy Technology Data Exchange (ETDEWEB)

    Nakashima, Hiroshi; Tanaka, Susumu; Konno, Chikara; Fukahori, Tokio; Hayashi, Katsumi [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment

    1998-11-01

    A survey work of recent benchmark experiments on an integral test for high energy nuclear data evaluation was carried out as one of the work of the Task Force on JENDL High Energy File Integral Evaluation (JHEFIE). In this paper the results are compiled and the status of recent benchmark experiments is described. (author)

  20. Calculations of EURACOS iron benchmark experiment using the HYBRID method

    International Nuclear Information System (INIS)

    In this paper, the HYBRID method is used in the calculations of the iron benchmark experiment at the EURACOS-II device. The saturation activities of the 32S(n,p)32P reaction at different depths in an iron block are computed with ENDF/B-IV data to compare with the measurements. At the outer layers of the iron block, the HYBRID calculation gives increasingly higher results than the VITAMIN-C multigroup calculation. With the adjustment of the two- to one-dimensional ratios, the HYBRID results agree with the measurements to within 10% at most penetration depths, a considerable improvement over the VITAMIN-C multigroup results. The development of a collapsing method for the HYBRID cross sections provides a more direct and practical way of using the HYBRID method in the two-dimensional calculations. It is observed that half of the window effect is smeared in the collapsing treatment, but it still provides a better cross-section set than the VITAMIN-C cross sections for the deep-penetration calculations

  1. MCNP simulation of the TRIGA Mark II benchmark experiment

    International Nuclear Information System (INIS)

    The complete 3D MCNP model of the TRIGA Mark II reactor is presented. It enables precise calculations of some quantities of interest in a steady-state mode of operation. Calculational results are compared to the experimental results gathered during reactor reconstruction in 1992. Since the operating conditions were well defined at that time, the experimental results can be used as a benchmark. It may be noted that this benchmark is one of very few high enrichment benchmarks available. In our simulations experimental conditions were thoroughly simulated: fuel elements and control rods were precisely modeled as well as entire core configuration and the vicinity of the core. ENDF/B-VI and ENDF/B-V libraries were used. Partial results of benchmark calculations are presented. Excellent agreement of core criticality, excess reactivity and control rod worths can be observed. (author)

  2. Benchmark validation by means of pulsed sphere experiment at OKTAVIAN

    Energy Technology Data Exchange (ETDEWEB)

    Ichihara, Chihiro [Kyoto Univ., Kumatori, Osaka (Japan). Research Reactor Inst.; Hayashi, S.A.; Kimura, Itsuro; Yamamoto, Junji; Takahashi, Akito

    1997-03-01

    The new version of Japanese nuclear data library JENDL-3.2 has recently been released. JENDL Fusion File which adopted DDX representations for secondary neutrons was also improved with the new evaluation method. On the other hand, FENDL nuclear data project to compile nuclear data library for fusion related research has been conducted partly under auspices of International Atomic Energy Agency (IAEA). The first version FENDL-1 consists of JENDL-3.1, ENDF/B-VI, BROND-2 and EFF-1 and has been released in 1995. The work for the second version FENDL-2 is now ongoing. The Bench mark validation of the nuclear data libraries have been performed to help selecting the candidate for the FENDL-2. The benchmark experiment have been conducted at OKTAVIAN of Osaka university. The sample spheres were constructed by filling the spherical shells with sample. The leakage neutron spectra from sphere piles were measured with a time-of-flight method. The measured spectra were compared with the theoretical calculation using MCNP 4A and the processed libraries from JENDL-3.1, JENDL-3.2, JENDL Fusion File, and FENDL-1. JENDL Fusion File and JENDL-3.2 gave almost the same prediction for the experiment. And both prediction are almost satisfying for Li, Cr, Mn, Cu, Zr, Nb and Mo, whereas for Al, LiF, CF2, Si, Ti, Co and W there is some discrepancy. However, they gave better prediction than the calculations using the library from FENDL-1, except for W. (author)

  3. Benchmarking the particle background in the LHC experiments

    International Nuclear Information System (INIS)

    The experiments for the Large Hadron Collider (LHC) have to work for 15 years in the presence of a very high particle background of photons in the energy range from 100 keV to 10 MeV and neutrons in the range from 0.025 keV to 20 MeV. The background is so high that it becomes a major design criterion for the ATLAS experiment, a general purpose experiment at LHC that will be operational in the year 2005. The exact level of this background is poorly known. At present an uncertainty factor of five has to be assumed to which the limited knowledge of shower processes in the absorber material and the ensuing neutron and photon production is one contribution. So far, the background has been estimated only through extensive Monte Carlo evaluation with the particle transport code FLUKA. The lack of relevant measurements, which were not done up to now, are to a large extent responsible for this uncertainty. This work describes in detail the benchmarking measurements and analysis of these backgrounds in an experimental arrangement that approaches rather closely the layout and shielding in the ATLAS detector. The absolute yield and energy of the particles emanating from the final stages of the hadronic shower were measured using a Bi(4)Ge(3)O(12) detector. In this study particular care was taken to guard against spurious effects which could mask the measurements of the photon background. Typically, we expect to measure a photon per 10000 incident hadrons which is equivalent to a reduction factor in energy of 100 millions. At first, calibration measurements with well known radioactive sources were carried out in order to evaluate the response to photons and neutrons of the used detector. The measurements show very good agreement with the simulations. The actual benchmarking task comprised measurements with different beam intensities and momenta, different positions and absorber thicknesses in order to reduce systematic effects and assess residual activities from other sources

  4. The European Union benchmarking experience. From euphoria to fatigue?

    Directory of Open Access Journals (Sweden)

    Michael Zängle

    2004-06-01

    Full Text Available Even if one may agree with the possible criticism of the Lisbon process as being too vague in com-mitment or as lacking appropriate statistical techniques and indicators, the benchmarking system pro-vided by EUROSTAT seems to be sufficiently effective in warning against imminent failure. The Lisbon objectives are very demanding. This holds true even if each of the objectives is looked at in isolation. But 'Lisbon' is more demanding than that, requiring a combination of several objectives to be achieved simultaneously (GDP growth, labour productivity, job-content of growth, higher quality of jobs and greater social cohesion. Even to countries like Ireland, showing exceptionally high performance in GDP growth and employment promotion during the period under investigation, achieving potentially conflicting objectives simultaneously seems to be beyond feasibility. The European Union benchmark-ing exercise is embedded in the context of the Open Method(s of Co-ordination (OMC. This context makes the benchmarking approach part and parcel of an overarching philosophy, which relates the benchmarking indicators to each other and assigns to them their role in corroborating the increasingly dominating project of the 'embedded neo-liberalism'. Against this background, the present paper is focussed on the following point. With the EU bench-marking system being effective enough to make the imminent under-achievement visible, there is a danger of disillusionment and 'benchmarking fatigue', which may provoke an ideological crisis. The dominant project being so deeply rooted, however, chances are high that this crisis will be solved im-manently in terms of embedded neo-liberalism by strengthening the neo-liberal branch of the Euro-pean project. Confining itself to the Europe of Fifteen, the analysis draws on EUROSTAT's database of Structural Indicators. ...

  5. A benchmark on computational simulation of a CT fracture experiment

    International Nuclear Information System (INIS)

    For a better understanding of the fracture behavior of cracked welds in piping, FRAMATOME, EDF and CEA have launched an important analytical research program. This program is mainly based on the analysis of the effects of the geometrical parameters (the crack size and the welded joint dimensions) and the yield strength ratio on the fracture behavior of several cracked configurations. Two approaches have been selected for the fracture analyses: on one hand, the global approach based on the concept of crack driving force J and on the other hand, a local approach of ductile fracture. In this approach the crack initiation and growth are modelized by the nucleation, growth and coalescence of cavities in front of the crack tip. The model selected in this study estimates only the growth of the cavities using the RICE and TRACEY relationship. The present study deals with a benchmark on computational simulation of CT fracture experiments using three computer codes : ALIBABA developed by EDF the CEA's code CASTEM 2000 and the FRAMATOME's code SYSTUS. The paper is split into three parts. At first, the authors present the experimental procedure for high temperature toughness testing of two CT specimens taken from a welded pipe, characteristic of pressurized water reactor primary piping. Secondly, considerations are outlined about the Finite Element analysis and the application procedure. A detailed description is given on boundary and loading conditions, on the mesh characteristics, on the numerical scheme involved and on the void growth computation. Finally, the comparisons between numerical and experimental results are presented up to the crack initiation, the tearing process being not taken into account in the present study. The variations of J and of the local variables used to estimate the damage around the crack tip (triaxiality and hydrostatic stresses, plastic deformations, void growth ...) are computed as a function of the increasing load

  6. Benchmark experiment on vanadium assembly with D-T neutrons. In-situ measurement

    Energy Technology Data Exchange (ETDEWEB)

    Maekawa, Fujio; Kasugai, Yoshimi; Konno, Chikara; Wada, Masayuki; Oyama, Yukio; Ikeda, Yujiro [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment; Murata, Isao; Kokooo; Takahashi, Akito

    1998-03-01

    Fusion neutronics benchmark experimental data on vanadium were obtained for neutrons in almost entire energies as well as secondary gamma-rays. Benchmark calculations for the experiment were performed to investigate validity of recent nuclear data files, i.e., JENDL Fusion File, FENDL/E-1.0 and EFF-3. (author)

  7. Validation of 232Th evaluated nuclear data through benchmark experiments

    International Nuclear Information System (INIS)

    Evaluated nuclear data files for nuclides of the thorium-uranium cycle were released as a result of a coordinated research project of the International Atomic Energy Agency. To validate the data a series of benchmark test cases were analysed, with emphasis on the most important nuclides 232Th and 233U. Remaining nuclides appearing in the benchmarks were taken from the ENDF/B-VII beta-2 library, except for a few sensitivity studies where the data were taken from other sources. The results show marked improvement in the agreement with measurements, much of it coming from the new evaluations. (author)

  8. Design of Pre-collimator System for Neutronics Benchmark Experiment

    Institute of Scientific and Technical Information of China (English)

    2011-01-01

    In order to carry out evaluation of neutron nuclear data, in the last "Five-Year" period, China Institute of Atomic Energy has developed a set of neutron nuclear data benchmarking test system, and used the time-of-flight technique to measure the neutron

  9. The European Union benchmarking experience. From euphoria to fatigue?

    OpenAIRE

    Michael Zängle

    2004-01-01

    Even if one may agree with the possible criticism of the Lisbon process as being too vague in com-mitment or as lacking appropriate statistical techniques and indicators, the benchmarking system pro-vided by EUROSTAT seems to be sufficiently effective in warning against imminent failure. The Lisbon objectives are very demanding. This holds true even if each of the objectives is looked at in isolation. But 'Lisbon' is more demanding than that, requiring a combination of several objectives to b...

  10. Benchmark experiment for the verification of radiation transport calculations for the radiotherapy dosimetry; Benchmark-Experiment zur Verifikation von Strahlungstransportrechnungen fuer die Dosimetrie in der Strahlentherapie

    Energy Technology Data Exchange (ETDEWEB)

    Renner, Franziska

    2014-10-02

    Monte Carlo simulations are regarded as the most accurate method of solving complex problems of radiation transport. Therefore, they have great potential to realize more exact dose calculations for treatment planning in radiation therapy. However, there is a lack of information on how correct the results of Monte Carlo calculations are on an absolute basis. A practical verification of the calculations can be performed by direct comparison with a benchmark experiment. Thereby, the uncertainties of the experimental result and of the simulation also have to be considered to make a meaningful comparison between the experiment and the simulation possible. This dissertation presents a benchmark experiment and its results, including the uncertainty, which can be used to test the accuracy of Monte Carlo calculations in the field of radiation therapy. The experiment was planned to have parallels to clinical radiation therapy, among other things, with respect to the radiation applied, the materials used and the manner of dose detection. The benchmark experiment aimed at an absolute comparison with a simulation result and because of this it was necessary to use a special research accelerator as a radiation source in the experiment. The accurate characterization of the accelerator beam was a precondition to define a realistic radiation source for the Monte Carlo simulation. Therefore, this work also deals with the characterization of the source and investigations regarding the X-ray target used. Additionally, the dissertation contains the verification of the widely used Monte Carlo program EGSnrc by the benchmark experiment. The simulation of the experiment by EGSnrc, the results and the estimation of the uncertainty related to the simulation are documented in this work.The results and findings of this dissertation end in a comparison between the results of the benchmark experiment and the corresponding calculations with EGSnrc. The benchmark experiment and the simulations

  11. Experiment vs simulation RT WFNDEC 2014 benchmark: CIVA results

    International Nuclear Information System (INIS)

    The French Atomic Energy Commission and Alternative Energies (CEA) has developed for years the CIVA software dedicated to simulation of NDE techniques such as Radiographic Testing (RT). RT modelling is achieved in CIVA using combination of a determinist approach based on ray tracing for transmission beam simulation and a Monte Carlo model for the scattered beam computation. Furthermore, CIVA includes various detectors models, in particular common x-ray films and a photostimulable phosphor plates. This communication presents the results obtained with the configurations proposed in the World Federation of NDEC 2014 RT modelling benchmark with the RT models implemented in the CIVA software

  12. Benchmark study of TRIPOLI-4 through experiment and MCNP codes

    Energy Technology Data Exchange (ETDEWEB)

    Michel, M. [CEA, LIST, Laboratoire Capteurs et Architectures Electroniques, F-91191 Gif-sur-Yvette (France); Coulon, R. [Canberra France, F-78182 Saint Quentin en Yvelines (France); Normand, S. [CEA, LIST, Laboratoire Capteurs et Architectures Electroniques, F-91191 Gif-sur-Yvette (France); Huot, N.; Petit, O. [CEA, DEN DANS, SERMA, F-91191 Gif-sur-Yvette (France)

    2011-07-01

    Reliability on simulation results is essential in nuclear physics. Although MCNP5 and MCNPX are the world widely used 3D Monte Carlo radiation transport codes, alternative Monte Carlo simulation tools exist to simulate neutral and charged particles' interactions with matter. Therefore, benchmark are required in order to validate these simulation codes. For instance, TRIPOLI-4.7, developed at the French Alternative Energies and Atomic Energy Commission for neutron and photon transport, now also provides the user with a full feature electron-photon electromagnetic shower. Whereas the reliability of TRIPOLI-4.7 for neutron and photon transport has been validated yet, the new development regarding electron-photon matter interaction needs additional validation benchmarks. We will thus demonstrate how accurately TRIPOLI-4's 'deposited spectrum' tally can simulate gamma spectrometry problems, compared to MCNP's 'F8' tally. The experimental setup is based on an HPGe detector measuring the decay spectrum of an {sup 152}Eu source. These results are then compared with those given by MCNPX 2.6d and TRIPOLI-4 codes. This paper deals with both the experimental aspect and simulation. We will demonstrate that TRIPOLI-4 is a potential alternative to both MCNPX and MCNP5 for gamma-electron interaction simulation. (authors)

  13. Benchmark study of TRIPOLI-4 through experiment and MCNP codes

    International Nuclear Information System (INIS)

    Reliability on simulation results is essential in nuclear physics. Although MCNP5 and MCNPX are the world widely used 3D Monte Carlo radiation transport codes, alternative Monte Carlo simulation tools exist to simulate neutral and charged particles' interactions with matter. Therefore, benchmark are required in order to validate these simulation codes. For instance, TRIPOLI-4.7, developed at the French Alternative Energies and Atomic Energy Commission for neutron and photon transport, now also provides the user with a full feature electron-photon electromagnetic shower. Whereas the reliability of TRIPOLI-4.7 for neutron and photon transport has been validated yet, the new development regarding electron-photon matter interaction needs additional validation benchmarks. We will thus demonstrate how accurately TRIPOLI-4's 'deposited spectrum' tally can simulate gamma spectrometry problems, compared to MCNP's 'F8' tally. The experimental setup is based on an HPGe detector measuring the decay spectrum of an 152Eu source. These results are then compared with those given by MCNPX 2.6d and TRIPOLI-4 codes. This paper deals with both the experimental aspect and simulation. We will demonstrate that TRIPOLI-4 is a potential alternative to both MCNPX and MCNP5 for gamma-electron interaction simulation. (authors)

  14. Current status of accelerator benchmarks in OECD/NEA radiation shielding experiments database SINBAD

    International Nuclear Information System (INIS)

    In order to preserve and make available the information on the performed radiation shielding benchmarks the Data Bank of the OECD Nuclear Energy Agency (OECD/NEADB) and the Radiation Safety Information Computational Center (RSICC) at Oak Ridge National Laboratory (ORNL) started the Shielding Integral Benchmark Archive and Database (SINBAD) Project in the early 1990's. The SINBAD database now comprises 100 shielding benchmarks, also covering accelerator shielding benchmarks in addition to fission reactor shielding and fusion blanket neutronics. Recently, a thorough revision of six of the 23 accelerator benchmark experiments was completed in order to provide detailed verification of the completeness and consistency of the benchmark information, in particular concerning the evaluation of the experimental sources of uncertainty. This review process is expected to provide users with an easier choice and help them make better use of the experimental information and is planned to be extended to other available benchmarks. The OECD NEA Working Party on Scientific Issues of Reactor Systems (WPRS) Expert Group on Radiation Transport and Shielding (EGRTS) was created in 2011 and its mandate is to monitor, steer and support the continued development of the SINBAD database. Co-ordination of the SATIF-11 Workshop is part of the mandate of EGRTS and the development of the accelerator benchmark database is done in a close co-operation with the SATIF expert group. Proposals and assistance in new benchmark compilations are welcome. SINBAD is available at no charge from RSICC and from the NEA Data Bank. Since its beginnings SINBAD has been used by nuclear data evaluators, computer code developers, experiment designers and university students. (authors)

  15. Data Assimilation of Benchmark Experiments for Homogenous Thermal / Epithermal Uranium Systems

    International Nuclear Information System (INIS)

    This presentation reports on the data assimilation of benchmark experiments for homogeneous thermal and epithermal uranium systems. The assimilation method is based on Kalman filters using integral parameters and sensitivity coefficients calculated with MONK9 and ENDF/B-VII data. The assimilation process results in an overall improvement of the calculation-benchmark agreement, and may help in the selection of nuclear data after analysis of adjustment trends

  16. Benchmark on the Kritz-2 Leu and MOX critical experiments

    International Nuclear Information System (INIS)

    In the framework of the joint activities of the OECD/NEA Working Party on the Physics of Plutonium Fuels and Innovative Fuel Cycles (WPPR)1 and the Task Force on Reactor-based Plutonium Disposition (TFRPD), an international benchmark exercise based on KRITZ UO2 and MOX critical configurations was launched in October 2000. The aim of this exercise was to investigate the capabilities of the current production codes and nuclear data libraries to analyse MOX-fuelled systems, and to compare the accuracy of the predictions for the MOX- and UO2-fuelled configurations. Institutions from 7 countries participated in this exercise, providing 13 solutions. The report provides comparative analyses of calculated and measured results, as well as intercomparisons of some of the results obtained by participants by calculation only. (author)

  17. Status of international benchmark experiment for effective delayed neutron fraction ({beta}eff)

    Energy Technology Data Exchange (ETDEWEB)

    Okajima, S.; Sakurai, T.; Mukaiyama, T. [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment

    1997-03-01

    To improve the prediction accuracy of the {beta}eff, the program of the international benchmark experiment (Beta Effect Reactor Experiment for a New International Collaborative Evaluation: BERNICE) was planned. This program composed of two parts; BERNICE-MASURCA and BERNICE-FCA. The former one was carried out in the fast critical facility MASURCA of CEA, FRANCE between 1993 and 1994. The latter one started in the FCA, JAERI in 1995 and still is going. In these benchmark experiments, various experimental techniques have been applied for in-pile measurements of the {beta}eff. The accuracy of the measurements was better than 3%. (author)

  18. The use of benchmark experiments for the validation of nuclear data

    International Nuclear Information System (INIS)

    The paper concerns the adjustment of cross-section data to give improved statistical consistency between calculation and experiment for nuclear reactors. Early work in the field is briefly outlined, as well as the analysis of benchmark experiments to validate nuclear data. Iron inelastic scattering cross-section results are discussed, with reference to inadequacies in nuclear data employed in calculations. (U.K.)

  19. Shielding benchmark calculations of selected spent fuel storage cask experiments

    Energy Technology Data Exchange (ETDEWEB)

    Broadhead, B.L.; Tang, J.S.; Parks, C.V. (Oak Ridge National Lab., TN (United States)); Taniuchi, H. (Kobe Steel Ltd. (Japan))

    1993-01-01

    This paper describes the application of the three-dimensional Monte Carlo code MORSE-SGC, as implemented in the SCALE system calculational sequence SAS4, to the analysis of a series of benchmark spent fuel storage cask measurements performed at the Idaho National Engineering Laboratory. A total of five storage cask problems were analyzed to determine the expected accuracies of computational analyses using well-established Monte Carlo codes. The results presented herein represent the current status of the work. Predicted neutron dose results generally compare very favorably (within 30%) with the measurements for the cask lid, bottom, and along the cask side. Gamma-ray dose rates exhibit differing trends, depending on the measurement location. For lid and bottom doses, as well as side doses near the endfittings, agreement is again within 30%, although several exceptions are seen. However, for gamma doses along the cask side and adjacent to the active fuel, a factor of 2 overprediction is noted. Investigations into the cause of these discrepancies are currently in progress.

  20. Shielding benchmark calculations of selected spent fuel storage cask experiments

    Energy Technology Data Exchange (ETDEWEB)

    Broadhead, B.L.; Tang, J.S.; Parks, C.V. [Oak Ridge National Lab., TN (United States); Taniuchi, H. [Kobe Steel Ltd. (Japan)

    1993-03-01

    This paper describes the application of the three-dimensional Monte Carlo code MORSE-SGC, as implemented in the SCALE system calculational sequence SAS4, to the analysis of a series of benchmark spent fuel storage cask measurements performed at the Idaho National Engineering Laboratory. A total of five storage cask problems were analyzed to determine the expected accuracies of computational analyses using well-established Monte Carlo codes. The results presented herein represent the current status of the work. Predicted neutron dose results generally compare very favorably (within 30%) with the measurements for the cask lid, bottom, and along the cask side. Gamma-ray dose rates exhibit differing trends, depending on the measurement location. For lid and bottom doses, as well as side doses near the endfittings, agreement is again within 30%, although several exceptions are seen. However, for gamma doses along the cask side and adjacent to the active fuel, a factor of 2 overprediction is noted. Investigations into the cause of these discrepancies are currently in progress.

  1. International Handbook of Evaluated Criticality Safety Benchmark Experiments - ICSBEP (DVD), Version 2013

    International Nuclear Information System (INIS)

    The Criticality Safety Benchmark Evaluation Project (CSBEP) was initiated in October of 1992 by the United States Department of Energy. The project quickly became an international effort as scientists from other interested countries became involved. The International Criticality Safety Benchmark Evaluation Project (ICSBEP) became an official activity of the Organisation for Economic Co-operation and Development (OECD) Nuclear Energy Agency (NEA) in 1995. This handbook contains criticality safety benchmark specifications that have been derived from experiments performed at various nuclear critical experiment facilities around the world. The benchmark specifications are intended for use by criticality safety engineers to validate calculational techniques used to establish minimum subcritical margins for operations with fissile material and to determine criticality alarm requirement and placement. Many of the specifications are also useful for nuclear data testing. Example calculations are presented; however, these calculations do not constitute a validation of the codes or cross section data. The evaluated criticality safety benchmark data are given in nine volumes. These volumes span nearly 66,000 pages and contain 558 evaluations with benchmark specifications for 4,798 critical, near critical or subcritical configurations, 24 criticality alarm placement/shielding configurations with multiple dose points for each and 200 configurations that have been categorised as fundamental physics measurements that are relevant to criticality safety applications. New to the Handbook are benchmark specifications for Critical, Bare, HEU(93.2)- Metal Sphere experiments referred to as ORSphere that were performed by a team of experimenters at Oak Ridge National Laboratory in the early 1970's. A photograph of this assembly is shown on the front cover

  2. Design of a pre-collimator system for neutronics benchmark experiment

    International Nuclear Information System (INIS)

    Benchmark experiment is an important means to inspect the reliability and accuracy of the evaluated nuclear data, the effect/background ratios are the important parameters to weight the quality of experimental data. In order to obtain higher effect/background ratios, a pre-collimator system was designed for benchmark experiment. This system mainly consists of a pre-collimator and a shadow cone, The MCNP-4C code was used to simulate the background spectra under various conditions, from the results we found that with the pre-collimator system have a very marked improvement in the effect/background ratios. (authors)

  3. Analysis of the impact of correlated benchmark experiments on the validation of codes for criticality safety analysis

    International Nuclear Information System (INIS)

    The validation of a code for criticality safety analysis requires the recalculation of benchmark experiments. The selected benchmark experiments are chosen such that they have properties similar to the application case that has to be assessed. A common source of benchmark experiments is the 'International Handbook of Evaluated Criticality Safety Benchmark Experiments' (ICSBEP Handbook) compiled by the 'International Criticality Safety Benchmark Evaluation Project' (ICSBEP). In order to take full advantage of the information provided by the individual benchmark descriptions for the application case, the recommended procedure is to perform an uncertainty analysis. The latter is based on the uncertainties of experimental results included in most of the benchmark descriptions. They can be performed by means of the Monte Carlo sampling technique. The consideration of uncertainties is also being introduced in the supplementary sheet of DIN 25478 'Application of computer codes in the assessment of criticality safety'. However, for a correct treatment of uncertainties taking into account the individual uncertainties of the benchmark experiments is insufficient. In addition, correlations between benchmark experiments have to be handled correctly. For example, these correlations can arise due to different cases of a benchmark experiment sharing the same components like fuel pins or fissile solutions. Thus, manufacturing tolerances of these components (e.g. diameter of the fuel pellets) have to be considered in a consistent manner in all cases of the benchmark experiment. At the 2012 meeting of the Expert Group on 'Uncertainty Analysis for Criticality Safety Assessment' (UACSA) of the OECD/NEA a benchmark proposal was outlined that aimed for the determination of the impact on benchmark correlations on the estimation of the computational bias of the neutron multiplication factor (keff). The analysis presented here is based on this proposal. (orig.)

  4. PINEX II - French contribution to an international benchmark experiment

    International Nuclear Information System (INIS)

    The aim of the PINEX-2 experiment was to demonstrate that it was possible to delay the failure threshold of an irradiated fuel pin during an overpower transient assuming a particular design of the pin. The preirradiation calculations were conducted with the mixed oxyde fuel reference code COREF 1. The transient thermal and mechanical calculations were performed with a special version of the PHYSURA IA code, using the COREF 1 results as input data. Although PHYSURA is made of simple models, it is able to interpret correctly a complex experiment like PINEX 2. Indeed, it has been established: a good agreement with thermal experimental results, a good agreement between calculated and experimental fuel redistribution out of the fissile zone and an agreement with experiment on clad rupture analysis

  5. Developing Evidence for Action on the Postgraduate Experience: An Effective Local Instrument to Move beyond Benchmarking

    Science.gov (United States)

    Sampson, K. A.; Johnston, L.; Comer, K.; Brogt, E.

    2016-01-01

    Summative and benchmarking surveys to measure the postgraduate student research experience are well reported in the literature. While useful, we argue that local instruments that provide formative resources with an academic development focus are also required. If higher education institutions are to move beyond the identification of issues and…

  6. The track rotation estimator and its application to shielding benchmark experiments

    International Nuclear Information System (INIS)

    A radically new approach is presented for the estimation of flux at a point by Monte Carlo methods. A mathematical review of the method is given followed by examples of isotropic and non-isotropic sources in spherical media which could be related to the interpretation of shielding benchmark experiments. (author)

  7. Benchmark Evaluation of the Medium-Power Reactor Experiment Program Critical Configurations

    Energy Technology Data Exchange (ETDEWEB)

    Margaret A. Marshall; John D. Bess

    2013-02-01

    A series of small, compact critical assembly (SCCA) experiments were performed in 1962-1965 at the Oak Ridge National Laboratory Critical Experiments Facility (ORCEF) for the Medium-Power Reactor Experiment (MPRE) program. The MPRE was a stainless-steel clad, highly enriched uranium (HEU)-O2 fuelled, BeO reflected reactor design to provide electrical power to space vehicles. Cooling and heat transfer were to be achieved by boiling potassium in the reactor core and passing vapor directly through a turbine. Graphite- and beryllium-reflected assemblies were constructed at ORCEF to verify the critical mass, power distribution, and other reactor physics measurements needed to validate reactor calculations and reactor physics methods. The experimental series was broken into three parts, with the third portion of the experiments representing the beryllium-reflected measurements. The latter experiments are of interest for validating current reactor design efforts for a fission surface power reactor. The entire series has been evaluated as acceptable benchmark experiments and submitted for publication in the International Handbook of Evaluated Criticality Safety Benchmark Experiments and in the International Handbook of Evaluated Reactor Physics Benchmark Experiments.

  8. Benchmark experiment for the verification of radiation transport calculations for the radiotherapy dosimetry

    International Nuclear Information System (INIS)

    Monte Carlo simulations are regarded as the most accurate method of solving complex problems of radiation transport. Therefore, they have great potential to realize more exact dose calculations for treatment planning in radiation therapy. However, there is a lack of information on how correct the results of Monte Carlo calculations are on an absolute basis. A practical verification of the calculations can be performed by direct comparison with a benchmark experiment. Thereby, the uncertainties of the experimental result and of the simulation also have to be considered to make a meaningful comparison between the experiment and the simulation possible. This dissertation presents a benchmark experiment and its results, including the uncertainty, which can be used to test the accuracy of Monte Carlo calculations in the field of radiation therapy. The experiment was planned to have parallels to clinical radiation therapy, among other things, with respect to the radiation applied, the materials used and the manner of dose detection. The benchmark experiment aimed at an absolute comparison with a simulation result and because of this it was necessary to use a special research accelerator as a radiation source in the experiment. The accurate characterization of the accelerator beam was a precondition to define a realistic radiation source for the Monte Carlo simulation. Therefore, this work also deals with the characterization of the source and investigations regarding the X-ray target used. Additionally, the dissertation contains the verification of the widely used Monte Carlo program EGSnrc by the benchmark experiment. The simulation of the experiment by EGSnrc, the results and the estimation of the uncertainty related to the simulation are documented in this work.The results and findings of this dissertation end in a comparison between the results of the benchmark experiment and the corresponding calculations with EGSnrc. The benchmark experiment and the simulations

  9. IRPhEP-handbook, International Handbook of Evaluated Reactor Physics Benchmark Experiments

    International Nuclear Information System (INIS)

    1 - Description: The purpose of the International Reactor Physics Experiment Evaluation Project (IRPhEP) is to provide an extensively peer-reviewed set of reactor physics-related integral data that can be used by reactor designers and safety analysts to validate the analytical tools used to design next-generation reactors and establish the safety basis for operation of these reactors. This work of the IRPhEP is formally documented in the 'International Handbook of Evaluated Reactor Physics Benchmark Experiments,' a single source of verified and extensively peer-reviewed reactor physics benchmark measurements data. The IRPhE Handbook is available on DVD. You may request a DVD by completing the DVD Request Form available at: http://irphep.inl.gov/handbook/hbrequest.shtml The evaluation process entails the following steps: 1. Identify a comprehensive set of reactor physics experimental measurements data, 2. Evaluate the data and quantify overall uncertainties through various types of sensitivity analysis to the extent possible, verify the data by reviewing original and subsequently revised documentation, and by talking with the experimenters or individuals who are familiar with the experimental facility, 3. Compile the data into a standardized format, 4. Perform calculations of each experiment with standard reactor physics codes where it would add information, 5. Formally document the work into a single source of verified and peer reviewed reactor physics benchmark measurements data. The International Handbook of Evaluated Reactor Physics Benchmark Experiments contains reactor physics benchmark specifications that have been derived from experiments that were performed at various nuclear experimental facilities around the world. The benchmark specifications are intended for use by reactor physics personal to validate calculational techniques. The 2008 Edition of the International Handbook of Evaluated Reactor Physics Experiments contains data from 25 different

  10. INTERNATIONAL EXPERIENCE OF HOSPITALITY INDUSTRY TRAINING AS A BASE-BENCHMARKING PROJECT

    OpenAIRE

    Наталья Николаевна Масюк; Петр Владимирович Петрищев

    2013-01-01

    Integration of universities in the international educational space includes not only the usual attraction of foreign students, but also the interaction with the best universities in the world in the field of knowledge with a view to the mutual agreement of all kinds. To do this, at the preliminary stage is implemented benchmarking project to study the experience of best practices.Objective: To study the international experience of training in the hospitality industry.Method or the methodology...

  11. Benchmarking of MCNP against B ampersand W LRC Core XI critical experiments

    International Nuclear Information System (INIS)

    The MCNP Monte Carlo code and its ENDF/B-V continuous-energy cross- section library previously has been benchmarked against a variety of critical experiments, and that benchmarking recently has been extended to include its ENDF/B-VI continuous-energy cross-section library and additional critical experiments. This study further extends the benchmarking of MCNP and its two continuous-energy libraries to 17 large-scale mockup experiments that closely resemble the core of a pressurized water reactor (PWR). The experiments were performed at Babcock ampersand Wilcox's Lynchburg Research Center in 1970 and 1971. The series was designated as Core XI, and the individual experiments were characterized as different ''loadings.'' The experiments were performed inside a large aluminum tank that contained borated water. The water height for each loading was exactly 145 cm, and the soluble boron concentration in the water was adjusted until the configuration was slightly supercritical, with a value of 1.0007 for keff. Pin-by-pin power distributions were measured for several of the loadings

  12. Shielding benchmark experiments through concrete and iron with high-energy proton and heavy ion accelerators

    International Nuclear Information System (INIS)

    The deep penetration of neutrons through thick shield has become a very serious problem in the shielding design of high-energy, high-intensity accelerator facility. In the design calculation, the Monte Carlo transport calculation through thick shields has large statistical errors and the basic nuclear data and model used in the existing Monte Carlo codes are not well evaluated because of very few experimental data. It is therefore strongly needed to do the deep penetration experiment as shielding benchmark for investigating the calculation accuracy. Under this circumference, we performed the following two shielding experiments through concrete and iron, one with a 800 MeV proton accelerator of the rutherford appleton laboratory (RAL), England and the other with a high energy heavy iron accelerator of the national institute of radiological sciences (NIRS), Japan. Here these two shielding benchmark experiments are outlined. (orig.)

  13. Linac code benchmarking of HALODYN and PARMILA based on beam experiments

    Science.gov (United States)

    Yin, X.; Bayer, W.; Hofmann, I.

    2016-01-01

    As part of the 'High Intensity Pulsed Proton Injector' (HIPPI) project in the European Framework Programme, a program for the comparison and benchmarking of 3D Particle-In-Cell (PIC) linac codes with experiment has been implemented. HALODYN and PARMILA are two of the codes involved in this program. In this study, the initial Twiss parameters were obtained from the results of beam experiments that were conducted using the GSI UNILAC in low-beam-current. Furthermore, beam dynamics simulations of the Alvarez Drift Tube Linac (DTL) section were performed by HALODYN and PARMILA codes and benchmarked for the same beam experiments. These simulation results exhibit some agreements with the experimental results for the low-beam-current case. The similarities and differences between the experimental and simulated results were analyzed quantitatively. In addition, various physical aspects of the simulation codes and the linac design strategy are also discussed.

  14. Shielding benchmark experiments through concrete and iron with high-energy proton and heavy ion accelerators

    Energy Technology Data Exchange (ETDEWEB)

    Nakamura, T.; Sasaki, M.; Nunomiya, T.; Iwase, H. [Tohoku Univ., Sendai (Japan). Dept. of Quantum Science and Energy Engineering; Nakao, N.; Shibata, T. [High Energy Accelerator Research Organization (KEK), Ibaraki (Japan); Kim, E. [Japan Atomic Energy Research Inst. (JAERI), Ibaraki (Japan). Tokai Establishment; Kurosawa, T. [Japan Synchrotron Radiation Research Inst. (JASRI), Hyogo (Japan); Taniguchi, S. [Electrotechnical Lab. (ETL), Tsukuba, Ibaraki (Japan); Uwamino, Y.; Ito, S. [The Inst. of Physical and Chemical Research (RIKEN), Saitama (Japan); Fukumura, A. [National Inst. of Radiological Sciences (NIRS), Chiba (Japan); Perry, D.R.; Wright, P. [Rutherford Appleton Lab. (RAL), Didcot, Oxfordshire (United Kingdom). Health and Safety Group

    2001-07-01

    The deep penetration of neutrons through thick shield has become a very serious problem in the shielding design of high-energy, high-intensity accelerator facility. In the design calculation, the Monte Carlo transport calculation through thick shields has large statistical errors and the basic nuclear data and model used in the existing Monte Carlo codes are not well evaluated because of very few experimental data. It is therefore strongly needed to do the deep penetration experiment as shielding benchmark for investigating the calculation accuracy. Under this circumference, we performed the following two shielding experiments through concrete and iron, one with a 800 MeV proton accelerator of the rutherford appleton laboratory (RAL), England and the other with a high energy heavy iron accelerator of the national institute of radiological sciences (NIRS), Japan. Here these two shielding benchmark experiments are outlined. (orig.)

  15. Nuclear knowledge management experience of the international criticality safety benchmark evaluation project

    International Nuclear Information System (INIS)

    Full text: The International Criticality Safety Benchmark Evaluation Project (ICSBEP) was initiated in October of 1992 by the Department of Energy Defence Programs, now NNSA. The U.S. effort to support and provide leadership for the ICSBEP has been funded by DOE-DP since that time. The project is managed through the Idaho National Engineering and Environmental Laboratory (INEEL), but involves nationally known criticality safety experts from Los Alamos National Laboratory, Lawrence Livermore National Laboratory, Savannah River Technology Center, Oak Ridge National Laboratory and the Y-12 Plant, Hanford, Argonne National Laboratory, and the Rocky Flat Plant. An International Criticality Safety Data Exchange component was added to the project during 1994. Representatives from the United Kingdom, France, Japan, the Russian Federation, Hungary, Republic of Korea, Slovenia, Yugoslavia, Kazakhstan, Spain, Israel, Brazil, and Poland are now participating on the project and China, South Africa, and the Czech Republic have indicated that they plan to contribute to the project. The ICSBEP is an official activity of the OECD-NEA. The United States is the lead country, providing most of the administrative support. The purpose of the ICSBEP is to: 1. Identify and evaluate a comprehensive set of criticality related benchmark data. 2. Verify the data, to the extent possible, by reviewing original and subsequently revised documentation, logbook data when possible, and by talking with the experimenters or individuals who are familiar with the experimenters or the experimental facility. 3. Compile the data into a standardized format. 4. Perform calculations of each experiment with standard criticality safety codes. 5. Formally document the work into a single source of verified and internationally peer reviewed benchmark critical data. Each experiment evaluation undergoes a thorough internal review by someone within the evaluator's organization. The internal reviewers verifies: 1. The

  16. Integral Data Benchmark of HENDL2.0/MG Compared with Neutronics Shielding Experiments

    Science.gov (United States)

    Jiang, Jieqiong; Xu, Dezheng; Zheng, Shanliang; He, Zhaozhong; Hu, Yanglin; Li, Jingjing; Zou, Jun; Zeng, Qin; Chen, Mingliang; Wang, Minghuang

    2009-10-01

    HENDL2.0, the latest version of the hybrid evaluated nuclear data library, was developed based upon some evaluated data from FENDL2.1 and ENDF/B-VII. To qualify and validate the working library, an integral test for the neutron production data of HENDL2.0 was performed with a series of existing spherical shell benchmark experiments (such as V, Be, Fe, Pb, Cr, Mn, Cu, Al, Si, Co, Zr, Nb, Mo, W and Ti). These experiments were simulated numerically using HENDL2.0/MG and a home-developed code VisualBUS. Calculations were conducted with both FENDL2.1/MG and FENDL2.1/MC, which are based on a continuous-energy Monte Carlo Code MCNP/4C. By comparison and analysis of the neutron leakage spectra and the integral test, benchmark results of neutron production data are presented in this paper.

  17. Integral Data Benchmark of HENDL2.0/MG Compared with Neutronics Shielding Experiments

    Institute of Scientific and Technical Information of China (English)

    JIANG Jieqiong; XU Dezheng; ZHENG Shanliang; HE Zhaozhong; HU Yanglin; LI Jingjing; ZOU Jun; ZENG Qin; CHEN Mingliang; WANG Minghuang

    2009-01-01

    HENDL2.0,the latest version of the hybrid evaluated nuclear data library,was developed based upon some evaluated data from FENDL2.1 and ENDF/B-VII.To qualify and validate the working library,an integral test for the neutron production data of HENDL2.0 was performed with a series of existing spherical shell benchmark experiments (such as V,Be,Fe,Pb,Cr,Mn,Cu,Al,Si,Co,Zr,Nb,Mo,W and Ti).These experiments were simulated numerically using HENDL2.0/MG and a home-developed code VisualBUS.Calculations were conducted with both FENDL2.1/MG and FENDL2.1/MC,which are based on a continuous-energy Monte Carlo Code MCNP/4C.By comparison and analysis of the neutron leakage spectra and the integral test,benchmark results of neutron production data are presented in this paper.

  18. Material Activation Benchmark Experiments at the NuMI Hadron Absorber Hall in Fermilab

    International Nuclear Information System (INIS)

    In our previous study, double and mirror symmetric activation peaks found for Al and Au arranged spatially on the back of the Hadron absorber of the NuMI beamline in Fermilab were considerably higher than those expected purely from muon-induced reactions. From material activation bench-mark experiments, we conclude that this activation is due to hadrons with energy greater than 3 GeV that had passed downstream through small gaps in the hadron absorber

  19. Benchmark experiment on vanadium assembly with D-T neutrons. Leakage neutron spectrum measurement

    Energy Technology Data Exchange (ETDEWEB)

    Kokooo; Murata, I.; Nakano, D.; Takahashi, A. [Osaka Univ., Suita (Japan); Maekawa, F.; Ikeda, Y.

    1998-03-01

    The fusion neutronics benchmark experiments have been done for vanadium and vanadium alloy by using the slab assembly and time-of-flight (TOF) method. The leakage neutron spectra were measured from 50 keV to 15 MeV and comparison were done with MCNP-4A calculations which was made by using evaluated nuclear data of JENDL-3.2, JENDL-Fusion File and FENDL/E-1.0. (author)

  20. Benchmark Experiments at VNIITF Test Facilities for Verification of Nuclear Data Libraries

    International Nuclear Information System (INIS)

    The paper describes test facilities used by the Russian Federal Nuclear Center, All-Russian Institute of Technical Physics (VNIITF) to perform benchmark experiments essential for the verification of nuclear data libraries. The key experiments discussed in the paper include critical mass measurements; the investigation of reaction rate distribution in critical and subcritical systems, in particular those with a 14-MeV neutron source; and studies on the spectra of neutrons and gamma quanta emitted from spheres and reflected by hemispheres with a central pulse source of 14-MeV neutrons. New experiments are proposed with a view to revising nuclear data essential for new nuclear developments

  1. Continuous energy Monte Carlo analysis of neutron shielding benchmark experiments with cross sections in JENDL-3

    Energy Technology Data Exchange (ETDEWEB)

    Ueki, Kohtaro; Ohashi, Atsuto (Ship Research Inst., Mitaka, Tokyo (Japan)); Kawai, Masayoshi

    1993-04-01

    The iron, carbon and beryllium cross sections in JENDL-3 have been tested by the continuous energy Monte Carlo analysis of the neutron shielding benchmark experiments. The iron cross sections have been tested with analysis of the ORNL and the Winfrith experiments using the fission neutron sources, and also the LLNL iron experiment using the D-T neutron source. The carbon and beryllium cross sections have been tested with the JAERI-FNS TOF experiments using the D-T neutron source. Revision of the subroutine TALLYD and an appropriate weight-window-parameter assignment have been accomplished in the MCNP code. In consequence, the FSD for each energy bin is reduced so small that the Monte Carlo results for neutron energy spectra could be recognized to be reliable. The Monte Carlo calculations with JENDL-3 indicate a good agreement with the benchmark experiments in a wide energy range, as a whole. Particularly, for the Winfrith iron experiment, the results with JENDL-3 give better agreement, just below the iron 24keV window, than that with ENDF/B-IV. For the JAERI-FNS TOF graphite experiment, the calculated angular fluxes with JENDL-3 give closer agreement than that with ENDF/B-IV at several peaks and dips caused by the inelastic scattering. However, distinct underestimation is observed in the calculated energy spectrum with JENDL-3 between 0.8 and 3.0 MeV for the two iron experiments using fission neutron sources. (author).

  2. Activation measurements for the E.C. bulk shield benchmark experiment

    Science.gov (United States)

    Angelone, M.; Arpesella, C.; Martone, M.; Pillon, Mario

    1995-03-01

    The use of the absolute radiometric techniques for the E. C. bulk shield experiment at the 14 MeV Frascati Neutron Generator (FNG) is reported. In this application, the activity level, in some cases, results too low to be measured at the Frascati counting station. In these cases the radiometric measurements are performed using the low background HPGe detectors located at the underground laboratory of Gran Sasso d'Italia. The use of these detectors enhances the FNG capability of performing bulk shield benchmark experiments allowing the measurements of very low activation levels.

  3. Benchmark analyses of criticality calculation codes based on the evaluated dissolver-type criticality experiment systems

    International Nuclear Information System (INIS)

    Criticality calculation codes/code systems MCNP, MVP, SCALE and JACS, which are currently typically used in Japan for nuclear criticality safety evaluation, were benchmarked for so called dissolver-typed systems, i.e., fuel rod arrays immersed in fuel solution. The benchmark analyses were made for the evaluated critical experiments published in the International Criticality Safety Benchmark Evaluation Project (ICSBEP) Handbook: one evaluation representing five critical configurations from heterogeneous core of low-enriched uranium dioxides at the Japan Atomic Energy Research Institute and two evaluations representing 16 critical configurations from heterogeneous core of mixed uranium and plutonium dioxides (MOXs) at the Battelle Pacific Northwest Laboratories of the U.S.A. The results of the analyses showed that the minimum values of the neutron multiplication factor obtained with MCNP, MVP, SCALE and JACS were 0.993, 0.990, 0.993, 0.972, respectively, which values are from 2% to 4% larger than the maximum permissible multiplication factor of 0.95. (author)

  4. Description of the Babcock and Wilcox owners group cavity dosimetry benchmark experiment

    International Nuclear Information System (INIS)

    The Babcock and Wilcox Owners Group (B and WOG) Cavity Dosimetry Benchmark experiment is the first step in the B and WOG program to develop measurement-based methodology for use in monitoring vessel fluence in the post-Reactor Vessel Surveillance Program timeframe. Ex-vessel dosimetry has been chosen as the vehicle to provide fluence measurements for use in this measurement-based methodology. (Fluence is measured indirectly by first measuring a relatable quantity and then applying the known correspondence between the measured quantity and the fluence, (e.g., Cs137 activity of a fission foil or tracks on an SSTR). The results of the In-Out Experiment will be used in refining the analytical models and benchmarking the final methodology. The experiment will provide neutron and gamma fluence measurements, at points both inside and outside the reactor vessel, through the use of numerous fluence measuring devices. Four different categories of ex-vessel monitors have been specified. The in-vessel fluence will be measured using an unirradiated, standard B and W reactor vessel surveillance capsule that will be installed in a spare holder tube at the same azimuthal position as the main ex-vessel dosimetry stringer. This paper presents a detailed description of the experiment

  5. Computer simulation of Masurca critical and subcritical experiments. Muse-4 benchmark. Final report

    International Nuclear Information System (INIS)

    The efficient and safe management of spent fuel produced during the operation of commercial nuclear power plants is an important issue. In this context, partitioning and transmutation (P and T) of minor actinides and long-lived fission products can play an important role, significantly reducing the burden on geological repositories of nuclear waste and allowing their more effective use. Various systems, including existing reactors, fast reactors and advanced systems have been considered to optimise the transmutation scheme. Recently, many countries have shown interest in accelerator-driven systems (ADS) due to their potential for transmutation of minor actinides. Much R and D work is still required in order to demonstrate their desired capability as a whole system, and the current analysis methods and nuclear data for minor actinide burners are not as well established as those for conventionally-fuelled systems. Recognizing a need for code and data validation in this area, the Nuclear Science Committee of the OECD/NEA has organised various theoretical benchmarks on ADS burners. Many improvements and clarifications concerning nuclear data and calculation methods have been achieved. However, some significant discrepancies for important parameters are not fully understood and still require clarification. Therefore, this international benchmark based on MASURCA experiments, which were carried out under the auspices of the EC 5. Framework Programme, was launched in December 2001 in co-operation with the CEA (France) and CIEMAT (Spain). The benchmark model was oriented to compare simulation predictions based on available codes and nuclear data libraries with experimental data related to TRU transmutation, criticality constants and time evolution of the neutronic flux following source variation, within liquid metal fast subcritical systems. A total of 16 different institutions participated in this first experiment based benchmark, providing 34 solutions. The large number

  6. Simulation of TRIGA Mark II Benchmark Experiment using WIMSD4 and CITATION codes; Simulacao com WIMSD4 e CITATION do Triga Mark II benchmark experiment

    Energy Technology Data Exchange (ETDEWEB)

    Dalle, Hugo Moura [Centro de Desenvolvimento da Tecnologia Nuclear (CDTN), Belo Horizonte, MG (Brazil); Pereira, Claubia [Minas Gerais Univ., Belo Horizonte, MG (Brazil). Dept. de Engenharia Nuclear

    2000-07-01

    This paper presents a simulation of the TRIGA Mark II Benchmark Experiment, Part I: Steady-State Operation and is part of the calculation methodology validation developed to the neutronic calculation of the CDTN's TRIGA IPR - R1 reactor. A version of the WIMSD4, obtained in the Centro de Tecnologia Nuclear, in Cuba, was used in the cells calculation. In the core calculations was adopted the diffusion code CITATION. Was adopted a 3D representation of the core and the calculations were carried out at two energy groups. Many of the experiments were simulated, including, K{sub eff}, control rods reactivity worth, fuel elements reactivity worth distribution and the fuel temperature reactivity coefficient. The comparison of the obtained results, with the experimental results, shows differences in the range of the accuracy of the measurements, to the control rods worth and fuel temperature reactivity coefficient, or on an acceptable range, following the literature, to the K{sub eff} and fuel elements reactivity worth distribution and the fuel temperature reactivity coefficient. The comparison of the obtained results, with the experimental. results, shows differences in the range of the accuracy of the measurements, to the control rods worth and fuel temperature reactivity coefficient, or in an acceptable range, following the literature, to the K{sub eff} and fuel elements reactivity worth distribution. (author)

  7. Data collection of fusion neutronics benchmark experiment conducted at FNS/JAERI

    Energy Technology Data Exchange (ETDEWEB)

    Maekawa, Fujio; Konno, Chikara; Kasugai, Yoshimi; Oyama, Yukio; Ikeda, Yujiro [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment

    1998-08-01

    Fusion neutronics benchmark experimental data have been continued at the Fusion Neutronics Source (FNS) facility in JAERI. This report compiles unpublished results of the in-situ measurement experiments conducted by the end of 1996. Experimental data of neutron spectra in entire energy range, dosimetry reaction rates, gamma-ray spectrum and gamma-ray heating rates are acquired for five materials of beryllium, vanadium, iron, copper and tungsten. These experimental data along with data previously reported are effective for validating cross section data stored in evaluated nuclear data files such as JENDL. (author)

  8. Data collection of fusion neutronics benchmark experiment conducted at FNS/JAERI

    International Nuclear Information System (INIS)

    Fusion neutronics benchmark experimental data have been continued at the Fusion Neutronics Source (FNS) facility in JAERI. This report compiles unpublished results of the in-situ measurement experiments conducted by the end of 1996. Experimental data of neutron spectra in entire energy range, dosimetry reaction rates, gamma-ray spectrum and gamma-ray heating rates are acquired for five materials of beryllium, vanadium, iron, copper and tungsten. These experimental data along with data previously reported are effective for validating cross section data stored in evaluated nuclear data files such as JENDL. (author)

  9. Fusion blanket benchmark experiments on a 60 cm-thick lithium-oxide cylindrical assembly

    International Nuclear Information System (INIS)

    Integral experiments on a Li2O cylindrical assembly have been carried out, using the FNS facility to provide benchmark data for verification of methods and data used in fusion neutronics research. The size of assembly was 63 cm (diameter) by 61 cm (length). Measurements included 6Li and 7Li tritium production rates ; 235U, 238U, 237Np and 232Th fission rates ; 27Al(n,α)24Na, 58Ni(n,2n)57Ni, 115In(n,n1)115mIn and 115In(n,γ)116In reaction rates. Neutron energy spectra in the assembly, as well as response rates of TLDs and PIN diodes, were also measured. Measured data are presented in tabular form together with estimated errors. A sample calculation using the DOT3.5 code is provided to facilitate the reader understanding of the experiments. Although several different measuring techniques are used in the experiment, the data are mutually consistent. This fact supports that present experimental data can be applied to the benchmark verification of methods and data. (author)

  10. Simulation of hydrogen deflagration experimentBenchmark exercise with lumped-parameter codes

    International Nuclear Information System (INIS)

    Highlights: • Blind and open simulations of hydrogen combustion experiment in large-scale containment-like facility with different lumped-parameter codes. • Simulation of axial as well as radial flame propagation. • Confirmation of adequacy of lumped-parameter codes for safety analyses of actual nuclear power plants. - Abstract: An experiment on hydrogen deflagration (Upward Flame Propagation Experiment – UFPE) was proposed by the Jozef Stefan Institute (Slovenia) and performed in the HYKA A2 facility at the Karlsruhe Institute of Technology (Germany). The experimental results were used to organize a benchmark exercise for lumped-parameter codes. Six organizations (JSI, AEP, LEI, NUBIKI, RSE and UJD SR) participated in the benchmark exercise, using altogether four different computer codes: ANGAR, ASTEC, COCOSYS and ECART. Both blind and open simulations were performed. In general, all the codes provided satisfactory results of the pressure increase, whereas the results of the temperature show a wider dispersal. Concerning the flame axial and radial velocities, the results may be considered satisfactory, given the inherent simplification of the lumped-parameter description compared to the local instantaneous description

  11. Benchmark experiment on titanium with DT neutron at JAEA/FNS

    International Nuclear Information System (INIS)

    Titanium is contained in lithium titanate which is a tritium breeding material candidate. In the nuclear design, accurate nuclear data are needed. However, few benchmark experiments had been performed for titanium. We performed a benchmark experiment with a titanium assembly and a DT neutron source at JAEA/FNS. The titanium assembly was covered with Li2O blocks in order to reduce background neutrons. Dosimetry reaction rates were measured with niobium, indium and gold foils inside the assembly. And fission rates of 235U were measured by using micro fission chambers. This experiment was analyzed by using the Monte Carlo neutron transport code MCNP5-1.40 with recent nuclear data libraries of ENDF/B-VII.0, ENDF/B-VII.1, JEFF-3.1.2, JENDL-4.0 and JENDL-4.0u1. The calculation results were compared with the measured one in order to validate the nuclear data libraries of titanium. The calculated results with ENDF/B-VII.1 agreed with the measured one the best because the (n,2n) and (n,n′cont) reaction cross section data and resonance parameters were improved

  12. Simulation of hydrogen deflagration experimentBenchmark exercise with lumped-parameter codes

    Energy Technology Data Exchange (ETDEWEB)

    Kljenak, Ivo, E-mail: ivo.kljenak@ijs.si [Jožef Stefan Institute, Jamova cesta 39, SI-1000 Ljubljana (Slovenia); Kuznetsov, Mikhail, E-mail: mike.kuznetsov@kit.edu [Karlsruhe Institute of Technology, Kaiserstraße 12, 76131 Karlsruhe (Germany); Kostka, Pal, E-mail: kostka@nubiki.hu [NUBIKI Nuclear Safety Research Institute, Konkoly-Thege Miklós út 29-33, 1121 Budapest (Hungary); Kubišova, Lubica, E-mail: lubica.kubisova@ujd.gov.sk [Nuclear Regulatory Authority of the Slovak Republic, Bajkalská 27, 82007 Bratislava (Slovakia); Maltsev, Mikhail, E-mail: maltsev_MB@aep.ru [JSC Atomenergoproekt, 1, st. Podolskykh Kursantov, Moscow (Russian Federation); Manzini, Giovanni, E-mail: giovanni.manzini@rse-web.it [Ricerca sul Sistema Energetico, Via Rubattino 54, 20134 Milano (Italy); Povilaitis, Mantas, E-mail: mantas.p@mail.lei.lt [Lithuania Energy Institute, Breslaujos g.3, 44403 Kaunas (Lithuania)

    2015-03-15

    Highlights: • Blind and open simulations of hydrogen combustion experiment in large-scale containment-like facility with different lumped-parameter codes. • Simulation of axial as well as radial flame propagation. • Confirmation of adequacy of lumped-parameter codes for safety analyses of actual nuclear power plants. - Abstract: An experiment on hydrogen deflagration (Upward Flame Propagation Experiment – UFPE) was proposed by the Jozef Stefan Institute (Slovenia) and performed in the HYKA A2 facility at the Karlsruhe Institute of Technology (Germany). The experimental results were used to organize a benchmark exercise for lumped-parameter codes. Six organizations (JSI, AEP, LEI, NUBIKI, RSE and UJD SR) participated in the benchmark exercise, using altogether four different computer codes: ANGAR, ASTEC, COCOSYS and ECART. Both blind and open simulations were performed. In general, all the codes provided satisfactory results of the pressure increase, whereas the results of the temperature show a wider dispersal. Concerning the flame axial and radial velocities, the results may be considered satisfactory, given the inherent simplification of the lumped-parameter description compared to the local instantaneous description.

  13. Benchmark experiments of effective delayed neutron fraction βeff at FCA

    International Nuclear Information System (INIS)

    Benchmark experiments of effective delayed neutron fraction βeff were performed at Fast Critical Assembly (FCA) in the Japan Atomic Energy Research Institute. The experiments were made in three cores providing systematic change of nuclide contribution to the βeff: XIX-1 core fueled with 93% enriched uranium, XIX-2 core fueled with plutonium and uranium (23% enrichment) and XIX-3 core fueled with plutonium (92% fissile Pu). Six organizations from five countries participated in these experiments and measured the βeff by using their own methods and instruments. Target accuracy in the βeff was achieved to be better than ±3% by averaging the βeff values measured using a wide variety of experimental methods. (author)

  14. International comparison of criticality accident evaluation methods. Evaluation plan of super-critical benchmark based on TRACY experiment

    International Nuclear Information System (INIS)

    In order to evaluate criticality accident analysis codes, a criticality accident benchmark problem was made based on the TRACY experiment. It is evaluated by the contributors of the expert group on criticality excursion analysis, a group of criticality safety WP of OECD/NEA/NSC. This paper reports the detail of TRACY Benchmark I and II, and preliminary results of its analysis using AGNES code. (author)

  15. Benchmark experiments for validation of reaction rates determination in reactor dosimetry

    International Nuclear Information System (INIS)

    The precision of Monte Carlo calculations of quantities of neutron dosimetry strongly depends on precision of reaction rates prediction. Research reactor represents a very useful tool for validation of the ability of a code to calculate such quantities as it can provide environments with various types of neutron energy spectra. Especially, a zero power research reactor with well-defined core geometry and neutronic properties enables precise comparison between experimental and calculated data. Thus, at the VR-1 zero power research reactor, a set of benchmark experiments were proposed and carried out to verify the MCNP Monte Carlo code ability to predict correctly the reaction rates. For that purpose two frequently used reactions were chosen: He-3(n,p)H-3 and Au-197(n,γ)Au-198. The benchmark consists of response measurement of small He-3 gas filled detector in various positions of reactor core and of activated gold wires placed inside the core or to its vicinity. The reaction rates were calculated in MCNP5 code utilizing a detailed model of VR-1 reactor which was validated for neutronic calculations at the reactor. The paper describes in detail the experimental set-up of the benchmark, the MCNP model of the VR-1 reactor and provides a comparison between experimental and calculated data. - Highlights: • Use of zero power reactor for validation of reaction rates calculations. • Reaction rates measurement in reactor core by He-3 detector and Au wires. • Validation of reaction rates calculation by MCNP5. • Comparison of measured and calculated RR for different positions in core

  16. Preparation of a criticality benchmark based on experiments performed at the RA-6 reactor

    International Nuclear Information System (INIS)

    The operation and fuel management of a reactor uses neutronic modeling to predict its behavior in operational and accidental conditions. This modeling uses computational tools and nuclear data that must be contrasted against benchmark experiments to ensure its accuracy. These benchmarks have to be simple enough to be possible to model with the desired computer code and have quantified and bound uncertainties. The start-up of the RA-6 reactor, final stage of the conversion and renewal project, allowed us to obtain experimental results with fresh fuel. In this condition the material composition of the fuel elements is precisely known, which contributes to a more precise modeling of the critical condition. These experimental results are useful to evaluate the precision of the models used to design the core, based on U3Si2 and cadmium wires as burnable poisons, for which no data was previously available. The analysis of this information can be used to validate models for the analysis of similar configurations, which is necessary to follow the operational history of the reactor and perform fuel management. The analysis of the results and the generation of the model were done following the methodology established by International Criticality Safety Benchmark Evaluation Project, which gathers and analyzes experimental data for critical systems. The results were very satisfactory resulting on a value for the multiplication factor of the model of 1.0000 ± 0.0044, and a calculated value of 0.9980 ± 0.0001 using MCNP 5 and ENDF/B-VI. The utilization of as-built dimensions and compositions, and the sensitivity analysis allowed us to review the design calculations and analyze their precision, accuracy and error compensation.

  17. INTERNATIONAL EXPERIENCE OF HOSPITALITY INDUSTRY TRAINING AS A BASE-BENCHMARKING PROJECT

    Directory of Open Access Journals (Sweden)

    Наталья Николаевна Масюк

    2013-09-01

    Full Text Available Integration of universities in the international educational space includes not only the usual attraction of foreign students, but also the interaction with the best universities in the world in the field of knowledge with a view to the mutual agreement of all kinds. To do this, at the preliminary stage is implemented benchmarking project to study the experience of best practices.Objective: To study the international experience of training in the hospitality industry.Method or the methodology of the work: desk research.Results: The list of the world's best schools of the hospitality industry and examples of effective cooperation.The area of application of the results: Management, Marketing.DOI: http://dx.doi.org/10.12731/2218-7405-2013-9-20

  18. Recent accelerator experiments updates in Shielding INtegral Benchmark Archive Database (SINBAD)

    Science.gov (United States)

    Kodeli, I.; Sartori, E.; Kirk, B.

    2006-06-01

    SINBAD is an internationally established set of radiation shielding and dosimetry data relative to experiments relevant in reactor shielding, fusion blanket neutronics and accelerator shielding. In addition to the characterization of the radiation source, it describes shielding materials and instrumentation and the relevant detectors. The experimental results, be it dose, reaction rates or unfolded spectra are presented in tabular ASCII form that can easily be exported to different computer environments for further use. Most sets in SINBAD also contain the computer model used for the interpretation of the experiment and, where available, results from uncertainty analysis. The set of primary documents used for the benchmark compilation and evaluation are provided in computer readable form. SINBAD is available free of charge from RSICC and from the NEA Data Bank.

  19. The bulk shielding benchmark experiment at the Frascati Neutron Generator (FNG)

    Energy Technology Data Exchange (ETDEWEB)

    Batistoni, P. [Associazione EURATOM-ENEA sulla Fusione, Frascati (Italy); Angelone, M. [Associazione EURATOM-ENEA sulla Fusione, Frascati (Italy); Martone, M. [Associazione EURATOM-ENEA sulla Fusione, Frascati (Italy); Pillon, M. [Associazione EURATOM-ENEA sulla Fusione, Frascati (Italy); Rado, V. [Associazione EURATOM-ENEA sulla Fusione, Frascati (Italy); Santamarina, A. [Commissariat al`Energie Atomique, Centre d`Etudes de Cadarache, F-13108 St. Paul-lez-Durance Cedex (France); Abidi, I. [Commissariat al`Energie Atomique, Centre d`Etudes de Cadarache, F-13108 St. Paul-lez-Durance Cedex (France); Gastaldi, B. [Commissariat al`Energie Atomique, Centre d`Etudes de Cadarache, F-13108 St. Paul-lez-Durance Cedex (France); Martini, M. [Commissariat al`Energie Atomique, Centre d`Etudes de Cadarache, F-13108 St. Paul-lez-Durance Cedex (France); Marquette, J.P. [Commissariat al`Energie Atomique, Centre d`Etudes de Cadarache, F-13108 St. Paul-lez-Durance Cedex (France)

    1995-03-01

    In the design of next-step fusion devices such as NET/ITER the nuclear performance of shielding blankets is of key importance in terms of nuclear heating of superconducting magnets and radiation damage. In the framework of the European Fusion Technology Program, ENEA Frascati and CEA Cadarache in collaboration performed a bulk shielding benchmark experiment using the 14MeV Frascati Neutron Generator (FNG), aimed at obtaining accurate experimental data for improving the nuclear database and methods used in shielding designs. The experiment consisted of the irradiation of a stainless steel block by 14MeV neutrons. The neutron reaction rates at various depths inside the block have been measured using fission chambers and activation foils characterized by different energy response ranges. The experimental results have been compared with numerical results calculated using both S{sub n} and Monte Carlo transport codes and the cross-section library EFF.1 (European Fusion File). (orig.).

  20. Recent accelerator experiments updates in Shielding INtegral Benchmark Archive Database (SINBAD)

    Energy Technology Data Exchange (ETDEWEB)

    Kodeli, I. [IAEA Representative at OECD/NEA Data Bank, 12 bd. des Iles, 92130 Issy-les-Moulinaux (France)]. E-mail: ivo.kodeli@oecd.org; Sartori, E. [OECD Nuclear Energy Agency, 12 bd des Iles, 92130 Issy les Moulineaux (France); Kirk, B. [RSICC, Oak Ridge National Laboratory, POB 2008, Oak Ridge, TN 37831-6362 (United States)

    2006-06-23

    SINBAD is an internationally established set of radiation shielding and dosimetry data relative to experiments relevant in reactor shielding, fusion blanket neutronics and accelerator shielding. In addition to the characterization of the radiation source, it describes shielding materials and instrumentation and the relevant detectors. The experimental results, be it dose, reaction rates or unfolded spectra are presented in tabular ASCII form that can easily be exported to different computer environments for further use. Most sets in SINBAD also contain the computer model used for the interpretation of the experiment and, where available, results from uncertainty analysis. The set of primary documents used for the benchmark compilation and evaluation are provided in computer readable form. SINBAD is available free of charge from RSICC and from the NEA Data Bank.

  1. Recent accelerator experiments updates in Shielding INtegral Benchmark Archive Database (SINBAD)

    International Nuclear Information System (INIS)

    SINBAD is an internationally established set of radiation shielding and dosimetry data relative to experiments relevant in reactor shielding, fusion blanket neutronics and accelerator shielding. In addition to the characterization of the radiation source, it describes shielding materials and instrumentation and the relevant detectors. The experimental results, be it dose, reaction rates or unfolded spectra are presented in tabular ASCII form that can easily be exported to different computer environments for further use. Most sets in SINBAD also contain the computer model used for the interpretation of the experiment and, where available, results from uncertainty analysis. The set of primary documents used for the benchmark compilation and evaluation are provided in computer readable form. SINBAD is available free of charge from RSICC and from the NEA Data Bank

  2. Quality in E-Learning--A Conceptual Framework Based on Experiences from Three International Benchmarking Projects

    Science.gov (United States)

    Ossiannilsson, E.; Landgren, L.

    2012-01-01

    Between 2008 and 2010, Lund University took part in three international benchmarking projects, "E-xcellence+," the "eLearning Benchmarking Exercise 2009," and the "First Dual-Mode Distance Learning Benchmarking Club." A comparison of these models revealed a rather high level of correspondence. From this finding and from desktop studies of the…

  3. Scope of STACY experiments for criticality benchmark data on low enriched uranium and plutonium solution system

    International Nuclear Information System (INIS)

    The Static Experiment Critical Facility, STACY was constructed in the Nuclear Fuel Cycle Safety Engineering Research Facility, NUCEF of the Japan Atomic Energy Research Institute in order to produce the fundamental critical data of uranyl nitrate solution, plutonium nitrate solution and their mixture. A series of experiments using single core tank have been performed using 10% enriched uranyl nitrate solution since the first criticality in 1995. Benchmark data of STACY are now used for verification of Japanese criticality safety code system and nuclear data libraries. Kinetic parameters, temperature coefficients and reflector effects of structural material are also measured using single homogeneous core. It is on schedule to make experiments for neutron interaction effect and for simulating the dissolving process with a heterogeneous core using low enriched uranyl nitrate solution. After these experiments, systematic critical and subcritical experiments on plutonium nitrate solution will start in five years. This paper reviews the main results of STACY since the initial criticality and describes the criticality properties of the experimental cores in the future program. (author)

  4. Benchmark study on neutron cross sections based on pulsed sphere experiment

    International Nuclear Information System (INIS)

    Benchmark validation of neutron cross sections was performed by comparing theoretical calculated leakage spectra with measured ones by means of pulsed sphere experiment conducted at OKTAVIAN of Osaka University, FZK, German and IPPE Obninsk, Russia. It was found out that the nuclear data of Be in both JENDL Fusion File and ENDF/B-VI had similar trend in each experiment. However, there exists some discrepancy among the three different experiments, which suggests further study is needed to validate Be nuclear data. Calculated spectra for Li, Cr, Mn, Cu, Zr, Nb and Mo using JENDL Fusion File predict the experiment fairly well. However, for LiF, (CF2)n, Si, Ti, Co and W, the calculated spectra are not in good agreement with the measurement. The prediction using FENDL-1.0 data gives NN agreement in case of Li, Cr, Mn, Cu and Mo, whereas in other case, the prediction gives insufficient result. The analysis of Fe and Pb experiment conducted at IPPE Obninsk showed that Fe data in JENDL Fusion File and JENDL-3.2 were much better than FENDL/E-1.0. Pb data in JENDL Fusion File appeared to have been much improved as compared with JENDL-3.2 evaluation. (author)

  5. Benchmark experiment on the model of fusion reactor blanket with uranium neutron multiplier

    International Nuclear Information System (INIS)

    Benchmark experiment on the model of thermonuclear reactor blanket with 14 MeV neutron source is described. The model design corresponds to the known concept of the fast hybrid blanket with 238U neutron multiplier and main tritium production on 6Li. Detailed measurements of the following process velocities were carried out: tritium production on lithium isotopes; reactions modelling tritium production; (n, γ) and (n, 2n) processes for 238U; fission reactions for 235,238U, 239Pu, 237Np. Neutron flux integral measurements were performed by a set of threshold detectors on the basic of the 115In(n, n'), 204Pb(n, n'), 64Zn(n, p), 27Al(n, p), 56Fe(n, p), 107Ag(n, 2n), 63Cu(n, 2n) and 64(n, 2n) reactions

  6. Benchmark experiment on a copper slab assembly bombarded by D-T neutrons

    International Nuclear Information System (INIS)

    Copper is a very important material for fusion reactor because it is used in superconducting magnets or first walls and so on. To verify nuclear data of copper, a benchmark experiment was performed using the D-T neutron source of the FNS facility in Japan Atomic Energy Research Institute. An cylindrical experimental assembly of 629 mm in diameter and 608 mm in thickness made of pure copper was located at 200 mm from the D-T neutron source. In the assembly, the following quantities were measured; i) neutron spectra in energy regions of MeV and keV, ii) neutron reaction rates, iii) prompt and decay gamma-ray spectra and iv) gamma-ray heating rates. The obtained experimental data were compiled in this report. (author)

  7. The ISMIP-HOM benchmark experiments performed using the Finite-Element code Elmer

    Directory of Open Access Journals (Sweden)

    O. Gagliardini

    2008-06-01

    Full Text Available The aim of this paper is to describe in detail how the benchmark tests ISMIP-HOM (Ice Sheet Model Intercomparison Project-Higher-Order ice-sheet Model have been performed using the open source finite element (FE code Elmer (http://www.csc.fi/elmer. The ISMIP-HOM setup consists of five diagnostic and one prognostic experiments, for both 2-D and 3-D geometries. For all the tests, the full-Stokes equations are solved. Some technical points concerning FE, such as mesh characteristics, stabilisation methods, numerical methods used to solve the linear system and parallel performance are discussed. For all these setups, the CPU time consumption in relation to the accuracy of the solution is analysed. Based on these findings, some general rules on optimising the computing time versus the accuracy of the results are deduced.

  8. The ISMIP-HOM benchmark experiments performed using the Finite-Element code Elmer

    Directory of Open Access Journals (Sweden)

    O. Gagliardini

    2008-02-01

    Full Text Available The aim of this paper is to describe in detail how the benchmark tests ISMIP-HOM (Ice Sheet Model Intercomparison Project – Higher-Order ice-sheet Model has been performed using the open-Source finite element code Elmer (http://www.csc.fi/elmer. The ISMIP-HOM setup consists of five diagnostic and one prognostic experiments, for both 2-D and 3-D geometries. For all the tests, the full-Stokes equations are solved. Some FE technical points, such as mesh characteristics, stabilisation methods, numerical methods used to solve the linear system and parallel performance are discussed. For all these setups, the CPU time consumption is analysed in comparison to the accuracy of the solution. Some general rules are then inferred that optimise the computing time versus the accuracy of the results.

  9. A parallel high-order accurate finite element nonlinear Stokes ice sheet model and benchmark experiments

    Energy Technology Data Exchange (ETDEWEB)

    Leng, Wei [Chinese Academy of Sciences; Ju, Lili [University of South Carolina; Gunzburger, Max [Florida State University; Price, Stephen [Los Alamos National Laboratory; Ringler, Todd [Los Alamos National Laboratory,

    2012-01-01

    The numerical modeling of glacier and ice sheet evolution is a subject of growing interest, in part because of the potential for models to inform estimates of global sea level change. This paper focuses on the development of a numerical model that determines the velocity and pressure fields within an ice sheet. Our numerical model features a high-fidelity mathematical model involving the nonlinear Stokes system and combinations of no-sliding and sliding basal boundary conditions, high-order accurate finite element discretizations based on variable resolution grids, and highly scalable parallel solution strategies, all of which contribute to a numerical model that can achieve accurate velocity and pressure approximations in a highly efficient manner. We demonstrate the accuracy and efficiency of our model by analytical solution tests, established ice sheet benchmark experiments, and comparisons with other well-established ice sheet models.

  10. Verification of HELIOS-MASTER system through benchmark of critical experiments

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Ha Yong; Kim, Kyo Yun; Cho, Byung Oh; Lee, Chung Chan; Zee, Sung Quun

    1999-03-01

    HELIOS-MASTER code system is verified through the benchmark of the critical experiments that were performed by RRC Kurchatov Institute with water moderated hexagonally pitched lattices of highly enriched uranium fuel rods (80w/o). We also used the same input by using MCNP code that was described in evaluation report, and compare our results with those of evaluation report. HELIOS developed by Scandpower A/S is a two-dimensional transport program for generation of group cross sections and MASTER developed by KAERI is a three-dimensional nuclear design and analysis code based on the two group diffusion theory. It solves neutronics model with AFEN (Analytic Function Expansion Nodal) method for hexagonal geometry. The results show that HELIOS-MASTER code system is fast and accurate enough so that this code system can be used as nuclear core analysis tool for hexagonal geometry. (author). 4 refs., 4 tabs., 10 figs.

  11. Verification of HELIOS-MASTER system through benchmark of critical experiments

    International Nuclear Information System (INIS)

    HELIOS-MASTER code system is verified through the benchmark of the critical experiments that were performed by RRC Kurchatov Institute with water moderated hexagonally pitched lattices of highly enriched uranium fuel rods (80w/o). We also used the same input by using MCNP code that was described in evaluation report, and compare our results with those of evaluation report. HELIOS developed by Scandpower A/S is a two-dimensional transport program for generation of group cross sections and MASTER developed by KAERI is a three-dimensional nuclear design and analysis code based on the two group diffusion theory. It solves neutronics model with AFEN (Analytic Function Expansion Nodal) method for hexagonal geometry. The results show that HELIOS-MASTER code system is fast and accurate enough so that this code system can be used as nuclear core analysis tool for hexagonal geometry. (author). 4 refs., 4 tabs., 10 figs

  12. Dose Rate Experiment at JET for Benchmarking the Calculation Direct One Step Method

    International Nuclear Information System (INIS)

    Neutrons produced by D-D and D-T plasmas induce the activation of tokamak materials and of components. The development of reliable methods to assess dose rates is a key issue for maintenance and operating nuclear machines, in normal and off-normal conditions. In the frame of the EFDA Fusion Technology work programme, a computational tool based upon MCNP Monte Carlo code has been developed to predict the dose rate after shutdown: it is called Direct One Step Method (D1S). The D1S is an innovative approach in which the decay gammas are coupled to the neutrons as in the prompt case and they are transported in one single step in the same run. Benchmarking of this new tool with experimental data taken in a complex geometry like that of a tokamak is a fundamental step to test the reliability of the D1S method. A dedicated benchmark experiment was proposed for the 2005-2006 experimental campaign of JET. Two irradiation positions have been selected for the benchmark: one inner position inside the vessel, not far from the plasma, called the 2 upper irradiation end (IE2), where neutron fluence is relatively high. The second position is just outside a vertical port in an external position (EX). Here the neutron flux is lower and the dose rate to be measured is not very far from the residual background. Passive detectors are used for in-vessel measurements: the high sensitivity Thermo Luminescent Dosimeters (TLDs) GR-200A (natural LiF), which ensure measurements down to environmental dose level. An active detector of Geiger-Muller (GM) type is used for out of vessel dose rate measurement. Before their use the detectors were calibrated in a secondary gamma-ray standard (Cs-137 and Co-60) facility in term of air-kerma. The background measurement was carried-out in the period July -September 2005 in the outside position EX using the GM tube and in September 2005 inside the vacuum vessel using TLD detectors located in the 2 Upper irradiation end IE2. In the present work

  13. Three-dimensional coupled kinetics/thermal- hydraulic benchmark TRIGA experiments

    International Nuclear Information System (INIS)

    This research project provides separate effects tests in order to benchmark neutron kinetics models coupled with thermal-hydraulic (T/H) models used in best-estimate codes such as the Nuclear Regulatory Commission's (NRC) RELAP and TRAC code series and industrial codes such as RETRAN. Before this research project was initiated, no adequate experimental data existed for reactivity initiated transients that could be used to assess coupled three-dimensional (3D) kinetics and 3D T/H codes which have been, or are being developed around the world. Using various Test Reactor Isotope General Atomic (TRIGA) reactor core configurations at the Penn State Breazeale Reactor (PSBR), it is possible to determine the level of neutronics modeling required to describe kinetics and T/H feedback interactions. This research demonstrates that the small compact PSBR TRIGA core does not necessarily behave as a point kinetics reactor, but that this TRIGA can provide actual test results for 3D kinetics code benchmark efforts. This research focused on developing in-reactor tests that exhibited 3D neutronics effects coupled with 3D T/H feedback. A variety of pulses were used to evaluate the level of kinetics modeling needed for prompt temperature feedback in the fuel. Ramps and square waves were used to evaluate the detail of modeling needed for the delayed T/H feedback of the coolant. A stepped ramp was performed to evaluate and verify the derived thermal constants for the specific PSBR TRIGA core loading pattern. As part of the analytical benchmark research, the STAR 3D kinetics code (, STAR: Space and time analysis of reactors, Version 5, Level 3, Users Guide, Yankee Atomic Electric Company, YEAC 1758, Bolton, MA) was used to model the transient experiments. The STAR models were coupled with the one-dimensional (1D) WIGL and LRA and 3D COBRA (, COBRA IIIC: A digital computer program for steady-state and transient thermal-hydraulic analysis of rod bundle nuclear fuel elements, Battelle

  14. The Benchmark experiment on stainless steel bulk shielding at the Frascati neutron generator

    International Nuclear Information System (INIS)

    In the framework of the European Technology Program for NET/ITER, ENEA (Italian Agency for New Technologies, Energy and Environment) - Frascati and CEA (Commissariat a L'Energie Atomique) - Cadarache collaborated on a Bulk Shield Benchmark Experiment using the 14-MeV Frascati Neutron Generator (FNG). The aim of the experiment was to obtain accurate experimental data for improving the nuclear database and methods used in shielding designs, through a rigorous analysis of the results. The experiment consisted of the irradiation of a stainless steel block by 14-MeV neutrons. The neutron reaction rates at different depths inside the block were measured by fission chambers and activation foils characterized by different energy response ranges. The experimental results have been compared with numerical results calculated using both SN and Monte Carlo transport codes and as transport cross section library the European Fusion File (EFF). In particular, the present report describes the experimental and numerical activity, including neutron measurements and Monte Carlo calculations, carried out by the ENEA Italian Agency for New Technologies, Energy and Environment) team

  15. Study on shielding design methods for fusion reactors using benchmark experiments

    International Nuclear Information System (INIS)

    In this study, a series of engineering benchmark experiments have been performed on the critical issues of shielding designs for DT fusion reactors. Based on the experiments, calculational accuracy of shielding design methods used in the ITER conceptual design, discrete ordinates code DOT3.5 and Monte Carlo code MCNP-3, have been estimated, and difficulties on calculational methods have been revealed. Furthermore, the feasibility for shielding designs have been examined with respect to a discrete ordinates code system BERMUDA which is developed to attain high accuracy of calculation. As for neutron streaming in an off-set narrow gap experimental assembly made of stainless steel, DOT3.5 and MCNP-3 codes reproduced the experiments within the accuracy presumed in the ITER conceptual design. DOT3.5 and MCNP-3 codes are available for secondary γ ray nuclear heating in a type 316L stainless steel assembly and neutron streaming in a multi-layered slit experimental assembly, respectively. Moreover, BERMUDA-2DN code is an effective tool as to neutron deep penetration in a type 316L stainless steel assembly and the neutron behavior in a large cavity experimental assembly. (author)

  16. The stainless steel bulk shielding benchmark experiment at the Frascati Neutron Generator (FNG)

    Energy Technology Data Exchange (ETDEWEB)

    Batistoni, P. (Associazione Euratom-ENEA sulla Fusione, CRE Frascati, I-00044 Frascati, Rome (Italy)); Angelone, M. (Associazione Euratom-ENEA sulla Fusione, CRE Frascati, I-00044 Frascati, Rome (Italy)); Martone, M. (Associazione Euratom-ENEA sulla Fusione, CRE Frascati, I-00044 Frascati, Rome (Italy)); Petrizzi, L. (Associazione Euratom-ENEA sulla Fusione, CRE Frascati, I-00044 Frascati, Rome (Italy)); Pillon, M. (Associazione Euratom-ENEA sulla Fusione, CRE Frascati, I-00044 Frascati, Rome (Italy)); Rado, V. (Associazione Euratom-ENEA sulla Fusione, CRE Frascati, I-00044 Frascati, Rome (Italy)); Santamarina, A. (Commissariat a l' Energie Atomique, Centre d' Etudes Nucleaires Cadarache, 13108, St.-Paul-lez-Durance Cedex (France)); Abidi, I. (Commissariat a l' Energie Atomique, Centre d' Etudes Nucleaires Cadarache, 13108, St.-Paul-lez-Durance Cedex (France)); Gastaldi, G. (Commissariat a l' Energie Atomique, Centre d' Etudes Nucleaires Cadarache, 13108, St.-Paul-lez-Durance Cedex

    1994-09-01

    In the framework of the European Technology Program for NET/ITER, ENEA (Ente Nazionale per le Nuove Tecnologie, l'Energia e l'Ambiente), Frascati and CEA (Commissariat a l'Energie Atomique), Cadarache, are collaborating on a bulk shielding benchmark experiment using the 14 MeV Frascati Neutron Generator (FNG). The aim of the experiment is to obtain accurate experimental data for improving the nuclear database and methods used in the shielding designs, through a rigorous analysis of the results. The experiment consists of the irradiation of a stainless steel block by 14 MeV neutrons. The neutron flux and spectra at different depths, up to 65 cm inside the block, are measured by fission chambers and activation foils characterized by different energy response ranges. The [gamma]-ray dose measurements are performed with ionization chambers and thermo-luminescent dosimeters (TLD). The first results are presented, as well as the comparison with calculations using the cross section library EFF (European Fusion File). ((orig.))

  17. Study on shielding design methods for fusion reactors using benchmark experiments

    Energy Technology Data Exchange (ETDEWEB)

    Nakashima, Hiroshi (Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment)

    1992-02-01

    In this study, a series of engineering benchmark experiments have been performed on the critical issues of shielding designs for DT fusion reactors. Based on the experiments, calculational accuracy of shielding design methods used in the ITER conceptual design, discrete ordinates code DOT3.5 and Monte Carlo code MCNP-3, have been estimated, and difficulties on calculational methods have been revealed. Furthermore, the feasibility for shielding designs have been examined with respect to a discrete ordinates code system BERMUDA which is developed to attain high accuracy of calculation. As for neutron streaming in an off-set narrow gap experimental assembly made of stainless steel, DOT3.5 and MCNP-3 codes reproduced the experiments within the accuracy presumed in the ITER conceptual design. DOT3.5 and MCNP-3 codes are available for secondary {gamma} ray nuclear heating in a type 316L stainless steel assembly and neutron streaming in a multi-layered slit experimental assembly, respectively. Moreover, BERMUDA-2DN code is an effective tool as to neutron deep penetration in a type 316L stainless steel assembly and the neutron behavior in a large cavity experimental assembly. (author).

  18. Benchmark experiment on stainless steel bulk shielding at Frascati neutron generator

    Energy Technology Data Exchange (ETDEWEB)

    Batistoni, P.; Angelone, M.; Martone, M.; Pillon, M.; Rado, V. [ENEA, Frascati (Italy). Centro Ricerche Energia - Area Energia e Innovazione

    1994-11-01

    In the framework of the European Technology Program for NET/ITER, ENEA (Italian Agency for New Technologies, Energy and Environment) - Frascati and CEA (Commissariat a L`Energie Atomique) - Cadarache collaborated on a Bulk Shield Benchmark Experiment using the 14-MeV Frascati Neutron Generator (FNG). The aim of the experiment was to obtain accurate experimental data for improving the nuclear database and methods used in shielding designs, through a rigorous analysis of the results. The experiment consisted of the irradiation of a stainless steel block by 14-MeV neutrons. The neutron reaction rates at different depths inside the block were measured by fission chambers and activation foils characterized by different energy response ranges. The experimental results have been compared with numerical results calculated using both S{sub N} and Monte Carlo transport codes and as transport cross section library the European Fusion File (EFF). In particular, the present report describes the experimental and numerical activity, including neutron measurements and Monte Carlo calculations, carried out by the (ENEA Italian Agency for New Technologies, Energy and Environment) team.

  19. The stainless steel bulk shielding benchmark experiment at the Frascati Neutron Generator (FNG)

    Science.gov (United States)

    Batistoni, P.; Angelone, M.; Martone, M.; Petrizzi, L.; Pillon, M.; Rado, V.; Santamarina, A.; Abidi, I.; Gastaldi, G.; Joyer, P.; Marquette, J. P.; Martini, M.

    1994-09-01

    In the framework of the European Technology Program for NET/ITER, ENEA (Ente Nazionale per le Nuove Tecnologie, l'Energia e l'Ambiente), Frascati and CEA (Commissariat à l'Energie Atomique), Cadarache, are collaborating on a bulk shielding benchmark experiment using the 14 MeV Frascati Neutron Generator (FNG). The aim of the experiment is to obtain accurate experimental data for improving the nuclear database and methods used in the shielding designs, through a rigorous analysis of the results. The experiment consists of the irradiation of a stainless steel block by 14 MeV neutrons. The neutron flux and spectra at different depths, up to 65 cm inside the block, are measured by fission chambers and activation foils characterized by different energy response ranges. The γ-ray dose measurements are performed with ionization chambers and thermo-luminescent dosimeters (TLD). The first results are presented, as well as the comparison with calculations using the cross section library EFF (European Fusion File).

  20. Experimental investigation of the natural and forced convection on solidification of Sn-3wt. %Pb alloy using a benchmark experiment

    OpenAIRE

    Hachani, Lakhdar; Boussaa, Redouane; Saadi, Bachir; X. D. Wang; Zaidat, Kader; Belgacem Bouzida, Aissaa; Henry, D.; Henry, Dominique; Botton, Val‚éry; Ben Hadid, H.; Fautrelle, Yves

    2012-01-01

    International audience We deal with the development of a solidification benchmark experiment in order to investigate the structure formation as well as solute macro-mesosegregation, by means of a well-controlled solidification experiment. The experiment consists in solidifying a rectangular ingot of Sn-3wt.%Pb alloy, by using two lateral heat exchangers which allow extraction of the heat flux from one or two vertical sides of the sample. The domain is a quasi two dimensional parallepipedic...

  1. Benchmarking atomic physics models for magnetically confined fusion plasma physics experiments

    International Nuclear Information System (INIS)

    In present magnetically confined fusion devices, high and intermediate Z impurities are either puffed into the plasma for divertor radiative cooling experiments or are sputtered from the high Z plasma facing armor. The beneficial cooling of the edge as well as the detrimental radiative losses from the core of these impurities can be properly understood only if the atomic physics used in the modeling of the cooling curves is very accurate. To this end, a comprehensive experimental and theoretical analysis of some relevant impurities is undertaken. Gases (Ne, Ar, Kr, and Xe) are puffed and nongases are introduced through laser ablation into the FTU tokamak plasma. The charge state distributions and total density of these impurities are determined from spatial scans of several photometrically calibrated vacuum ultraviolet and x-ray spectrographs (3 - 1600 Angstrom), the multiple ionization state transport code transport code (MIST) and a collisional radiative model. The radiative power losses are measured with bolometery, and the emissivity profiles were measured by a visible bremsstrahlung array. The ionization balance, excitation physics, and the radiative cooling curves are computed from the Hebrew University Lawrence Livermore atomic code (HULLAC) and are benchmarked by these experiments. (Supported by U.S. DOE Grant No. DE-FG02-86ER53214 at JHU and Contract No. W-7405-ENG-48 at LLNL.) copyright 1999 American Institute of Physics

  2. Benchmarking atomic physics models for magnetically confined fusion plasma physics experiments

    Science.gov (United States)

    May, M. J.; Finkenthal, M.; Soukhanovskii, V.; Stutman, D.; Moos, H. W.; Pacella, D.; Mazzitelli, G.; Fournier, K.; Goldstein, W.; Gregory, B.

    1999-01-01

    In present magnetically confined fusion devices, high and intermediate Z impurities are either puffed into the plasma for divertor radiative cooling experiments or are sputtered from the high Z plasma facing armor. The beneficial cooling of the edge as well as the detrimental radiative losses from the core of these impurities can be properly understood only if the atomic physics used in the modeling of the cooling curves is very accurate. To this end, a comprehensive experimental and theoretical analysis of some relevant impurities is undertaken. Gases (Ne, Ar, Kr, and Xe) are puffed and nongases are introduced through laser ablation into the FTU tokamak plasma. The charge state distributions and total density of these impurities are determined from spatial scans of several photometrically calibrated vacuum ultraviolet and x-ray spectrographs (3-1600 Å), the multiple ionization state transport code transport code (MIST) and a collisional radiative model. The radiative power losses are measured with bolometery, and the emissivity profiles were measured by a visible bremsstrahlung array. The ionization balance, excitation physics, and the radiative cooling curves are computed from the Hebrew University Lawrence Livermore atomic code (HULLAC) and are benchmarked by these experiments. (Supported by U.S. DOE Grant No. DE-FG02-86ER53214 at JHU and Contract No. W-7405-ENG-48 at LLNL.)

  3. Criticality and reactor physics benchmark experiments. Influence of nuclear data uncertainties

    International Nuclear Information System (INIS)

    A number of LWR-type criticality and reactor physics experiments, mainly from the ICSBEP and IRPhEP Handbooks, many of which were already used in the past for international benchmarks in the framework of OECD/NEA working groups, are being evaluated with respect to uncertainties in the basic nuclear data. For this, the sampling based uncertainty and sensitivity analysis tool XSUSA along with the Monte Carlo code KE-NO-V.a as transport solver is employed. Particular emphasis is put on experiments where differential quantities, mainly reaction rate distributions, were measured; the uncertainties of such quantities are not directly accessible to tools based on first order perturbation theory. With respect to multiplication factors and reactivity differences, all results are compared with corresponding results obtained with TSUNAMI-3D from the SCALE 6.1 system; the agreement is very good for all assemblies under consideration. With respect to fission rate distributions, the uncertainty analyses yield only moderate uncertainties from nuclear data; therefore, in general the total uncertainty is dominated by measurement uncertainties, which include the uncertainties of technological parameters. The work is continuously being extended; in the future, also non-LWR specific assemblies, mainly relevant for GEN-IV reactors, will be investigated. (author)

  4. Set of benchmark experiments on slit shielding compositions of thermonuclear reactors

    Energy Technology Data Exchange (ETDEWEB)

    Andreev, M.I.; Afanasiev, V.V.; Belevitin, A.G.; Karaulov, A.V.; Romodanov, V.L. E-mail: rom@lng.mephi.msk.su; Sakharov, V.K.; Tikhomirov, G.V.; Vasiliev, A.P.; Kandiev, Ya. Z.; Lyutov, V.D.; Sokolov, Yu. A.; Terekhin, V.A.; Shmakov, V.M.; Androsenko, P.A.; Semenov, V.P.; Trykov, L.A.; Lopatkin, A.V.; Muratov, V.G

    2001-09-01

    The paper is based on the results of the ISTC project no. 180 that has recently been completed. The aim of the project was the development of methodical, hardware and design basis to carry out computational and experimental research on non-uniform shieldings of thermonuclear reactors. As a result a set of benchmark experiments were created. On their basis verification of the domestic and foreign computational codes with the nuclear data estimated was realized. For these purposes the iron hollow slits shielding compositions irradiated with 14.8 MeV energy neutrons were studied. The experimental installations allowed research of the shielding compositions with the following characteristics: a solid structure, a structure with one slit of a central symmetry, and the structures with asymmetric slits and with two slits. The thickness of shielding compositions in this research was 500 mm. The results of experiments were compared to the results of calculations by means of the MCNP-4a and PRIZMA computing codes with use of the FENDL-1.1, FENDL-2, JENDL-3.2 and BAS-78 libraries of nuclear data. The results of comparison made it possible to obtain the recommendations for use of these nuclear data.

  5. Benchmark analysis of TRIGA mark II reactivity experiment using a continuous energy Monte Carlo code MCNP

    International Nuclear Information System (INIS)

    The benchmark analysis of reactivity experiments in the TRIGA-II core at the Musashi Institute of Technology Research Reactor (Musashi reactor; 100 kW) was performed by a three-dimensional continuous-energy Monte Carlo code MCNP4A. The reactivity worth and integral reactivity curves of the control rods as well as the reactivity worth distributions of fuel and graphite elements were used in the validation process of the physical model and neutron cross section data from the ENDF/B-V evaluation. The calculated values of integral reactivity curves of the control rods were in agreement with the experimental data obtained by the period method. The integral worth measured by the rod drop method was also consistent with the calculation. The calculated values of the fuel and the graphite element worth distributions were consistent with the measured ones within the statistical error estimates. These results showed that the exact core configuration including the control rod positions to reproduce the fission source distribution in the experiment must be introduced into the calculation core for obtaining the precise solution. It can be concluded that our simulation model of the TRIGA-II core is precise enough to reproduce the control rod worth, fuel and graphite elements reactivity worth distributions. (author)

  6. Simulation of TRIGA Mark II Benchmark Experiment using WIMSD4 and CITATION codes

    International Nuclear Information System (INIS)

    This paper presents a simulation of the TRIGA Mark II Benchmark Experiment, Part I: Steady-State Operation and is part of the calculation methodology validation developed to the neutronic calculation of the CDTN's TRIGA IPR - R1 reactor. A version of the WIMSD4, obtained in the Centro de Tecnologia Nuclear, in Cuba, was used in the cells calculation. In the core calculations was adopted the diffusion code CITATION. Was adopted a 3D representation of the core and the calculations were carried out at two energy groups. Many of the experiments were simulated, including, Keff, control rods reactivity worth, fuel elements reactivity worth distribution and the fuel temperature reactivity coefficient. The comparison of the obtained results, with the experimental results, shows differences in the range of the accuracy of the measurements, to the control rods worth and fuel temperature reactivity coefficient, or on an acceptable range, following the literature, to the Keff and fuel elements reactivity worth distribution and the fuel temperature reactivity coefficient. The comparison of the obtained results, with the experimental. results, shows differences in the range of the accuracy of the measurements, to the control rods worth and fuel temperature reactivity coefficient, or in an acceptable range, following the literature, to the Keff and fuel elements reactivity worth distribution. (author)

  7. Benchmark analysis of the DeCART MOC code with the VENUS-2 critical experiment

    International Nuclear Information System (INIS)

    Computational benchmarks based on well-defined problems with a complete set of input and a unique solution are often used as a means of verifying the reliability of numerical solutions. VENUS is a widely used MOX benchmark problem for the validation of numerical methods and nuclear data set. In this paper, the results of benchmarking the DeCART (Deterministic Core Analysis based on Ray Tracing) integral transport code is reported using the OECD/NEA VENUS-2 MOX benchmark problem. Both 2-D and 3-D DeCART calculations were performed and comparisons are reported with measured data, as well as with the results of other benchmark participants. In general the DeCART results agree well with both the experimental data as well as those of other participants. (authors)

  8. Recent benchmarking experience of the OECD/Nea expert group on three-dimensional radiation transport

    International Nuclear Information System (INIS)

    Experience of the OECD/NEA Expert Group on three-dimensional radiation transport pertaining to the C5 MOX fuel problem is examined, and the group's activity since the project's initiation in the fall of 1999 is reviewed. Twenty groups from seven nations submitted solutions to the two- and/or three-dimensional forms of the problem. Their solution methods are compared and their results analyzed. Observations are drawn from the benchmarking experience to assess the strengths and weaknesses of current methods and to better understand the challenges encountered by those who seek to obtain accurate solutions to large-scale multidimensional neutron transport problems. Drawing on the tabulated results, our e-mail correspondence and telephone conversations with participants, and on our own parametric studies, we are able to share some insights concerning space-angle transport approximations. For light water reactor physics problems without spatial homogenization like this one, refinement of angular approximations proved to be a greater challenge than refinement of the spatial approximation. Either a stair-step representation of the fuel-coolant interface or a polygonal representation was sufficient to describe the pin cell geometry, but only if great care was take to preserve the fuel volume exactly. The use of a high order angular approximation, such as S32 or P31, was required to obtain an accurate pin power and eigenvalue solution. The dangers of employing just one level of space-angle approximation became apparent to a number of participants during the course of this benchmark exercise. Frequently, eigenvalue errors resulting from coarse angular and spatial approximations have opposing effects on the eigenvalue. Thus, by cancellation of error, an accurate eigenvalue can be obtained using a coarse space-angle approximation while the flux solution is quite inaccurate. As a result, refinement of the spatial or angular approximation in such situations can cause the

  9. Research Reactor Benchmarks

    International Nuclear Information System (INIS)

    A criticality benchmark experiment performed at the Jozef Stefan Institute TRIGA Mark II research reactor is described. This experiment and its evaluation are given as examples of benchmark experiments at research reactors. For this reason the differences and possible problems compared to other benchmark experiments are particularly emphasized. General guidelines for performing criticality benchmarks in research reactors are given. The criticality benchmark experiment was performed in a normal operating reactor core using commercially available fresh 20% enriched fuel elements containing 12 wt% uranium in uranium-zirconium hydride fuel material. Experimental conditions to minimize experimental errors and to enhance computer modeling accuracy are described. Uncertainties in multiplication factor due to fuel composition and geometry data are analyzed by sensitivity analysis. The simplifications in the benchmark model compared to the actual geometry are evaluated. Sample benchmark calculations with the MCNP and KENO Monte Carlo codes are given

  10. Monte Carlo Simulation of the TRIGA Mark II Benchmark Experiment with Burned Fuel

    International Nuclear Information System (INIS)

    Monte Carlo calculations of a criticality experiment with burned fuel on the TRIGA Mark II research reactor are presented. The main objective was to incorporate burned fuel composition calculated with the WIMSD4 deterministic code into the MCNP4B Monte Carlo code and compare the calculated keff with the measurements. The criticality experiment was performed in 1998 at the ''Jozef Stefan'' Institute TRIGA Mark II reactor in Ljubljana, Slovenia, with the same fuel elements and loading pattern as in the TRIGA criticality benchmark experiment with fresh fuel performed in 1991. The only difference was that in 1998, the fuel elements had on average burnup of ∼3%, corresponding to 1.3-MWd energy produced in the core in the period between 1991 and 1998. The fuel element burnup accumulated during 1991-1998 was calculated with the TRIGLAV in-house-developed fuel management two-dimensional multigroup diffusion code. The burned fuel isotopic composition was calculated with the WIMSD4 code and compared to the ORIGEN2 calculations. Extensive comparison of burned fuel material composition was performed for both codes for burnups up to 20% burned 235U, and the differences were evaluated in terms of reactivity. The WIMSD4 and ORIGEN2 results agreed well for all isotopes important in reactivity calculations, giving increased confidence in the WIMSD4 calculation of the burned fuel material composition. The keff calculated with the combined WIMSD4 and MCNP4B calculations showed good agreement with the experimental values. This shows that linking of WIMSD4 with MCNP4B for criticality calculations with burned fuel is feasible and gives reliable results

  11. 2-D Circulation Control Airfoil Benchmark Experiments Intended for CFD Code Validation

    Science.gov (United States)

    Englar, Robert J.; Jones, Gregory S.; Allan, Brian G.; Lin, Johb C.

    2009-01-01

    A current NASA Research Announcement (NRA) project being conducted by Georgia Tech Research Institute (GTRI) personnel and NASA collaborators includes the development of Circulation Control (CC) blown airfoils to improve subsonic aircraft high-lift and cruise performance. The emphasis of this program is the development of CC active flow control concepts for both high-lift augmentation, drag control, and cruise efficiency. A collaboration in this project includes work by NASA research engineers, whereas CFD validation and flow physics experimental research are part of NASA s systematic approach to developing design and optimization tools for CC applications to fixed-wing aircraft. The design space for CESTOL type aircraft is focusing on geometries that depend on advanced flow control technologies that include Circulation Control aerodynamics. The ability to consistently predict advanced aircraft performance requires improvements in design tools to include these advanced concepts. Validation of these tools will be based on experimental methods applied to complex flows that go beyond conventional aircraft modeling techniques. This paper focuses on recent/ongoing benchmark high-lift experiments and CFD efforts intended to provide 2-D CFD validation data sets related to NASA s Cruise Efficient Short Take Off and Landing (CESTOL) study. Both the experimental data and related CFD predictions are discussed.

  12. Benchmark experiments of fusion neutron induced gamma-ray radioactivity in various structural materials

    International Nuclear Information System (INIS)

    The fusion reactor inventory code FISPACT, together with the European Activation File EAF, is the European reference software for calculating the neutron-induced activation of fusion reactor relevant materials. Experimental verifications (benchmarks) of the code predictions have been performed at ENEA Frascati by means of an irradiation facility consisting of a D-T neutron generator and a moderator/reflector structure which is employed to mimic the neutron spectrum at the a fusion device first wall. Various materials (vanadium alloy, SiC, AlSI 316, martensitic steel F82H, copper, tungsten, iron, niobium), candidates to e used in a fusion reactor, have been exposed to neutrons produced in the facility (about 109 n x cm-2 x s-1) and the short and medium-lived induced radioactivity has been measured by gamma-ray spectroscopy. The experimental results have been used to validate the inventory code FISPACT, the physical database EAF, including its uncertainty predictions, and the composition of the material irradiated in particular for its minor elements and impurities. The comparison between calculated (C) and experimental results (E) is reported as C/E values and shows a satisfactory agreement for almost all radionuclides. Radionuclides for which there is not agreement between calculations and experiments are also discussed and an analysis of the causes of the lack of agreement is carried out. (author)

  13. Shielding benchmark experiments and sensitivity studies in progress at some European laboratories

    Energy Technology Data Exchange (ETDEWEB)

    Hehn, G. (Univ., Stuttgart); Mattes, M.; Matthes, W.; Nicks, R.; Rief, H.

    1975-10-01

    A 100 group standard library based on ENDF/B3 has been prepared by IKE and JRC. This library is used for the analysis of the current European and Japanese iron benchmark experiments. Further measurements are planned for checking the data sets for graphite, sodium and water. In a cooperation between the IKE and JRC groups coupled neutron-photon cross section sets will be produced. Point data are processed at IKE by the modular program system RSYST (CDC 6600) for elaborating the ENDFB data, whereas the JRC group, apart from using standard codes such as SUPERTOG 3, GAMLEG etc., has developed a series of auxiliary programs (IBM 360) for handling the DLC 2D and POPOP libraries and for producing the combined neutron-plus gamma library EL4 (119 groups). Sensitivity studies (in progress at IKE) make possible improvements in methods and optimization of calculation efforts for establishing group data. A tentative sensitivity study for a 3 dimensional MC approach is in progress at Ispra. As for nuclear data evaluation, the JRC group is calculating barium cross sections and their associated gamma spectra. 6 figures. (auth)

  14. Assessment of DeCART for Numerical Benchmark Problems Based on the Compact Nuclear Power Source Critical Experiments

    International Nuclear Information System (INIS)

    DeCART is a 3-dimensional whole-core code based on the synthesis of 2-D radial MOC (Method Of Characteristics) transport and 1-D axial nodal diffusion methods. This code has been applied to the PWR physics analysis. Recently its' geometry treatment capability has been extended to deal with the hexagonal meshes for the VHTR (Very High Temperature gas-cooled Reactor) physics analysis, which requires a verification of the applicability to the VHTR fuels. The Argonne national laboratory has developed the numerical benchmark problems based on the Compact Nuclear Power Source (CNPS) experiments conducted at the Los Alamos National Laboratory (LANL) in the late 1980s in order to support the validation and verification work for the VHTR physics codes. Development of the numerical benchmarks was required from a lack of experimental information on the design data uncertainties and the inconsistency in the design data from different sources. Two- and three-dimensional numerical benchmarks based on the CNPS experiment are specified for the verification of the VHTR physics. In this study the DeCART code was assessed by performing the CNPS benchmark calculations and comparing the results with the MCNP ones

  15. Assessment of DeCART for Numerical Benchmark Problems Based on the Compact Nuclear Power Source Critical Experiments

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Kyung Hoon; Cho, Jin Young; Kim, Kang Seog; Joo, Hyung Kook; Lee, Chung Chan [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    2006-07-01

    DeCART is a 3-dimensional whole-core code based on the synthesis of 2-D radial MOC (Method Of Characteristics) transport and 1-D axial nodal diffusion methods. This code has been applied to the PWR physics analysis. Recently its' geometry treatment capability has been extended to deal with the hexagonal meshes for the VHTR (Very High Temperature gas-cooled Reactor) physics analysis, which requires a verification of the applicability to the VHTR fuels. The Argonne national laboratory has developed the numerical benchmark problems based on the Compact Nuclear Power Source (CNPS) experiments conducted at the Los Alamos National Laboratory (LANL) in the late 1980s in order to support the validation and verification work for the VHTR physics codes. Development of the numerical benchmarks was required from a lack of experimental information on the design data uncertainties and the inconsistency in the design data from different sources. Two- and three-dimensional numerical benchmarks based on the CNPS experiment are specified for the verification of the VHTR physics. In this study the DeCART code was assessed by performing the CNPS benchmark calculations and comparing the results with the MCNP ones.

  16. IKS Deliverable - D1.1 Report: Design of the Semantic Benchmark Experiment

    OpenAIRE

    Kowatsch, Tobias; Maass, Wolfgang; Damjanovic, Violeta; Behrendt, Wernher; Gruber, Andreas; Nagel, Benjamin; Sauer, Stefan; Engels, Gregor

    2009-01-01

    Public Deliverable - The objective of this deliverable is to design a benchmark model for Content Management Systems (CMSs) in order to identify relevant requirements for the Interactive Knowledge Stack (IKS). The IKS will be a layered set of software components and specifications with the goal to improve the interaction with knowledge objects of CMSs by using Semantic Web technologies. In contrast to projects that benchmarked rather technical aspects of CMSs, we propose a model that evaluate...

  17. Benchmarking and Regulation of Electricity Transmission and Distribution Utilities: Lessons from International Experience

    OpenAIRE

    Jamasb, Tooraj; Michael G. Pollitt

    2001-01-01

    Since the early 1980's, many countries have implemented electricity sector reform, many of which have bundled generation, transmission, distribution and supply activities, and have introduced competition in generation and supply. An increasing number of countries are also adopting incentive regulation to promote efficiency improvement in the natural monopoly activities - transmission and distribution. Incentive regulation almost invariably involves benchmarking or comparison of actual vs...

  18. Benchmarking FENDL libraries through analysis of bulk shielding experiments on large SS316 assemblies for verification of ITER shielding characteristics

    International Nuclear Information System (INIS)

    FENDL-1 data base has been developed recently for use in ITER/EDA phase and other fusion-related design activities. It is now undergoing extensive testing and benchmarking using experimental data of differential and integral measured parameters obtained from fusion-oriented experiments. As part of co-operation between UCLA (U.S.) with JAERI (Japan) on executing the required neutronics R ampersand D tasks for ITER shield design, two bulk shielding experiments on large SS316 assemblies were selected for benchmarking FENDL/MG-1 multigroup data base and FENDL/MC-1 continous energy data base. The analyses with the multigroup data (performed with S8, P5, DORT calculations with shielded and unshielded data) also included library derived from ENDF/B-VI data base for comparison purposes. The MCNP Monte Carlo code was used by JAERI with the FENDL/MC-1 data. The results of this benchmarking is reported in this paper along with the observed deficiencies and discrepancies. 20 refs., 27 figs., 1 tab

  19. Overview of the 2014 Edition of the International Handbook of Evaluated Reactor Physics Benchmark Experiments (IRPhEP Handbook)

    International Nuclear Information System (INIS)

    The International Reactor Physics Experiment Evaluation Project (IRPhEP) is a widely recognized world class program. The work of the IRPhEP is documented in the International Handbook of Evaluated Reactor Physics Benchmark Experiments (IRPhEP Handbook). Integral data from the IRPhEP Handbook is used by reactor safety and design, nuclear data, criticality safety, and analytical methods development specialists, worldwide, to perform necessary validations of their calculational techniques. The IRPhEP Handbook is among the most frequently quoted reference in the nuclear industry and is expected to be a valuable resource for future decades.

  20. Overview of the 2014 Edition of the International Handbook of Evaluated Reactor Physics Benchmark Experiments (IRPhEP Handbook)

    Energy Technology Data Exchange (ETDEWEB)

    John D. Bess; J. Blair Briggs; Jim Gulliford; Ian Hill

    2014-10-01

    The International Reactor Physics Experiment Evaluation Project (IRPhEP) is a widely recognized world class program. The work of the IRPhEP is documented in the International Handbook of Evaluated Reactor Physics Benchmark Experiments (IRPhEP Handbook). Integral data from the IRPhEP Handbook is used by reactor safety and design, nuclear data, criticality safety, and analytical methods development specialists, worldwide, to perform necessary validations of their calculational techniques. The IRPhEP Handbook is among the most frequently quoted reference in the nuclear industry and is expected to be a valuable resource for future decades.

  1. Overview of the 2014 edition of the international handbook of evaluated reactor physics benchmark experiments (IRPhEP Handbook)

    International Nuclear Information System (INIS)

    The International Reactor Physics Experiment Evaluation Project (IRPhEP) is a widely recognized world class program. The work of the IRPhEP is documented in the International Handbook of Evaluated Reactor Physics Benchmark Experiments (IRPhEP Handbook). Integral data from the IRPhEP Handbook is used by reactor safety and design, nuclear data, criticality safety, and analytical methods development specialists, worldwide, to perform necessary validations of their calculational techniques. The IRPhEP Handbook is among the most frequently quoted reference in the nuclear industry and is expected to be a valuable resource for future decades. (author)

  2. Integral benchmark experiments of the Japanese Evaluated Nuclear Data Library (JENDL)-3.3 for the fusion reactor design

    International Nuclear Information System (INIS)

    JENDL-3.3 is a neutron cross section data library of 337 nuclei evaluated from the latest experimental data. JENDL-3.3 introduces double differential cross sections, which are energy- and angle-dependent ones of the scattered secondary neutrons, and are important for anisotropic neutron transport calculations for the fusion reactors design. This paper overviews benchmark experiments carried out for key fusion related nuclei such as Iron and Vanadium, and the results of analyses with JENDL-3.3, together with JENDL-3.2 and FENDL-2 for a comparison purpose. The experiments have been carried out at the Fusion Neutron Source (FNS) of JAERI. During the neutron injection into the assemblies, neutron and secondary gamma-ray spectra have been measured inside and outside the assemblies. For the test assemblies, we have used Iron, Copper, Vanadium and Tungsten as a single element material, and LiAlO2 and SiC as a compound material. From the integral benchmark experiments it was confirmed that the accuracy of JENDL-3.3 has been improved well compared with JENDL- 3.2 and FENDL-2 by the re-evaluation using latest experimental data, and JENDL-3.3 is suitable for the nuclear analysis of the fusion reactor. (author)

  3. IAEA benchmark calculations on control rod withdrawal test performed during PHENIX end-of-life experiments. JAEA's calculation results

    International Nuclear Information System (INIS)

    This paper describes details of the IAEA/CRP benchmark calculation by JAEA on the control rod withdrawal test in the Phenix End-of-Life Experiments. The power distribution deviation by the control rod insertion/withdrawal, which is the major target of the benchmark, is well simulated by calculation. In addition to the CRP activities, neutron and photon heat transport effect is evaluated in the nuclear heating calculation of the benchmark analysis. It is confirmed that the neutron and photon heat transport effect contributes to the improvement of the absolute power calculation results in the breeder blanket region. (author)

  4. Neutronic Analysis of the 3 MW TRIGA MARK II Research Reactor, Part II: Benchmark Analysis of TRIGA Experiments

    International Nuclear Information System (INIS)

    The three-dimensional continuous-energy Monte Carlo code MCNP4C was used to develop a versatile and accurate full-core model of the TRIGA MARK II research reactor at AERE, Savar. Thr consistency and accuracy of both the Monte Carlo simulation and neutron transport physics was established by benchmarking the TRIGA experiments. Analysis of neutron flux and reactivity experiments comprising control rod worths, critical rod height, excess reactivity and shutdown margin were used in the validation process. Calculations of fast neutron flux, and fuel and graphite element worths distribution are also presented. Good agreement between the experiments and MCNP calculations indicate that the simulation of TRIGA reactor is treated adequately. (author)

  5. Integrating CFD, CAA, and Experiments Towards Benchmark Datasets for Airframe Noise Problems

    Science.gov (United States)

    Choudhari, Meelan M.; Yamamoto, Kazuomi

    2012-01-01

    Airframe noise corresponds to the acoustic radiation due to turbulent flow in the vicinity of airframe components such as high-lift devices and landing gears. The combination of geometric complexity, high Reynolds number turbulence, multiple regions of separation, and a strong coupling with adjacent physical components makes the problem of airframe noise highly challenging. Since 2010, the American Institute of Aeronautics and Astronautics has organized an ongoing series of workshops devoted to Benchmark Problems for Airframe Noise Computations (BANC). The BANC workshops are aimed at enabling a systematic progress in the understanding and high-fidelity predictions of airframe noise via collaborative investigations that integrate state of the art computational fluid dynamics, computational aeroacoustics, and in depth, holistic, and multifacility measurements targeting a selected set of canonical yet realistic configurations. This paper provides a brief summary of the BANC effort, including its technical objectives, strategy, and selective outcomes thus far.

  6. Coupled fast-thermal core 'HERBE', as the benchmark experiment at the RB reactor

    International Nuclear Information System (INIS)

    Validation of the well-known Monte Carlo code MCNPTM against measured criticality data for the coupled fast-thermal HERBE. System at the RB research reactor is shown in this paper. Experimental data are obtained for regular HERBE core and for the cases of controlled flooding of the neutron converter zone by heavy water. Earlier calculations of these criticality parameters, done by combination of transport and diffusion codes using 2D geometry model are also compared to new calculations carried out by the MCNP code in 3D geometry, applying new detailed 3D model of the HEU fuel slug, developed recently. Satisfactory agreements in comparison of the HERBE criticality calculation results with experimental data, in spite complex heterogeneous composition of the HERBE core, are obtained and confirmed that HERBE core could be used as a criticality benchmark for coupled fast-thermal core. (author)

  7. Analysis of BFS-62-3A critical experiment benchmark model - IGCAR results

    International Nuclear Information System (INIS)

    The BFS 62-3A assembly is a full scale model of BN-600 hybrid core. The MOX zone is represented as a ring between medium enriched (MEZ) and high enriched zones (HEZ). The hybrid core with steel reflector is represented in a 120 deg sector of BFS. For a homogenised 3-D core of BFS, equivalent experimental data of keff and SVRE values were derived by including the following corrections to the actually obtained experimental results: (a) heterogeneity effect and (b) 3-D model simplification effect. The nuclear data used was XSET-98. It is a 26 group set with ABBN type self-shielding factor table. The benchmark models were analysed by diffusion theory. 3-D calculations were done by TREDFR code in 26 groups with 6 triangular meshes per fuel assembly. The number of triangles was 24414. Axial mesh size corrections were estimated for some cases. The convergence criteria for were 0.000001 for keff and 0.0001 for point wise fission source. The multiplication factor of the reference core of the benchmark is compared with measured. The multiplication factor is predicted with in the uncertainty margin. The SVRE values were computed as Δk/k1k2 and compared to measured values. It is found that the predictions are with in the uncertainty margin except in the MOX region. Reason for this needs to be investigated. As a first step, axial mesh size effect was estimated for MOX SVRE (sodium void reactivity effect) case with use finer meshes in the reference core as well the MOX voided core. By increasing the axial meshe from 35 to 54 both the keff reduced by the same amount leaving the MOX SVRE worth unchanged

  8. Benchmarking HRD.

    Science.gov (United States)

    Ford, Donald J.

    1993-01-01

    Discusses benchmarking, the continuous process of measuring one's products, services, and practices against those recognized as leaders in that field to identify areas for improvement. Examines ways in which benchmarking can benefit human resources functions. (JOW)

  9. Soil Shielding Benchmark Experiment and its Simulation with Mars using Secondary Particles Produced by 12 GeV Protons

    Energy Technology Data Exchange (ETDEWEB)

    Suzuki, T.; Numajiri, M.; Ban, S.; Kanda, Y.; Oki, Y.; Namito, Y.; Miura, T.; Hirayama, H.; Shibata, T.; Kondo, K.; Takasaki, M.; Tanaka, K.H.; Y; Yamanoi; Minakawa, M.; Noumi, H.; Ieiri, M.; Kato, Y.; Ishii, H.; Suzuki, Y.; Nishikawa, K.; Mokhov, N

    1998-07-01

    A soil-shielding benchmark experiment was conducted using secondary particles produced by 12 GeV protons, which were injected into an iron rod surrounded by soil. Induced activities of {sup 22}Na in aluminium (Al) and soil samples were measured and the experiment was simulated by the MARS Monte Carlo code. The induced activities in Al samples were calculated using spallation cross sections and fluence, where the fluence was calculated by the MARS code. The relation between flux density and induced activities in soil was investigated using calculated flux densities: the distribution of the ratio of induced activities in soil samples to the flux densities showed the radial and axial independence. Both saturated activities and distribution coincide with experimental data within the estimated errors. These results indicate the successful simulation by the MARS code. (author)

  10. Benchmark experiments and numerical modelling of the columnar-equiaxed dendritic growth in the transparent alloy Neopentylglycol-(d)Camphor

    Science.gov (United States)

    Sturz, L.; Wu, M.; Zimmermann, G.; Ludwig, A.; Ahmadein, M.

    2015-06-01

    Solidification benchmark experiments on columnar and equiaxed dendritic growth, as well as the columnar-equiaxed transition have been carried out under diffusion-dominated conditions for heat and mass transfer in a low-gravity environment. The system under investigation is the transparent organic alloy system Neopentylglycol-37.5wt.-%(d)Camphor, processed aboard a TEXUS sounding rocket flight. Solidifications was observed by standard optical methods in addition to measurements of the thermal fields within the sheet like experimental cells of 1 mm thickness. The dendrite tip kinetic, primary dendrite arm spacing, temporal and spatial temperature evolution, columnar tip velocity and the critical parameters at the CET have been analysed. Here we focus on a detailed comparison of the experiment “TRACE” with a 5-phase volume averaging model to validate the numerical model and to give insight into the corresponding physical mechanisms and parameters leading to CET. The results are discussed in terms of sensitivity versus numerical parameters.

  11. Multidimensional benchmarking

    OpenAIRE

    Campbell, Akiko

    2016-01-01

    Benchmarking is a process of comparison between performance characteristics of separate, often competing organizations intended to enable each participant to improve its own performance in the marketplace (Kay, 2007). Benchmarking sets organizations’ performance standards based on what “others” are achieving. Most widely adopted approaches are quantitative and reveal numerical performance gaps where organizations lag behind benchmarks; however, quantitative benchmarking on its own rarely yi...

  12. ENDF/B-VII.1 Neutron Cross Section Data Testing with Critical Assembly Benchmarks and Reactor Experiments

    International Nuclear Information System (INIS)

    The ENDF/B-VII.1 library is the latest revision to the United States' Evaluated Nuclear Data File (ENDF). The ENDF library is currently in its seventh generation, with ENDF/B-VII.0 being released in 2006. This revision expands upon that library, including the addition of new evaluated files (was 393 neutron files previously, now 418 including replacement of elemental vanadium and zinc evaluations with isotopic evaluations) and extension or updating of many existing neutron data files. Complete details are provided in the companion paper [1]. This paper focuses on how accurately application libraries may be expected to perform in criticality calculations with these data. Continuous energy cross section libraries, suitable for use with the MCNP Monte Carlo transport code, have been generated and applied to a suite of nearly one thousand critical benchmark assemblies defined in the International Criticality Safety Benchmark Evaluation Project's International Handbook of Evaluated Criticality Safety Benchmark Experiments. This suite covers uranium and plutonium fuel systems in a variety of forms such as metallic, oxide or solution, and under a variety of spectral conditions, including unmoderated (i.e., bare), metal reflected and water or other light element reflected. Assembly eigenvalues that were accurately predicted with ENDF/B-VII.0 cross sections such as unmoderated and uranium reflected 235U and 239Pu assemblies, HEU solution systems and LEU oxide lattice systems that mimic commercial PWR configurations continue to be accurately calculated with ENDF/B-VII.1 cross sections, and deficiencies in predicted eigenvalues for assemblies containing selected materials, including titanium, manganese, cadmium and tungsten are greatly reduced. Improvements are also confirmed for selected actinide reaction rates such as 236U capture. Other deficiencies, such as the overprediction of Pu solution system critical eigenvalues and a decreasing trend in calculated eigenvalue for

  13. Benchmark analysis of reactivity experiment in the TRIGA Mark 2 reactor using a continuous energy Monte Carlo code MCNP

    International Nuclear Information System (INIS)

    A good model on experimental data (criticality, control rod worth, and fuel element worth distributions) is encouraged to provide from the Musashi-TRIGA Mark 2 reactor. In the previous paper, as the keff values for different fuel loading patterns had been provided ranging from the minimum core to the full one, the data would be candidate for an ICSBEP evaluation. Evaluation of the control rod worth and fuel element worth distributions presented in this paper could be an excellent benchmark data applicable for validation of calculation technique used in the field of modern research reactor. As a result of simulation on the TRIGA-2 benchmark experiment, which was performed by three-dimensional continuous-energy Monte Carlo code (MCNP4A), it was found that the MCNP calculated values of control rod worth were consisted to the experimental data for both rod-drop and period methods. And for the fuel and the graphite element worth distributions, the MCNP calculated values agreed well with the measured ones though consideration of real control rod positions was needed for calculating fuel element reactivity positioned in inner ring. (G.K.)

  14. Attempt of the joint analysis of the entire set of the HEU-SOL type experiments from the ''International Handbook of Evaluated Criticality Safety Benchmark Experiments''

    International Nuclear Information System (INIS)

    This paper provides a description of work on the joint analysis of the entire set of the experiments with the solutions of highly enriched uranium in light water from the ''International Handbook of Evaluated Criticality Safety Benchmark Experiments''. The purpose of the work was to analyze the experiments for interconsistency, discover and evaluate possible correlations between them, discover and eliminate systematic errors and disagreements, and get a consistent set of evaluated experiments for future use in validation of calculations of critical mass of solutions of highly enriched uranium of different concentrations in light water and evaluation of uncertainty of these calculations. The paper describes in details how the correlations between the experimental uncertainties were determined as well as how systematic errors were discovered. (author)

  15. OECD EGBUC Benchmark VIII. Comparison of calculation codes and methods for the analysis of small-sample reactivity experiments

    International Nuclear Information System (INIS)

    Small-sample reactivity experiments are relevant to provide accurate information on the integral cross sections of materials. One of the specificities of these experiments is that the measured reactivity worth generally ranges between 1 and 10 pcm, which precludes the use of Monte Carlo for the analysis. As a consequence, several papers have been devoted to deterministic calculation routes, implying spatial and/or energetic discretization which could involve calculation bias. Within the Expert Group on Burn-Up Credit of the OECD/NEA, a benchmark was proposed to compare different calculation codes and methods for the analysis of these experiments. In four Sub-Phases with geometries ranging from a single cell to a full 3D core model, participants were asked to evaluate the reactivity worth due to the addition of small quantities of separated fission products and actinides into a UO2 fuel. Fourteen institutes using six different codes have participated in the Benchmark. For reactivity worth of more than a few tens of pcm, the Monte-Carlo approach based on the eigen-value difference method appears clearly as the reference method. However, in the case of reactivity worth as low as 1 pcm, it is concluded that the deterministic approach based on the exact perturbation formalism is more accurate and should be preferred. Promising results have also been reported using the newly available exact perturbation capability, developed in the Monte Carlo code TRIPOLI4, based on the calculation of a continuous energy adjoint flux in the reference situation, convoluted to the forward flux of the perturbed situation. (author)

  16. Criticality experiments: analysis, evaluation, and programs. 1. Preview of the 2001 Edition of the 'International Handbook of Criticality Safety Benchmark Experiments'

    International Nuclear Information System (INIS)

    The International Criticality Safety Benchmark Evaluation Project (ICSBEP) was initiated in 1992 and has become a major internationally recognized program. The purpose of the ICSBEP is to identify, evaluate, verify, and formally document a comprehensive and internationally peer-reviewed set of criticality safety benchmark data. The work of the ICSBEP is published as an Organization for Economic Cooperation and Development (OECD) handbook entitled 'International Handbook of Criticality Safety Benchmark Experiments'. More than 150 scientists from around the world have combined their efforts to produce this handbook, which currently spans more than 19 000 pages and contains benchmark specifications for more than 2352 critical configurations. The handbook is intended for use by criticality safety analysts to perform necessary validations of their calculational techniques. The 2001 edition of the 'International Handbook of Criticality Safety Benchmark Experiments' is scheduled for publication in September of 2001 and could contain as many as 30 new evaluations of experimental data. Included in the list of 'in-progress' evaluations are: 1. the ZPPR-21 experiments entitled 'Criticality Studies for Integral Fast Reactors'; 2. RAPSODIE mixed plutonium/uranium fuel rods in water (IPSN, France); 3. mixed uranium/plutonium (29.87%) nitrate solutions poisoned with gadolinium (IPSN, France); 4. PuO2-UO2-polystyrene cubes with poison plates (Westinghouse SMS, United States); 5. highly enriched uranyl nitrate solution in steel containers with 'pipe' intersections (Westinghouse SMS, United States); 6. HEU metal in oil (Westinghouse SMS, United States); 7. an evaluation of experiments with liquid mixtures of HEU hexafluoride and hydrofluoric acid (IPSN, France); 8. critical experiments of stainless steel clad SPERT fuel in water (INEEL, United States); 9. HEU foils reflected by SiO2 and polyethylene (LANL, United States); 10. Un-reflected highly enriched uranyl nitrate subcritical

  17. Benchmark of the FLUKA model of crystal channeling against the UA9-H8 experiment

    International Nuclear Information System (INIS)

    Channeling in bent crystals is increasingly considered as an option for the collimation of high-energy particle beams. The installation of crystals in the LHC has taken place during this past year and aims at demonstrating the feasibility of crystal collimation and a possible cleaning efficiency improvement. The performance of CERN collimation insertions is evaluated with the Monte Carlo code FLUKA, which is capable to simulate energy deposition in collimators as well as beam loss monitor signals. A new model of crystal channeling was developed specifically so that similar simulations can be conducted in the case of crystal-assisted collimation. In this paper, most recent results of this model are brought forward in the framework of a joint activity inside the UA9 collaboration to benchmark the different simulation tools available. The performance of crystal STF 45, produced at INFN Ferrara, was measured at the H8 beamline at CERN in 2010 and serves as the basis to the comparison. Distributions of deflected particles are shown to be in very good agreement with experimental data. Calculated dechanneling lengths and crystal performance in the transition region between amorphous regime and volume reflection are also close to the measured ones

  18. Analysis of OKTAVIAN Shielding Benchmark Experiments by ENDF/B-VII, JEFF-3.1, and JENDL-3.3

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Do-Heon; Gil, Choong-Sup; Lee, Young-Ouk [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)

    2007-07-01

    International collaborations for the ITER Project and other fusion-related development projects have been conducted to create a reference fusion nuclear data library such as FENDL, which was a collection of the best cross section data from national nuclear data libraries. Recent release of newly evaluated nuclear data libraries requires an extensive and intensive benchmarking of the updated transport libraries to become a candidate for the future collection. In this study, the pulsed sphere experiments for leakage neutron and gamma-ray spectra at the D-T neutron source facility of Osaka University, OKTAVIAN were employed to test the ENDF/B-VII beta 1, JEFF-3.1, and JENDL-3.3 libraries. The continuous energy Monte Carlo transport code MCNPX-2.5 was used along with the ACE format libraries processed by a modified version of the NJOY99.90 code.

  19. ENDF/B VI iron validation onpca-replica (H2O/FE) shielding benchmark experiment

    Energy Technology Data Exchange (ETDEWEB)

    Pescarini, M. [ENEA, Bologna (Italy). Centro Ricerche Energia `E. Clementel` - Area Energia e Innovazione

    1994-05-01

    The PCA-REPLICA (H2O/Fe) neutron shielding benchmark experiment is analysed using the SN 2-D DOT 3.5 code and the 3-D-equivalent flux synthesis method. This engineering benchmark reproduces the ex-core radial geometry of a PWR, including a mild steel reactor pressure vessel (RPV) simulator, and is dsigned to test the accuracy of the calculation of the in-vessel neutron exposure parameters (fast fluence and iron displacement rates). This accuracy is strongly dependent on the quality of the iron neutron cross section used to describe the nuclear reactions within the RPV simulator. In particular, in this report, the cross sections based on the ENDF/B VI iron data files are tested, through a comparison of the calculated integral and spectral results with the corresponding experimental data. In addition, the present results are compared, on the same benchmark experiment, with those of a preceding ENEA (Italian Agency for Energy, New Technologies and Environment)-Bologna validation of the JEF-2.1 iron cross sections. The integral result comparison indicates that, for all the thresold detectors considered (Rh-103 (n,n) Rh-103m, In-115 (n,n) In-115 (n,n) In-115m and S-32 (n.p) P-32), the ENDF/B VI iron data produce better results than the JEF-2.1 iron data. In particular, in the ENDF/B VI calcultaions, an improvement of the in-vessel C/E (Calculated/Experimental) activity ratios for the lower energy threshold detectors, Rh-103 and In-115, is observed. This improvement becomes more evident with increasing neutron penetration depth in the vessel. This is probably attributable to the fact that the inelastic scattering cross section values of the ENDF/B VI Fe-56 data file, approximately in the 0.86 - 1.5 MeV energy range, are lower then the corresponding values of the JEF-2.1 data file.

  20. Results and experiences from the analysis of the OECD PWR MSLB and BWR TT benchmarks by the coupled code system ATHLET-QUABOX/CUBBOX

    International Nuclear Information System (INIS)

    The GRS results of the two QECD/NRC benchmarks for coupled codes - PWR Main Steam Line Break (MSLB) and BWR Turbine Trip (TT) are presented in the paper. Both benchmarks have been calculated for all exercises by the coupled code system ATHLET - QUABOX/CUBBOX developed by GRS. The OECD PWR Main Steam Line Break Benchmark has been defined to validate the coupled code systems with 3D neutronics by comparing solutions of different codes (code-to-code comparison). The BWR TT Benchmark is based on measurements. Both benchmarks proved to be a valuable source for coupled code validation. The sensitivity of results on modelling features is also discussed in the paper. In addition, the effect of different mapping schemes between fuel assemblies of the core loading and the thermal-fluid dynamics on the accuracy of 3D neutronics solutions is shown. The results for the MSLB transient are also evaluated to compare 3D neutronics and point-kinetics solutions in view of integral and local parameters. Thus, the experiences with the coupled code system ATHLET - QUABOX/CUBBOX during the MSLB and TT benchmarks activities are summarised. (author)

  1. Financial benchmarking

    OpenAIRE

    Boldyreva, Anna

    2014-01-01

    This bachelor's thesis is focused on financial benchmarking of TULIPA PRAHA s.r.o. The aim of this work is to evaluate financial situation of the company, identify its strengths and weaknesses and to find out how efficient is the performance of this company in comparison with top companies within the same field by using INFA benchmarking diagnostic system of financial indicators. The theoretical part includes the characteristic of financial analysis, which financial benchmarking is based on a...

  2. Pore-scale and Continuum Simulations of Solute Transport Micromodel Benchmark Experiments

    Energy Technology Data Exchange (ETDEWEB)

    Oostrom, Martinus; Mehmani, Yashar; Romero Gomez, Pedro DJ; Tang, Y.; Liu, H.; Yoon, Hongkyu; Kang, Qinjun; Joekar Niasar, Vahid; Balhoff, Matthew; Dewers, T.; Tartakovsky, Guzel D.; Leist, Emily AE; Hess, Nancy J.; Perkins, William A.; Rakowski, Cynthia L.; Richmond, Marshall C.; Serkowski, John A.; Werth, Charles J.; Valocchi, Albert J.; Wietsma, Thomas W.; Zhang, Changyong

    2016-08-01

    Four sets of micromodel nonreactive solute transport experiments were conducted with flow velocity, grain diameter, pore-aspect ratio, and flow focusing heterogeneity as the variables. The data sets were offered to pore-scale modeling groups to test their simulators. Each set consisted of two learning experiments, for which all results was made available, and a challenge experiment, for which only the experimental description and base input parameters were provided. The experimental results showed a nonlinear dependence of the dispersion coefficient on the Peclet number, a negligible effect of the pore-aspect ratio on transverse mixing, and considerably enhanced mixing due to flow focusing. Five pore-scale models and one continuum-scale model were used to simulate the experiments. Of the pore-scale models, two used a pore-network (PN) method, two others are based on a lattice-Boltzmann (LB) approach, and one employed a computational fluid dynamics (CFD) technique. The learning experiments were used by the PN models to modify the standard perfect mixing approach in pore bodies into approaches to simulate the observed incomplete mixing. The LB and CFD models used these experiments to appropriately discretize the grid representations. The continuum model use published non-linear relations between transverse dispersion coefficients and Peclet numbers to compute the required dispersivity input values. Comparisons between experimental and numerical results for the four challenge experiments show that all pore-scale models were all able to satisfactorily simulate the experiments. The continuum model underestimated the required dispersivity values and, resulting in less dispersion. The PN models were able to complete the simulations in a few minutes, whereas the direct models needed up to several days on supercomputers to resolve the more complex problems.

  3. Benchmark physics experiment of metallic-fueled LMFBR at FCA. 2

    International Nuclear Information System (INIS)

    An availability of data and method for a design of metallic-fueled LMFBR is examined by using the experiment results of FCA assembly XVI-1. Experiment included criticality and reactivity coefficients such as Doppler, sodium void, fuel shifting and fuel expansion. Reaction rate ratios, sample worth and control rod worth were also measured. Analysis was made by using three-dimensional diffusion calculations and JENDL-2 cross sections. Predictions of assembly XVI-1 reactor physics parameters agree reasonably well with the measured values, but for some reactivity coefficients such as Doppler, large zone sodium void and fuel shifting further improvement of calculation method was need. (author)

  4. Shielding benchmark problems, (2)

    International Nuclear Information System (INIS)

    Shielding benchmark problems prepared by Working Group of Assessment of Shielding Experiments in the Research Committee on Shielding Design in the Atomic Energy Society of Japan were compiled by Shielding Laboratory in Japan Atomic Energy Research Institute. Fourteen shielding benchmark problems are presented newly in addition to twenty-one problems proposed already, for evaluating the calculational algorithm and accuracy of computer codes based on discrete ordinates method and Monte Carlo method and for evaluating the nuclear data used in codes. The present benchmark problems are principally for investigating the backscattering and the streaming of neutrons and gamma rays in two- and three-dimensional configurations. (author)

  5. Benchmark analysis of criticality experiments in the TRIGA mark II using a continuous energy Monte Carlo code MCNP

    International Nuclear Information System (INIS)

    The criticality analysis of the TRIGA-II benchmark experiment at the Musashi Institute of Technology Research Reactor (MuITR, 100kW) was performed by the three-dimensional continuous-energy Monte Carlo code (MCNP4A). To minimize errors due to an inexact geometry model, all fresh fuels and control rods as well as vicinity of the core were precisely modeled. Effective multiplication factors (keff) in the initial core critical experiment and in the excess reactivity adjustment for the several fuel-loading patterns as well as the fuel element reactivity worth distributions were used in the validation process of the physical model and neutron cross section data from the ENDF/B-V evaluation. The calculated keff overestimated the experimental data by about 1.0%Δk/k for both the initial core and the several fuel-loading arrangements (fuels or graphite elements were added only to the outer-ring), but the discrepancy increased to 1.8%Δk/k for the some fuel-loading patterns (graphite elements were inserted into the inner-ring). The comparison result of the fuel element worth distribution showed above tendency. All in all, the agreement between the MCNP predictions and the experimentally determined values is good, which indicates that the Monte Carlo model is enough to simulate criticality of the TRIGA-II reactor. (author)

  6. Verification and validation benchmarks.

    Energy Technology Data Exchange (ETDEWEB)

    Oberkampf, William Louis; Trucano, Timothy Guy

    2007-02-01

    Verification and validation (V&V) are the primary means to assess the accuracy and reliability of computational simulations. V&V methods and procedures have fundamentally improved the credibility of simulations in several high-consequence fields, such as nuclear reactor safety, underground nuclear waste storage, and nuclear weapon safety. Although the terminology is not uniform across engineering disciplines, code verification deals with assessing the reliability of the software coding, and solution verification deals with assessing the numerical accuracy of the solution to a computational model. Validation addresses the physics modeling accuracy of a computational simulation by comparing the computational results with experimental data. Code verification benchmarks and validation benchmarks have been constructed for a number of years in every field of computational simulation. However, no comprehensive guidelines have been proposed for the construction and use of V&V benchmarks. For example, the field of nuclear reactor safety has not focused on code verification benchmarks, but it has placed great emphasis on developing validation benchmarks. Many of these validation benchmarks are closely related to the operations of actual reactors at near-safety-critical conditions, as opposed to being more fundamental-physics benchmarks. This paper presents recommendations for the effective design and use of code verification benchmarks based on manufactured solutions, classical analytical solutions, and highly accurate numerical solutions. In addition, this paper presents recommendations for the design and use of validation benchmarks, highlighting the careful design of building-block experiments, the estimation of experimental measurement uncertainty for both inputs and outputs to the code, validation metrics, and the role of model calibration in validation. It is argued that the understanding of predictive capability of a computational model is built on the level of

  7. Benchmark selection

    DEFF Research Database (Denmark)

    Hougaard, Jens Leth; Tvede, Mich

    2002-01-01

    Within a production theoretic framework, this paper considers an axiomatic approach to benchmark selection. It is shown that two simple and weak axioms; efficiency and comprehensive monotonicity characterize a natural family of benchmarks which typically becomes unique. Further axioms are added in...... order to obtain a unique selection...

  8. Interactive benchmarking

    DEFF Research Database (Denmark)

    Lawson, Lartey; Nielsen, Kurt

    2005-01-01

    We discuss individual learning by interactive benchmarking using stochastic frontier models. The interactions allow the user to tailor the performance evaluation to preferences and explore alternative improvement strategies by selecting and searching the different frontiers using directional...... suggested benchmarking tool. The study investigates how different characteristics on dairy farms influences the technical efficiency....

  9. Precious benchmarking

    International Nuclear Information System (INIS)

    Recently, there has been a new word added to our vocabulary - benchmarking. Because of benchmarking, our colleagues travel to power plants all around the world and guests from the European power plants visit us. We asked Marek Niznansky from the Nuclear Safety Department in Jaslovske Bohunice NPP to explain us this term. (author)

  10. 500-MeV electron beam bench-mark experiments and calculations

    International Nuclear Information System (INIS)

    Experiments measuring the energy deposited by electron beams were performed to provide bench marks against which to evaluate our HANDYL76 electron beam computer code. The experiments, done at Stanford's Mk III accelerator, measured dose vs depth and dose vs radius profiles induced in layered aluminum targets by 500-MeV electrons. The dose was measured by passive thermoluminescence and photographic film placed between aluminum plates. The calculations predict a dose vs radius profile that forward-peaks on axis after the beam passes through a 200-cm air gap; the experimental measurements do not show this peak. This discrepancy indicates there may be a problem in using HANDYL76 to calculate deep penetration of a target with a large gap

  11. Validation of the Serpent 2 code on TRIGA Mark II benchmark experiments.

    Science.gov (United States)

    Ćalić, Dušan; Žerovnik, Gašper; Trkov, Andrej; Snoj, Luka

    2016-01-01

    The main aim of this paper is the development and validation of a 3D computational model of TRIGA research reactor using Serpent 2 code. The calculated parameters were compared to the experimental results and to calculations performed with the MCNP code. The results show that the calculated normalized reaction rates and flux distribution within the core are in good agreement with MCNP and experiment, while in the reflector the flux distribution differ up to 3% from the measurements. PMID:26516989

  12. Analysis of the previous and preparation of new experiments on fast multiplying assemblies for obtaining benchmark data on criticality

    International Nuclear Information System (INIS)

    The JIPNR-Sosny of the NAS of Belarus created and explored a number of uranium-containing critical assemblies of the BTS series in designing fast BRIG-300 reactor with N2O4 ↔ 2NO2 ↔ 2NO + O2 coolant and the PVER fast-resonance neutron spectrum reactor with a steam-water coolant. Research in the physics of these reactors was performed on fast-thermal critical assemblies at the critical facility Roza. Structurally, these critical assemblies consisted of fast and thermal reactor cores and the buffer zones located between them, intended for leakage spectrum neutron conversion from a thermal zone to a spectrum of neutrons of the modelled fast reactor. Fast zones are a non-uniform hexagonal lattice of cylindrical fuel rods with fuel composition based on metal U (90% U-235), UO2 (36% U-235), depleted U (0.4% U-235), rods with SiO2; a buffer zone is a non-uniform hexagonal lattice of cylindrical fuel rods based on UO2 (36% U-235), natural U and depleted U (0.4% U-235), rods with B4C and made from stainless steel; a thermal zone is a uniform rectangular uranium-polyethylene lattice of cylindrical fuel rods based on the fuel composition UO2+Mg (10% U-235). For obtaining benchmark data on the criticality, computational models have been developed and the analysis of experiments has been carried out to estimate the experimental results as criticality benchmark data. The JIPNR-Sosny of the NAS of Belarus also prepared experiments on the criticality of multiplying systems simulating some physical features of the core of fast low power small-size gas-cooled reactors with UZrCN nuclear fuel. For these purposes, the critical assemblies P-20 were developed at the critical facility “Giacint”. These assemblies represent a uniform hexagonal lattice of fuel cassette: the central area is based on cylindrical fuel rods with UZrCN (19.75% U-235), the peripheral area is based on the cylindrical fuel rods with metallic U (90% U-235), UO2 (36% U-235) and natural U; and the reflector on

  13. Electron-impact ionization of helium: A comprehensive experiment benchmarks theory

    International Nuclear Information System (INIS)

    Single ionization of helium by 70.6-eV electron impact is studied in a comprehensive experiment covering a major part of the entire collision kinematics and the full 4π solid angle for the emitted electron. The absolutely normalized triple-differential experimental cross sections are compared with results from the convergent close-coupling (CCC) and the time-dependent close-coupling (TDCC) theories. Whereas excellent agreement with the TDCC prediction is only found for equal energy sharing, the CCC calculations are in excellent agreement with essentially all experimentally observed dynamical features, including the absolute magnitude of the cross sections.

  14. Benchmark of SIMULATE5 thermal hydraulics against the Frigg and NUPEC full bundle test experiments

    International Nuclear Information System (INIS)

    SIMULATE5 is Studsvik Scandpower's next generation nodal code. The core portion of the thermal hydraulic models of PWR and BWRs are treated as essentially identical, with each assembly having an active channel and a number of parallel water channels. In addition, the BWR assembly may be divided into four radial sub-assemblies. For natural circulation reactors, the BWR thermal hydraulic model is capable of modeling an entire vessel loop: core, chimney, upper plenum, standpipes, steam separators, downcomer, recirculation pumps, and lower plenum. This paper presents results of the validation of the BWR thermal hydraulic model against: (1) pressure drop data measured in the Frigg and NUPEC test facilities; (2) void fraction distribution measured in the Frigg and NUPEC loops; (3) quarter-assembly void fraction measured in the NUPEC experiments and (4) natural and forced circulation flow measurements in the Frigg loop. (author)

  15. WLUP benchmarks

    International Nuclear Information System (INIS)

    The IAEA-WIMS Library Update Project (WLUP) is on the end stage. The final library will be released on 2002. It is a result of research and development made by more than ten investigators during 10 years. The organization of benchmarks for testing and choosing the best set of data has been coordinated by the author of this paper. It is presented the organization, name conventions, contents and documentation of WLUP benchmarks, and an updated list of the main parameters for all cases. First, the benchmarks objectives and types are given. Then, comparisons of results from different WIMSD libraries are included. Finally it is described the program QVALUE for analysis and plot of results. Some examples are given. The set of benchmarks implemented on this work is a fundamental tool for testing new multigroup libraries. (author)

  16. Selecting benchmarks for reactor calculations

    OpenAIRE

    Alhassan, Erwin; Sjöstrand, Henrik; Duan, Junfeng; Helgesson, Petter; Pomp, Stephan; Österlund, Michael; Rochman, Dimitri; Koning, Arjan J.

    2014-01-01

    Criticality, reactor physics, fusion and shielding benchmarks are expected to play important roles in GENIV design, safety analysis and in the validation of analytical tools used to design these reactors. For existing reactor technology, benchmarks are used to validate computer codes and test nuclear data libraries. However the selection of these benchmarks are usually done by visual inspection which is dependent on the expertise and the experience of the user and there by resulting in a user...

  17. Shielding Benchmark Computational Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Hunter, H.T.; Slater, C.O.; Holland, L.B.; Tracz, G.; Marshall, W.J.; Parsons, J.L.

    2000-09-17

    Over the past several decades, nuclear science has relied on experimental research to verify and validate information about shielding nuclear radiation for a variety of applications. These benchmarks are compared with results from computer code models and are useful for the development of more accurate cross-section libraries, computer code development of radiation transport modeling, and building accurate tests for miniature shielding mockups of new nuclear facilities. When documenting measurements, one must describe many parts of the experimental results to allow a complete computational analysis. Both old and new benchmark experiments, by any definition, must provide a sound basis for modeling more complex geometries required for quality assurance and cost savings in nuclear project development. Benchmarks may involve one or many materials and thicknesses, types of sources, and measurement techniques. In this paper the benchmark experiments of varying complexity are chosen to study the transport properties of some popular materials and thicknesses. These were analyzed using three-dimensional (3-D) models and continuous energy libraries of MCNP4B2, a Monte Carlo code developed at Los Alamos National Laboratory, New Mexico. A shielding benchmark library provided the experimental data and allowed a wide range of choices for source, geometry, and measurement data. The experimental data had often been used in previous analyses by reputable groups such as the Cross Section Evaluation Working Group (CSEWG) and the Organization for Economic Cooperation and Development/Nuclear Energy Agency Nuclear Science Committee (OECD/NEANSC).

  18. Co-ordinated research programme on benchmark study for the seismic analysis and testing of WWER-type nuclear power plants. V. 5C. Experience data. Working material

    International Nuclear Information System (INIS)

    In August 1991, following the SMiRT-11 Conference in Tokyo, a Technical Committee Meeting was held on the 'Seismic safety issues relating to existing NPPs'. The Proceedings of this TCM was subsequently compiled in an IAEA Working Material. One of the main recommendations of this TCM, called for the harmonization of criteria and methods used in Member States in seismic reassessment and upgrading of existing NPPs. Twenty four institutions from thirteen countries participated in the CRP named 'Benchmark study for the seismic analysis and testing of WWER type NPPs'. Two types of WWER reactors (WWER-1000 and WWER-440/213) selected for benchmarking. Kozloduy NPP Units 5/6 and Paks NPP represented these respectively as prototypes. Consistent with the recommendations of the TCM and the working paper prepared by the subsequent Consultants' Meeting, the focal activity of the CRP was the benchmarking exercises. A similar methodology was followed both for Paks NPP and Kozloduy NPP Unit 5. Firstly, the NPP (mainly the reactor building) was tested using a blast loading generated by a series of explosions from buried TNT charges. Records from this test were obtained at several free field locations (both downhole and surface), foundation mat, various elevations of structures as well as some tanks and the stack. Then the benchmark participants were provided with structural drawings, soil data and the free field record of the blast experiment. Their task was to make a blind prediction of the response at preselected locations. The analytical results from these participants were then compared with the results from the test. Although the benchmarking exercises constituted the focus of the CRP, there were many other interesting problems related to the seismic safety of WWER type NPPs which were addressed by the participants. These involved generic studies, i.e. codes and standards used in original WWER designs and their comparison with current international practice; seismic analysis

  19. Inter-code comparison exercise for criticality excursion analysis. Benchmarks phase 1: pulse mode experiments with uranyl nitrate solution using the Tracy and Silene experimental facilities

    International Nuclear Information System (INIS)

    The NEA Working Party on Nuclear Criticality Safety established an Expert Group on Criticality Excursion Analysis in 2001 to explore the performance of various transient codes to evaluate criticality accidents in a fissile solution. Inter-code comparison exercises among four transient codes (AGNES, CRITEX, INCTAC and TRACE) have been carried out with typical transient experiments using uranyl nitrate fuel solution. Two sets of benchmarks were carried out based on experimental programmes performed in the Tracy reactor in Japan, and the Silene reactor in France. Tracy and Silene have the same geometrical features: an annular cylinder with a central void tube for a transient rod and similar operational modes for reactivity insertion. The experiments selected are representative benchmarks for low- and high-enriched uranyl nitrate solution, about 10 wt% for Tracy and 93 wt% for the Silene core. This report provides an analysis of the benchmark results obtained with four different codes. It will be of particular interest to criticality safety practitioners developing transient codes, notably since little experimental data is available and the existing transient codes are presently unavailable to the public. (authors)

  20. Benchmark exercise

    International Nuclear Information System (INIS)

    The motivation to conduct this benchmark exercise, a summary of the results, and a discussion of and conclusions from the intercomparison are given in Section 5.2. This section contains further details of the results of the calculations and intercomparisons, illustrated by tables and figures, but avoiding repetition of Section 5.2 as far as possible. (author)

  1. Orientation of Oblique Airborne Image Sets - Experiences from the Isprs/eurosdr Benchmark on Multi-Platform Photogrammetry

    Science.gov (United States)

    Gerke, M.; Nex, F.; Remondino, F.; Jacobsen, K.; Kremer, J.; Karel, W.; Hu, H.; Ostrowski, W.

    2016-06-01

    During the last decade the use of airborne multi camera systems increased significantly. The development in digital camera technology allows mounting several mid- or small-format cameras efficiently onto one platform and thus enables image capture under different angles. Those oblique images turn out to be interesting for a number of applications since lateral parts of elevated objects, like buildings or trees, are visible. However, occlusion or illumination differences might challenge image processing. From an image orientation point of view those multi-camera systems bring the advantage of a better ray intersection geometry compared to nadir-only image blocks. On the other hand, varying scale, occlusion and atmospheric influences which are difficult to model impose problems to the image matching and bundle adjustment tasks. In order to understand current limitations of image orientation approaches and the influence of different parameters such as image overlap or GCP distribution, a commonly available dataset was released. The originally captured data comprises of a state-of-the-art image block with very high overlap, but in the first stage of the so-called ISPRS/EUROSDR benchmark on multi-platform photogrammetry only a reduced set of images was released. In this paper some first results obtained with this dataset are presented. They refer to different aspects like tie point matching across the viewing directions, influence of the oblique images onto the bundle adjustment, the role of image overlap and GCP distribution. As far as the tie point matching is concerned we observed that matching of overlapping images pointing to the same cardinal direction, or between nadir and oblique views in general is quite successful. Due to the quite different perspective between images of different viewing directions the standard tie point matching, for instance based on interest points does not work well. How to address occlusion and ambiguities due to different views onto

  2. Evaluation of the concrete shield compositions from the 2010 criticality accident alarm system benchmark experiments at the CEA Valduc SILENE facility

    Energy Technology Data Exchange (ETDEWEB)

    Miller, Thomas Martin [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Celik, Cihangir [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Dunn, Michael E [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Wagner, John C [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); McMahan, Kimberly L [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Authier, Nicolas [French Atomic Energy Commission (CEA), Centre de Valduc, Is-sur-Tille (France); Jacquet, Xavier [French Atomic Energy Commission (CEA), Centre de Valduc, Is-sur-Tille (France); Rousseau, Guillaume [French Atomic Energy Commission (CEA), Centre de Valduc, Is-sur-Tille (France); Wolff, Herve [French Atomic Energy Commission (CEA), Centre de Valduc, Is-sur-Tille (France); Savanier, Laurence [French Atomic Energy Commission (CEA), Centre de Valduc, Is-sur-Tille (France); Baclet, Nathalie [French Atomic Energy Commission (CEA), Centre de Valduc, Is-sur-Tille (France); Lee, Yi-kang [French Atomic Energy Commission (CEA), Centre de Saclay, Gif sur Yvette (France); Trama, Jean-Christophe [French Atomic Energy Commission (CEA), Centre de Saclay, Gif sur Yvette (France); Masse, Veronique [French Atomic Energy Commission (CEA), Centre de Saclay, Gif sur Yvette (France); Gagnier, Emmanuel [French Atomic Energy Commission (CEA), Centre de Saclay, Gif sur Yvette (France); Naury, Sylvie [French Atomic Energy Commission (CEA), Centre de Saclay, Gif sur Yvette (France); Blanc-Tranchant, Patrick [French Atomic Energy Commission (CEA), Centre de Saclay, Gif sur Yvette (France); Hunter, Richard [Babcock International Group (United Kingdom); Kim, Soon [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Dulik, George Michael [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Reynolds, Kevin H. [Y-12 National Security Complex, Oak Ridge, TN (United States)

    2015-01-01

    In October 2010, a series of benchmark experiments were conducted at the French Commissariat a l'Energie Atomique et aux Energies Alternatives (CEA) Valduc SILENE facility. These experiments were a joint effort between the United States Department of Energy Nuclear Criticality Safety Program and the CEA. The purpose of these experiments was to create three benchmarks for the verification and validation of radiation transport codes and evaluated nuclear data used in the analysis of criticality accident alarm systems. This series of experiments consisted of three single-pulsed experiments with the SILENE reactor. For the first experiment, the reactor was bare (unshielded), whereas in the second and third experiments, it was shielded by lead and polyethylene, respectively. The polyethylene shield of the third experiment had a cadmium liner on its internal and external surfaces, which vertically was located near the fuel region of SILENE. During each experiment, several neutron activation foils and thermoluminescent dosimeters (TLDs) were placed around the reactor. Nearly half of the foils and TLDs had additional high-density magnetite concrete, high-density barite concrete, standard concrete, and/or BoroBond shields. CEA Saclay provided all the concrete, and the US Y-12 National Security Complex provided the BoroBond. Measurement data from the experiments were published at the 2011 International Conference on Nuclear Criticality (ICNC 2011) and the 2013 Nuclear Criticality Safety Division (NCSD 2013) topical meeting. Preliminary computational results for the first experiment were presented in the ICNC 2011 paper, which showed poor agreement between the computational results and the measured values of the foils shielded by concrete. Recently the hydrogen content, boron content, and density of these concrete shields were further investigated within the constraints of the previously available data. New computational results for the first experiment are now available

  3. Gaming in a benchmarking environment. A non-parametric analysis of benchmarking in the water sector

    OpenAIRE

    De Witte, Kristof; Marques, Rui

    2009-01-01

    This paper discusses the use of benchmarking in general and its application to the drinking water sector. It systematizes the various classifications on performance measurement, discusses some of the pitfalls of benchmark studies and provides some examples of benchmarking in the water sector. After presenting in detail the institutional framework of the water sector of the Belgian region of Flanders (without benchmarking experiences), Wallonia (recently started a public benchmark) and the Net...

  4. MCNP neutron benchmarks

    International Nuclear Information System (INIS)

    More than 50 neutron benchmark calculations have recently been completed as part of an ongoing program to validate the MCNP Monte Carlo radiation transport code. The benchmark calculations reported here are part of an ongoing multiyear, multiperson effort to benchmark version 4 of the MCNP code. The MCNP is a Monte Carlo three-dimensional general-purpose, continuous-energy neutron, photon, and electron transport code. It is used around the world for many applications including aerospace, oil-well logging, physics experiments, criticality safety, reactor analysis, medical imaging, defense applications, accelerator design, radiation hardening, radiation shielding, health physics, fusion research, and education. The first phase of the benchmark project consisted of analytic and photon problems. The second phase consists of the ENDF/B-V neutron problems reported in this paper and in more detail in the comprehensive report. A cooperative program being carried out a General Electric, San Jose, consists of light water reactor benchmark problems. A subsequent phase focusing on electron problems is planned

  5. FLOWTRAN benchmarking with onset of flow instability data from 1988 Columbia University single-tube OFI experiment

    International Nuclear Information System (INIS)

    Benchmarking FLOWTRAN, Version 16.2, with an Onset of Significant Voiding (OSV) criterion against measured Onset of Flow Instability (OFI) data from the 1988--89 Columbia University downflow tests has shown that FLOWTRAN with OSV is a conservative OFI predictor. Calculated limiting flow rates based on the Savannah River Site (SRS) OSV criterion were always higher than the measured flow rates at OFI. This work supplements recent FLOWTRAN benchmarking against 1963 downflow tests at Columbia University and 1988 downflow tests at the Heat Transfer Laboratory. These studies provide confidence that using FLOWTRAN with an OSV based criterion for SRS reactor limits analyses will generate operating limits that are conservative with respect to OFI, the criterion selected to prevent fuel damage

  6. Numerical benchmarks TRIPOLI - MCNP with use of MCAM on FNG ITER bulk shield and FNG HCLL TBM mock-up experiments

    International Nuclear Information System (INIS)

    3D Monte Carlo (MC) transport codes are of first importance for the assessment of breeding blankets neutronic performances. This article supported by the EFDA Goal Oriented Training Program Eurobreed presents the difference in results between the CEA MC code TRIPOLI-4 and MCNP on two fusion neutronics benchmarks, assessing therefore TRIPOLI-4 calculation capabilities on shielding and tritium production rate (TPR). The first selected benchmark, assessing the shielding capability, is the Frascati neutron generator (FNG) ITER bulk shield experiment whereas the second benchmark, assessing the TPR calculation, is the preliminary design of the FNG helium cooled lithium-lead (HCLL) test blanket module (TBM) mock-up. To ensure the consistency of the geometry description, MCAM tool is used for automatic TRIPOLI - MCNP geometry conversions and check. A good coherence between TRIPOLI-4 and MCNP for neutron flux, reaction rates and TPR calculations is obtained. Moreover, it appears that MCAM performs fast, automatic and appropriate TRIPOLI - MCNP geometry conversions and finally that the tabulated FNG neutron source model from KIT is appropriate for TRIPOLI-4 calculations.

  7. Numerical benchmarks TRIPOLI - MCNP with use of MCAM on FNG ITER bulk shield and FNG HCLL TBM mock-up experiments

    Energy Technology Data Exchange (ETDEWEB)

    Fausser, Clement, E-mail: clement.fausser@cea.fr [CEA, DEN, Saclay, DANS/DM2S/SERMA, F-91191 Gif-sur-Yvette (France); Lee, Yi-Kang [CEA, DEN, Saclay, DANS/DM2S/SERMA, F-91191 Gif-sur-Yvette (France); Villari, Rosaria [Associazione EURATOM-ENEA sulla Fusione, Via Enrico Fermi 45, 00044 Frascati, Rome (Italy); Zeng Qin; Zhang Junjun [Institute of Plasma Physics, Chinese Academy of Sciences, Hefei, Anhui 230031 (China); Serikov, Arkady [Karlsruhe Institute of Technology, Institute for Neutron Physics and Reactor Technology (Germany); Trama, Jean-Christophe; Gabriel, Franck [CEA, DEN, Saclay, DANS/DM2S/SERMA, F-91191 Gif-sur-Yvette (France)

    2011-10-15

    3D Monte Carlo (MC) transport codes are of first importance for the assessment of breeding blankets neutronic performances. This article supported by the EFDA Goal Oriented Training Program Eurobreed presents the difference in results between the CEA MC code TRIPOLI-4 and MCNP on two fusion neutronics benchmarks, assessing therefore TRIPOLI-4 calculation capabilities on shielding and tritium production rate (TPR). The first selected benchmark, assessing the shielding capability, is the Frascati neutron generator (FNG) ITER bulk shield experiment whereas the second benchmark, assessing the TPR calculation, is the preliminary design of the FNG helium cooled lithium-lead (HCLL) test blanket module (TBM) mock-up. To ensure the consistency of the geometry description, MCAM tool is used for automatic TRIPOLI - MCNP geometry conversions and check. A good coherence between TRIPOLI-4 and MCNP for neutron flux, reaction rates and TPR calculations is obtained. Moreover, it appears that MCAM performs fast, automatic and appropriate TRIPOLI - MCNP geometry conversions and finally that the tabulated FNG neutron source model from KIT is appropriate for TRIPOLI-4 calculations.

  8. Quantum benchmarks for Gaussian states

    CERN Document Server

    Chiribella, Giulio

    2014-01-01

    Teleportation and storage of continuous variable states of light and atoms are essential building blocks for the realization of large scale quantum networks. Rigorous validation of these implementations require identifying, and surpassing, benchmarks set by the most effective strategies attainable without the use of quantum resources. Such benchmarks have been established for special families of input states, like coherent states and particular subclasses of squeezed states. Here we solve the longstanding problem of defining quantum benchmarks for general pure Gaussian states with arbitrary phase, displacement, and squeezing, randomly sampled according to a realistic prior distribution. As a special case, we show that the fidelity benchmark for teleporting squeezed states with totally random phase and squeezing degree is 1/2, equal to the corresponding one for coherent states. We discuss the use of entangled resources to beat the benchmarks in experiments.

  9. Jendl-3.1 iron validation on the PCA-REPLICA (H{sub 2}O/Fe) shielding benchmark experiment

    Energy Technology Data Exchange (ETDEWEB)

    Pescarini, M.; Borgia, M. G. [ENEA, Centro Ricerche ``Ezio Clementel``, Bologna (Italy). Dipt. Energia

    1997-03-01

    The PCA-REPLICA (H{sub 2}O/Fe) neutron shielding benchmarks experiment is analysed using the SN 2-D DOT 3.5-E code and the 3-D-equivalent flux synthesis method. This engineering benchmark reproduces the ex-core radial geometry of a PWR, including a mild steel reactor pressure vessel (RPV) simulator, and is designed to test the accuracy of the calculation of the in-vessel neutron exposure parameters. This accuracy is strongly dependent on the quality of the iron neutron cross sections used to describe the nuclear reactions within the RPV simulator. In particular, in this report, the cross sections based on the JENDL-3.1 iron data files are tested, through a comparison of the calculated integral and spectral results with the corresponding experimental data. In addition, the present results are compared, on the same benchmark experiment, with those of a preceding ENEA-Bologna validation of the ENDF/B VI iron cross sections. The integral result comparison indicates that, for all the threshold detectors considered (Rh-103 (n, n`) Rh-103m, In-115 (n, n`) In-115m and S-32 (n, p) P-32), the JENDL-3.1 natural iron data produce satisfactory results similar to those obtained with the ENDF/B VI iron data. On the contrary, when the JENDL/3.1 Fe-56 data file is used, strongly underestimated results are obtained for the lower energy threshold detectors, Rh-103 and In-115. This fact, in particular, becomes more evident with increasing the neutron penetration depth in the RPV simulator.

  10. CEC thermal-hydraulic benchmark exercise on Fiploc verification experiment F2 in Battelle model containment. Experimental phases 2, 3 and 4. Results of comparisons

    International Nuclear Information System (INIS)

    The present final report comprises the major results of Phase II of the CEC thermal-hydraulic benchmark exercise on Fiploc verification experiment F2 in the Battelle model containment, experimental phases 2, 3 and 4, which was organized and sponsored by the Commission of the European Communities for the purpose of furthering the understanding and analysis of long-term thermal-hydraulic phenomena inside containments during and after severe core accidents. This benchmark exercise received high European attention with eight organizations from six countries participating with eight computer codes during phase 2. Altogether 18 results from computer code runs were supplied by the participants and constitute the basis for comparisons with the experimental data contained in this publication. This reflects both the high technical interest in, as well as the complexity of, this CEC exercise. Major comparison results between computations and data are reported on all important quantities relevant for containment analyses during long-term transients. These comparisons comprise pressure, steam and air content, velocities and their directions, heat transfer coefficients and saturation ratios. Agreements and disagreements are discussed for each participating code/institution, conclusions drawn and recommendations provided. The phase 2 CEC benchmark exercise provided an up-to-date state-of-the-art status review of the thermal-hydraulic capabilities of present computer codes for containment analyses. This exercise has shown that all of the participating codes can simulate the important global features of the experiment correctly, like: temperature stratification, pressure and leakage, heat transfer to structures, relative humidity, collection of sump water. Several weaknesses of individual codes were identified, and this may help to promote their development. As a general conclusion it may be said that while there is still a wide area of necessary extensions and improvements, the

  11. Selecting benchmarks for reactor calculations

    International Nuclear Information System (INIS)

    Criticality, reactor physics, fusion and shielding benchmarks are expected to play important roles in GENIV design, safety analysis and in the validation of analytical tools used to design these reactors. For existing reactor technology, benchmarks are used to validate computer codes and test nuclear data libraries. However the selection of these benchmarks are usually done by visual inspection which is dependent on the expertise and the experience of the user and thereby resulting in a user bias in the process. In this paper we present a method for the selection of these benchmarks for reactor applications and uncertainty reduction based on Total Monte Carlo (TMC) method. Similarities between an application case and one or several benchmarks are quantified using the correlation coefficient. Based on the method, we also propose two approaches for reducing nuclear data uncertainty using integral benchmark experiments as an additional constrain in the TMC method: a binary accept/reject method and a method of uncertainty reduction using weights. Finally, the methods were applied to a full Lead Fast Reactor core and a set of criticality benchmarks. (author)

  12. Benchmark experiments with global climate models applicable to extra-solar gas giant planets in the shallow atmosphere approximation

    CERN Document Server

    Bending, V L; Kolb, U

    2012-01-01

    The growing field of exoplanetary atmospheric modelling has seen little work on standardised benchmark tests for its models, limiting understanding of the dependence of results on specific models and conditions. With spatially resolved observations as yet difficult to obtain, such a test is invaluable. Although an intercomparison test for models of tidally locked gas giant planets has previously been suggested and carried out, the data provided were limited in terms of comparability. Here, the shallow PUMA model is subjected to such a test, and detailed statistics produced to facilitate comparison, with both time means and the associated standard deviations displayed, removing the time dependence and providing a measure of the variability. Model runs have been analysed to determine the variability between resolutions, and the effect of resolution on the energy spectra studied. Superrotation is a robust and reproducible feature at all resolutions.

  13. A Global Vision over Benchmarking Process: Benchmarking Based Enterprises

    OpenAIRE

    Catalina SITNIKOV; Giurca Vasilescu, Laura

    2008-01-01

    Benchmarking uses the knowledge and the experience of others to improve the enterprise. Starting from the analysis of the performance and underlying the strengths and weaknesses of the enterprise it should be assessed what must be done in order to improve its activity. Using benchmarking techniques, an enterprise looks at how processes in the value chain are performed. The approach based on the vision “from the whole towards the parts” (a fragmented image of the enterprise’s value chain) redu...

  14. The PRISM Benchmark Suite

    OpenAIRE

    Kwiatkowsa, Marta; Norman, Gethin; Parker, David

    2012-01-01

    We present the PRISM benchmark suite: a collection of probabilistic models and property specifications, designed to facilitate testing, benchmarking and comparisons of probabilistic verification tools and implementations.

  15. Kvantitativ benchmark - Produktionsvirksomheder

    DEFF Research Database (Denmark)

    Sørensen, Ole H.; Andersen, Vibeke

    Rapport med resultatet af kvantitativ benchmark over produktionsvirksomhederne i VIPS projektet.......Rapport med resultatet af kvantitativ benchmark over produktionsvirksomhederne i VIPS projektet....

  16. Benchmarking in Student Affairs.

    Science.gov (United States)

    Mosier, Robert E.; Schwarzmueller, Gary J.

    2002-01-01

    Discusses the use of benchmarking in student affairs, focusing on issues related to student housing. Provides examples of how benchmarking has influenced administrative practice at many institutions. (EV)

  17. Laser-plasma interaction in ignition relevant plasmas: benchmarking our 3D modelling capabilities versus recent experiments

    Energy Technology Data Exchange (ETDEWEB)

    Divol, L; Froula, D H; Meezan, N; Berger, R; London, R A; Michel, P; Glenzer, S H

    2007-09-27

    We have developed a new target platform to study Laser Plasma Interaction in ignition-relevant condition at the Omega laser facility (LLE/Rochester)[1]. By shooting an interaction beam along the axis of a gas-filled hohlraum heated by up to 17 kJ of heater beam energy, we were able to create a millimeter-scale underdense uniform plasma at electron temperatures above 3 keV. Extensive Thomson scattering measurements allowed us to benchmark our hydrodynamic simulations performed with HYDRA [1]. As a result of this effort, we can use with much confidence these simulations as input parameters for our LPI simulation code pF3d [2]. In this paper, we show that by using accurate hydrodynamic profiles and full three-dimensional simulations including a realistic modeling of the laser intensity pattern generated by various smoothing options, fluid LPI theory reproduces the SBS thresholds and absolute reflectivity values and the absence of measurable SRS. This good agreement was made possible by the recent increase in computing power routinely available for such simulations.

  18. Electron induced break-up of helium. Benchmark experiments on a dynamical four-body Coulomb system

    International Nuclear Information System (INIS)

    This work presents an experimental study of fragmentation of helium by electron impact, in which absolute fully differential cross sections for single ionization, ionization-excitation and double ionization were determined. By applying a charged-particle imaging technique, the so-called 'reaction microscope', a large fraction of the final-state momentum space is covered, and the major limitations of previous experimental methods applied in this field could be overcome. Decisive modifications of the previous reaction microscope were undertaken, the most important one being the arrangement of the projectile beam parallel to the imaging-fields. For single ionization on helium an enhanced electron emission outside the projectile scattering plane is observed at both considered impact energies (102 eV and 1 keV), which is similar to the result found for ion-impact (M. Schulz et al., Nature (London) 422, 48 (2003)). The angle resolved cross sections obtained for double ionization at 105 eV impact energy reveal, that the process is dominated by the mutual repulsion of the three final-state continuum electrons. However, signatures of more complex dynamics are also observed. The data provide an ultimate benchmark for recently developed theories treating the dynamical three- and four-body Coulomb problem. (orig.)

  19. Assessment of the available 233U cross-section evaluations in the calculation of critical benchmark experiments

    International Nuclear Information System (INIS)

    In this report we investigate the adequacy of the available 233U cross-section data for calculation of experimental critical systems. The 233U evaluations provided in two evaluated nuclear data libraries, the U.S. Data Bank [ENDF/B (Evaluated Nuclear Data Files)] and the Japanese Data Bank [JENDL (Japanese Evaluated Nuclear Data Library)] are examined. Calculations were performed for six thermal and ten fast experimental critical systems using the Sn transport XSDRNPM code. To verify the performance of the 233U cross-section data for nuclear criticality safety application in which the neutron energy spectrum is predominantly in the epithermal energy range, calculations of four numerical benchmark systems with energy spectra in the intermediate energy range were done. These calculations serve only as an indication of the difference in calculated results that may be expected when the two 233U cross-section evaluations are used for problems with neutron spectra in the intermediate energy range. Additionally, comparisons of experimental and calculated central fission rate ratios were also made. The study has suggested that an ad hoc 233U evaluation based on the JENDL library provides better overall results for both fast and thermal experimental critical systems

  20. Assessment of the Available (Sup 233)U Cross Sections Evaluations in the Calculation of Critical Benchmark Experiments

    Energy Technology Data Exchange (ETDEWEB)

    Leal, L.C.

    1993-01-01

    In this report we investigate the adequacy of the available {sup 233}U cross-section data for calculation of experimental critical systems. The {sup 233}U evaluations provided in two evaluated nuclear data libraries, the U. S. Data Bank [ENDF/B (Evaluated Nuclear Data Files)] and the Japanese Data Bank [JENDL (Japanese Evaluated Nuclear Data Library)] are examined. Calculations were performed for six thermal and ten fast experimental critical systems using the Sn transport XSDRNPM code. To verify the performance of the {sup 233}U cross-section data for nuclear criticality safety application in which the neutron energy spectrum is predominantly in the epithermal energy range, calculations of four numerical benchmark systems with energy spectra in the intermediate energy range were done. These calculations serve only as an indication of the difference in calculated results that may be expected when the two {sup 233}U cross-section evaluations are used for problems with neutron spectra in the intermediate energy range. Additionally, comparisons of experimental and calculated central fission rate ratios were also made. The study has suggested that an ad hoc {sup 233}U evaluation based on the JENDL library provides better overall results for both fast and thermal experimental critical systems.

  1. Electron induced break-up of helium. Benchmark experiments on a dynamical four-body Coulomb system

    Energy Technology Data Exchange (ETDEWEB)

    Duerr, M.

    2006-07-05

    This work presents an experimental study of fragmentation of helium by electron impact, in which absolute fully differential cross sections for single ionization, ionization-excitation and double ionization were determined. By applying a charged-particle imaging technique, the so-called 'reaction microscope', a large fraction of the final-state momentum space is covered, and the major limitations of previous experimental methods applied in this field could be overcome. Decisive modifications of the previous reaction microscope were undertaken, the most important one being the arrangement of the projectile beam parallel to the imaging-fields. For single ionization on helium an enhanced electron emission outside the projectile scattering plane is observed at both considered impact energies (102 eV and 1 keV), which is similar to the result found for ion-impact (M. Schulz et al., Nature (London) 422, 48 (2003)). The angle resolved cross sections obtained for double ionization at 105 eV impact energy reveal, that the process is dominated by the mutual repulsion of the three final-state continuum electrons. However, signatures of more complex dynamics are also observed. The data provide an ultimate benchmark for recently developed theories treating the dynamical three- and four-body Coulomb problem. (orig.)

  2. Benchmarking of the 3-D CAD-based Discrete Ordinates code “ATTILA” for dose rate calculations against experiments and Monte Carlo calculations

    International Nuclear Information System (INIS)

    Shutdown dose rate (SDDR) inside and around the diagnostics ports of ITER is performed at PPPL/UCLA using the 3-D, FEM, Discrete Ordinates code, ATTILA, along with its updated FORNAX transmutation/decay gamma library. Other ITER partners assess SDDR using codes based on the Monte Carlo (MC) approach (e.g. MCNP code) for transport calculation and the radioactivity inventory code FISPACT or other equivalent decay data libraries for dose rate assessment. To reveal the range of discrepancies in the results obtained by various analysts, an extensive experimental and calculation benchmarking effort has been undertaken to validate the capability of ATTILA for dose rate assessment. On the experimental validation front, the comparison was performed using the measured data from two SDDR experiments performed at the FNG facility, Italy. Comparison was made to the experimental data and to MC results obtained by other analysts. On the calculation validation front, the ATTILA's predictions were compared to other results at key locations inside a calculation benchmark whose configuration duplicates an upper diagnostics port plug (UPP) in ITER. Both serial and parallel version of ATTILA-7.1.0 are used in the PPPL/UCLA analysis performed with FENDL-2.1/FORNAX databases. In the FNG 1st experimental, it was shown that ATTILA's dose rates are largely over estimated (by ∼30–60%) with the ANSI/ANS-6.1.1 flux-to-dose factors whereas the ICRP-74 factors give better agreement (10–20%) with the experimental data and with the MC results at all cooling times. In the 2nd experiment, there is an under estimation in SDDR calculated by both MCNP and ATTILA based on ANSI/ANS-6.1.1 for cooling times up to ∼4 days after irradiation. Thereafter, an over estimation is observed (∼5–10% with MCNP and ∼10–15% with ATTILA). As for the calculation benchmark, the agreement is much better based on ICRP-74 1996 data. The divergence among all dose rate results at ∼11 days cooling time is no

  3. Automatic generation of 3D fine mesh geometries for the analysis of the venus-3 shielding benchmark experiment with the Tort code

    Energy Technology Data Exchange (ETDEWEB)

    Pescarini, M.; Orsi, R.; Martinelli, T. [ENEA, Ente per le Nuove Tecnologie, l' Energia e l' Ambiente, Centro Ricerche Ezio Clementel Bologna (Italy)

    2003-07-01

    In many practical radiation transport applications today the cost for solving refined, large size and complex multi-dimensional problems is not so much computing but is linked to the cumbersome effort required by an expert to prepare a detailed geometrical model, verify and validate that it is correct and represents, to a specified tolerance, the real design or facility. This situation is, in particular, relevant and frequent in reactor core criticality and shielding calculations, with three-dimensional (3D) general purpose radiation transport codes, requiring a very large number of meshes and high performance computers. The need for developing tools that make easier the task to the physicist or engineer, by reducing the time required, by facilitating through effective graphical display the verification of correctness and, finally, that help the interpretation of the results obtained, has clearly emerged. The paper shows the results of efforts in this field through detailed simulations of a complex shielding benchmark experiment. In the context of the activities proposed by the OECD/NEA Nuclear Science Committee (NSC) Task Force on Computing Radiation Dose and Modelling of Radiation-Induced Degradation of Reactor Components (TFRDD), the ENEA-Bologna Nuclear Data Centre contributed with an analysis of the VENUS-3 low-flux neutron shielding benchmark experiment (SCK/CEN-Mol, Belgium). One of the targets of the work was to test the BOT3P system, originally developed at the Nuclear Data Centre in ENEA-Bologna and actually released to OECD/NEA Data Bank for free distribution. BOT3P, ancillary system of the DORT (2D) and TORT (3D) SN codes, permits a flexible automatic generation of spatial mesh grids in Cartesian or cylindrical geometry, through combinatorial geometry algorithms, following a simplified user-friendly approach. This system demonstrated its validity also in core criticality analyses, as for example the Lewis MOX fuel benchmark, permitting to easily

  4. Automatic generation of 3D fine mesh geometries for the analysis of the venus-3 shielding benchmark experiment with the Tort code

    International Nuclear Information System (INIS)

    In many practical radiation transport applications today the cost for solving refined, large size and complex multi-dimensional problems is not so much computing but is linked to the cumbersome effort required by an expert to prepare a detailed geometrical model, verify and validate that it is correct and represents, to a specified tolerance, the real design or facility. This situation is, in particular, relevant and frequent in reactor core criticality and shielding calculations, with three-dimensional (3D) general purpose radiation transport codes, requiring a very large number of meshes and high performance computers. The need for developing tools that make easier the task to the physicist or engineer, by reducing the time required, by facilitating through effective graphical display the verification of correctness and, finally, that help the interpretation of the results obtained, has clearly emerged. The paper shows the results of efforts in this field through detailed simulations of a complex shielding benchmark experiment. In the context of the activities proposed by the OECD/NEA Nuclear Science Committee (NSC) Task Force on Computing Radiation Dose and Modelling of Radiation-Induced Degradation of Reactor Components (TFRDD), the ENEA-Bologna Nuclear Data Centre contributed with an analysis of the VENUS-3 low-flux neutron shielding benchmark experiment (SCK/CEN-Mol, Belgium). One of the targets of the work was to test the BOT3P system, originally developed at the Nuclear Data Centre in ENEA-Bologna and actually released to OECD/NEA Data Bank for free distribution. BOT3P, ancillary system of the DORT (2D) and TORT (3D) SN codes, permits a flexible automatic generation of spatial mesh grids in Cartesian or cylindrical geometry, through combinatorial geometry algorithms, following a simplified user-friendly approach. This system demonstrated its validity also in core criticality analyses, as for example the Lewis MOX fuel benchmark, permitting to easily

  5. Development of the 4S and related technologies (8). An application of physics benchmark experiment results to safety analyses of small fast reactors. An analysis of delayed neutron reaction benchmark results using nuclear design methodology

    International Nuclear Information System (INIS)

    Various benchmark cores were analyzed with the 4S nuclear design methodology (in combination of transport calculation methods and JENDL-3.3 base libraries), in order to enhance the data base for evaluation of uncertainties of nuclear characteristics. Delayed neutron fraction analyses were made for support of the safety analyses of the 4S core. Using critical benchmark data obtained in leakage dominant small ZPR and ZPPR cores, low enriched uranium fast spectra cores have been analyzed using JENDL-3.3 and ENDF/B-VII.0 libraries to understand the results obtained from the recent reflector control physics benchmark FCA XXIII cores. The evaluation showed delayed neutron fractions and criticality were reproduced in good agreement with the 4S nuclear design methodology. (author)

  6. Benchmark experiments on a lead reflected system and calculations on the geometry of the experimental facility using most of the commonly available nuclear data libraries

    International Nuclear Information System (INIS)

    A series of criticality benchmark experiments with a small LWR-type core, reflected by 30 cm of lead, was defined jointly by SEC (Service d'Etude de Criticite), Fontenay-aux-Roses, and SRD (Safety and Reliability Directorate). These experiments are very representative of the reflecting effect of lead, since the contribution of the lead to the reactivity was assessed as about 30% in Δ K. The experiments were carried out by SRSC (Service de Recherche en Surete et Criticite), Valduc, in December 1983 in the sub-critical facility called APPARATUS B. In addition, they confirmed and measured the effect on reactivity of a water gap between the core and the lead reflector; with a water gap of less than 1 cm, the reactivity can be greater than that of the core directly reflected the lead or by over 20 cm of water. The experimental results were to a large extent made use of by SRD with the aid of the MONK Monte Carlo code and to some extent by SEC with the aid of the MORET Monte Carlo Code. All the results obtained are presented in the summary tables. These experiments allowed to compare the different libraries of cross sections available

  7. Benchmarking the cad-based attila discrete ordinates code with experimental data of fusion experiments and to the results of MCNP code in simulating ITER

    International Nuclear Information System (INIS)

    Attila is a newly developed finite element code based on Sn neutron, gamma, and charged particle transport in 3-D geometry in which unstructured tetrahedral meshes are generated to describe complex geometry that is based on CAD input (Solid Works, Pro/Engineer, etc). In the present work we benchmark its calculation accuracy by comparing its prediction to the measured data inside two experimental mock-ups bombarded with 14 MeV neutrons. The results are also compared to those based on MCNP calculations. The experimental mock-ups simulate parts of the International Thermonuclear Experimental Reactor (ITER) in-vessel components, namely: (1) the Tungsten mockup configuration (54.3 cm x 46.8 cm x 45 cm), and (2) the ITER shielding blanket followed by the SCM region (simulated by alternating layers of SS316 and copper). In the latter configuration, a high aspect ratio rectangular streaming channel was introduced (to simulate steaming paths between ITER blanket modules) which ends with a rectangular cavity. The experiments on these two fusion-oriented integral experiments were performed at the Fusion Neutron Generator (FNG) facility, Frascati, Italy. In addition, the nuclear performance of the ITER MCNP 'Benchmark' CAD model has been performed with Attila to compare its results to those obtained with CAD-based MCNP approach developed by several ITER participants. The objective of this paper is to compare results based on two distinctive 3-D calculation tools using the same nuclear data, FENDL2.1, and the same response functions of several reaction rates measured in ITER mock-ups and to enhance confidence from the international neutronics community in the Attila code and how it can precisely quantify the nuclear field in large and complex systems, such as ITER. Attila has the advantage of providing a full flux mapping visualization everywhere in one run where components subjected to excessive radiation level and strong streaming paths can be identified. In addition, the

  8. Benchmarking and Performance Measurement.

    Science.gov (United States)

    Town, J. Stephen

    This paper defines benchmarking and its relationship to quality management, describes a project which applied the technique in a library context, and explores the relationship between performance measurement and benchmarking. Numerous benchmarking methods contain similar elements: deciding what to benchmark; identifying partners; gathering…

  9. Benchmark data set for wheat growth models: field experiments and AgMIP multi-model simulations

    NARCIS (Netherlands)

    Supit, I.; Wolf, J.; Asseng, S.; Ewert, F.; Martre, P.; Rosenzweig, C.

    2015-01-01

    The data set includes a current representative management treatment from detailed, quality-tested sentinel field experiments with wheat from four contrasting environments including Australia, The Netherlands, India and Argentina. Measurements include local daily climate data (solar radiation, maximu

  10. Implementation of Benchmarking Transportation Logistics Practices and Future Benchmarking Organizations

    International Nuclear Information System (INIS)

    The purpose of the Office of Civilian Radioactive Waste Management's (OCRWM) Logistics Benchmarking Project is to identify established government and industry practices for the safe transportation of hazardous materials which can serve as a yardstick for design and operation of OCRWM's national transportation system for shipping spent nuclear fuel and high-level radioactive waste to the proposed repository at Yucca Mountain, Nevada. The project will present logistics and transportation practices and develop implementation recommendations for adaptation by the national transportation system. This paper will describe the process used to perform the initial benchmarking study, highlight interim findings, and explain how these findings are being implemented. It will also provide an overview of the next phase of benchmarking studies. The benchmarking effort will remain a high-priority activity throughout the planning and operational phases of the transportation system. The initial phase of the project focused on government transportation programs to identify those practices which are most clearly applicable to OCRWM. These Federal programs have decades of safe transportation experience, strive for excellence in operations, and implement effective stakeholder involvement, all of which parallel OCRWM's transportation mission and vision. The initial benchmarking project focused on four business processes that are critical to OCRWM's mission success, and can be incorporated into OCRWM planning and preparation in the near term. The processes examined were: transportation business model, contract management/out-sourcing, stakeholder relations, and contingency planning. More recently, OCRWM examined logistics operations of AREVA NC's Business Unit Logistics in France. The next phase of benchmarking will focus on integrated domestic and international commercial radioactive logistic operations. The prospective companies represent large scale shippers and have vast experience in

  11. Integral test of JENDL-3PR1 through benchmark experiments on Li/sub 2/O slab assemblies

    International Nuclear Information System (INIS)

    Two types of benchmarck experiments on Li/sub 2/O assemblies have been carried out. They were analyzed by using three transport codes with JENDL-3PR1 and ENDF/B-4. The calculation using JENDL-3PR1 predicted tritium production rates of /sup 6/Li and /sup 7/Li better than those using ENDF/B-4

  12. SAS Code for Calculating Intraclass Correlation Coefficients and Effect Size Benchmarks for Site-Randomized Education Experiments

    Science.gov (United States)

    Brandon, Paul R.; Harrison, George M.; Lawton, Brian E.

    2013-01-01

    When evaluators plan site-randomized experiments, they must conduct the appropriate statistical power analyses. These analyses are most likely to be valid when they are based on data from the jurisdictions in which the studies are to be conducted. In this method note, we provide software code, in the form of a SAS macro, for producing statistical…

  13. Benchmark test of TRIPOLI-4 code through simple model calculation and analysis of fusion neutronics experiments at JAEA/FNS

    Energy Technology Data Exchange (ETDEWEB)

    Ohta, Masayuki, E-mail: ohta.masayuki@jaea.go.jp [Japan Atomic Energy Agency, Tokai-mura, Naka-gun, Ibaraki-ken 319-1195 (Japan); Takakura, Kosuke; Ochiai, Kentaro; Sato, Satoshi; Konno, Chikara [Japan Atomic Energy Agency, Tokai-mura, Naka-gun, Ibaraki-ken 319-1195 (Japan)

    2013-10-15

    In order to examine a basic performance of the TRIPOLI code, two types of analyses were carried out with TRIPOLI-4.4 and MCNP5-1.40; one is a simple model calculation and the other is an analysis of iron fusion neutronics experiments with DT neutrons at the Fusion Neutronics Source (FNS) facility in Japan Atomic Energy Agency (JAEA). In the simple model calculation, we adopted a sphere of 0.5 m in radius with a 20 MeV neutron source in the center and calculated leakage neutron spectra from the sphere. We also analyzed in situ and Time-of-Flight (TOF) experiments for iron at JAEA/FNS. For the in situ experiment, neutron spectra and reaction rates for dosimetry reactions were calculated for several points inside the assembly. For the TOF experiment, angular neutron leakage spectra from the assembly were calculated. Results with TRIPOLI were comparable to those with MCNP in most calculations, but a difference between TRIPOLI and MCNP calculation results, probably caused by inadequate treatment of inelastic scattering data in TRIPOLI, appears in some calculations.

  14. Benchmark experiment for the cross section of the 100Mo(p,2n)99mTc and 100Mo(p,pn)99Mo reactions

    Science.gov (United States)

    Takács, S.; Ditrói, F.; Aikawa, M.; Haba, H.; Otuka, N.

    2016-05-01

    As nuclear medicine community has shown an increasing interest in accelerator produced 99mTc radionuclide, the possible alternative direct production routes for producing 99mTc were investigated intensively. One of these accelerator production routes is based on the 100Mo(p,2n)99mTc reaction. The cross section of this nuclear reaction was studied by several laboratories earlier but the available data-sets are not in good agreement. For large scale accelerator production of 99mTc based on the 100Mo(p,2n)99mTc reaction, a well-defined excitation function is required to optimise the production process effectively. One of our recent publications pointed out that most of the available experimental excitation functions for the 100Mo(p,2n)99mTc reaction have the same general shape while their amplitudes are different. To confirm the proper amplitude of the excitation function, results of three independent experiments were presented (Takács et al., 2015). In this work we present results of a thick target count rate measurement of the Eγ = 140.5 keV gamma-line from molybdenum irradiated by Ep = 17.9 MeV proton beam, as an integral benchmark experiment, to prove the cross section data reported for the 100Mo(p,2n)99mTc and 100Mo(p,pn)99Mo reactions in Takács et al. (2015).

  15. Shielding benchmark test

    International Nuclear Information System (INIS)

    Iron data in JENDL-2 have been tested by analyzing shielding benchmark experiments for neutron transmission through iron block performed at KFK using CF-252 neutron source and at ORNL using collimated neutron beam from reactor. The analyses are made by a shielding analysis code system RADHEAT-V4 developed at JAERI. The calculated results are compared with the measured data. As for the KFK experiments, the C/E values are about 1.1. For the ORNL experiments, the calculated values agree with the measured data within an accuracy of 33% for the off-center geometry. The d-t neutron transmission measurements through carbon sphere made at LLNL are also analyzed preliminarily by using the revised JENDL data for fusion neutronics calculation. (author)

  16. New Benchmarks from Tokamak Experiments for Theoretical Calculations of the Dielectronic Satellite Spectra of Helium-like Ions

    International Nuclear Information System (INIS)

    Dielectronic satellite spectra of helium-like argon, recorded with a high-resolution X-ray crystal spectrometer at the National Spherical Torus Experiment, were found to be inconsistent with existing predictions resulting in unacceptable values for the power balance and suggesting the unlikely existence of non-Maxwellian electron energy distributions. These problems were resolved with calculations from a new atomic code. It is now possible to perform reliable electron temperature measurements and to eliminate the uncertainties associated with determinations of non-Maxwellian distributions

  17. Lower hybrid current drive: an overview of simulation models, benchmarking with experiment, and predictions for future devices

    International Nuclear Information System (INIS)

    This paper reviews the status of lower hybrid current drive (LHCD) simulation and modeling. We first discuss modules used for wave propagation, absorption, and current drive with particular emphasis placed on comparing exact numerical solutions of the Fokker Planck equation in 2-dimension with solution methods that employ 1-dimensional and adjoint approaches. We also survey model predictions for LHCD in past and present experiments showing detailed comparisons between simulated and observed current drive efficiencies and hard X-ray profiles. Finally we discuss several model predictions for lower hybrid current profile control in proposed next step reactor options. (authors)

  18. New integral experiments for large angle scattering cross section data benchmarking with DT neutron beam at JAEA/FNS

    Energy Technology Data Exchange (ETDEWEB)

    Ohnishi, Seiki, E-mail: ohnishi@nmri.go.jp [National Maritime Research Institute, Mitaka, Tokyo (Japan); Kondo, Keitaro [Japan Atomic Energy Agency, Tokai-mura, Ibaraki (Japan); Azuma, Tetsushi [Kyoto University, Kumatori-cho, Osaka (Japan); Sato, Satoshi; Ochiai, Kentaro; Takakura, Kosuke [Japan Atomic Energy Agency, Tokai-mura, Ibaraki (Japan); Murata, Isao [Osaka University, Suita, Osaka (Japan); Konno, Chikara [Japan Atomic Energy Agency, Tokai-mura, Ibaraki (Japan)

    2012-08-15

    Highlights: Black-Right-Pointing-Pointer We constructed a deuteron-triton fusion neutron beam. Black-Right-Pointing-Pointer A large SS316 cylinder assembly was irradiated by the neutron beam. Black-Right-Pointing-Pointer The reaction rates of activation foil were measured inside it. Black-Right-Pointing-Pointer The C/Es became smaller with the distance from the beam axis. Black-Right-Pointing-Pointer The large angle scattering cross section around 90 Degree-Sign affects that. - Abstract: A new integral experiment with a deuteron-triton fusion (DT) neutron beam started in order to validate scattering cross section data. First the DT neutron beam was constructed with a collimator. The performance of the collimator system and the characteristics of the DT neutron beam were measured. Second a new integral experiment for type 316 stainless steel (SS316) was carried out with this DT neutron beam. The DT neutron beam of 3.5 cm in diameter was injected to the front surface center of an SS316 cylindrical assembly. Reaction rates of the {sup 93}Nb(n,2n){sup 92m}Nb reaction in the assembly were measured with the activation foil method and were calculated with the Monte Carlo transport calculation code. The measurement points were located at three positions, on the center of the beam axis and at 15 cm and 30 cm apart from the axis. The ratio of calculation to experiment of the {sup 93}Nb(n,2n){sup 92m}Nb reaction rate became smaller than 1 with the distance from the beam axis. Then, the dependency of each reaction rate on scattering angle was calculated. It was proved that at off-axis positions, where C/E is smaller than 1, 90 Degree-Sign scattering contribute relatively larger than at on-axis positions and backward scattering made little contribution to the results in this experiment. The reasons of the discrepancy between the measured and calculated data will be investigated.

  19. A Blind Test Experiment in Volcano Geodesy: a Benchmark for Inverse Methods of Ground Deformation and Gravity Data

    Science.gov (United States)

    D'Auria, Luca; Fernandez, Jose; Puglisi, Giuseppe; Rivalta, Eleonora; Camacho, Antonio; Nikkhoo, Mehdi; Walter, Thomas

    2016-04-01

    The inversion of ground deformation and gravity data is affected by an intrinsic ambiguity because of the mathematical formulation of the inverse problem. Current methods for the inversion of geodetic data rely on both parametric (i.e. assuming a source geometry) and non-parametric approaches. The former are able to catch the fundamental features of the ground deformation source but, if the assumptions are wrong or oversimplified, they could provide misleading results. On the other hand, the latter class of methods, even if not relying on stringent assumptions, could suffer from artifacts, especially when dealing with poor datasets. In the framework of the EC-FP7 MED-SUV project we aim at comparing different inverse approaches to verify how they cope with basic goals of Volcano Geodesy: determining the source depth, the source shape (size and geometry), the nature of the source (magmatic/hydrothermal) and hinting the complexity of the source. Other aspects that are important in volcano monitoring are: volume/mass transfer toward shallow depths, propagation of dikes/sills, forecasting the opening of eruptive vents. On the basis of similar experiments already done in the fields of seismic tomography and geophysical imaging, we have devised a bind test experiment. Our group was divided into one model design team and several inversion teams. The model design team devised two physical models representing volcanic events at two distinct volcanoes (one stratovolcano and one caldera). They provided the inversion teams with: the topographic reliefs, the calculated deformation field (on a set of simulated GPS stations and as InSAR interferograms) and the gravity change (on a set of simulated campaign stations). The nature of the volcanic events remained unknown to the inversion teams until after the submission of the inversion results. Here we present the preliminary results of this comparison in order to determine which features of the ground deformation and gravity source

  20. Benchmark Data Set for Wheat Growth Models: Field Experiments and AgMIP Multi-Model Simulations.

    Science.gov (United States)

    Asseng, S.; Ewert, F.; Martre, P.; Rosenzweig, C.; Jones, J. W.; Hatfield, J. L.; Ruane, A. C.; Boote, K. J.; Thorburn, P.J.; Rotter, R. P.

    2015-01-01

    The data set includes a current representative management treatment from detailed, quality-tested sentinel field experiments with wheat from four contrasting environments including Australia, The Netherlands, India and Argentina. Measurements include local daily climate data (solar radiation, maximum and minimum temperature, precipitation, surface wind, dew point temperature, relative humidity, and vapor pressure), soil characteristics, frequent growth, nitrogen in crop and soil, crop and soil water and yield components. Simulations include results from 27 wheat models and a sensitivity analysis with 26 models and 30 years (1981-2010) for each location, for elevated atmospheric CO2 and temperature changes, a heat stress sensitivity analysis at anthesis, and a sensitivity analysis with soil and crop management variations and a Global Climate Model end-century scenario.

  1. Benchmark Analyses on the Control Rod Withdrawal Tests Performed During the PHÉNIX End-of-Life Experiments. Report of a Coordinated Research Project 2008–2011

    International Nuclear Information System (INIS)

    The IAEA supports Member State activities in advanced fast reactor technology development by providing a major fulcrum for information exchange and collaborative research programmes. The IAEA’s activities in this field are mainly carried out within the framework of the Technical Working Group on Fast Reactors (TWG-FR), which assists in the implementation of corresponding IAEA activities and ensures that all technical activities are in line with the expressed needs of Member States. In the broad range of activities, the IAEA proposes and establishes coordinated research projects (CRPs) aimed at improving Member States’ capabilities in fast reactor design and analysis. An important opportunity to conduct collaborative research activities was provided by the experimental campaign run by the French Alternative Energies and Atomic Energy Commission (CEA, Commissariat à l’énergie atomique et aux énergies alternatives) at the PHÉNIX, a prototype sodium cooled fast reactor. Before the definitive shutdown in 2009, end-of-life tests were conducted to gather additional experience on the operation of sodium cooled reactors. Thanks to the CEA opening the experiments to international cooperation, the IAEA decided in 2007 to launch the CRP entitled Control Rod Withdrawal and Sodium Natural Circulation Tests Performed during the PHÉNIX End-of-Life Experiments. The CRP, together with institutes from seven States, contributed to improving capabilities in sodium cooled fast reactor simulation through code verification and validation, with particular emphasis on temperature and power distribution calculations and the analysis of sodium natural circulation phenomena. The objective of this publication is to document the results and main achievements of the benchmark analyses on the control rod withdrawal test performed within the framework of the PHÉNIX end-of-life experimental campaign

  2. Benchmark Analyses on the Natural Circulation Test Performed During the PHENIX End-of-Life Experiments. Final Report of a Co-ordinated Research Project 2008-2011

    International Nuclear Information System (INIS)

    The International Atomic Energy Agency (IAEA) supports Member State activities in the area of advanced fast reactor technology development by providing a forum for information exchange and collaborative research programmes. The Agency's activities in this field are mainly carried out within the framework of the Technical Working Group on Fast Reactors (TWG-FR), which assists in the implementation of corresponding IAEA activities and ensures that all technical activities are in line with the expressed needs of Member States. Among its broad range of activities, the IAEA proposes and establishes coordinated research projects (CRPs) aimed at the improvement of Member State capabilities in the area of fast reactor design and analysis. An important opportunity to undertake collaborative research was provided by the experimental campaign of the French Alternative Energies and Atomic Energy Commission (CEA) in the prototype sodium fast reactor PHENIX before it was shut down in 2009. The overall purpose of the end of life tests was to gather additional experience on the operation of sodium cooled reactors. As the CEA opened the experiments to international cooperation, in 2007 the IAEA launched a CRP on ''Control Rod Withdrawal and Sodium Natural Circulation Tests Performed during the PHENIX End-of-Life Experiments''. The CRP, with the participation of institutes from eight countries, contributed to improving capabilities in sodium cooled reactor simulation through code verification and validation, with particular emphasis on temperature and power distribution calculations and the analysis of sodium natural circulation phenomena. The objective of this report is to document the results and main achievements of the benchmark analyses on the natural circulation test performed in the framework of the PHENIX end of life experimental campaign

  3. Aquatic Life Benchmarks

    Data.gov (United States)

    U.S. Environmental Protection Agency — The Aquatic Life Benchmarks is an EPA-developed set of criteria for freshwater species. These benchmarks are based on toxicity values reviewed by EPA and used in...

  4. Validity assessment of shielding design tools for ITER through analysis of benchmark experiment on SS316/water shield conducted at FNS/JAERI

    Energy Technology Data Exchange (ETDEWEB)

    Maekawa, Fujio; Ikeda, Yujiro; Verzilov, Y.M.; Konno, Chikara; Wada, Masayuki; Maekawa, Hiroshi; Oyama, Yukio; Uno, Yoshitomo [Japan Atomic Energy Research Inst., Ibaraki (Japan)

    1996-12-31

    To assess validity of the shielding design tools for ITER, the benchmark experiment on SS316/water shield conducted at FNS/JAERI is analyzed. As far as a simple bulk shield of SS316/water is concerned, the followings are found assuming that no uncertainty is involved in the response functions of the design parameters. Nuclear data bases of JENDL Fusion File and FENDL/E-1.0 are valid to predict all the design parameters with uncertainties less than a factor of 1.25. At the connection legs between shield blanket modules and back plates, both MCNP and DOT calculations can predict helium production rate with uncertainties less than 10%. For the toroidal field coils on the midplane, all the nuclear parameters can be predicted with uncertainties less than a factor of 1.25 by MCNP and DOT with consideration of self-shielding correction of cross sections and energy group structure of 125-n and 40-{gamma}. The uncertainties for toroidal field coils are considerably smaller than the design margins secured to the shielding designs under ITER/EDA. 22 refs., 8 figs.

  5. Atomic Energy Research benchmark activity

    International Nuclear Information System (INIS)

    The test problems utilized in the validation and verification process of computer programs in Atomic Energie Research are collected into one bunch. This is the first step towards issuing a volume in which tests for VVER are collected, along with reference solutions and a number of solutions. The benchmarks do not include the ZR-6 experiments because they have been published along with a number of comparisons in the Final reports of TIC. The present collection focuses on operational and mathematical benchmarks which cover almost the entire range of reaktor calculation. (Author)

  6. Status of the international criticality safety benchmark evaluation project (ICSBEP)

    International Nuclear Information System (INIS)

    Since ICNC'99, four new editions of the International Handbook of Evaluated Criticality Safety Benchmark Experiments have been published. The number of benchmark specifications in the Handbook has grown from 2157 in 1999 to 3073 in 2003, an increase of nearly 1000 specifications. These benchmarks are used to validate neutronics codes and nuclear cross-section data. Twenty evaluations representing 192 benchmark specifications were added to the Handbook in 2003. The status of the International Criticality Safety Benchmark Evaluation Project (ICSBEP) is provided in this paper along with a summary of the newly added benchmark specifications that appear in the 2003 Edition of the Handbook. (author)

  7. Benchmarking semantic web technology

    CERN Document Server

    García-Castro, R

    2009-01-01

    This book addresses the problem of benchmarking Semantic Web Technologies; first, from a methodological point of view, proposing a general methodology to follow in benchmarking activities over Semantic Web Technologies and, second, from a practical point of view, presenting two international benchmarking activities that involved benchmarking the interoperability of Semantic Web technologies using RDF(S) as the interchange language in one activity and OWL in the other.The book presents in detail how the different resources needed for these interoperability benchmarking activities were defined:

  8. Benchmark calculations for EGS5

    International Nuclear Information System (INIS)

    In the past few years, EGS4 has undergone an extensive upgrade to EGS5, in particularly in the areas of low-energy electron physics, low-energy photon physics, PEGS cross section generation, and the coding from Mortran to Fortran programming. Benchmark calculations have been made to assure the accuracy, reliability and high quality of the EGS5 code system. This study reports three benchmark examples that show the successful upgrade from EGS4 to EGS5 based on the excellent agreements among EGS4, EGS5 and measurements. The first benchmark example is the 1969 Crannell Experiment to measure the three-dimensional distribution of energy deposition for 1-GeV electrons shower in water and aluminum tanks. The second example is the 1995 Compton-scattered spectra measurements for 20-40 keV, linearly polarized photon by Namito et. al., in KEK, which was a main part of the low-energy photon expansion work for both EGS4 and EGS5. The third example is the 1986 heterogeneity benchmark experiment by Shortt et. al., who used a monoenergetic 20-MeV electron beam to hit the front face of a water tank containing both air and aluminum cylinders and measured spatial depth dose distribution using a small solid-state detector. (author)

  9. The Conic Benchmark Format

    DEFF Research Database (Denmark)

    Friberg, Henrik A.

    This document constitutes the technical reference manual of the Conic Benchmark Format with le extension: .cbf or .CBF. It unies linear, second-order cone (also known as conic quadratic) and semidenite optimization with mixed-integer variables. The format has been designed with benchmark libraries...... in mind, and therefore focuses on compact and easily parsable representations. The problem structure is separated from the problem data, and the format moreover facilitate benchmarking of hotstart capability through sequences of changes....

  10. Bayesian Benchmark Dose Analysis

    OpenAIRE

    Fang, Qijun; Piegorsch, Walter W.; Barnes, Katherine Y.

    2014-01-01

    An important objective in environmental risk assessment is estimation of minimum exposure levels, called Benchmark Doses (BMDs) that induce a pre-specified Benchmark Response (BMR) in a target population. Established inferential approaches for BMD analysis typically involve one-sided, frequentist confidence limits, leading in practice to what are called Benchmark Dose Lower Limits (BMDLs). Appeal to Bayesian modeling and credible limits for building BMDLs is far less developed, however. Indee...

  11. Risk Management with Benchmarking

    OpenAIRE

    Suleyman Basak; Alex Shapiro; Lucie Teplá

    2005-01-01

    Portfolio theory must address the fact that, in reality, portfolio managers are evaluated relative to a benchmark, and therefore adopt risk management practices to account for the benchmark performance. We capture this risk management consideration by allowing a prespecified shortfall from a target benchmark-linked return, consistent with growing interest in such practice. In a dynamic setting, we demonstrate how a risk-averse portfolio manager optimally under- or overperforms a target benchm...

  12. The role of benchmarking for yardstick competition

    International Nuclear Information System (INIS)

    With the increasing interest in yardstick regulation, there is a need to understand the most appropriate method for realigning tariffs at the outset. Benchmarking is the tool used for such realignment and is therefore a necessary first-step in the implementation of yardstick competition. A number of concerns have been raised about the application of benchmarking, making some practitioners reluctant to move towards yardstick based regimes. We assess five of the key concerns often discussed and find that, in general, these are not as great as perceived. The assessment is based on economic principles and experiences with applying benchmarking to regulated sectors, e.g. in the electricity and water industries in the UK, The Netherlands, Austria and Germany in recent years. The aim is to demonstrate that clarity on the role of benchmarking reduces the concern about its application in different regulatory regimes. We find that benchmarking can be used in regulatory settlements, although the range of possible benchmarking approaches that are appropriate will be small for any individual regulatory question. Benchmarking is feasible as total cost measures and environmental factors are better defined in practice than is commonly appreciated and collusion is unlikely to occur in environments with more than 2 or 3 firms (where shareholders have a role in monitoring and rewarding performance). Furthermore, any concern about companies under-recovering costs is a matter to be determined through the regulatory settlement and does not affect the case for using benchmarking as part of that settlement. (author)

  13. MCNP neutron benchmarks

    International Nuclear Information System (INIS)

    Over 50 neutron benchmark calculations have recently been completed as part of an ongoing program to validate the MCNP Monte Carlo radiation transport code. The new and significant aspects of this work are as follows: These calculations are the first attempt at a validation program for MCNP and the first official benchmarking of version 4 of the code. We believe the chosen set of benchmarks is a comprehensive set that may be useful for benchmarking other radiation transport codes and data libraries. These calculations provide insight into how well neutron transport calculations can be expected to model a wide variety of problems

  14. How Activists Use Benchmarks

    DEFF Research Database (Denmark)

    Seabrooke, Leonard; Wigan, Duncan

    2015-01-01

    Non-governmental organisations use benchmarks as a form of symbolic violence to place political pressure on firms, states, and international organisations. The development of benchmarks requires three elements: (1) salience, that the community of concern is aware of the issue and views it as...... important; (2) will, that activists and issue entrepreneurs will carry the message forward; and (3) expertise, that benchmarks created can be defended as accurate representations of what is happening on the issue of concern. We contrast two types of benchmarking cycles where salience, will, and expertise...

  15. Benchmarking in Identifying Priority Directions of Development of Telecommunication Operators

    Directory of Open Access Journals (Sweden)

    Zaharchenko Lolita A.

    2013-12-01

    Full Text Available The article analyses evolution of development and possibilities of application of benchmarking in the telecommunication sphere. It studies essence of benchmarking on the basis of generalisation of approaches of different scientists to definition of this notion. In order to improve activity of telecommunication operators, the article identifies the benchmarking technology and main factors, that determine success of the operator in the modern market economy, and the mechanism of benchmarking and component stages of carrying out benchmarking by a telecommunication operator. It analyses the telecommunication market and identifies dynamics of its development and tendencies of change of the composition of telecommunication operators and providers. Having generalised the existing experience of benchmarking application, the article identifies main types of benchmarking of telecommunication operators by the following features: by the level of conduct of (branch, inter-branch and international benchmarking; by relation to participation in the conduct (competitive and joint; and with respect to the enterprise environment (internal and external.

  16. Vver-1000 Mox core computational benchmark

    International Nuclear Information System (INIS)

    The NEA Nuclear Science Committee has established an Expert Group that deals with the status and trends of reactor physics, fuel performance and fuel cycle issues related to disposing of weapons-grade plutonium in mixed-oxide fuel. The objectives of the group are to provide NEA member countries with up-to-date information on, and to develop consensus regarding, core and fuel cycle issues associated with burning weapons-grade plutonium in thermal water reactors (PWR, BWR, VVER-1000, CANDU) and fast reactors (BN-600). These issues concern core physics, fuel performance and reliability, and the capability and flexibility of thermal water reactors and fast reactors to dispose of weapons-grade plutonium in standard fuel cycles. The activities of the NEA Expert Group on Reactor-based Plutonium Disposition are carried out in close co-operation (jointly, in most cases) with the NEA Working Party on Scientific Issues in Reactor Systems (WPRS). A prominent part of these activities include benchmark studies. At the time of preparation of this report, the following benchmarks were completed or in progress: VENUS-2 MOX Core Benchmarks: carried out jointly with the WPRS (formerly the WPPR) (completed); VVER-1000 LEU and MOX Benchmark (completed); KRITZ-2 Benchmarks: carried out jointly with the WPRS (formerly the WPPR) (completed); Hollow and Solid MOX Fuel Behaviour Benchmark (completed); PRIMO MOX Fuel Performance Benchmark (ongoing); VENUS-2 MOX-fuelled Reactor Dosimetry Calculation (ongoing); VVER-1000 In-core Self-powered Neutron Detector Calculational Benchmark (started); MOX Fuel Rod Behaviour in Fast Power Pulse Conditions (started); Benchmark on the VENUS Plutonium Recycling Experiments Configuration 7 (started). This report describes the detailed results of the benchmark investigating the physics of a whole VVER-1000 reactor core using two-thirds low-enriched uranium (LEU) and one-third MOX fuel. It contributes to the computer code certification process and to the

  17. Benchmark af erhvervsuddannelserne

    DEFF Research Database (Denmark)

    Bogetoft, Peter; Wittrup, Jesper

    I dette arbejdspapir diskuterer vi, hvorledes de danske erhvervsskoler kan benchmarkes, og vi præsenterer resultaterne af en række beregningsmodeller. Det er begrebsmæssigt kompliceret at benchmarke erhvervsskolerne. Skolerne udbyder en lang række forskellige uddannelser. Det gør det vanskeligt at...

  18. Thermal Performance Benchmarking (Presentation)

    Energy Technology Data Exchange (ETDEWEB)

    Moreno, G.

    2014-11-01

    This project will benchmark the thermal characteristics of automotive power electronics and electric motor thermal management systems. Recent vehicle systems will be benchmarked to establish baseline metrics, evaluate advantages and disadvantages of different thermal management systems, and identify areas of improvement to advance the state-of-the-art.

  19. Internet based benchmarking

    DEFF Research Database (Denmark)

    Bogetoft, Peter; Nielsen, Kurt

    2005-01-01

    We discuss the design of interactive, internet based benchmarking using parametric (statistical) as well as nonparametric (DEA) models. The user receives benchmarks and improvement potentials. The user is also given the possibility to search different efficiency frontiers and hereby to explore...

  20. Numerical methods: Analytical benchmarking in transport theory

    International Nuclear Information System (INIS)

    Numerical methods applied to reactor technology have reached a high degree of maturity. Certainly one- and two-dimensional neutron transport calculations have become routine, with several programs available on personal computer and the most widely used programs adapted to workstation and minicomputer computational environments. With the introduction of massive parallelism and as experience with multitasking increases, even more improvement in the development of transport algorithms can be expected. Benchmarking an algorithm is usually not a very pleasant experience for the code developer. Proper algorithmic verification by benchmarking involves the following considerations: (1) conservation of particles, (2) confirmation of intuitive physical behavior, and (3) reproduction of analytical benchmark results. By using today's computational advantages, new basic numerical methods have been developed that allow a wider class of benchmark problems to be considered

  1. Co-ordinated research programme on benchmark study for the seismic analysis and testing of WWER-type nuclear power plants. V. 5B. Experience data. Working material

    International Nuclear Information System (INIS)

    The Co-ordinated research programme on the benchmark study for the seismic analysis and testing of WWER-type nuclear power plants was initiated subsequent to the request from representatives of Member States. The conclusions adopted at the Technical Committee Meeting on Seismic Issues related to existing nuclear power plants held in Tokyo in 1991 called for the harmonization of methods and criteria used in Member States in issues related to seismic safety. The Consulltants' Meeting which followed resulted in producing a working document for CRP. It was decided that a benchmark study is the most effective way to achieve the principal objective. Two types of WWER reactors (WWER-440/213 and WWER-1000) were selected as prototypes for the benchmark exercise to be tested on a full scale using explosions and/or vibration generators. The two prototypes are Kozloduy Units 5/6 for WWER-1000 and Paks for WWER-440/213 nuclear power plants. This volume of Working material contains reports on the effects of Armenia earthquakes on selected power, industry and commercial facilities and seismic functional qualification of active mechanical and electrical components tested on shaking table

  2. ICSBEP Benchmarks For Nuclear Data Applications

    Science.gov (United States)

    Briggs, J. Blair

    2005-05-01

    The International Criticality Safety Benchmark Evaluation Project (ICSBEP) was initiated in 1992 by the United States Department of Energy. The ICSBEP became an official activity of the Organization for Economic Cooperation and Development (OECD) — Nuclear Energy Agency (NEA) in 1995. Representatives from the United States, United Kingdom, France, Japan, the Russian Federation, Hungary, Republic of Korea, Slovenia, Serbia and Montenegro (formerly Yugoslavia), Kazakhstan, Spain, Israel, Brazil, Poland, and the Czech Republic are now participating. South Africa, India, China, and Germany are considering participation. The purpose of the ICSBEP is to identify, evaluate, verify, and formally document a comprehensive and internationally peer-reviewed set of criticality safety benchmark data. The work of the ICSBEP is published as an OECD handbook entitled "International Handbook of Evaluated Criticality Safety Benchmark Experiments." The 2004 Edition of the Handbook contains benchmark specifications for 3331 critical or subcritical configurations that are intended for use in validation efforts and for testing basic nuclear data. New to the 2004 Edition of the Handbook is a draft criticality alarm / shielding type benchmark that should be finalized in 2005 along with two other similar benchmarks. The Handbook is being used extensively for nuclear data testing and is expected to be a valuable resource for code and data validation and improvement efforts for decades to come. Specific benchmarks that are useful for testing structural materials such as iron, chromium, nickel, and manganese; beryllium; lead; thorium; and 238U are highlighted.

  3. A performance geodynamo benchmark

    Science.gov (United States)

    Matsui, H.; Heien, E. M.

    2014-12-01

    In the last ten years, a number of numerical dynamo models have successfully represented basic characteristics of the geomagnetic field. However, to approach the parameters regime of the Earth's outer core, we need massively parallel computational environment for extremely large spatial resolutions. Local methods are expected to be more suitable for massively parallel computation because the local methods needs less data communication than the spherical harmonics expansion, but only a few groups have reported results of the dynamo benchmark using local methods (Harder and Hansen, 2005; Matsui and Okuda, 2005; Chan et al., 2007) because of the difficulty treating magnetic boundary conditions based on the local methods. On the other hand, some numerical dynamo models using spherical harmonics expansion has performed successfully with thousands of processes. We perform benchmark tests to asses various numerical methods to asses the next generation of geodynamo simulations. The purpose of the present benchmark test is to assess numerical geodynamo models on a massively parallel computational platform. To compare among many numerical methods as possible, we consider the model with the insulated magnetic boundary by Christensen et al. (2001) and with the pseudo vacuum magnetic boundary, because the pseudo vacuum boundaries are implemented easier by using the local method than the magnetic insulated boundaries. In the present study, we consider two kinds of benchmarks, so-called accuracy benchmark and performance benchmark. In the present study, we will report the results of the performance benchmark. We perform the participated dynamo models under the same computational environment (XSEDE TACC Stampede), and investigate computational performance. To simplify the problem, we choose the same model and parameter regime as the accuracy benchmark test, but perform the simulations with much finer spatial resolutions as possible to investigate computational capability (e

  4. EPRI depletion benchmark calculations using PARAGON

    International Nuclear Information System (INIS)

    Highlights: • PARAGON depletion calculations are benchmarked against the EPRI reactivity decrement experiments. • Benchmarks cover a wide range of enrichments, burnups, cooling times, and burnable absorbers, and different depletion and storage conditions. • Results from PARAGON-SCALE scheme are more conservative relative to the benchmark data. • ENDF/B-VII based data reduces the excess conservatism and brings the predictions closer to benchmark reactivity decrement values. - Abstract: In order to conservatively apply burnup credit in spent fuel pool criticality analyses, code validation for both fresh and used fuel is required. Fresh fuel validation is typically done by modeling experiments from the “International Handbook.” A depletion validation can determine a bias and bias uncertainty for the worth of the isotopes not found in the fresh fuel critical experiments. Westinghouse’s burnup credit methodology uses PARAGON™ (Westinghouse 2-D lattice physics code) and its 70-group cross-section library, which have been benchmarked, qualified, and licensed both as a standalone transport code and as a nuclear data source for core design simulations. A bias and bias uncertainty for the worth of depletion isotopes, however, are not available for PARAGON. Instead, the 5% decrement approach for depletion uncertainty is used, as set forth in the Kopp memo. Recently, EPRI developed a set of benchmarks based on a large set of power distribution measurements to ascertain reactivity biases. The depletion reactivity has been used to create 11 benchmark cases for 10, 20, 30, 40, 50, and 60 GWd/MTU and 3 cooling times 100 h, 5 years, and 15 years. These benchmark cases are analyzed with PARAGON and the SCALE package and sensitivity studies are performed using different cross-section libraries based on ENDF/B-VI.3 and ENDF/B-VII data to assess that the 5% decrement approach is conservative for determining depletion uncertainty

  5. Benchmarking expert system tools

    Science.gov (United States)

    Riley, Gary

    1988-01-01

    As part of its evaluation of new technologies, the Artificial Intelligence Section of the Mission Planning and Analysis Div. at NASA-Johnson has made timing tests of several expert system building tools. Among the production systems tested were Automated Reasoning Tool, several versions of OPS5, and CLIPS (C Language Integrated Production System), an expert system builder developed by the AI section. Also included in the test were a Zetalisp version of the benchmark along with four versions of the benchmark written in Knowledge Engineering Environment, an object oriented, frame based expert system tool. The benchmarks used for testing are studied.

  6. Toxicological Benchmarks for Wildlife

    Energy Technology Data Exchange (ETDEWEB)

    Sample, B.E. Opresko, D.M. Suter, G.W.

    1993-01-01

    Ecological risks of environmental contaminants are evaluated by using a two-tiered process. In the first tier, a screening assessment is performed where concentrations of contaminants in the environment are compared to no observed adverse effects level (NOAEL)-based toxicological benchmarks. These benchmarks represent concentrations of chemicals (i.e., concentrations presumed to be nonhazardous to the biota) in environmental media (water, sediment, soil, food, etc.). While exceedance of these benchmarks does not indicate any particular level or type of risk, concentrations below the benchmarks should not result in significant effects. In practice, when contaminant concentrations in food or water resources are less than these toxicological benchmarks, the contaminants may be excluded from further consideration. However, if the concentration of a contaminant exceeds a benchmark, that contaminant should be retained as a contaminant of potential concern (COPC) and investigated further. The second tier in ecological risk assessment, the baseline ecological risk assessment, may use toxicological benchmarks as part of a weight-of-evidence approach (Suter 1993). Under this approach, based toxicological benchmarks are one of several lines of evidence used to support or refute the presence of ecological effects. Other sources of evidence include media toxicity tests, surveys of biota (abundance and diversity), measures of contaminant body burdens, and biomarkers. This report presents NOAEL- and lowest observed adverse effects level (LOAEL)-based toxicological benchmarks for assessment of effects of 85 chemicals on 9 representative mammalian wildlife species (short-tailed shrew, little brown bat, meadow vole, white-footed mouse, cottontail rabbit, mink, red fox, and whitetail deer) or 11 avian wildlife species (American robin, rough-winged swallow, American woodcock, wild turkey, belted kingfisher, great blue heron, barred owl, barn owl, Cooper's hawk, and red

  7. Cross Section Evaluation Group shielding benchmark compilation. Volume II

    Energy Technology Data Exchange (ETDEWEB)

    Rose, P.F.; Roussin, R.W.

    1983-12-01

    At the time of the release of ENDF/B-IV in 1974, the Shielding Subcommittee had identified a series of 12 shielding data testing benchmarks (the SDT series). Most were used in the ENDF/B-IV data testing effort. A new concept and series was begun in the interim, the so-called Shielding Benchmark (SB) series. An effort was made to upgrade the SDT series as far as possible and to add new SB benchmarks. In order to be designated in the SB class, both an experiment and analysis must have been performed. The current recommended benchmark for Shielding Data Testing are listed. Until recently, the philosophy has been to include only citations to published references for shielding benchmarks. It is now our intention to provide adequate information in this volume for proper analysis of any new benchmarks added to the collection. These compilations appear in Section II, with the SB5 Fusion Reactor Shielding Benchmark as the first entry.

  8. Reactor fuel depletion benchmark of TINDER

    International Nuclear Information System (INIS)

    Highlights: • A reactor burnup benchmark of TINDER, coupling MCNP6 to CINDER2008, was performed. • TINDER is a poor candidate for fuel depletion calculations using its current libraries. • Data library modification is necessary if fuel depletion is desired from TINDER. - Abstract: Accurate burnup calculations are key to proper nuclear reactor design, fuel cycle modeling, and disposal estimations. The TINDER code, originally designed for activation analyses, has been modified to handle full burnup calculations, including the widely used predictor–corrector feature. In order to properly characterize the performance of TINDER for this application, a benchmark calculation was performed. Although the results followed the trends of past benchmarked codes for a UO2 PWR fuel sample from the Takahama-3 reactor, there were obvious deficiencies in the final result, likely in the nuclear data library that was used. Isotopic comparisons versus experiment and past code benchmarks are given, as well as hypothesized areas of deficiency and future work

  9. Validation of the BUGJEFF311.BOLIB, BUGENDF70.BOLIB and BUGLE-B7 broad-group libraries on the PCA-Replica (H2O/Fe) neutron shielding benchmark experiment

    OpenAIRE

    Pescarini Massimo; Orsi Roberto; Frisoni Manuela

    2016-01-01

    The PCA-Replica 12/13 (H2O/Fe) neutron shielding benchmark experiment was analysed using the TORT-3.2 3D SN code. PCA-Replica reproduces a PWR ex-core radial geometry with alternate layers of water and steel including a pressure vessel simulator. Three broad-group coupled neutron/photon working cross section libraries in FIDO-ANISN format with the same energy group structure (47 n + 20 γ) and based on different nuclear data were alternatively used: the ENEA BUGJEFF311.BOLIB (JEFF-3.1.1) and U...

  10. Three-Dimensional (X,Y,Z) Deterministic Analysis of the PCA-Replica Neutron Shielding Benchmark Experiment using the TORT-3.2 Code and Group Cross Section Libraries for LWR Shielding and Pressure Vessel Dosimetry

    OpenAIRE

    Pescarini Massimo; Orsi Roberto; Frisoni Manuela

    2016-01-01

    The PCA-Replica 12/13 (H2O/Fe) neutron shielding benchmark experiment was analysed using the ORNL TORT-3.2 3D SN code. PCA-Replica, specifically conceived to test the accuracy of nuclear data and transport codes employed in LWR shielding and radiation damage calculations, reproduces a PWR ex-core radial geometry with alternate layers of water and steel including a PWR pressure vessel simulator. Three broad-group coupled neutron/photon working cross section libraries in FIDO-ANISN format with ...

  11. Diagnostic Algorithm Benchmarking

    Science.gov (United States)

    Poll, Scott

    2011-01-01

    A poster for the NASA Aviation Safety Program Annual Technical Meeting. It describes empirical benchmarking on diagnostic algorithms using data from the ADAPT Electrical Power System testbed and a diagnostic software framework.

  12. GeodeticBenchmark_GEOMON

    Data.gov (United States)

    Vermont Center for Geographic Information — The GeodeticBenchmark_GEOMON data layer consists of geodetic control monuments (points) that have a known position or spatial reference. The locations of these...

  13. Benchmarking in University Toolbox

    OpenAIRE

    Katarzyna Kuźmicz

    2015-01-01

    In the face of global competition and rising challenges that higher education institutions (HEIs) meet, it is imperative to increase innovativeness and efficiency of their management. Benchmarking can be the appropriate tool to search for a point of reference necessary to assess institution’s competitive position and learn from the best in order to improve. The primary purpose of the paper is to present in-depth analysis of benchmarking application in HEIs worldwide. The study involves indica...

  14. Accelerator shielding benchmark problems

    International Nuclear Information System (INIS)

    Accelerator shielding benchmark problems prepared by Working Group of Accelerator Shielding in the Research Committee on Radiation Behavior in the Atomic Energy Society of Japan were compiled by Radiation Safety Control Center of National Laboratory for High Energy Physics. Twenty-five accelerator shielding benchmark problems are presented for evaluating the calculational algorithm, the accuracy of computer codes and the nuclear data used in codes. (author)

  15. Benchmarking conflict resolution algorithms

    OpenAIRE

    Vanaret, Charlie; Gianazza, David; Durand, Nicolas; Gotteland, Jean-Baptiste

    2012-01-01

    Applying a benchmarking approach to conflict resolution problems is a hard task, as the analytical form of the constraints is not simple. This is especially the case when using realistic dynamics and models, considering accelerating aircraft that may follow flight paths that are not direct. Currently, there is a lack of common problems and data that would allow researchers to compare the performances of several conflict resolution algorithms. The present paper introduces a benchmarking approa...

  16. Benchmarking and regulation

    OpenAIRE

    Agrell, Per Joakim; Bogetoft, Peter

    2013-01-01

    Benchmarking methods, and in particular Data Envelopment Analysis (DEA), have become well-established and informative tools for economic regulation. DEA is now routinely used by European regulators to set reasonable revenue caps for energy transmission and distribution system operators. The application of benchmarking in regulation, however, requires specific steps in terms of data validation, model specification and outlier detection that are not systematically documented in open publication...

  17. Accelerator shielding benchmark problems

    Energy Technology Data Exchange (ETDEWEB)

    Hirayama, H.; Ban, S.; Nakamura, T. [and others

    1993-01-01

    Accelerator shielding benchmark problems prepared by Working Group of Accelerator Shielding in the Research Committee on Radiation Behavior in the Atomic Energy Society of Japan were compiled by Radiation Safety Control Center of National Laboratory for High Energy Physics. Twenty-five accelerator shielding benchmark problems are presented for evaluating the calculational algorithm, the accuracy of computer codes and the nuclear data used in codes. (author).

  18. The MCNP6 Analytic Criticality Benchmark Suite

    Energy Technology Data Exchange (ETDEWEB)

    Brown, Forrest B. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States). Monte Carlo Codes Group

    2016-06-16

    Analytical benchmarks provide an invaluable tool for verifying computer codes used to simulate neutron transport. Several collections of analytical benchmark problems [1-4] are used routinely in the verification of production Monte Carlo codes such as MCNP® [5,6]. Verification of a computer code is a necessary prerequisite to the more complex validation process. The verification process confirms that a code performs its intended functions correctly. The validation process involves determining the absolute accuracy of code results vs. nature. In typical validations, results are computed for a set of benchmark experiments using a particular methodology (code, cross-section data with uncertainties, and modeling) and compared to the measured results from the set of benchmark experiments. The validation process determines bias, bias uncertainty, and possibly additional margins. Verification is generally performed by the code developers, while validation is generally performed by code users for a particular application space. The VERIFICATION_KEFF suite of criticality problems [1,2] was originally a set of 75 criticality problems found in the literature for which exact analytical solutions are available. Even though the spatial and energy detail is necessarily limited in analytical benchmarks, typically to a few regions or energy groups, the exact solutions obtained can be used to verify that the basic algorithms, mathematics, and methods used in complex production codes perform correctly. The present work has focused on revisiting this benchmark suite. A thorough review of the problems resulted in discarding some of them as not suitable for MCNP benchmarking. For the remaining problems, many of them were reformulated to permit execution in either multigroup mode or in the normal continuous-energy mode for MCNP. Execution of the benchmarks in continuous-energy mode provides a significant advance to MCNP verification methods.

  19. Thermal-hydraulic–iodine chemistry coupling: Insights gained from the SARNET benchmark on the THAI experiments Iod-11 and Iod-12

    International Nuclear Information System (INIS)

    Highlights: • The I2 transport in two multi-compartment THAI tests was analyzed. • In a benchmark 4 different codes were applied by 7 organizations. • The I2 concentrations were mostly overestimated, up to a factor 100. • Inadequate iodine models and inaccurate thermal-hydraulic parameters were detected. • The user effect on the quality of the iodine results was large. - Abstract: In the SARNET2 WP8.3 THAI Benchmark the capability of current accident codes to simulate the iodine transport and behavior in sub-divided containments has been assessed. In THAI test Iod-11 and Iod-12, made available for the benchmark, the distribution of molecular iodine (I2) in the five compartments of the 60 m3 vessel under stratified and well mixed conditions was measured. The main processes addressed are the I2 transport with the atmospheric flows and the interaction of I2 with the steel surface. During test Iod-11 the surfaces in contact with the containment atmosphere were dry. In Iod-12, steam was released, which condensed on the walls. Nine post-test calculations were conducted for Iod-11 and eight for Iod-12 by seven organizations using four different codes: ASTEC-IODE (CIEMAT, GRS and TUS), COCOSYS-AIM (AREVA, FZ-Jülich and GRS), ECART (Pisa University) and MELCOR (Pisa University and VTT). Different nodalizations of the THAI vessel with 20–65 zones were applied. Generally, for both tests the analytical thermal-hydraulic results are in a fairly good agreement with the measurements. Only the calculated local relative humidity deviates significantly from the measured values in all calculations. The results in Iod-11 for the local I2 concentration in the gaseous phase are quite diverse. Three calculations show only minor deviations from the measurement, whereas the others are substantially different from the measured I2 concentrations. For Iod-12, no calculation delivers a satisfactory evolution of the I2 concentration in all five compartments of the vessel. There are

  20. The KMAT: Benchmarking Knowledge Management.

    Science.gov (United States)

    de Jager, Martha

    Provides an overview of knowledge management and benchmarking, including the benefits and methods of benchmarking (e.g., competitive, cooperative, collaborative, and internal benchmarking). Arthur Andersen's KMAT (Knowledge Management Assessment Tool) is described. The KMAT is a collaborative benchmarking tool, designed to help organizations make…

  1. Shielding Integral Benchmark Archive and Database (SINBAD)

    Energy Technology Data Exchange (ETDEWEB)

    Kirk, Bernadette Lugue [ORNL; Grove, Robert E [ORNL; Kodeli, I. [International Atomic Energy Agency (IAEA); Sartori, Enrico [ORNL; Gulliford, J. [OECD Nuclear Energy Agency

    2011-01-01

    The Shielding Integral Benchmark Archive and Database (SINBAD) collection of benchmarks was initiated in the early 1990 s. SINBAD is an international collaboration between the Organization for Economic Cooperation and Development s Nuclear Energy Agency Data Bank (OECD/NEADB) and the Radiation Safety Information Computational Center (RSICC) at Oak Ridge National Laboratory (ORNL). SINBAD is a major attempt to compile experiments and corresponding computational models with the goal of preserving institutional knowledge and expertise that need to be handed down to future scientists. SINBAD is also a learning tool for university students and scientists who need to design experiments or gain expertise in modeling and simulation. The SINBAD database is currently divided into three categories fission, fusion, and accelerator benchmarks. Where possible, each experiment is described and analyzed using deterministic or probabilistic (Monte Carlo) radiation transport software.

  2. Shielding Integral Benchmark Archive and Database (SINBAD)

    International Nuclear Information System (INIS)

    The Shielding Integral Benchmark Archive and Database (SINBAD) collection of benchmarks was initiated in the early 1990s. SINBAD is an international collaboration between the Organization for Economic Cooperation and Development's Nuclear Energy Agency Data Bank (OECD/NEADB) and the Radiation Safety Information Computational Center (RSICC) at Oak Ridge National Laboratory (ORNL). SINBAD is a major attempt to compile experiments and corresponding computational models with the goal of preserving institutional knowledge and expertise that need to be handed down to future scientists. SINBAD is also a learning tool for university students and scientists who need to design experiments or gain expertise in modeling and simulation. The SINBAD database is currently divided into three categories fission, fusion, and accelerator benchmarks. Where possible, each experiment is described and analyzed using deterministic or probabilistic (Monte Carlo) radiation transport software.

  3. Generation of a WIMS-D/4 multigroup constants library based on the JENDL-3.2 nuclear data and its validation through some benchmark experiments analysis

    Energy Technology Data Exchange (ETDEWEB)

    Rahman, M. [Institute of Nuclear Science and Technology, Savar, Dacca (Bangladesh); Takano, Hideki

    1996-11-01

    A new 69 group library of multigroup constants for the lattice code WIMS-D/4 has been generated with an improved resonance treatment, processing nuclear data from JENDL-3.2 by NJOY91.108. A parallel ENDF/B-VI based library has also been constructed for intercomparison of results. Benchmark calculations for a number of thermal reactor critical assemblies of both uranium and plutonium fuels have been performed with the code WIMS-D/4.1 with its three different libraries: the original WIMS library (NEA-0329/10) and the new ENDF/B-VI and JENDL-3.2 based libraries. The results calculated with both ENDF and JENDL based libraries show similar tendency and are found in better agreement with the experimental values. Benchmark parameters are further calculated with the comprehensive lattice code SRAC95. The results from SRAC95 and WIMS-D/4.1 (both with JENDL-3.2 based libraries) agrees well to each other as well as to the other previously published values. (author)

  4. Benchmarking in Mobarakeh Steel Company

    OpenAIRE

    Sasan Ghasemi; Mohammad Nazemi; Mehran Nejati

    2008-01-01

    Benchmarking is considered as one of the most effective ways of improving performance in companies. Although benchmarking in business organizations is a relatively new concept and practice, it has rapidly gained acceptance worldwide. This paper introduces the benchmarking project conducted in Esfahan's Mobarakeh Steel Company, as the first systematic benchmarking project conducted in Iran. It aims to share the process deployed for the benchmarking project in this company and illustrate how th...

  5. Benchmarking the Netherlands. Benchmarking for growth

    International Nuclear Information System (INIS)

    This is the fourth edition of the Ministry of Economic Affairs' publication 'Benchmarking the Netherlands', which aims to assess the competitiveness of the Dutch economy. The methodology and objective of the benchmarking remain the same. The basic conditions for economic activity (institutions, regulation, etc.) in a number of benchmark countries are compared in order to learn from the solutions found by other countries for common economic problems. This publication is devoted entirely to the potential output of the Dutch economy. In other words, its ability to achieve sustainable growth and create work over a longer period without capacity becoming an obstacle. This is important because economic growth is needed to increase prosperity in the broad sense and meeting social needs. Prosperity in both a material (per capita GDP) and immaterial (living environment, environment, health, etc) sense, in other words. The economy's potential output is determined by two structural factors: the growth of potential employment and the structural increase in labour productivity. Analysis by the Netherlands Bureau for Economic Policy Analysis (CPB) shows that in recent years the increase in the capacity for economic growth has been realised mainly by increasing the supply of labour and reducing the equilibrium unemployment rate. In view of the ageing of the population in the coming years and decades the supply of labour is unlikely to continue growing at the pace we have become accustomed to in recent years. According to a number of recent studies, to achieve a respectable rate of sustainable economic growth the aim will therefore have to be to increase labour productivity. To realise this we have to focus on for six pillars of economic policy: (1) human capital, (2) functioning of markets, (3) entrepreneurship, (4) spatial planning, (5) innovation, and (6) sustainability. These six pillars determine the course for economic policy aiming at higher productivity growth. Throughout

  6. NFS Tricks and Benchmarking Traps

    OpenAIRE

    Seltzer, Margo; Ellard, Daniel

    2003-01-01

    We describe two modi cations to the FreeBSD 4.6 NFS server to increase read throughput by improving the read-ahead heuristic to deal with reordered requests and stride access patterns. We show that for some stride access patterns, our new heuristics improve end-to-end NFS throughput by nearly a factor of two. We also show that benchmarking and experimenting with changes to an NFS server can be a subtle and challenging task, and that it is often difficult to distinguish the impact of a new ...

  7. Experiences with NJOY and ENDF pre-processing system on PC-486 and benchmarks of WIMS library generated from ENDF/B-VI

    International Nuclear Information System (INIS)

    The objective of the present study is to develop WIMS-CITATION computation system using PC-486 for nuclear analysis of advanced Pressurized Water Reactors, whose burnups are designed to achieve more than 60,000 MWD/MTU, and CANDU reactors. In order to achieve the objective, it is essential to generate the proper WIMS library. The possibility of employing NJOY code and ENDF Pre-Processing System are under investigation. In order to verify the acceptability of the WIMS library generated from ENDF/B-VI on PC-486, new libraries are generated and compared with the libraries generated using CYBER and SUN Work Station for testing benchmark results of five light water and five heavy water lattices. The result shows that PC-486-based WIMS library is acceptable even though there exist some minor problems. (author). 11 refs, 16 figs, 9 tabs

  8. Deviating From the Benchmarks

    DEFF Research Database (Denmark)

    Rocha, Vera; Van Praag, Mirjam; Carneiro, Anabela

    This paper studies three related questions: To what extent otherwise similar startups employ different quantities and qualities of human capital at the moment of entry? How persistent are initial human capital choices over time? And how does deviating from human capital benchmarks influence firm...... survival? The analysis is based on a matched employer-employee dataset and covers about 17,500 startups in manufacturing and services. We adopt a new procedure to estimate individual benchmarks for the quantity and quality of initial human resources, acknowledging correlations between hiring decisions......, founders human capital, and the ownership structure of startups (solo entrepreneurs versus entrepreneurial teams). We then study the survival implications of exogenous deviations from these benchmarks, based on spline models for survival data. Our results indicate that (especially negative) deviations from...

  9. Benchmarking for Best Practice

    CERN Document Server

    Zairi, Mohamed

    1998-01-01

    Benchmarking for Best Practice uses up-to-the-minute case-studies of individual companies and industry-wide quality schemes to show how and why implementation has succeeded. For any practitioner wanting to establish best practice in a wide variety of business areas, this book makes essential reading. .It is also an ideal textbook on the applications of TQM since it describes concepts, covers definitions and illustrates the applications with first-hand examples. Professor Mohamed Zairi is an international expert and leading figure in the field of benchmarking. His pioneering work in this area l

  10. Remote Sensing Segmentation Benchmark

    Czech Academy of Sciences Publication Activity Database

    Mikeš, Stanislav; Haindl, Michal; Scarpa, G.

    Piscataway, NJ : IEEE Press, 2012, s. 1-4. ISBN 978-1-4673-4960-4. [IAPR Workshop on Pattern Recognition in Remote Sensing (PRRS). Tsukuba Science City (JP), 11.11.2012] R&D Projects: GA ČR GAP103/11/0335; GA ČR GA102/08/0593 Grant ostatní: CESNET(CZ) 409/2011 Keywords : remote sensing * segmentation * benchmark Subject RIV: BD - Theory of Information http://library.utia.cas.cz/separaty/2013/RO/mikes-remote sensing segmentation benchmark.pdf

  11. Features and technology of enterprise internal benchmarking

    Directory of Open Access Journals (Sweden)

    A.V. Dubodelova

    2013-06-01

    Full Text Available The aim of the article. The aim of the article is to generalize characteristics, objectives, advantages of internal benchmarking. The stages sequence of internal benchmarking technology is formed. It is focused on continuous improvement of process of the enterprise by implementing existing best practices.The results of the analysis. Business activity of domestic enterprises in crisis business environment has to focus on the best success factors of their structural units by using standard research assessment of their performance and their innovative experience in practice. Modern method of those needs satisfying is internal benchmarking. According to Bain & Co internal benchmarking is one the three most common methods of business management.The features and benefits of benchmarking are defined in the article. The sequence and methodology of implementation of individual stages of benchmarking technology projects are formulated.The authors define benchmarking as a strategic orientation on the best achievement by comparing performance and working methods with the standard. It covers the processes of researching, organization of production and distribution, management and marketing methods to reference objects to identify innovative practices and its implementation in a particular business.Benchmarking development at domestic enterprises requires analysis of theoretical bases and practical experience. Choice best of experience helps to develop recommendations for their application in practice.Also it is essential to classificate species, identify characteristics, study appropriate areas of use and development methodology of implementation. The structure of internal benchmarking objectives includes: promoting research and establishment of minimum acceptable levels of efficiency processes and activities which are available at the enterprise; identification of current problems and areas that need improvement without involvement of foreign experience

  12. Benchmarking the World's Best

    Science.gov (United States)

    Tucker, Marc S.

    2012-01-01

    A century ago, the United States was a world leader in industrial benchmarking. However, after World War II, once no one could compete with the U.S., it became complacent. Many industrialized countries now have higher student achievement and more equitable and efficient education systems. A higher proportion of young people in their workforces…

  13. Benchmarks: WICHE Region 2012

    Science.gov (United States)

    Western Interstate Commission for Higher Education, 2013

    2013-01-01

    Benchmarks: WICHE Region 2012 presents information on the West's progress in improving access to, success in, and financing of higher education. The information is updated annually to monitor change over time and encourage its use as a tool for informed discussion in policy and education communities. To establish a general context for the…

  14. Benchmark problem proposal

    International Nuclear Information System (INIS)

    The meeting of the Radiation Energy Spectra Unfolding Workshop organized by the Radiation Shielding Information Center is discussed. The plans of the unfolding code benchmarking effort to establish methods of standardization for both the few channel neutron and many channel gamma-ray and neutron spectroscopy problems are presented

  15. Benchmarking and Performance Management

    Directory of Open Access Journals (Sweden)

    Adrian TANTAU

    2010-12-01

    Full Text Available The relevance of the chosen topic is explained by the meaning of the firm efficiency concept - the firm efficiency means the revealed performance (how well the firm performs in the actual market environment given the basic characteristics of the firms and their markets that are expected to drive their profitability (firm size, market power etc.. This complex and relative performance could be due to such things as product innovation, management quality, work organization, some other factors can be a cause even if they are not directly observed by the researcher. The critical need for the management individuals/group to continuously improve their firm/company’s efficiency and effectiveness, the need for the managers to know which are the success factors and the competitiveness determinants determine consequently, what performance measures are most critical in determining their firm’s overall success. Benchmarking, when done properly, can accurately identify both successful companies and the underlying reasons for their success. Innovation and benchmarking firm level performance are critical interdependent activities. Firm level variables, used to infer performance, are often interdependent due to operational reasons. Hence, the managers need to take the dependencies among these variables into account when forecasting and benchmarking performance. This paper studies firm level performance using financial ratio and other type of profitability measures. It uses econometric models to describe and then propose a method to forecast and benchmark performance.

  16. CCF benchmark test

    International Nuclear Information System (INIS)

    A benchmark test on common cause failures (CCF) was performed giving interested institutions in Germany the opportunity of demonstrating and justifying their interpretations of events, their methods and models for analyzed CCF. The participants of this benchmark test belonged to expert and consultant organisations and to industrial institutions. The task for the benchmark test was to analyze two typical groups of motor-operated valves in German nuclear power plants. The benchmark test was carried out in two steps. In the first step the participants were to assess in a qualitative way some 200 event-reports on isolation valves. They then were to establish, quantitatively, the reliability parameters for the CCF in the two groups of motor-operated valves using their own methods and their own calculation models. In a second step the reliability parameters were to be recalculated on the basis of a common reference of well defined events, chosen from all given events, in order to analyze the influence of the calculation models on the reliability parameters. (orig.)

  17. Benchmarking Public Procurement 2016

    OpenAIRE

    World Bank Group

    2015-01-01

    Benchmarking Public Procurement 2016 Report aims to develop actionable indicators which will help countries identify and monitor policies and regulations that impact how private sector companies do business with the government. The project builds on the Doing Business methodology and was initiated at the request of the G20 Anti-Corruption Working Group.

  18. NAS Parallel Benchmarks Results

    Science.gov (United States)

    Subhash, Saini; Bailey, David H.; Lasinski, T. A. (Technical Monitor)

    1995-01-01

    The NAS Parallel Benchmarks (NPB) were developed in 1991 at NASA Ames Research Center to study the performance of parallel supercomputers. The eight benchmark problems are specified in a pencil and paper fashion i.e. the complete details of the problem to be solved are given in a technical document, and except for a few restrictions, benchmarkers are free to select the language constructs and implementation techniques best suited for a particular system. In this paper, we present new NPB performance results for the following systems: (a) Parallel-Vector Processors: Cray C90, Cray T'90 and Fujitsu VPP500; (b) Highly Parallel Processors: Cray T3D, IBM SP2 and IBM SP-TN2 (Thin Nodes 2); (c) Symmetric Multiprocessing Processors: Convex Exemplar SPP1000, Cray J90, DEC Alpha Server 8400 5/300, and SGI Power Challenge XL. We also present sustained performance per dollar for Class B LU, SP and BT benchmarks. We also mention NAS future plans of NPB.

  19. The Snowmass points and slopes: Benchmarks for SUSY searches

    International Nuclear Information System (INIS)

    The ''Snowmass Points and Slopes'' (SPS) are a set of benchmark points and parameter lines in the MSSM parameter space corresponding to different scenarios in the search for Supersymmetry at present and future experiments. This set of benchmarks was agreed upon at the 2001 ''Snowmass Workshop on the Future of Particle Physics'' as a consensus based on different existing proposals

  20. The Snowmass points and slopes: benchmarks for SUSY searches

    International Nuclear Information System (INIS)

    The ''snowmass points and slopes'' (SPS) are a set of benchmark points and parameter lines in the MSSM parameter space corresponding to different scenarios in the search for Supersymmetry at present and future experiments. This set of benchmarks was agreed upon at the 2001 ''Snowmass Workshop on the Future of Particle Physics'' as a consensus based on different existing proposals. (orig.)

  1. The Snowmass points and slopes : benchmarks for SUSY searches

    International Nuclear Information System (INIS)

    The ''Snowmass Points and Slopes'' (SPS) are a set of benchmark points and parameter lines in the MSSM parameter space corresponding to different scenarios in the search for Supersymmetry at present and future experiments. This set of benchmarks was agreed upon at the 2001 ''Snowmass Workshop on the Future of Particle Physics'' as a consensus based on different existing proposals

  2. The Snowmass Points and Slopes: benchmarks for SUSY searches

    International Nuclear Information System (INIS)

    The ''Snowmass Points and Slopes'' (SPS) are a set of benchmark points and parameter lines in the MSSM parameter space corresponding to different scenarios in the search for Supersymmetry at present and future experiments. This set of benchmarks was agreed upon at the 2001 ''Snowmass Workshop on the Future of Particle Physics'' as a consensus based on different existing proposals. (orig.)

  3. The Snowmass Points and Slopes: Benchmarks for SUSY Studies

    International Nuclear Information System (INIS)

    The ''Snowmass Points and Slopes'' (SPS) are a set of benchmark points and parameter lines in the MSSM parameter space corresponding to different scenarios in the search for Supersymmetry at present and future experiments. This set of benchmarks was agreed upon at the 2001 ''Snowmass Workshop on the Future of Particle Physics'' as a consensus based on different existing proposals

  4. Storage-Intensive Supercomputing Benchmark Study

    Energy Technology Data Exchange (ETDEWEB)

    Cohen, J; Dossa, D; Gokhale, M; Hysom, D; May, J; Pearce, R; Yoo, A

    2007-10-30

    : SuperMicro X7DBE Xeon Dual Socket Blackford Server Motherboard; 2 Intel Xeon Dual-Core 2.66 GHz processors; 1 GB DDR2 PC2-5300 RAM (2 x 512); 80GB Hard Drive (Seagate SATA II Barracuda). The Fusion board is presently capable of 4X in a PCIe slot. The image resampling benchmark was run on a dual Xeon workstation with NVIDIA graphics card (see Chapter 5 for full specification). An XtremeData Opteron+FPGA was used for the language classification application. We observed that these benchmarks are not uniformly I/O intensive. The only benchmark that showed greater that 50% of the time in I/O was the graph algorithm when it accessed data files over NFS. When local disk was used, the graph benchmark spent at most 40% of its time in I/O. The other benchmarks were CPU dominated. The image resampling benchmark and language classification showed order of magnitude speedup over software by using co-processor technology to offload the CPU-intensive kernels. Our experiments to date suggest that emerging hardware technologies offer significant benefit to boosting the performance of data-intensive algorithms. Using GPU and FPGA co-processors, we were able to improve performance by more than an order of magnitude on the benchmark algorithms, eliminating the processor bottleneck of CPU-bound tasks. Experiments with a prototype solid state nonvolative memory available today show 10X better throughput on random reads than disk, with a 2X speedup on a graph processing benchmark when compared to the use of local SATA disk.

  5. Reactor group constants and benchmark test

    International Nuclear Information System (INIS)

    The evaluated nuclear data files such as JENDL, ENDF/B-VI and JEF-2 are validated by analyzing critical mock-up experiments for various type reactors and assessing applicability for nuclear characteristics such as criticality, reaction rates, reactivities, etc. This is called Benchmark Testing. In the nuclear calculations, the diffusion and transport codes use the group constant library which is generated by processing the nuclear data files. In this paper, the calculation methods of the reactor group constants and benchmark test are described. Finally, a new group constants scheme is proposed. (author)

  6. Shielding integral benchmark archive and database

    International Nuclear Information System (INIS)

    SINBAD (Shielding integral benchmark archive and database) is a new electronic database developed to store a variety of radiation shielding benchmark data so that users can easily and incorporate the data into their calculations. SINBAD is an excellent data source for users who require the quality assurance necessary in developing cross-section libraries or radiation transport codes. The future needs of the scientific community are best served by the electronic database format of SINBAD and its user-friendly interface, combined with its data accuracy and integrity. It has been designed to be able to include data from nuclear reactor shielding, fusion blankets and accelerator shielding experiments. (authors)

  7. Benchmark testing of 233U evaluations

    International Nuclear Information System (INIS)

    In this paper we investigate the adequacy of available 233U cross-section data (ENDF/B-VI and JENDL-3) for calculation of critical experiments. An ad hoc revised 233U evaluation is also tested and appears to give results which are improved relative to those obtained with either ENDF/B-VI or JENDL-3 cross sections. Calculations of keff were performed for ten fast benchmarks and six thermal benchmarks using the three cross-section sets. Central reaction-rate-ratio calculations were also performed

  8. Benchmarking i den offentlige sektor

    DEFF Research Database (Denmark)

    Bukh, Per Nikolaj; Dietrichson, Lars; Sandalgaard, Niels

    2008-01-01

    I artiklen vil vi kort diskutere behovet for benchmarking i fraværet af traditionelle markedsmekanismer. Herefter vil vi nærmere redegøre for, hvad benchmarking er med udgangspunkt i fire forskellige anvendelser af benchmarking. Regulering af forsyningsvirksomheder vil blive behandlet, hvorefter...

  9. Benchmark Evaluation of HTR-PROTEUS Pebble Bed Experimental Program

    International Nuclear Information System (INIS)

    Benchmark models were developed to evaluate 11 critical core configurations of the HTR-PROTEUS pebble bed experimental program. Various additional reactor physics measurements were performed as part of this program; currently only a total of 37 absorber rod worth measurements have been evaluated as acceptable benchmark experiments for Cores 4, 9, and 10. Dominant uncertainties in the experimental keff for all core configurations come from uncertainties in the 235U enrichment of the fuel, impurities in the moderator pebbles, and the density and impurity content of the radial reflector. Calculations of keff with MCNP5 and ENDF/B-VII.0 neutron nuclear data are greater than the benchmark values but within 1% and also within the 3σ uncertainty, except for Core 4, which is the only randomly packed pebble configuration. Repeated calculations of keff with MCNP6.1 and ENDF/B-VII.1 are lower than the benchmark values and within 1% (~3σ) except for Cores 5 and 9, which calculate lower than the benchmark eigenvalues within 4σ. The primary difference between the two nuclear data libraries is the adjustment of the absorption cross section of graphite. Simulations of the absorber rod worth measurements are within 3σ of the benchmark experiment values. The complete benchmark evaluation details are available in the 2014 edition of the International Handbook of Evaluated Reactor Physics Benchmark Experiments

  10. Texture Segmentation Benchmark

    Czech Academy of Sciences Publication Activity Database

    Haindl, Michal; Mikeš, Stanislav

    Los Alamitos : IEEE Press, 2008, s. 2933-2936. ISBN 978-1-4244-2174-9. [19th International Conference on Pattern Recognition. Tampa (US), 07.12.2008-11.12.2008] R&D Projects: GA AV ČR 1ET400750407; GA MŠk 1M0572; GA ČR GA102/07/1594; GA ČR GA102/08/0593 Grant ostatní: GA MŠk(CZ) 2C06019 Institutional research plan: CEZ:AV0Z10750506 Keywords : texture segmentation * image segmentation * benchmark Subject RIV: BD - Theory of Information http://library.utia.cas.cz/separaty/2008/RO/haindl-texture segmentation benchmark.pdf

  11. Radiography benchmark 2014

    International Nuclear Information System (INIS)

    The purpose of the 2014 WFNDEC RT benchmark study was to compare predictions of various models of radiographic techniques, in particular those that predict the contribution of scattered radiation. All calculations were carried out for homogenous materials and a mono-energetic X-ray point source in the energy range between 100 keV and 10 MeV. The calculations were to include the best physics approach available considering electron binding effects. Secondary effects like X-ray fluorescence and bremsstrahlung production were to be taken into account if possible. The problem to be considered had two parts. Part I examined the spectrum and the spatial distribution of radiation behind a single iron plate. Part II considered two equally sized plates, made of iron and aluminum respectively, only evaluating the spatial distribution. Here we present the results of above benchmark study, comparing them to MCNP as the assumed reference model. The possible origins of the observed deviations are discussed

  12. Benchmarking of LSTM Networks

    OpenAIRE

    Breuel, Thomas M.

    2015-01-01

    LSTM (Long Short-Term Memory) recurrent neural networks have been highly successful in a number of application areas. This technical report describes the use of the MNIST and UW3 databases for benchmarking LSTM networks and explores the effect of di?erent architectural and hyperparameter choices on performance. Significant ?ndings include: (1) LSTM performance depends smoothly on learning rates, (2) batching and momentum has no significant effect on performance, (3) softmax training outperfor...

  13. Texture Fidelity Benchmark

    Czech Academy of Sciences Publication Activity Database

    Haindl, Michal; Kudělka, Miloš

    Los Alamitos, USA: IEEE Computer Society CPS, 2014. ISBN 978-1-4799-7971-4. [International Workshop on Computational Intelligence for Multimedia Understanding 2014 (IWCIM). Paris (FR), 01.11.2014-02.11.2014] R&D Projects: GA ČR(CZ) GA14-10911S Institutional support: RVO:67985556 Keywords : Benchmark testing * fidelity criteria * texture Subject RIV: BD - Theory of Information http://library.utia.cas.cz/separaty/2014/RO/haindl-0439654.pdf

  14. Cloud benchmarking for performance

    OpenAIRE

    Varghese, Blesson; Akgun, Ozgur; Miguel, Ian; Thai, Long; Barker, Adam

    2014-01-01

    How can applications be deployed on the cloud to achieve maximum performance? This question has become significant and challenging with the availability of a wide variety of Virtual Machines (VMs) with different performance capabilities in the cloud. The above question is addressed by proposing a six step benchmarking methodology in which a user provides a set of four weights that indicate how important each of the following groups: memory, processor, computation and storage are to the applic...

  15. The NAS Parallel Benchmarks

    Energy Technology Data Exchange (ETDEWEB)

    Bailey, David H.

    2009-11-15

    The NAS Parallel Benchmarks (NPB) are a suite of parallel computer performance benchmarks. They were originally developed at the NASA Ames Research Center in 1991 to assess high-end parallel supercomputers. Although they are no longer used as widely as they once were for comparing high-end system performance, they continue to be studied and analyzed a great deal in the high-performance computing community. The acronym 'NAS' originally stood for the Numerical Aeronautical Simulation Program at NASA Ames. The name of this organization was subsequently changed to the Numerical Aerospace Simulation Program, and more recently to the NASA Advanced Supercomputing Center, although the acronym remains 'NAS.' The developers of the original NPB suite were David H. Bailey, Eric Barszcz, John Barton, David Browning, Russell Carter, LeoDagum, Rod Fatoohi, Samuel Fineberg, Paul Frederickson, Thomas Lasinski, Rob Schreiber, Horst Simon, V. Venkatakrishnan and Sisira Weeratunga. The original NAS Parallel Benchmarks consisted of eight individual benchmark problems, each of which focused on some aspect of scientific computing. The principal focus was in computational aerophysics, although most of these benchmarks have much broader relevance, since in a much larger sense they are typical of many real-world scientific computing applications. The NPB suite grew out of the need for a more rational procedure to select new supercomputers for acquisition by NASA. The emergence of commercially available highly parallel computer systems in the late 1980s offered an attractive alternative to parallel vector supercomputers that had been the mainstay of high-end scientific computing. However, the introduction of highly parallel systems was accompanied by a regrettable level of hype, not only on the part of the commercial vendors but even, in some cases, by scientists using the systems. As a result, it was difficult to discern whether the new systems offered any fundamental

  16. VHTRC temperature coefficient benchmark problem

    International Nuclear Information System (INIS)

    As an activity of IAEA Coordinated Research Programme, a benchmark problem is proposed for verifications of neutronic calculation codes for a low enriched uranium fuel high temperature gas-cooled reactor. Two problems are given on the base of heating experiments at the VHTRC which is a pin-in-block type core critical assembly loaded mainly with 4% enriched uranium coated particle fuel. One problem, VH1-HP, asks to calculate temperature coefficient of reactivity from the subcritical reactivity values at five temperature steps between an room temperature where the assembly is nearly at critical state and 200degC. The other problem, VH1-HC, asks to calculate the effective multiplication factor of nearly critical loading cores at the room temperature and 200degC. Both problems further ask to calculate cell parameters such as migration area and spectral indices. Experimental results corresponding to main calculation items are also listed for comparison. (author)

  17. ICSBEP criticality benchmarking for nuclear data validations, KAMINI, PURNIMA-II and PURNIMA-I

    International Nuclear Information System (INIS)

    India has contributed three experimental benchmarks to the International handbook of the International Criticality safety Benchmark Evaluation Project (ICSBEP) of the US-DOE/NEA-DB. This presentation describes the interesting experience in creating these three Indian experimental benchmarks for nuclear data and code validation studies. The concept of definition of benchmark is also reviewed for convenience. Series of sensitivity studies are performed to assess the various uncertainties that arise in knowledge of the description of the actual system

  18. Co-ordinated research programme on benchmark study for the seismic analysis and testing of WWER-type nuclear power plants. V. 5A. Experience data. Working material. Experience database of Romanian facilities subjected to the last three Vrancea earthquakes. Final report from November 1994 - October 1995

    International Nuclear Information System (INIS)

    The Co-ordinated research programme on the benchmark study for the seismic analysis and testing of WWER-type nuclear power plants was initiated subsequent to the request from representatives of Member States. The conclusions adopted at the Technical Committee Meeting on Seismic Issues related to existing nuclear power plants held in Tokyo in 1991 called for the harmonization of methods and criteria used in Member States in issues related to seismic safety. The Consulltants' Meeting which followed resulted in producing a working document for CRP. It was decided that a benchmark study is the most effective way to achieve the principal objective. Two types of WWER reactors (WWER-440/213 and WWER-1000) were selected as prototypes for the benchmark exercise to be tested on a full scale using explosions and/or vibration generators. The two prototypes are Kozloduy Units 5/6 for WWER-1000 and Paks for WWER-440/213 nuclear power plants. This volume of Working material contains the report experience database for Romanian facilities contingent to the three Vrancea earthquakes

  19. Self-benchmarking Guide for Data Centers: Metrics, Benchmarks, Actions

    Energy Technology Data Exchange (ETDEWEB)

    Mathew, Paul; Ganguly, Srirupa; Greenberg, Steve; Sartor, Dale

    2009-07-13

    This guide describes energy efficiency metrics and benchmarks that can be used to track the performance of and identify potential opportunities to reduce energy use in data centers. This guide is primarily intended for personnel who have responsibility for managing energy use in existing data centers - including facilities managers, energy managers, and their engineering consultants. Additionally, data center designers may also use the metrics and benchmarks described in this guide for goal-setting in new construction or major renovation. This guide provides the following information: (1) A step-by-step outline of the benchmarking process. (2) A set of performance metrics for the whole building as well as individual systems. For each metric, the guide provides a definition, performance benchmarks, and potential actions that can be inferred from evaluating this metric. (3) A list and descriptions of the data required for computing the metrics. This guide is complemented by spreadsheet templates for data collection and for computing the benchmarking metrics. This guide builds on prior data center benchmarking studies supported by the California Energy Commission. Much of the benchmarking data are drawn from the LBNL data center benchmarking database that was developed from these studies. Additional benchmark data were obtained from engineering experts including facility designers and energy managers. This guide also builds on recent research supported by the U.S. Department of Energy's Save Energy Now program.

  20. A new numerical benchmark of a freshwater lens

    Science.gov (United States)

    Stoeckl, L.; Walther, M.; Graf, T.

    2016-04-01

    A numerical benchmark for 2-D variable-density flow and solute transport in a freshwater lens is presented. The benchmark is based on results of laboratory experiments conducted by Stoeckl and Houben (2012) using a sand tank on the meter scale. This benchmark describes the formation and degradation of a freshwater lens over time as it can be found under real-world islands. An error analysis gave the appropriate spatial and temporal discretization of 1 mm and 8.64 s, respectively. The calibrated parameter set was obtained using the parameter estimation tool PEST. Comparing density-coupled and density-uncoupled results showed that the freshwater-saltwater interface position is strongly dependent on density differences. A benchmark that adequately represents saltwater intrusion and that includes realistic features of coastal aquifers or freshwater lenses was lacking. This new benchmark was thus developed and is demonstrated to be suitable to test variable-density groundwater models applied to saltwater intrusion investigations.

  1. Quantum benchmarks for pure single-mode Gaussian states.

    Science.gov (United States)

    Chiribella, Giulio; Adesso, Gerardo

    2014-01-10

    Teleportation and storage of continuous variable states of light and atoms are essential building blocks for the realization of large-scale quantum networks. Rigorous validation of these implementations require identifying, and surpassing, benchmarks set by the most effective strategies attainable without the use of quantum resources. Such benchmarks have been established for special families of input states, like coherent states and particular subclasses of squeezed states. Here we solve the longstanding problem of defining quantum benchmarks for general pure Gaussian single-mode states with arbitrary phase, displacement, and squeezing, randomly sampled according to a realistic prior distribution. As a special case, we show that the fidelity benchmark for teleporting squeezed states with totally random phase and squeezing degree is 1/2, equal to the corresponding one for coherent states. We discuss the use of entangled resources to beat the benchmarks in experiments. PMID:24483875

  2. Intra and inter-organizational learning from benchmarking IS services

    DEFF Research Database (Denmark)

    Mengiste, Shegaw Anagaw; Kræmmergaard, Pernille; Hansen, Bettina

    benchmarking their IS services and functions since 2006. Particularly, this research tackled existing IS benchmarking approaches and methods by turning to a learning-oriented perspective and by empirically exploring the dynamic process of intra and inter-organizational learning from benchmarking IS/IT services......This paper reports a case study of benchmarking IS services in Danish municipalities. Drawing on Holmqvist’s (2004) organizational learning model of exploration and exploitation, the paper explores intra and inter-organizational learning dynamics among Danish municipalities that are involved in....... The paper also makes a contribution by emphasizing the importance of informal cross-municipality consortiums to facilitate learning and experience sharing across municipalities. The findings of the case study demonstrated that the IS benchmarking scheme is relatively successful in sharing good...

  3. Entropy-based benchmarking methods

    OpenAIRE

    Temurshoev, Umed

    2012-01-01

    We argue that benchmarking sign-volatile series should be based on the principle of movement and sign preservation, which states that a bench-marked series should reproduce the movement and signs in the original series. We show that the widely used variants of Denton (1971) method and the growth preservation method of Causey and Trager (1981) may violate this principle, while its requirements are explicitly taken into account in the pro-posed entropy-based benchmarking methods. Our illustrati...

  4. Benchmarking in the Semantic Web

    OpenAIRE

    García-Castro, Raúl; Gómez-Pérez, A.

    2009-01-01

    The Semantic Web technology needs to be thoroughly evaluated for providing objective results and obtaining massive improvement in its quality; thus, the transfer of this technology from research to industry will speed up. This chapter presents software benchmarking, a process that aims to improve the Semantic Web technology and to find the best practices. The chapter also describes a specific software benchmarking methodology and shows how this methodology has been used to benchmark the inter...

  5. HPC Benchmark Suite NMx Project

    Data.gov (United States)

    National Aeronautics and Space Administration — Intelligent Automation Inc., (IAI) and University of Central Florida (UCF) propose to develop a comprehensive numerical test suite for benchmarking current and...

  6. KENO-IV code benchmark calculation, (6)

    International Nuclear Information System (INIS)

    A series of benchmark tests has been undertaken in JAERI in order to examine the capability of JAERI's criticality safety evaluation system consisting of the Monte Carlo calculation code KENO-IV and the newly developed multigroup constants library MGCL. The present report describes the results of a benchmark test using criticality experiments about Plutonium fuel in various shape. In all, 33 cases of experiments have been calculated for Pu(NO3)4 aqueous solution, Pu metal or PuO2-polystyrene compact in various shape (sphere, cylinder, rectangular parallelepiped). The effective multiplication factors calculated for the 33 cases distribute widely between 0.955 and 1.045 due to wide range of system variables. (author)

  7. Shielding benchmark test for JENDL-3T

    Energy Technology Data Exchange (ETDEWEB)

    Hasegawa, Akira (Japan Atomic Energy Research Inst., Tokai, Ibaraki. Tokai Research Establishment)

    1988-03-01

    The results of the shielding benchmark tests for JENDL-3T (testing stage version of JENDL-3), performed by JNDC Shielding Sub-working group, are summarized. Especially, problems of total cross-section in MeV range for O, Na, Fe, revealed from the analysis of the Broomstick's experiment, are discussed in details. For the deep penetration profiles of Fe, which is very important feature in shielding calculation, ASPIS benchmark experiment is analysed and discussed. From the study overall applicability of JENDL-3T data for the shielding calculation is confirmed. At the same time some problems still remained are also pointed out. By the reflection of this feedback information applicability of JENDL-3, forth coming official version, will be greatly improved.

  8. CEC thermal-hydraulic benchmark exercise on Fiploc verification experiment F2 in Battelle model containment long-term heat-up phase. Results for phase I

    International Nuclear Information System (INIS)

    The major objective of the F2 experiment was to investigate the thermal-hydraulic long-term phenomena with special emphasis on natural convection phenomena in a loop-type geometry affected by variations of steam and air injections at different locations as well as dry energy supply into various compartments. The open post-test exercise is being performed in two consecutive phases, with Phase I covering the initial long-term heat-up phase. The exercise received widespread international attention with nine organizations from six European countries participating with seven different computer codes (FUMO, Jericho2, Fiploc, Wavco, Contain, Melcor, Cobra/Fathoms). These codes cover a broad spectrum of presently known European computational tools in severe accident containment analyses. The participants used either the specified mass flow or pressure control boundary conditions. Some exercised their codes for both. In total, 14 different computations were officially provided by the participants indicating strong interests and cooperative efforts by various institutions

  9. CEC thermal-hydraulic benchmark exercise on Fiploc verification experiment F2 in Battelle model containment long-term heat-up phase. Results for phase I

    Energy Technology Data Exchange (ETDEWEB)

    Fischer, K.; Schall, M.; Wolf, L. [Battelle Institut E.V., Frankfurt am Main (DE)

    1991-12-31

    The major objective of the F2 experiment was to investigate the thermal-hydraulic long-term phenomena with special emphasis on natural convection phenomena in a loop-type geometry affected by variations of steam and air injections at different locations as well as dry energy supply into various compartments. The open post-test exercise is being performed in two consecutive phases, with Phase I covering the initial long-term heat-up phase. The exercise received widespread international attention with nine organizations from six European countries participating with seven different computer codes (FUMO, Jericho2, Fiploc, Wavco, Contain, Melcor, Cobra/Fathoms). These codes cover a broad spectrum of presently known European computational tools in severe accident containment analyses. The participants used either the specified mass flow or pressure control boundary conditions. Some exercised their codes for both. In total, 14 different computations were officially provided by the participants indicating strong interests and cooperative efforts by various institutions.

  10. Regional Competitive Intelligence: Benchmarking and Policymaking

    OpenAIRE

    Huggins, Robert

    2010-01-01

    Im Bereich der Regionalpolitik erfreuen sich Benchmarking-Untersuchungen wachsender Beliebtheit. In diesem Beitrag werden das Konzept des regionalen Benchmarking sowie seine Verbindungen mit den regionalpolitischen Gestaltungsprozessen analysiert. Ich entwickle eine Typologie der regionalen Benchmarking-Untersuchungen und Benchmarker und unterziehe die Literatur einer kritischen Uumlberpruumlfung. Ich argumentiere, dass die Kritiker des regionalen Benchmarking nicht die Vielfalt und Entwicklu...

  11. NASA Software Engineering Benchmarking Study

    Science.gov (United States)

    Rarick, Heather L.; Godfrey, Sara H.; Kelly, John C.; Crumbley, Robert T.; Wifl, Joel M.

    2013-01-01

    was its software assurance practices, which seemed to rate well in comparison to the other organizational groups and also seemed to include a larger scope of activities. An unexpected benefit of the software benchmarking study was the identification of many opportunities for collaboration in areas including metrics, training, sharing of CMMI experiences and resources such as instructors and CMMI Lead Appraisers, and even sharing of assets such as documented processes. A further unexpected benefit of the study was the feedback on NASA practices that was received from some of the organizations interviewed. From that feedback, other potential areas where NASA could improve were highlighted, such as accuracy of software cost estimation and budgetary practices. The detailed report contains discussion of the practices noted in each of the topic areas, as well as a summary of observations and recommendations from each of the topic areas. The resulting 24 recommendations from the topic areas were then consolidated to eliminate duplication and culled into a set of 14 suggested actionable recommendations. This final set of actionable recommendations, listed below, are items that can be implemented to improve NASA's software engineering practices and to help address many of the items that were listed in the NASA top software engineering issues. 1. Develop and implement standard contract language for software procurements. 2. Advance accurate and trusted software cost estimates for both procured and in-house software and improve the capture of actual cost data to facilitate further improvements. 3. Establish a consistent set of objectives and expectations, specifically types of metrics at the Agency level, so key trends and models can be identified and used to continuously improve software processes and each software development effort. 4. Maintain the CMMI Maturity Level requirement for critical NASA projects and use CMMI to measure organizations developing software for NASA. 5

  12. Benchmarking monthly homogenization algorithms

    Directory of Open Access Journals (Sweden)

    V. K. C. Venema

    2011-08-01

    Full Text Available The COST (European Cooperation in Science and Technology Action ES0601: Advances in homogenization methods of climate series: an integrated approach (HOME has executed a blind intercomparison and validation study for monthly homogenization algorithms. Time series of monthly temperature and precipitation were evaluated because of their importance for climate studies and because they represent two important types of statistics (additive and multiplicative. The algorithms were validated against a realistic benchmark dataset. The benchmark contains real inhomogeneous data as well as simulated data with inserted inhomogeneities. Random break-type inhomogeneities were added to the simulated datasets modeled as a Poisson process with normally distributed breakpoint sizes. To approximate real world conditions, breaks were introduced that occur simultaneously in multiple station series within a simulated network of station data. The simulated time series also contained outliers, missing data periods and local station trends. Further, a stochastic nonlinear global (network-wide trend was added.

    Participants provided 25 separate homogenized contributions as part of the blind study as well as 22 additional solutions submitted after the details of the imposed inhomogeneities were revealed. These homogenized datasets were assessed by a number of performance metrics including (i the centered root mean square error relative to the true homogeneous value at various averaging scales, (ii the error in linear trend estimates and (iii traditional contingency skill scores. The metrics were computed both using the individual station series as well as the network average regional series. The performance of the contributions depends significantly on the error metric considered. Contingency scores by themselves are not very informative. Although relative homogenization algorithms typically improve the homogeneity of temperature data, only the best ones improve

  13. SSI and structural benchmarks

    International Nuclear Information System (INIS)

    This paper presents the latest results of the ongoing program entitled, Standard Problems for Structural Computer Codes, currently being worked on at BNL for the USNRC, Office of Nuclear Regulatory Research. During FY 1986, efforts were focussed on three tasks, namely, (1) an investigation of ground water effects on the response of Category I structures, (2) the Soil-Structure Interaction Workshop and (3) studies on structural benchmarks associated with Category I structures. The objective of the studies on ground water effects is to verify the applicability and the limitations of the SSI methods currently used by the industry in performing seismic evaluations of nuclear plants which are located at sites with high water tables. In a previous study by BNL (NUREG/CR-4588), it has been concluded that the pore water can influence significantly the soil-structure interaction process. This result, however, is based on the assumption of fully saturated soil profiles. Consequently, the work was further extended to include cases associated with variable water table depths. In this paper, results related to cut-off depths beyond which the pore water effects can be ignored in seismic calculations, are addressed. Comprehensive numerical data are given for soil configurations typical to those encountered in nuclear plant sites. These data were generated by using a modified version of the SLAM code which is capable of handling problems related to the dynamic response of saturated soils. Further, the paper presents some key aspects of the Soil-Structure Interaction Workshop (NUREG/CP-0054) which was held in Bethesda, MD on June 1, 1986. Finally, recent efforts related to the task on the structural benchmarks are described

  14. Benchmarking foreign electronics technologies

    Energy Technology Data Exchange (ETDEWEB)

    Bostian, C.W.; Hodges, D.A.; Leachman, R.C.; Sheridan, T.B.; Tsang, W.T.; White, R.M.

    1994-12-01

    This report has been drafted in response to a request from the Japanese Technology Evaluation Center`s (JTEC) Panel on Benchmarking Select Technologies. Since April 1991, the Competitive Semiconductor Manufacturing (CSM) Program at the University of California at Berkeley has been engaged in a detailed study of quality, productivity, and competitiveness in semiconductor manufacturing worldwide. The program is a joint activity of the College of Engineering, the Haas School of Business, and the Berkeley Roundtable on the International Economy, under sponsorship of the Alfred P. Sloan Foundation, and with the cooperation of semiconductor producers from Asia, Europe and the United States. Professors David A. Hodges and Robert C. Leachman are the project`s Co-Directors. The present report for JTEC is primarily based on data and analysis drawn from that continuing program. The CSM program is being conducted by faculty, graduate students and research staff from UC Berkeley`s Schools of Engineering and Business, and Department of Economics. Many of the participating firms are represented on the program`s Industry Advisory Board. The Board played an important role in defining the research agenda. A pilot study was conducted in 1991 with the cooperation of three semiconductor plants. The research plan and survey documents were thereby refined. The main phase of the CSM benchmarking study began in mid-1992 and will continue at least through 1997. reports are presented on the manufacture of integrated circuits; data storage; wireless technology; human-machine interfaces; and optoelectronics. Selected papers are indexed separately for inclusion in the Energy Science and Technology Database.

  15. Internal Benchmarking for Institutional Effectiveness

    Science.gov (United States)

    Ronco, Sharron L.

    2012-01-01

    Internal benchmarking is an established practice in business and industry for identifying best in-house practices and disseminating the knowledge about those practices to other groups in the organization. Internal benchmarking can be done with structures, processes, outcomes, or even individuals. In colleges or universities with multicampuses or a…

  16. Validation of the BUGJEFF311.BOLIB, BUGENDF70.BOLIB and BUGLE-B7 broad-group libraries on the PCA-Replica (H2O/Fe) neutron shielding benchmark experiment

    Science.gov (United States)

    Pescarini, Massimo; Orsi, Roberto; Frisoni, Manuela

    2016-03-01

    The PCA-Replica 12/13 (H2O/Fe) neutron shielding benchmark experiment was analysed using the TORT-3.2 3D SN code. PCA-Replica reproduces a PWR ex-core radial geometry with alternate layers of water and steel including a pressure vessel simulator. Three broad-group coupled neutron/photon working cross section libraries in FIDO-ANISN format with the same energy group structure (47 n + 20 γ) and based on different nuclear data were alternatively used: the ENEA BUGJEFF311.BOLIB (JEFF-3.1.1) and UGENDF70.BOLIB (ENDF/B-VII.0) libraries and the ORNL BUGLE-B7 (ENDF/B-VII.0) library. Dosimeter cross sections derived from the IAEA IRDF-2002 dosimetry file were employed. The calculated reaction rates for the Rh-103(n,n')Rh-103m, In-115(n,n')In-115m and S-32(n,p)P-32 threshold activation dosimeters and the calculated neutron spectra are compared with the corresponding experimental results.

  17. Three-Dimensional (X,Y,Z) Deterministic Analysis of the PCA-Replica Neutron Shielding Benchmark Experiment using the TORT-3.2 Code and Group Cross Section Libraries for LWR Shielding and Pressure Vessel Dosimetry

    Science.gov (United States)

    Pescarini, Massimo; Orsi, Roberto; Frisoni, Manuela

    2016-02-01

    The PCA-Replica 12/13 (H2O/Fe) neutron shielding benchmark experiment was analysed using the ORNL TORT-3.2 3D SN code. PCA-Replica, specifically conceived to test the accuracy of nuclear data and transport codes employed in LWR shielding and radiation damage calculations, reproduces a PWR ex-core radial geometry with alternate layers of water and steel including a PWR pressure vessel simulator. Three broad-group coupled neutron/photon working cross section libraries in FIDO-ANISN format with the same energy group structure (47 n + 20 γ) and based on different nuclear data were alternatively used: the ENEA BUGJEFF311.BOLIB (JEFF-3.1.1) and BUGENDF70.BOLIB (ENDF/B-VII.0) libraries and the ORNL BUGLE-96 (ENDF/B-VI.3) library. Dosimeter cross sections derived from the IAEA IRDF-2002 dosimetry file were employed. The calculated reaction rates for the Rh-103(n,n')Rh-103 m, In-115(n,n')In-115m and S-32(n,p)P-32 threshold activation dosimeters and the calculated neutron spectra are compared with the corresponding experimental results.

  18. Three-Dimensional (X,Y,Z Deterministic Analysis of the PCA-Replica Neutron Shielding Benchmark Experiment using the TORT-3.2 Code and Group Cross Section Libraries for LWR Shielding and Pressure Vessel Dosimetry

    Directory of Open Access Journals (Sweden)

    Pescarini Massimo

    2016-01-01

    Full Text Available The PCA-Replica 12/13 (H2O/Fe neutron shielding benchmark experiment was analysed using the ORNL TORT-3.2 3D SN code. PCA-Replica, specifically conceived to test the accuracy of nuclear data and transport codes employed in LWR shielding and radiation damage calculations, reproduces a PWR ex-core radial geometry with alternate layers of water and steel including a PWR pressure vessel simulator. Three broad-group coupled neutron/photon working cross section libraries in FIDO-ANISN format with the same energy group structure (47 n + 20 γ and based on different nuclear data were alternatively used: the ENEA BUGJEFF311.BOLIB (JEFF-3.1.1 and BUGENDF70.BOLIB (ENDF/B-VII.0 libraries and the ORNL BUGLE-96 (ENDF/B-VI.3 library. Dosimeter cross sections derived from the IAEA IRDF-2002 dosimetry file were employed. The calculated reaction rates for the Rh-103(n,n′Rh-103 m, In-115(n,n′In-115m and S-32(n,pP-32 threshold activation dosimeters and the calculated neutron spectra are compared with the corresponding experimental results.

  19. Validation of the BUGJEFF311.BOLIB, BUGENDF70.BOLIB and BUGLE-B7 broad-group libraries on the PCA-Replica (H2O/Fe neutron shielding benchmark experiment

    Directory of Open Access Journals (Sweden)

    Pescarini Massimo

    2016-01-01

    Full Text Available The PCA-Replica 12/13 (H2O/Fe neutron shielding benchmark experiment was analysed using the TORT-3.2 3D SN code. PCA-Replica reproduces a PWR ex-core radial geometry with alternate layers of water and steel including a pressure vessel simulator. Three broad-group coupled neutron/photon working cross section libraries in FIDO-ANISN format with the same energy group structure (47 n + 20 γ and based on different nuclear data were alternatively used: the ENEA BUGJEFF311.BOLIB (JEFF-3.1.1 and UGENDF70.BOLIB (ENDF/B-VII.0 libraries and the ORNL BUGLE-B7 (ENDF/B-VII.0 library. Dosimeter cross sections derived from the IAEA IRDF-2002 dosimetry file were employed. The calculated reaction rates for the Rh-103(n,n′Rh-103m, In-115(n,n′In-115m and S-32(n,pP-32 threshold activation dosimeters and the calculated neutron spectra are compared with the corresponding experimental results.

  20. Benchmarking Open-Source Tree Learners in R/RWeka

    OpenAIRE

    Schauerhuber, Michael; Zeileis, Achim; Meyer, David; Hornik, Kurt

    2007-01-01

    The two most popular classification tree algorithms in machine learning and statistics - C4.5 and CART - are compared in a benchmark experiment together with two other more recent constant-fit tree learners from the statistics literature (QUEST, conditional inference trees). The study assesses both misclassification error and model complexity on bootstrap replications of 18 different benchmark datasets. It is carried out in the R system for statistical computing, made possible by means of the...

  1. Towards a Core Model for Higher Education IT Management Benchmarking

    OpenAIRE

    Markus Juult, Janne

    2013-01-01

    This study evaluates three European higher education IT benchmarking projects by applying a custom comparison framework that is based on benchmarking literature and IT manager experience. The participating projects are Bencheit (Finland), UCISA (The United Kingdom) and UNIVERSITIC (Spain). EDUCAUSE (The United States of America) is also included as a project outside our geographical focus area due to its size and prominence in North America. Each of these projects is examined to map the data ...

  2. Benchmark Evaluation of the NRAD Reactor LEU Core Startup Measurements

    Energy Technology Data Exchange (ETDEWEB)

    J. D. Bess; T. L. Maddock; M. A. Marshall

    2011-09-01

    The Neutron Radiography (NRAD) reactor is a 250-kW TRIGA-(Training, Research, Isotope Production, General Atomics)-conversion-type reactor at the Idaho National Laboratory; it is primarily used for neutron radiography analysis of irradiated and unirradiated fuels and materials. The NRAD reactor was converted from HEU to LEU fuel with 60 fuel elements and brought critical on March 31, 2010. This configuration of the NRAD reactor has been evaluated as an acceptable benchmark experiment and is available in the 2011 editions of the International Handbook of Evaluated Criticality Safety Benchmark Experiments (ICSBEP Handbook) and the International Handbook of Evaluated Reactor Physics Benchmark Experiments (IRPhEP Handbook). Significant effort went into precisely characterizing all aspects of the reactor core dimensions and material properties; detailed analyses of reactor parameters minimized experimental uncertainties. The largest contributors to the total benchmark uncertainty were the 234U, 236U, Er, and Hf content in the fuel; the manganese content in the stainless steel cladding; and the unknown level of water saturation in the graphite reflector blocks. A simplified benchmark model of the NRAD reactor was prepared with a keff of 1.0012 {+-} 0.0029 (1s). Monte Carlo calculations with MCNP5 and KENO-VI and various neutron cross section libraries were performed and compared with the benchmark eigenvalue for the 60-fuel-element core configuration; all calculated eigenvalues are between 0.3 and 0.8% greater than the benchmark value. Benchmark evaluations of the NRAD reactor are beneficial in understanding biases and uncertainties affecting criticality safety analyses of storage, handling, or transportation applications with LEU-Er-Zr-H fuel.

  3. Benchmarking biofuels; Biobrandstoffen benchmarken

    Energy Technology Data Exchange (ETDEWEB)

    Croezen, H.; Kampman, B.; Bergsma, G.

    2012-03-15

    A sustainability benchmark for transport biofuels has been developed and used to evaluate the various biofuels currently on the market. For comparison, electric vehicles, hydrogen vehicles and petrol/diesel vehicles were also included. A range of studies as well as growing insight are making it ever clearer that biomass-based transport fuels may have just as big a carbon footprint as fossil fuels like petrol or diesel, or even bigger. At the request of Greenpeace Netherlands, CE Delft has brought together current understanding on the sustainability of fossil fuels, biofuels and electric vehicles, with particular focus on the performance of the respective energy carriers on three sustainability criteria, with the first weighing the heaviest: (1) Greenhouse gas emissions; (2) Land use; and (3) Nutrient consumption [Dutch] Greenpeace Nederland heeft CE Delft gevraagd een duurzaamheidsmeetlat voor biobrandstoffen voor transport te ontwerpen en hierop de verschillende biobrandstoffen te scoren. Voor een vergelijk zijn ook elektrisch rijden, rijden op waterstof en rijden op benzine of diesel opgenomen. Door onderzoek en voortschrijdend inzicht blijkt steeds vaker dat transportbrandstoffen op basis van biomassa soms net zoveel of zelfs meer broeikasgassen veroorzaken dan fossiele brandstoffen als benzine en diesel. CE Delft heeft voor Greenpeace Nederland op een rijtje gezet wat de huidige inzichten zijn over de duurzaamheid van fossiele brandstoffen, biobrandstoffen en elektrisch rijden. Daarbij is gekeken naar de effecten van de brandstoffen op drie duurzaamheidscriteria, waarbij broeikasgasemissies het zwaarst wegen: (1) Broeikasgasemissies; (2) Landgebruik; en (3) Nutriëntengebruik.

  4. Cleanroom energy benchmarking results

    Energy Technology Data Exchange (ETDEWEB)

    Tschudi, William; Xu, Tengfang

    2001-09-01

    A utility market transformation project studied energy use and identified energy efficiency opportunities in cleanroom HVAC design and operation for fourteen cleanrooms. This paper presents the results of this work and relevant observations. Cleanroom owners and operators know that cleanrooms are energy intensive but have little information to compare their cleanroom's performance over time, or to others. Direct comparison of energy performance by traditional means, such as watts/ft{sup 2}, is not a good indicator with the wide range of industrial processes and cleanliness levels occurring in cleanrooms. In this project, metrics allow direct comparison of the efficiency of HVAC systems and components. Energy and flow measurements were taken to determine actual HVAC system energy efficiency. The results confirm a wide variation in operating efficiency and they identify other non-energy operating problems. Improvement opportunities were identified at each of the benchmarked facilities. Analysis of the best performing systems and components is summarized, as are areas for additional investigation.

  5. Issues in Benchmark Metric Selection

    Science.gov (United States)

    Crolotte, Alain

    It is true that a metric can influence a benchmark but will esoteric metrics create more problems than they will solve? We answer this question affirmatively by examining the case of the TPC-D metric which used the much debated geometric mean for the single-stream test. We will show how a simple choice influenced the benchmark and its conduct and, to some extent, DBMS development. After examining other alternatives our conclusion is that the “real” measure for a decision-support benchmark is the arithmetic mean.

  6. Benchmarking & European Sustainable Transport Policies

    DEFF Research Database (Denmark)

    Gudmundsson, H.

    2003-01-01

    support Sustainable European Transport Policies. The key message is that transport benchmarking has not yet been developed to cope with the challenges of this task. Rather than backing down completely, the paper suggests some critical conditions for applying and adopting benchmarking for this purpose. One...... way forward is to ensure a higher level of environmental integration in transport policy benchmarking. To this effect the paper will discuss the possible role of the socalled Transport and Environment Reporting Mechanism developed by the European Environment Agency. The paper provides an independent...

  7. Benchmark simulation models, quo vadis?

    DEFF Research Database (Denmark)

    Jeppsson, U.; Alex, J; Batstone, D. J.;

    2013-01-01

    As the work of the IWA Task Group on Benchmarking of Control Strategies for wastewater treatment plants (WWTPs) is coming to an end, it is essential to disseminate the knowledge gained. For this reason, all authors of the IWA Scientific and Technical Report on benchmarking have come together to...... and spatial extension, process modifications within the WWTP, the realism of models, control strategy extensions and the potential for new evaluation tools within the existing benchmark system. We find that there are major opportunities for application within all of these areas, either from existing...

  8. California commercial building energy benchmarking

    Energy Technology Data Exchange (ETDEWEB)

    Kinney, Satkartar; Piette, Mary Ann

    2003-07-01

    Building energy benchmarking is the comparison of whole-building energy use relative to a set of similar buildings. It provides a useful starting point for individual energy audits and for targeting buildings for energy-saving measures in multiple-site audits. Benchmarking is of interest and practical use to a number of groups. Energy service companies and performance contractors communicate energy savings potential with ''typical'' and ''best-practice'' benchmarks while control companies and utilities can provide direct tracking of energy use and combine data from multiple buildings. Benchmarking is also useful in the design stage of a new building or retrofit to determine if a design is relatively efficient. Energy managers and building owners have an ongoing interest in comparing energy performance to others. Large corporations, schools, and government agencies with numerous facilities also use benchmarking methods to compare their buildings to each other. The primary goal of Task 2.1.1 Web-based Benchmarking was the development of a web-based benchmarking tool, dubbed Cal-Arch, for benchmarking energy use in California commercial buildings. While there were several other benchmarking tools available to California consumers prior to the development of Cal-Arch, there were none that were based solely on California data. Most available benchmarking information, including the Energy Star performance rating, were developed using DOE's Commercial Building Energy Consumption Survey (CBECS), which does not provide state-level data. Each database and tool has advantages as well as limitations, such as the number of buildings and the coverage by type, climate regions and end uses. There is considerable commercial interest in benchmarking because it provides an inexpensive method of screening buildings for tune-ups and retrofits. However, private companies who collect and manage consumption data are concerned that the

  9. International Criticality Safety Benchmark Evaluation Project (ICSBEP) - ICSBEP 2015 Handbook

    International Nuclear Information System (INIS)

    The Criticality Safety Benchmark Evaluation Project (CSBEP) was initiated in October of 1992 by the United States Department of Energy (DOE). The project quickly became an international effort as scientists from other interested countries became involved. The International Criticality Safety Benchmark Evaluation Project (ICSBEP) became an official activity of the Nuclear Energy Agency (NEA) in 1995. This handbook contains criticality safety benchmark specifications that have been derived from experiments performed at various critical facilities around the world. The benchmark specifications are intended for use by criticality safety engineers to validate calculation techniques used to establish minimum subcritical margins for operations with fissile material and to determine criticality alarm requirements and placement. Many of the specifications are also useful for nuclear data testing. Example calculations are presented; however, these calculations do not constitute a validation of the codes or cross-section data. The evaluated criticality safety benchmark data are given in nine volumes. These volumes span approximately 69000 pages and contain 567 evaluations with benchmark specifications for 4874 critical, near-critical or subcritical configurations, 31 criticality alarm placement/shielding configurations with multiple dose points for each, and 207 configurations that have been categorised as fundamental physics measurements that are relevant to criticality safety applications. New to the handbook are benchmark specifications for neutron activation foil and thermoluminescent dosimeter measurements performed at the SILENE critical assembly in Valduc, France as part of a joint venture in 2010 between the US DOE and the French Alternative Energies and Atomic Energy Commission (CEA). A photograph of this experiment is shown on the front cover. Experiments that are found unacceptable for use as criticality safety benchmark experiments are discussed in these

  10. The International Criticality Safety Benchmark Evaluation Project (ICSBEP)

    International Nuclear Information System (INIS)

    The International Criticality Safety Benchmark Evaluation Project (ICSBEP) was initiated in 1992 by the United States Department of Energy. The ICSBEP became an official activity of the Organisation for Economic Cooperation and Development (OECD) - Nuclear Energy Agency (NEA) in 1995. Representatives from the United States, United Kingdom, France, Japan, the Russian Federation, Hungary, Republic of Korea, Slovenia, Yugoslavia, Kazakhstan, Israel, Spain, and Brazil are now participating. The purpose of the ICSBEP is to identify, evaluate, verify, and formally document a comprehensive and internationally peer-reviewed set of criticality safety benchmark data. The work of the ICSBEP is published as an OECD handbook entitled 'International Handbook of Evaluated Criticality Safety Benchmark Experiments.' The 2003 Edition of the Handbook contains benchmark model specifications for 3070 critical or subcritical configurations that are intended for validating computer codes that calculate effective neutron multiplication and for testing basic nuclear data. (author)

  11. Issues in benchmarking human reliability analysis methods: A literature review

    International Nuclear Information System (INIS)

    There is a diversity of human reliability analysis (HRA) methods available for use in assessing human performance within probabilistic risk assessments (PRA). Due to the significant differences in the methods, including the scope, approach, and underlying models, there is a need for an empirical comparison investigating the validity and reliability of the methods. To accomplish this empirical comparison, a benchmarking study comparing and evaluating HRA methods in assessing operator performance in simulator experiments is currently underway. In order to account for as many effects as possible in the construction of this benchmarking study, a literature review was conducted, reviewing past benchmarking studies in the areas of psychology and risk assessment. A number of lessons learned through these studies is presented in order to aid in the design of future HRA benchmarking endeavors.

  12. Performance Targets and External Benchmarking

    DEFF Research Database (Denmark)

    Friis, Ivar; Hansen, Allan; Vámosi, Tamás S.

    Research on relative performance measures, transfer pricing, beyond budgeting initiatives, target costing, piece rates systems and value based management has for decades underlined the importance of external benchmarking in performance management. Research conceptualises external benchmarking as a...... market mechanism that can be brought inside the firm to provide incentives for continuous improvement and the development of competitive advances. However, whereas extant research primarily has focused on the importance and effects of using external benchmarks, less attention has been directed towards...... the conditions upon which the market mechanism is performing within organizations. This paper aims to contribute to research by providing more insight to the conditions for the use of external benchmarking as an element in performance management in organizations. Our study explores a particular type...

  13. Benchmarking and Sustainable Transport Policy

    DEFF Research Database (Denmark)

    Gudmundsson, Henrik; Wyatt, Andrew; Gordon, Lucy

    2004-01-01

    Order to learn from the best. In 2000 the European Commission initiated research to explore benchmarking as a tool to promote policies for ‘sustainable transport’. This paper reports findings and recommendations on how to address this challenge. The findings suggest that benchmarking is a valuable...... tool that may indeed help to move forward the transport policy agenda. However, there are major conditions and limitations. First of all it is not always so straightforward to delimit, measure and compare transport services in order to establish a clear benchmark. Secondly ‘sustainable transport......’ evokes a broad range of concerns that are hard to address fully at the level of specific practices. Thirdly policies are not directly comparable across space and context. For these reasons attempting to benchmark ‘sustainable transport policies’ against one another would be a highly complex task, which...

  14. Benchmarking Developing Asia's Manufacturing Sector

    OpenAIRE

    Felipe, Jesus; Gemma ESTRADA

    2007-01-01

    This paper documents the transformation of developing Asia's manufacturing sector during the last three decades and benchmarks its share in GDP with respect to the international regression line by estimating a logistic regression.

  15. Water Level Superseded Benchmark Sheets

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — Images of National Coast & Geodetic Survey (now NOAA's National Geodetic Survey/NGS) tidal benchmarks which have been superseded by new markers or locations....

  16. Shielding integral benchmark archive and database (SINBAD)

    Energy Technology Data Exchange (ETDEWEB)

    Kirk, B.L.; Grove, R.E. [Radiation Safety Information Computational Center RSICC, Oak Ridge National Laboratory, P.O. Box 2008, Oak Ridge, TN 37831-6171 (United States); Kodeli, I. [Josef Stefan Inst., Jamova 39, 1000 Ljubljana (Slovenia); Gulliford, J.; Sartori, E. [OECD NEA Data Bank, Bd des Iles, 92130 Issy-les-Moulineaux (France)

    2011-07-01

    The shielding integral benchmark archive and database (SINBAD) collection of experiments descriptions was initiated in the early 1990s. SINBAD is an international collaboration between the Organization for Economic Cooperation and Development's Nuclear Energy Agency Data Bank (OECD/NEADB) and the Radiation Safety Information Computational Center (RSICC) at Oak Ridge National Laboratory (ORNL). SINBAD was designed to compile experiments and corresponding computational models with the goal of preserving institutional knowledge and expertise that need to be handed down to future scientists. SINBAD can serve as a learning tool for university students and scientists who need to design experiments or gain expertise in modeling and simulation. The SINBAD database is currently divided into three categories - fission, fusion, and accelerator experiments. Many experiments are described and analyzed using deterministic or stochastic (Monte Carlo) radiation transport software. The nuclear cross sections also play an important role as they are necessary in performing computational analysis. (authors)

  17. Shielding integral benchmark archive and database (SINBAD)

    International Nuclear Information System (INIS)

    The shielding integral benchmark archive and database (SINBAD) collection of experiments descriptions was initiated in the early 1990s. SINBAD is an international collaboration between the Organization for Economic Cooperation and Development's Nuclear Energy Agency Data Bank (OECD/NEADB) and the Radiation Safety Information Computational Center (RSICC) at Oak Ridge National Laboratory (ORNL). SINBAD was designed to compile experiments and corresponding computational models with the goal of preserving institutional knowledge and expertise that need to be handed down to future scientists. SINBAD can serve as a learning tool for university students and scientists who need to design experiments or gain expertise in modeling and simulation. The SINBAD database is currently divided into three categories - fission, fusion, and accelerator experiments. Many experiments are described and analyzed using deterministic or stochastic (Monte Carlo) radiation transport software. The nuclear cross sections also play an important role as they are necessary in performing computational analysis. (authors)

  18. Benchmarking hypercube hardware and software

    Science.gov (United States)

    Grunwald, Dirk C.; Reed, Daniel A.

    1986-01-01

    It was long a truism in computer systems design that balanced systems achieve the best performance. Message passing parallel processors are no different. To quantify the balance of a hypercube design, an experimental methodology was developed and the associated suite of benchmarks was applied to several existing hypercubes. The benchmark suite includes tests of both processor speed in the absence of internode communication and message transmission speed as a function of communication patterns.

  19. Strategic Behaviour under Regulation Benchmarking

    OpenAIRE

    Jamasb, Tooraj; Nillesen, Paul; Michael G. Pollitt

    2003-01-01

    Liberalisation of generation and supply activities in the electricity sectors is often followed by regulatory reform of distribution networks. In order to improve the efficiency of distribution utilities, some regulators have adopted incentive regulation schemes that rely on performance benchmarking. Although regulation benchmarking can influence the ?regulation game?, the subject has received limited attention. This paper discusses how strategic behaviour can result in inefficient behav...

  20. Research on computer systems benchmarking

    Science.gov (United States)

    Smith, Alan Jay (Principal Investigator)

    1996-01-01

    This grant addresses the topic of research on computer systems benchmarking and is more generally concerned with performance issues in computer systems. This report reviews work in those areas during the period of NASA support under this grant. The bulk of the work performed concerned benchmarking and analysis of CPUs, compilers, caches, and benchmark programs. The first part of this work concerned the issue of benchmark performance prediction. A new approach to benchmarking and machine characterization was reported, using a machine characterizer that measures the performance of a given system in terms of a Fortran abstract machine. Another report focused on analyzing compiler performance. The performance impact of optimization in the context of our methodology for CPU performance characterization was based on the abstract machine model. Benchmark programs are analyzed in another paper. A machine-independent model of program execution was developed to characterize both machine performance and program execution. By merging these machine and program characterizations, execution time can be estimated for arbitrary machine/program combinations. The work was continued into the domain of parallel and vector machines, including the issue of caches in vector processors and multiprocessors. All of the afore-mentioned accomplishments are more specifically summarized in this report, as well as those smaller in magnitude supported by this grant.

  1. Benchmark test of JENDL High Energy File with MCNP

    International Nuclear Information System (INIS)

    Benchmark tests of the preliminary version of JENDL High Energy File with the MCNP code have been continued. A problem in JENDL High Energy File is pointed out that the angular distribution of elastically scattered neutrons by iron-56 is too emphasized toward 0 degree. Good results are obtained in the analyses of concrete shield experiments and iron secondary gamma-ray experiments. (author)

  2. A Simplified HTTR Diffusion Theory Benchmark

    International Nuclear Information System (INIS)

    The Georgia Institute of Technology (GA-Tech) recently developed a transport theory benchmark based closely on the geometry and the features of the HTTR reactor that is operational in Japan. Though simplified, the benchmark retains all the principal physical features of the reactor and thus provides a realistic and challenging test for the codes. The purpose of this paper is two-fold. The first goal is an extension of the benchmark to diffusion theory applications by generating the additional data not provided in the GA-Tech prior work. The second goal is to use the benchmark on the HEXPEDITE code available to the INL. The HEXPEDITE code is a Green's function-based neutron diffusion code in 3D hexagonal-z geometry. The results showed that the HEXPEDITE code accurately reproduces the effective multiplication factor of the reference HELIOS solution. A secondary, but no less important, conclusion is that in the testing against actual HTTR data of a full sequence of codes that would include HEXPEDITE, in the apportioning of inevitable discrepancies between experiment and models, the portion of error attributable to HEXPEDITE would be expected to be modest. If large discrepancies are observed, they would have to be explained by errors in the data fed into HEXPEDITE. Results based on a fully realistic model of the HTTR reactor are presented in a companion paper. The suite of codes used in that paper also includes HEXPEDITE. The results shown here should help that effort in the decision making process for refining the modeling steps in the full sequence of codes.

  3. HELIOS-2: Benchmarking against hexagonal lattices

    International Nuclear Information System (INIS)

    The critical experiments performed at the Hungarian ZR-6 reactor and experiments performed at General Purpose P critical facility at RRC 'Kurchatov Institute' are used to benchmark HELIOS-2, as part of its ongoing validation and verification activities. The comparisons presented in this paper are based on, ZR6-WWER-EXP-001, Vol. 1 (2007), LEU-COMP-THERM-015, Vol. 4 (2005), and LEU-COMP-THERM-061, Vol. 4 (2002). On ZR-6, single cell, macro cell, and 2D calculations on selected experiments, regular and perturbed, are made. In the 2D calculations, the radial leakage is treated by including the reflector in the calculations, representing only axial leakage by a measured axial buckling. keff and RMS reaction rates comparisons are presented. On General Purpose P critical facility, the entire experiment is modelled as 2D. Comparisons of keff and fission rates are presented and the effect of the axial buckling on keff is investigated. (Author)

  4. Benchmark Calculations of OECD/NEA Reactivity-Initiated Accidents

    International Nuclear Information System (INIS)

    The benchmark- Phase I was done from 2011 to 2013 with a consistent set of four experiments on very similar highly irradiated fuel rods tested under different experimental conditions: low temperature, low pressure, stagnant water coolant, very short power pulse (NSRR VA-1), high temperature, medium pressure, stagnant water coolant, very short power pulse (NSRR VA-3), high temperature, low pressure, flowing sodium coolant, larger power pulse (CABRI CIP0-1), high temperature, high pressure, flowing water coolant, medium width power pulse (CABRI CIP3-1). Based on the importance of the thermal-hydraulics aspects revealed during the Phase I, the specifications of the benchmark-Phase II was elaborated in 2014. The benchmark-Phase II focused on the deeper understanding of the differences in modeling of the different codes. The work on the benchmark- Phase II program will last the end of 2015. The benchmark cases for RIA are simulated with the code of FRAPTRAN 1.5, in order to understand the phenomena during RIA and to check the capacity of the code itself. The results of enthalpy, cladding strain and outside temperature among 21 parameters asked by the benchmark program are summarized, and they seem to reasonably reflect the actual phenomena, except for them of case 6

  5. Integral Benchmark Data for Nuclear Data Testing Through the ICSBEP & IRPhEP

    Energy Technology Data Exchange (ETDEWEB)

    J. Blair Briggs; John D. Bess; Jim Gulliford; Ian Hill

    2013-10-01

    The status of the International Criticality Safety Benchmark Evaluation Project (ICSBEP) and International Reactor Physics Experiment Evaluation Project (IRPhEP) was last discussed directly with the nuclear data community at ND2007. Since ND2007, integral benchmark data that are available for nuclear data testing have increased significantly. The status of the ICSBEP and the IRPhEP is discussed and selected benchmark configurations that have been added to the ICSBEP and IRPhEP Handbooks since ND2007 are highlighted.

  6. Two dimensional shielding benchmark analysis for sodium

    International Nuclear Information System (INIS)

    Results of the analysis of a shielding benchmark experiment on 'fast reactor source' neutron transport through 1.8 metres of sodium is presented in this paper. The two dimensional discrete ordinates code DOT and DLC 37 coupled neutron-gamma multigroup cross sections were used in the analyses. These calculations are compared with measurements on: (i) neutron spectral distribution given by activation detector response, and (ii) gamma ray doses. The agreement is found to be within ± 30 per cent in the fast spectrum region, and within a factor 3.5 in thermal region. For gammas these calculations overpredict the dose rate by a factor of four. (author)

  7. Benchmarks in Tacit Knowledge Skills Instruction

    DEFF Research Database (Denmark)

    Tackney, Charles T.; Strömgren, Ole; Sato, Toyoko

    2006-01-01

    While the knowledge management literature has addressed the explicit and tacit skills needed for successful performance in the modern enterprise, little attention has been paid to date in this particular literature as to how these wide-ranging skills may be suitably acquired during the course of an...... experience more empowering of essential tacit knowledge skills than that found in educational institutions in other national settings. We specify the program forms and procedures for consensus-based governance and group work (as benchmarks) that demonstrably instruct undergraduates in the tacit skill...

  8. Closed-Loop Neuromorphic Benchmarks

    Science.gov (United States)

    Stewart, Terrence C.; DeWolf, Travis; Kleinhans, Ashley; Eliasmith, Chris

    2015-01-01

    Evaluating the effectiveness and performance of neuromorphic hardware is difficult. It is even more difficult when the task of interest is a closed-loop task; that is, a task where the output from the neuromorphic hardware affects some environment, which then in turn affects the hardware's future input. However, closed-loop situations are one of the primary potential uses of neuromorphic hardware. To address this, we present a methodology for generating closed-loop benchmarks that makes use of a hybrid of real physical embodiment and a type of “minimal” simulation. Minimal simulation has been shown to lead to robust real-world performance, while still maintaining the practical advantages of simulation, such as making it easy for the same benchmark to be used by many researchers. This method is flexible enough to allow researchers to explicitly modify the benchmarks to identify specific task domains where particular hardware excels. To demonstrate the method, we present a set of novel benchmarks that focus on motor control for an arbitrary system with unknown external forces. Using these benchmarks, we show that an error-driven learning rule can consistently improve motor control performance across a randomly generated family of closed-loop simulations, even when there are up to 15 interacting joints to be controlled. PMID:26696820

  9. Benchmarking of human resources management

    Directory of Open Access Journals (Sweden)

    David M. Akinnusi

    2008-12-01

    Full Text Available This paper reviews the role of human resource management (HRM which, today, plays a strategic partnership role in management. The focus of the paper is on HRM in the public sector, where much hope rests on HRM as a means of transforming the public service and achieving much needed service delivery. However, a critical evaluation of HRM practices in the public sector reveals that these services leave much to be desired. The paper suggests the adoption of benchmarking as a process to revamp HRM in the public sector so that it is able to deliver on its promises. It describes the nature and process of benchmarking and highlights the inherent difficulties in applying benchmarking in HRM. It concludes with some suggestions for a plan of action. The process of identifying “best” practices in HRM requires the best collaborative efforts of HRM practitioners and academicians. If used creatively, benchmarking has the potential to bring about radical and positive changes in HRM in the public sector. The adoption of the benchmarking process is, in itself, a litmus test of the extent to which HRM in the public sector has grown professionally.

  10. ZZ IHEAS-BENCHMARKS, High-Energy Accelerator Shielding Benchmarks

    International Nuclear Information System (INIS)

    Description of program or function: Six kinds of Benchmark problems were selected for evaluating the model codes and the nuclear data for the intermediate and high energy accelerator shielding by the Shielding Subcommittee in the Research Committee on Reactor Physics. The benchmark problems contain three kinds of neutron production data from thick targets due to proton, alpha and electron, and three kinds of shielding data for secondary neutron and photon generated by proton. Neutron and photo-neutron reaction cross section data are also provided for neutrons up to 500 MeV and photons up to 300 MeV, respectively

  11. Compilation of benchmark results for fusion related Nuclear Data

    International Nuclear Information System (INIS)

    This report compiles results of benchmark tests for validation of evaluated nuclear data to be used in nuclear designs of fusion reactors. Parts of results were obtained under activities of the Fusion Neutronics Integral Test Working Group organized by the members of both Japan Nuclear Data Committee and the Reactor Physics Committee. The following three benchmark experiments were employed used for the tests: (i) the leakage neutron spectrum measurement experiments from slab assemblies at the D-T neutron source at FNS/JAERI, (ii) in-situ neutron and gamma-ray measurement experiments (so-called clean benchmark experiments) also at FNS, and (iii) the pulsed sphere experiments for leakage neutron and gamma-ray spectra at the D-T neutron source facility of Osaka University, OKTAVIAN. Evaluated nuclear data tested were JENDL-3.2, JENDL Fusion File, FENDL/E-1.0 and newly selected data for FENDL/E-2.0. Comparisons of benchmark calculations with the experiments for twenty-one elements, i.e., Li, Be, C, N, O, F, Al, Si, Ti, V, Cr, Mn, Fe, Co, Ni, Cu, Zr, Nb, Mo, W and Pb, are summarized. (author). 65 refs

  12. Radiation Detection Computational Benchmark Scenarios

    Energy Technology Data Exchange (ETDEWEB)

    Shaver, Mark W.; Casella, Andrew M.; Wittman, Richard S.; McDonald, Ben S.

    2013-09-24

    Modeling forms an important component of radiation detection development, allowing for testing of new detector designs, evaluation of existing equipment against a wide variety of potential threat sources, and assessing operation performance of radiation detection systems. This can, however, result in large and complex scenarios which are time consuming to model. A variety of approaches to radiation transport modeling exist with complementary strengths and weaknesses for different problems. This variety of approaches, and the development of promising new tools (such as ORNL’s ADVANTG) which combine benefits of multiple approaches, illustrates the need for a means of evaluating or comparing different techniques for radiation detection problems. This report presents a set of 9 benchmark problems for comparing different types of radiation transport calculations, identifying appropriate tools for classes of problems, and testing and guiding the development of new methods. The benchmarks were drawn primarily from existing or previous calculations with a preference for scenarios which include experimental data, or otherwise have results with a high level of confidence, are non-sensitive, and represent problem sets of interest to NA-22. From a technical perspective, the benchmarks were chosen to span a range of difficulty and to include gamma transport, neutron transport, or both and represent different important physical processes and a range of sensitivity to angular or energy fidelity. Following benchmark identification, existing information about geometry, measurements, and previous calculations were assembled. Monte Carlo results (MCNP decks) were reviewed or created and re-run in order to attain accurate computational times and to verify agreement with experimental data, when present. Benchmark information was then conveyed to ORNL in order to guide testing and development of hybrid calculations. The results of those ADVANTG calculations were then sent to PNNL for

  13. PageRank Pipeline Benchmark: Proposal for a Holistic System Benchmark for Big-Data Platforms

    OpenAIRE

    Dreher, Patrick; Byun, Chansup; Hill, Chris; Gadepally, Vijay; Kuszmaul, Bradley; Kepner, Jeremy

    2016-01-01

    The rise of big data systems has created a need for benchmarks to measure and compare the capabilities of these systems. Big data benchmarks present unique scalability challenges. The supercomputing community has wrestled with these challenges for decades and developed methodologies for creating rigorous scalable benchmarks (e.g., HPC Challenge). The proposed PageRank pipeline benchmark employs supercomputing benchmarking methodologies to create a scalable benchmark that is reflective of many...

  14. 3-D neutron transport benchmarks

    International Nuclear Information System (INIS)

    A set of 3-D neutron transport benchmark problems proposed by the Osaka University to NEACRP in 1988 has been calculated by many participants and the corresponding results are summarized in this report. The results of Keff, control rod worth and region-averaged fluxes for the four proposed core models, calculated by using various 3-D transport codes are compared and discussed. The calculational methods used were: Monte Carlo, Discrete Ordinates (Sn), Spherical Harmonics (Pn), Nodal Transport and others. The solutions of the four core models are quite useful as benchmarks for checking the validity of 3-D neutron transport codes

  15. Benchmarking of the FENDL-3 Neutron Cross-section Data Starter Library for Fusion Applications

    Energy Technology Data Exchange (ETDEWEB)

    Fischer, U., E-mail: ulrich.fischer@kit.edu [Association KIT-Euratom, Karlsruhe Institute of Technology, Hermann-von-Helmholtz-Platz 1, 76344 Eggenstein-Leopoldshafen (Germany); Angelone, M. [Associazione ENEA-Euratom, ENEA Fusion Division, Via E. Fermi 27, I-00044 Frascati (Italy); Bohm, T. [University of Wisconsin-Madison, 1500 Engineering Dr, Madison, WI 53706 (United States); Kondo, K. [Association KIT-Euratom, Karlsruhe Institute of Technology, Hermann-von-Helmholtz-Platz 1, 76344 Eggenstein-Leopoldshafen (Germany); Konno, C. [Japan Atomic Energy Agency, Tokai-mura, Naka-gun, Ibaraki-ken 319-1195 (Japan); Sawan, M. [University of Wisconsin-Madison, 1500 Engineering Dr, Madison, WI 53706 (United States); Villari, R. [Associazione ENEA-Euratom, ENEA Fusion Division, Via E. Fermi 27, I-00044 Frascati (Italy); Walker, B. [University of Wisconsin-Madison, 1500 Engineering Dr, Madison, WI 53706 (United States)

    2014-06-15

    This paper summarizes the benchmark analyses performed in a joint effort of ENEA (Italy), JAEA (Japan), KIT (Germany), and the University of Wisconsin (USA) on a computational ITER benchmark and a series of 14 MeV neutron benchmark experiments. The computational benchmark revealed a modest increase of the neutron flux levels in the deep penetration regions and a substantial increase of the gas production in steel components. The comparison to experimental results showed good agreement with no substantial differences between FENDL-3.0 and FENDL-2.1 for most of the responses. In general, FENDL-3 shows an improved performance for fusion neutronics applications.

  16. Benchmarking of the FENDL-3 Neutron Cross-section Data Starter Library for Fusion Applications

    International Nuclear Information System (INIS)

    This paper summarizes the benchmark analyses performed in a joint effort of ENEA (Italy), JAEA (Japan), KIT (Germany), and the University of Wisconsin (USA) on a computational ITER benchmark and a series of 14 MeV neutron benchmark experiments. The computational benchmark revealed a modest increase of the neutron flux levels in the deep penetration regions and a substantial increase of the gas production in steel components. The comparison to experimental results showed good agreement with no substantial differences between FENDL-3.0 and FENDL-2.1 for most of the responses. In general, FENDL-3 shows an improved performance for fusion neutronics applications

  17. Criticality safety benchmark evaluation project: Recovering the past

    Energy Technology Data Exchange (ETDEWEB)

    Trumble, E.F.

    1997-06-01

    A very brief summary of the Criticality Safety Benchmark Evaluation Project of the Westinghouse Savannah River Company is provided in this paper. The purpose of the project is to provide a source of evaluated criticality safety experiments in an easily usable format. Another project goal is to search for any experiments that may have been lost or contain discrepancies, and to determine if they can be used. Results of evaluated experiments are being published as US DOE handbooks.

  18. Benchmarking biodiversity performances of farmers

    NARCIS (Netherlands)

    Snoo, de G.R.; Lokhorst, A.M.; Dijk, van J.; Staats, H.; Musters, C.J.M.

    2010-01-01

    Farmers are the key players when it comes to the enhancement of farmland biodiversity. In this study, a benchmark system that focuses on improving farmers’ nature conservation was developed and tested among Dutch arable farmers in different social settings. The results show that especially tailored

  19. Nominal GDP: Target or Benchmark?

    OpenAIRE

    Hetzel, Robert L.

    2015-01-01

    Some observers have argued that the Federal Reserve would best fulfill its mandate by adopting a target for nominal gross domestic product (GDP). Insights from the monetarist tradition suggest that nominal GDP targeting could be destabilizing. However, adopting benchmarks for both nominal and real GDP could offer useful information about when monetary policy is too tight or too loose.

  20. Monte Carlo photon benchmark problems

    International Nuclear Information System (INIS)

    Photon benchmark calculations have been performed to validate the MCNP Monte Carlo computer code. These are compared to both the COG Monte Carlo computer code and either experimental or analytic results. The calculated solutions indicate that the Monte Carlo method, and MCNP and COG in particular, can accurately model a wide range of physical problems. 8 refs., 5 figs

  1. Benchmarked Library Websites Comparative Study

    KAUST Repository

    Ramli, Rindra M.

    2015-01-01

    This presentation provides an analysis of services provided by the benchmarked library websites. The exploratory study includes comparison of these websites against a list of criterion and presents a list of services that are most commonly deployed by the selected websites. In addition to that, the investigators proposed a list of services that could be provided via the KAUST library website.

  2. PRISMATIC CORE COUPLED TRANSIENT BENCHMARK

    Energy Technology Data Exchange (ETDEWEB)

    J. Ortensi; M.A. Pope; G. Strydom; R.S. Sen; M.D. DeHart; H.D. Gougar; C. Ellis; A. Baxter; V. Seker; T.J. Downar; K. Vierow; K. Ivanov

    2011-06-01

    The Prismatic Modular Reactor (PMR) is one of the High Temperature Reactor (HTR) design concepts that have existed for some time. Several prismatic units have operated in the world (DRAGON, Fort St. Vrain, Peach Bottom) and one unit is still in operation (HTTR). The deterministic neutronics and thermal-fluids transient analysis tools and methods currently available for the design and analysis of PMRs have lagged behind the state of the art compared to LWR reactor technologies. This has motivated the development of more accurate and efficient tools for the design and safety evaluations of the PMR. In addition to the work invested in new methods, it is essential to develop appropriate benchmarks to verify and validate the new methods in computer codes. The purpose of this benchmark is to establish a well-defined problem, based on a common given set of data, to compare methods and tools in core simulation and thermal hydraulics analysis with a specific focus on transient events. The benchmark-working group is currently seeking OECD/NEA sponsorship. This benchmark is being pursued and is heavily based on the success of the PBMR-400 exercise.

  3. National benchmarking against GLOBALGAP : Case studies of Good Agricultural Practices in Kenya, Malaysia, Mexico and Chile

    NARCIS (Netherlands)

    Valk, van der O.M.C.; Roest, van der J.G.

    2009-01-01

    This desk study examines the experiences and lessons learned from four case studies of countries aiming at the GLOBALGAP benchmarking procedure for national Good Agricultural Practices, namely Chile, Kenya, Malaysia, and Mexico. Aspects that determine the origin and character of the benchmarking pro

  4. Benchmarking in Czech Higher Education: The Case of Schools of Economics

    Science.gov (United States)

    Placek, Michal; Ochrana, František; Pucek, Milan

    2015-01-01

    This article describes the use of benchmarking in universities in the Czech Republic and academics' experiences with it. It is based on research conducted among academics from economics schools in Czech public and private universities. The results identified several issues regarding the utilisation and understanding of benchmarking in the Czech…

  5. How do I know if my forecasts are better? Using benchmarks in hydrological ensemble prediction

    Science.gov (United States)

    Pappenberger, F.; Ramos, M. H.; Cloke, H. L.; Wetterhall, F.; Alfieri, L.; Bogner, K.; Mueller, A.; Salamon, P.

    2015-03-01

    The skill of a forecast can be assessed by comparing the relative proximity of both the forecast and a benchmark to the observations. Example benchmarks include climatology or a naïve forecast. Hydrological ensemble prediction systems (HEPS) are currently transforming the hydrological forecasting environment but in this new field there is little information to guide researchers and operational forecasters on how benchmarks can be best used to evaluate their probabilistic forecasts. In this study, it is identified that the forecast skill calculated can vary depending on the benchmark selected and that the selection of a benchmark for determining forecasting system skill is sensitive to a number of hydrological and system factors. A benchmark intercomparison experiment is then undertaken using the continuous ranked probability score (CRPS), a reference forecasting system and a suite of 23 different methods to derive benchmarks. The benchmarks are assessed within the operational set-up of the European Flood Awareness System (EFAS) to determine those that are 'toughest to beat' and so give the most robust discrimination of forecast skill, particularly for the spatial average fields that EFAS relies upon. Evaluating against an observed discharge proxy the benchmark that has most utility for EFAS and avoids the most naïve skill across different hydrological situations is found to be meteorological persistency. This benchmark uses the latest meteorological observations of precipitation and temperature to drive the hydrological model. Hydrological long term average benchmarks, which are currently used in EFAS, are very easily beaten by the forecasting system and the use of these produces much naïve skill. When decomposed into seasons, the advanced meteorological benchmarks, which make use of meteorological observations from the past 20 years at the same calendar date, have the most skill discrimination. They are also good at discriminating skill in low flows and for all

  6. Collection of experimental data for fusion neutronics benchmark

    International Nuclear Information System (INIS)

    During the recent ten years or more, many benchmark experiments for fusion neutronics have been carried out at two principal D-T neutron sources, FNS at JAERI and OKTAVIAN at Osaka University, and precious experimental data have been accumulated. Under an activity of Fusion Reactor Physics Subcommittee of Reactor Physics Committee, these experimental data are compiled in this report. (author)

  7. Internal Quality Assurance Benchmarking. ENQA Workshop Report 20

    Science.gov (United States)

    Blackstock, Douglas; Burquel, Nadine; Comet, Nuria; Kajaste, Matti; dos Santos, Sergio Machado; Marcos, Sandra; Moser, Marion; Ponds, Henri; Scheuthle, Harald; Sixto, Luis Carlos Velon

    2012-01-01

    The Internal Quality Assurance group of ENQA (IQA Group) has been organising a yearly seminar for its members since 2007. The main objective is to share experiences concerning the internal quality assurance of work processes in the participating agencies. The overarching theme of the 2011 seminar was how to use benchmarking as a tool for…

  8. A Meta-Theory of Boundary Detection Benchmarks

    OpenAIRE

    Hou, Xiaodi; Yuille, Alan; Koch, Christof

    2012-01-01

    Human labeled datasets, along with their corresponding evaluation algorithms, play an important role in boundary detection. We here present a psychophysical experiment that addresses the reliability of such benchmarks. To find better remedies to evaluate the performance of any boundary detection algorithm, we propose a computational framework to remove inappropriate human labels and estimate the instrinsic properties of boundaries.

  9. A Uranium Bioremediation Reactive Transport Benchmark

    Energy Technology Data Exchange (ETDEWEB)

    Yabusaki, Steven B.; Sengor, Sevinc; Fang, Yilin

    2015-06-01

    A reactive transport benchmark problem set has been developed based on in situ uranium bio-immobilization experiments that have been performed at a former uranium mill tailings site in Rifle, Colorado, USA. Acetate-amended groundwater stimulates indigenous microorganisms to catalyze the reduction of U(VI) to a sparingly soluble U(IV) mineral. The interplay between the flow, acetate loading periods and rates, microbially-mediated and geochemical reactions leads to dynamic behavior in metal- and sulfate-reducing bacteria, pH, alkalinity, and reactive mineral surfaces. The benchmark is based on an 8.5 m long one-dimensional model domain with constant saturated flow and uniform porosity. The 159-day simulation introduces acetate and bromide through the upgradient boundary in 14-day and 85-day pulses separated by a 10 day interruption. Acetate loading is tripled during the second pulse, which is followed by a 50 day recovery period. Terminal electron accepting processes for goethite, phyllosilicate Fe(III), U(VI), and sulfate are modeled using Monod-type rate laws. Major ion geochemistry modeled includes mineral reactions, as well as aqueous and surface complexation reactions for UO2++, Fe++, and H+. In addition to the dynamics imparted by the transport of the acetate pulses, U(VI) behavior involves the interplay between bioreduction, which is dependent on acetate availability, and speciation-controlled surface complexation, which is dependent on pH, alkalinity and available surface complexation sites. The general difficulty of this benchmark is the large number of reactions (74), multiple rate law formulations, a multisite uranium surface complexation model, and the strong interdependency and sensitivity of the reaction processes. Results are presented for three simulators: HYDROGEOCHEM, PHT3D, and PHREEQC.

  10. Two benchmarks for qualification of pressure vessel fluence calculational methodology

    International Nuclear Information System (INIS)

    Two benchmarks for the qualification of the pressure vessel fluence calculational methodology were formulated and are briefly described. The Pool Critical Assembly (PCA) benchmark is based on the experiments performed at the PCA in Oak Ridge. The measured quantities to be compared against the calculated values are the equivalent fission fluxes at several locations in front, behind, and inside the pressure-vessel wall simulator. This benchmark is particularly suitable to test the capabilities of the calculational methodology and cross-section libraries to predict in-vessel gradients because only a few approximations are necessary in the analysis. The HBR-2 benchmark is based on the data for the H.B. Robinson-2 plant, which is a 2,300 MW (thermal) pressurized light-water reactor. The benchmark provides the reactor geometry, the material compositions, the core power distributions, and the power historical data. The quantities to be calculated are the specific activities of the radiometric monitors that were irradiated in the surveillance capsule and in the cavity location during one fuel cycle. The HBR-2 benchmark requires modeling approximations, power-to-neutron source conversion, and treatment of time dependant variations. It can therefore be used to test the overall performance and adequacy of the calculational methodology for power-reactor pressure-vessel flux calculations. Both benchmarks were analyzed with the DORT code and the BUGLE-96 cross-section library that is based on ENDF/B-VI evaluations. The calculations agreed with the measurements within 10%, and the calculations underpredicted the measurements in all the cases. This indicates that the ENDF/B-VI cross sections resolve most of the discrepancies between the measurements and calculations. The decrease of the CIM ratios with increased thickness of iron, which was typical for pre-ENDF/B-VI libraries, is almost completely removed

  11. BONFIRE: benchmarking computers and computer networks

    OpenAIRE

    Bouckaert, Stefan; Vanhie-Van Gerwen, Jono; Moerman, Ingrid; Phillips, Stephen; Wilander, Jerker

    2011-01-01

    The benchmarking concept is not new in the field of computing or computer networking. With “benchmarking tools”, one usually refers to a program or set of programs, used to evaluate the performance of a solution under certain reference conditions, relative to the performance of another solution. Since the 1970s, benchmarking techniques have been used to measure the performance of computers and computer networks. Benchmarking of applications and virtual machines in an Infrastructure-as-a-Servi...

  12. How Benchmarking and Higher Education Came Together

    Science.gov (United States)

    Levy, Gary D.; Ronco, Sharron L.

    2012-01-01

    This chapter introduces the concept of benchmarking and how higher education institutions began to use benchmarking for a variety of purposes. Here, benchmarking is defined as a strategic and structured approach whereby an organization compares aspects of its processes and/or outcomes to those of another organization or set of organizations to…

  13. A Framework for Urban Transport Benchmarking

    OpenAIRE

    Henning, Theuns; Essakali, Mohammed Dalil; Oh, Jung Eun

    2011-01-01

    This report summarizes the findings of a study aimed at exploring key elements of a benchmarking framework for urban transport. Unlike many industries where benchmarking has proven to be successful and straightforward, the multitude of the actors and interactions involved in urban transport systems may make benchmarking a complex endeavor. It was therefore important to analyze what has bee...

  14. Benchmarking: Achieving the best in class

    Energy Technology Data Exchange (ETDEWEB)

    Kaemmerer, L

    1996-05-01

    Oftentimes, people find the process of organizational benchmarking an onerous task, or, because they do not fully understand the nature of the process, end up with results that are less than stellar. This paper presents the challenges of benchmarking and reasons why benchmarking can benefit an organization in today`s economy.

  15. The LDBC Social Network Benchmark: Interactive Workload

    NARCIS (Netherlands)

    Erling, O.; Averbuch, A.; Larriba-Pey, J.; Chafi, H.; Gubichev, A.; Prat, A.; Pham, M.D.; Boncz, P.A.

    2015-01-01

    The Linked Data Benchmark Council (LDBC) is now two years underway and has gathered strong industrial participation for its mission to establish benchmarks, and benchmarking practices for evaluating graph data management systems. The LDBC introduced a new choke-point driven methodology for developin

  16. SARNET benchmark on QUENCH-11. Final report

    International Nuclear Information System (INIS)

    The QUENCH out-of-pile experiments at Forschungszentrum Karlsruhe (Karlsruhe Research Center) are set up to investigate the hydrogen source term that results from the water or steam injection into an uncovered core of a Light-Water Reactor, to examine the behavior of overheated fuel elements under different flooding conditions, and to create a database for model development and improvement of Severe Fuel Damage (SFD) code packages. The boil-off experiment QUENCH-11 was performed on December 8, 2005 as the second of two experiments in the frame of the EC-supported LACOMERA program. It was to simulate ceasing pumps in case of a small break LOCA or a station blackout with a late depressurization of the primary system, starting with boil-down of a test bundle that was partially filled with water. It is the first test to investigate the whole sequence of an anticipated reactor accident from the boil-off phase to delayed reflood of the bundle with a low water injection rate. The test is characterized by an interaction of thermal-hydraulics and material interactions that is even stronger than in previous QUENCH tests. It was proposed by INRNE Sofia (Bulgarian Academy of Sciences) and defined together with Forschungszentrum Karlsruhe. After the test, QUENCH-11 was chosen as a SARNET code benchmark exercise. Its task is a comparison between experimental data and analytical results to assess the reliability of the code prediction for different phases of an accident and the experiment. The SFD codes used were ASTEC, ATHLET-CD, ICARE-CATHARE, MELCOR, RATEG/SVECHA, RELAP/-SCDAPSIM, and SCDAP/RELAP5. The INRNE took responsibility as benchmark coordinator to compare the code results with the experimental data. As a basis of the present work, histories of temperatures, hydrogen production and other important variables were used. Besides, axial profiles at quench initiation and the final time of 7000 s, above all of temperatures, are presented. For most variables a mainstream of

  17. SARNET benchmark on QUENCH-11. Final report

    Energy Technology Data Exchange (ETDEWEB)

    Stefanova, A. [Bylgarska Akademiya na Naukite, Sofia (Bulgaria). Inst. for Nuclear Research and Nuclear Energy; Drath, T. [Ruhr-Univ. Bochum (Germany). Energy Systems and Energy Economics; Duspiva, J. [Nuclear Research Inst., Rez (CZ). Dept. of Reactor Technology] (and others)

    2008-03-15

    The QUENCH out-of-pile experiments at Forschungszentrum Karlsruhe (Karlsruhe Research Center) are set up to investigate the hydrogen source term that results from the water or steam injection into an uncovered core of a Light-Water Reactor, to examine the behavior of overheated fuel elements under different flooding conditions, and to create a database for model development and improvement of Severe Fuel Damage (SFD) code packages. The boil-off experiment QUENCH-11 was performed on December 8, 2005 as the second of two experiments in the frame of the EC-supported LACOMERA program. It was to simulate ceasing pumps in case of a small break LOCA or a station blackout with a late depressurization of the primary system, starting with boil-down of a test bundle that was partially filled with water. It is the first test to investigate the whole sequence of an anticipated reactor accident from the boil-off phase to delayed reflood of the bundle with a low water injection rate. The test is characterized by an interaction of thermal-hydraulics and material interactions that is even stronger than in previous QUENCH tests. It was proposed by INRNE Sofia (Bulgarian Academy of Sciences) and defined together with Forschungszentrum Karlsruhe. After the test, QUENCH-11 was chosen as a SARNET code benchmark exercise. Its task is a comparison between experimental data and analytical results to assess the reliability of the code prediction for different phases of an accident and the experiment. The SFD codes used were ASTEC, ATHLET-CD, ICARE-CATHARE, MELCOR, RATEG/SVECHA, RELAP/-SCDAPSIM, and SCDAP/RELAP5. The INRNE took responsibility as benchmark coordinator to compare the code results with the experimental data. As a basis of the present work, histories of temperatures, hydrogen production and other important variables were used. Besides, axial profiles at quench initiation and the final time of 7000 s, above all of temperatures, are presented. For most variables a mainstream of

  18. Critical benchmark results for a modified 16O evaluation

    International Nuclear Information System (INIS)

    The effect of a uniform reduction in the elastic scattering cross-section for 16O on critical benchmarks is quantified and discussed. It is hypothesised that current evaluations for 16O systematically overestimate elastic scattering by about 3% due to a normalisation error in various experimental data. Selected critical benchmarks from the HEU-SOL-THERM (HST) series of the International Handbook of Evaluated Criticality Safety Benchmark Experiments were simulated using the MC21 Monte Carlo code. The benchmark results show that a decrease in the elastic scattering cross-section to agree with high-precision experimental measurements leads to higher leakage and lower benchmark eigenvalues. Additionally, a trend with the above-thermal leakage fraction was observed. The sensitivity of this trend to the first Legendre polynomial coefficient of the elastic scattering angular distribution was calculated. Based on the observed sensitivity, a 35% decrease in the first-order Legendre polynomial coefficient would be required to eliminate the trend with above-thermal leakage fraction. (authors)

  19. Methodical aspects of benchmarking using in Consumer Cooperatives trade enterprises activity

    Directory of Open Access Journals (Sweden)

    Yu.V. Dvirko

    2013-03-01

    benchmarking; by direction vertical, horizontal, inner, combined, outer benchmarking; by type common, functional, branching, interbranching, competitive benchmarking. Despite the drawbacks of informal types of benchmarking, on our opinion the most expedient idea in Consumer Cooperatives trade enterprises activity is using the formalized benchmarking unions. Among them we should point out the following ones: partnership, cooperative, associative and institutional benchmarking. Partnership benchmarking cooperation where relations are main condition, which are oriented upon current problems decision taking into account various practical experience of other subject of benchmarking activity. Cooperative benchmarking is comparative analysis of key factors of enterprise success, which is considered to be ideal partner with the purpose of qualitative experience of business running borrowing. Associative benchmarking it is benchmarking conducted by enterprises which are in the united marketing alliance. Corporative benchmarking forms, under which we understand the process of market indicators improvement in the activity of several enterprises at the basis of partnership relations creation while exchanging and distributing qualitative experience. Institutional benchmarking foresees appellations for the help conducting appropriate events to the third party, we follow the mechanism of outsorting implementation.Conclusions and directions of further researches. Thus, the article defines goals, advantages and drawbacks of benchmarking using; there is given main benchmarking types classification; there is flashlighted authors understanding of main forms of Consumer Cooperatives trade enterprises benchmarking organization. The using of the suggested theoretical principles in practice will bring positive changes as to development of cooperative trade enterprises, the concept of which doesnt coincide with tendencies and peculiarities of market relations.

  20. A proposal to Asian countries with operating research reactors for making nuclear criticality safety benchmark evaluations

    International Nuclear Information System (INIS)

    The Criticality Safety Benchmark Evaluation Project (CSBEP) was initiated in 1992 by the United States Department of Energy. The project quickly became an international effort as scientists from other interested countries became involved. The International Criticality Safety Benchmark Evaluation Project (ICSBEP) is now an official activity of the Organization for Economic Cooperation and Development-Nuclear Energy Agency (OECD-NEA). 'International Handbook of Criticality Safety Benchmark Experiments' was prepared and is updated yearly by the working group of the project. This handbook contains criticality safety benchmark specifications that have been derived from experiments that were performed at various nuclear criticality facilities around the world. However, the handbook lacks criticality data of 20 wt%-enriched uranium fuel. The author proposes to make benchmark specifications derived from modern research reactors in Asia. Future evaluations of these reactors will facilitate to fill the 'enrichment gap'. (author)

  1. Methodology for Benchmarking IPsec Gateways

    Directory of Open Access Journals (Sweden)

    Adam Tisovský

    2012-08-01

    Full Text Available The paper analyses forwarding performance of IPsec gateway over the rage of offered loads. It focuses on the forwarding rate and packet loss particularly at the gateway’s performance peak and at the state of gateway’s overload. It explains possible performance degradation when the gateway is overloaded by excessive offered load. The paper further evaluates different approaches for obtaining forwarding performance parameters – a widely used throughput described in RFC 1242, maximum forwarding rate with zero packet loss and us proposed equilibrium throughput. According to our observations equilibrium throughput might be the most universal parameter for benchmarking security gateways as the others may be dependent on the duration of test trials. Employing equilibrium throughput would also greatly shorten the time required for benchmarking. Lastly, the paper presents methodology and a hybrid step/binary search algorithm for obtaining value of equilibrium throughput.

  2. Adapting benchmarking to project management : an analysis of project management processes, metrics, and benchmarking process models

    OpenAIRE

    Emhjellen, Kjetil

    1997-01-01

    Since the first publication on benchmarking in 1989 by Robert C. Camp of “Benchmarking: The search for Industry Best Practices that Lead to Superior Performance”, the improvement technique benchmarking has been established as an important tool in the process focused manufacturing or production environment. The use of benchmarking has expanded to other types of industry. Benchmarking has past the doorstep and is now in early trials in the project and construction environment....

  3. HS06 benchmark for an ARM server

    International Nuclear Information System (INIS)

    We benchmarked an ARM cortex-A9 based server system with a four-core CPU running at 1.1 GHz. The system used Ubuntu 12.04 as operating system and the HEPSPEC 2006 (HS06) benchmarking suite was compiled natively with gcc-4.4 on the system. The benchmark was run for various settings of the relevant gcc compiler options. We did not find significant influence from the compiler options on the benchmark result. The final HS06 benchmark result is 10.4.

  4. Restaurant Energy Use Benchmarking Guideline

    Energy Technology Data Exchange (ETDEWEB)

    Hedrick, R.; Smith, V.; Field, K.

    2011-07-01

    A significant operational challenge for food service operators is defining energy use benchmark metrics to compare against the performance of individual stores. Without metrics, multiunit operators and managers have difficulty identifying which stores in their portfolios require extra attention to bring their energy performance in line with expectations. This report presents a method whereby multiunit operators may use their own utility data to create suitable metrics for evaluating their operations.

  5. Local Innovation Systems and Benchmarking

    OpenAIRE

    Cantner, Uwe

    2008-01-01

    This paper reviews approaches used for evaluating the performance of local or regional innovation systems. This evaluation is performed by a benchmarking approach in which a frontier production function can be determined, based on a knowledge production function relating innovation inputs and innovation outputs. In analyses on the regional level and especially when acknowledging regional innovation systems those approaches have to take into account cooperative invention and innovation - the c...

  6. RISKIND verification and benchmark comparisons

    International Nuclear Information System (INIS)

    This report presents verification calculations and benchmark comparisons for RISKIND, a computer code designed to estimate potential radiological consequences and health risks to individuals and the population from exposures associated with the transportation of spent nuclear fuel and other radioactive materials. Spreadsheet calculations were performed to verify the proper operation of the major options and calculational steps in RISKIND. The program is unique in that it combines a variety of well-established models into a comprehensive treatment for assessing risks from the transportation of radioactive materials. Benchmark comparisons with other validated codes that incorporate similar models were also performed. For instance, the external gamma and neutron dose rate curves for a shipping package estimated by RISKIND were compared with those estimated by using the RADTRAN 4 code and NUREG-0170 methodology. Atmospheric dispersion of released material and dose estimates from the GENII and CAP88-PC codes. Verification results have shown the program to be performing its intended function correctly. The benchmark results indicate that the predictions made by RISKIND are within acceptable limits when compared with predictions from similar existing models

  7. Thermal Performance Benchmarking: Annual Report

    Energy Technology Data Exchange (ETDEWEB)

    Moreno, Gilbert

    2016-04-08

    The goal for this project is to thoroughly characterize the performance of state-of-the-art (SOA) automotive power electronics and electric motor thermal management systems. Information obtained from these studies will be used to: Evaluate advantages and disadvantages of different thermal management strategies; establish baseline metrics for the thermal management systems; identify methods of improvement to advance the SOA; increase the publicly available information related to automotive traction-drive thermal management systems; help guide future electric drive technologies (EDT) research and development (R&D) efforts. The performance results combined with component efficiency and heat generation information obtained by Oak Ridge National Laboratory (ORNL) may then be used to determine the operating temperatures for the EDT components under drive-cycle conditions. In FY15, the 2012 Nissan LEAF power electronics and electric motor thermal management systems were benchmarked. Testing of the 2014 Honda Accord Hybrid power electronics thermal management system started in FY15; however, due to time constraints it was not possible to include results for this system in this report. The focus of this project is to benchmark the thermal aspects of the systems. ORNL's benchmarking of electric and hybrid electric vehicle technology reports provide detailed descriptions of the electrical and packaging aspects of these automotive systems.

  8. Prismatic VHTR neutronic benchmark problems

    Energy Technology Data Exchange (ETDEWEB)

    Connolly, Kevin John, E-mail: connolly@gatech.edu [Nuclear and Radiological Engineering and Medical Physics Programs, George W. Woodruff School, Georgia Institute of Technology, Atlanta, GA (United States); Rahnema, Farzad, E-mail: farzad@gatech.edu [Nuclear and Radiological Engineering and Medical Physics Programs, George W. Woodruff School, Georgia Institute of Technology, Atlanta, GA (United States); Tsvetkov, Pavel V. [Department of Nuclear Engineering, Texas A& M University, College Station, TX (United States)

    2015-04-15

    Highlights: • High temperature gas-cooled reactor neutronics benchmark problems. • Description of a whole prismatic VHTR core in its full heterogeneity. • Modeled using continuous energy nuclear data at a representative hot operating temperature. • Benchmark results for core eigenvalue, block-averaged power, and some selected pin fission density results. - Abstract: This paper aims to fill an apparent scarcity of benchmarks based on high temperature gas-cooled reactors. Within is a description of a whole prismatic VHTR core in its full heterogeneity and modeling using continuous energy nuclear data at a representative hot operating temperature. Also included is a core which has been simplified for ease in modeling while attempting to preserve as faithfully as possible the neutron physics of the core. Fuel and absorber pins have been homogenized from the particle level, however, the blocks which construct the core remain strongly heterogeneous. A six group multigroup (discrete energy) cross section set has been developed via Monte Carlo using the original heterogeneous core as a basis. Several configurations of the core have been solved using these two cross section sets; eigenvalue results, block-averaged power results, and some selected pin fission density results are presented in this paper, along with the six-group cross section data, so that method developers may use these problems as a standard reference point.

  9. Prismatic VHTR neutronic benchmark problems

    International Nuclear Information System (INIS)

    Highlights: • High temperature gas-cooled reactor neutronics benchmark problems. • Description of a whole prismatic VHTR core in its full heterogeneity. • Modeled using continuous energy nuclear data at a representative hot operating temperature. • Benchmark results for core eigenvalue, block-averaged power, and some selected pin fission density results. - Abstract: This paper aims to fill an apparent scarcity of benchmarks based on high temperature gas-cooled reactors. Within is a description of a whole prismatic VHTR core in its full heterogeneity and modeling using continuous energy nuclear data at a representative hot operating temperature. Also included is a core which has been simplified for ease in modeling while attempting to preserve as faithfully as possible the neutron physics of the core. Fuel and absorber pins have been homogenized from the particle level, however, the blocks which construct the core remain strongly heterogeneous. A six group multigroup (discrete energy) cross section set has been developed via Monte Carlo using the original heterogeneous core as a basis. Several configurations of the core have been solved using these two cross section sets; eigenvalue results, block-averaged power results, and some selected pin fission density results are presented in this paper, along with the six-group cross section data, so that method developers may use these problems as a standard reference point

  10. An introduction to benchmarking in healthcare.

    Science.gov (United States)

    Benson, H R

    1994-01-01

    Benchmarking--the process of establishing a standard of excellence and comparing a business function or activity, a product, or an enterprise as a whole with that standard--will be used increasingly by healthcare institutions to reduce expenses and simultaneously improve product and service quality. As a component of total quality management, benchmarking is a continuous process by which an organization can measure and compare its own processes with those of organizations that are leaders in a particular area. Benchmarking should be viewed as a part of quality management programs, not as a replacement. There are four kinds of benchmarking: internal, competitive, functional and generic. With internal benchmarking, functions within an organization are compared with each other. Competitive benchmarking partners do business in the same market and provide a direct comparison of products or services. Functional and generic benchmarking are performed with organizations which may have a specific similar function, such as payroll or purchasing, but which otherwise are in a different business. Benchmarking must be a team process because the outcome will involve changing current practices, with effects felt throughout the organization. The team should include members who have subject knowledge; communications and computer proficiency; skills as facilitators and outside contacts; and sponsorship of senior management. Benchmarking requires quantitative measurement of the subject. The process or activity that you are attempting to benchmark will determine the types of measurements used. Benchmarking metrics usually can be classified in one of four categories: productivity, quality, time and cost-related. PMID:10139084

  11. Isospin-Violating Dark Matter Benchmarks for Snowmass 2013

    CERN Document Server

    Feng, Jonathan L; Marfatia, Danny; Sanford, David

    2013-01-01

    Isospin-violating dark matter (IVDM) generalizes the standard spin-independent scattering parameter space by introducing one additional parameter, the neutron-to-proton coupling ratio f_n/f_p. In IVDM the implications of direct detection experiments can be altered significantly. We review the motivations for considering IVDM and present benchmark models that illustrate some of the qualitatively different possibilities. IVDM strongly motivates the use of a variety of target nuclei in direct detection experiments.

  12. Analytical benchmarks for the kinetics of accelerator-driven systems

    International Nuclear Information System (INIS)

    The work presented in this paper illustrates the analytical benchmark philosophy with applications to subcritical source-driven system dynamics. Results for different complexity problems are presented in the frame of multigroup diffusion theory. The analysis performed on systems having physical characteristics typical of the Yalina Booster experiment are presented. The comparisons with the results of numerical calculations enlighten the convergence trend of discretized schemes and the limits of applicability to the analysis of experiments. (authors)

  13. Benchmark Evaluation of Start-Up and Zero-Power Measurements at the High-Temperature Engineering Test Reactor

    International Nuclear Information System (INIS)

    Benchmark models were developed to evaluate six cold-critical and two warm-critical, zero-power measurements of the HTTR. Additional measurements of a fully-loaded subcritical configuration, core excess reactivity, shutdown margins, six isothermal temperature coefficients, and axial reaction-rate distributions were also evaluated as acceptable benchmark experiments. Insufficient information is publicly available to develop finely-detailed models of the HTTR as much of the design information is still proprietary. However, the uncertainties in the benchmark models are judged to be of sufficient magnitude to encompass any biases and bias uncertainties incurred through the simplification process used to develop the benchmark models. Dominant uncertainties in the experimental keff for all core configurations come from uncertainties in the impurity content of the various graphite blocks that comprise the HTTR. Monte Carlo calculations of keff are between approximately 0.9 % and 2.7 % greater than the benchmark values. Reevaluation of the HTTR models as additional information becomes available could improve the quality of this benchmark and possibly reduce the computational biases. High-quality characterization of graphite impurities would significantly improve the quality of the HTTR benchmark assessment. Simulation of the other reactor physics measurements are in good agreement with the benchmark experiment values. The complete benchmark evaluation details are available in the 2014 edition of the International Handbook of Evaluated Reactor Physics Benchmark Experiments

  14. Benchmark on the Venus-2 MOX core measurements

    International Nuclear Information System (INIS)

    In the framework of joint activities carried out by the OECD/NEA Working Party on the Physics of Plutonium Fuels and Innovative Fuel Cycles (WPPR) and the Task Force on Reactor-based Plutonium Disposition (TFRPD), a blind international benchmark exercise for the prediction of power distribution in the two-dimensional VENUS-2 MOX core experiment was launched in May 1999. The aim of this exercise was to investigate the predictive capability of the current production codes with the latest nuclear data libraries used for calculating MOX-fuelled systems. Twelve participants from ten countries participated in this first international experiment-based benchmark. This report provides details of the comparative analysis between calculated and measured results. Comparison with experimental results identified the origins of discrepancies between calculations and measurements and enabled the quantitative comparison of the relative merits of the different.calculation methods. As the benchmark was two-dimensional, a benchmark based on three dimensional VENUS-2 experimental results will be launched for a more thorough investigation into the calculation methods used for MOX-fuelled systems. (author)

  15. Benchmarks

    Data.gov (United States)

    Earth Data Analysis Center, University of New Mexico — The National Flood Hazard Layer (NFHL) data incorporates all Digital Flood Insurance Rate Map(DFIRM) databases published by FEMA, and any Letters Of Map Revision...

  16. PageRank Pipeline Benchmark: Proposal for a Holistic System Benchmark for Big-Data Platforms

    CERN Document Server

    Dreher, Patrick; Hill, Chris; Gadepally, Vijay; Kuszmaul, Bradley; Kepner, Jeremy

    2016-01-01

    The rise of big data systems has created a need for benchmarks to measure and compare the capabilities of these systems. Big data benchmarks present unique scalability challenges. The supercomputing community has wrestled with these challenges for decades and developed methodologies for creating rigorous scalable benchmarks (e.g., HPC Challenge). The proposed PageRank pipeline benchmark employs supercomputing benchmarking methodologies to create a scalable benchmark that is reflective of many real-world big data processing systems. The PageRank pipeline benchmark builds on existing prior scalable benchmarks (Graph500, Sort, and PageRank) to create a holistic benchmark with multiple integrated kernels that can be run together or independently. Each kernel is well defined mathematically and can be implemented in any programming environment. The linear algebraic nature of PageRank makes it well suited to being implemented using the GraphBLAS standard. The computations are simple enough that performance predictio...

  17. MC benchmarks for GERDA LAr veto designs

    International Nuclear Information System (INIS)

    The Germanium Detector Array (GERDA) experiment is designed to search for neutrinoless beta decay in 76Ge and is able to directly test the present claim by parts of the Heidelberg-Moscow Collaboration. The experiment started recently its first physics phase with eight enriched detectors, after a 17 month long commissioning period. GERDA operates an array of HPGe detectors in liquid argon (LAr), which acts both as a shield for external backgrounds and as a cryogenic cooling. Furthermore, LAr has the potential to be instrumented and therefore be used as an active veto for background events through the detection of the produced scintillation light. In this talk, Monte Carlo studies for benchmarking and optimizing different LAr veto designs will be presented. LAr scintillates at 128 nm which, combined with the cryogenic temperature in which the detector is operated and its optical properties, poses many challenges in the design of an efficient veto that would help the experiment to reduce the total background level by one order of magnitude, as it is the goal for the second physics phase of the experiment.

  18. Shielding benchmark test of JENDL-3

    Energy Technology Data Exchange (ETDEWEB)

    Kawai, M. (Toshiba Corp., Kawasaki, Kanagawa (Japan)); Hasegawa, A.; Ueki, K. (and others)

    1990-02-01

    The integral test of JENDL-3 for shielding application was made on a cross sections of carbon, sodium and iron by analyzing the various shielding benchmark experiments: Broomstick experiments at ORNL for iron and sodium, neutron transmission experiments for sodium at ORNL, iron and carbon, ASPIS deep penetration experiments for iron, measurements of leakage spectrum from iron spheres at KfK, angular neutron spectrum measurements in graphite block at RPI. Analyses were made with radiation transport codes ANISN (1D, Sn), DIAC (1D, Sn), DOT-3.5 (2D, Sn) and MCNP (3D, point Monte Carlo). It was observed that revising JENDL-3T iron data resulted in an improvement in reproducing the experimental data, particularly in the MeV neutron energy region. For sodium, JENDL-3 gave better results than JENDL-2 and ENDF/B-IV. For carbon, JENDL-3 gave better agreement, compared to ENDF/B-IV. The conclusion is that JENDL-3 is highly applicable to the reactor shielding analyses. (author).

  19. Gaia FGK benchmark stars: Metallicity

    Science.gov (United States)

    Jofré, P.; Heiter, U.; Soubiran, C.; Blanco-Cuaresma, S.; Worley, C. C.; Pancino, E.; Cantat-Gaudin, T.; Magrini, L.; Bergemann, M.; González Hernández, J. I.; Hill, V.; Lardo, C.; de Laverny, P.; Lind, K.; Masseron, T.; Montes, D.; Mucciarelli, A.; Nordlander, T.; Recio Blanco, A.; Sobeck, J.; Sordo, R.; Sousa, S. G.; Tabernero, H.; Vallenari, A.; Van Eck, S.

    2014-04-01

    Context. To calibrate automatic pipelines that determine atmospheric parameters of stars, one needs a sample of stars, or "benchmark stars", with well-defined parameters to be used as a reference. Aims: We provide detailed documentation of the iron abundance determination of the 34 FGK-type benchmark stars that are selected to be the pillars for calibration of the one billion Gaia stars. They cover a wide range of temperatures, surface gravities, and metallicities. Methods: Up to seven different methods were used to analyze an observed spectral library of high resolutions and high signal-to-noise ratios. The metallicity was determined by assuming a value of effective temperature and surface gravity obtained from fundamental relations; that is, these parameters were known a priori and independently from the spectra. Results: We present a set of metallicity values obtained in a homogeneous way for our sample of benchmark stars. In addition to this value, we provide detailed documentation of the associated uncertainties. Finally, we report a value of the metallicity of the cool giant ψ Phe for the first time. Based on NARVAL and HARPS data obtained within the Gaia DPAC (Data Processing and Analysis Consortium) and coordinated by the GBOG (Ground-Based Observations for Gaia) working group and on data retrieved from the ESO-ADP database.Tables 6-76 are only available at the CDS via anonymous ftp to http://cdsarc.u-strasbg.fr (ftp://130.79.128.5) or via http://cdsarc.u-strasbg.fr/viz-bin/qcat?J/A+A/564/A133

  20. NASA Software Engineering Benchmarking Effort

    Science.gov (United States)

    Godfrey, Sally; Rarick, Heather

    2012-01-01

    Benchmarking was very interesting and provided a wealth of information (1) We did see potential solutions to some of our "top 10" issues (2) We have an assessment of where NASA stands with relation to other aerospace/defense groups We formed new contacts and potential collaborations (1) Several organizations sent us examples of their templates, processes (2) Many of the organizations were interested in future collaboration: sharing of training, metrics, Capability Maturity Model Integration (CMMI) appraisers, instructors, etc. We received feedback from some of our contractors/ partners (1) Desires to participate in our training; provide feedback on procedures (2) Welcomed opportunity to provide feedback on working with NASA

  1. Re-evaluation and continued development of shielding benchmark database SINBAD

    International Nuclear Information System (INIS)

    The Expert Group on Radiation Transport and Shielding was started at the OECD/NEA in June 2011 with the mandate to, among others, monitor, steer and support the continued development of the Shielding Integral Benchmark Archive and Database (SINBAD), in cooperation with RSICC. The present status of this database is presented, which grew since its beginnings in the 1990’s to an international reference database containing at present 100 benchmarks relevant for fission, fusion and accelerator shielding applications. As part of its activities a thorough revision of 45 benchmark experiments was completed recently in order to verify in details the completeness and consistency of the benchmark information, in particular concerning the evaluation of the experimental sources of uncertainty. This review process is expected to provide users with a proper choice and help them make better use of the experimental information and is planned to be extended to other available benchmarks. (author)

  2. The ORSphere Benchmark Evaluation and Its Potential Impact on Nuclear Criticality Safety

    Energy Technology Data Exchange (ETDEWEB)

    John D. Bess; Margaret A. Marshall; J. Blair Briggs

    2013-10-01

    In the early 1970’s, critical experiments using an unreflected metal sphere of highly enriched uranium (HEU) were performed with the focus to provide a “very accurate description…as an ideal benchmark for calculational methods and cross-section data files.” Two near-critical configurations of the Oak Ridge Sphere (ORSphere) were evaluated as acceptable benchmark experiments for inclusion in the International Handbook of Evaluated Criticality Safety Benchmark Experiments (ICSBEP Handbook). The results from those benchmark experiments were then compared with additional unmoderated and unreflected HEU metal benchmark experiment configurations currently found in the ICSBEP Handbook. For basic geometries (spheres, cylinders, and slabs) the eigenvalues calculated using MCNP5 and ENDF/B-VII.0 were within 3 of their respective benchmark values. There appears to be generally a good agreement between calculated and benchmark values for spherical and slab geometry systems. Cylindrical geometry configurations tended to calculate low, including more complex bare HEU metal systems containing cylinders. The ORSphere experiments do not calculate within their 1s uncertainty and there is a possibility that the effect of the measured uncertainties for the GODIVA I benchmark may need reevaluated. There is significant scatter in the calculations for the highly-correlated ORCEF cylinder experiments, which are constructed from close-fitting HEU discs and annuli. Selection of a nuclear data library can have a larger impact on calculated eigenvalue results than the variation found within calculations of a given experimental series, such as the ORCEF cylinders, using a single nuclear data set.

  3. A Privacy-Preserving Benchmarking Platform

    OpenAIRE

    Kerschbaum, Florian

    2010-01-01

    A privacy-preserving benchmarking platform is practically feasible, i.e. its performance is tolerable to the user on current hardware while fulfilling functional and security requirements. This dissertation designs, architects, and evaluates an implementation of such a platform. It contributes a novel (secure computation) benchmarking protocol, a novel method for computing peer groups, and a realistic evaluation of the first ever privacy-preserving benchmarking platform.

  4. Rethinking benchmark dates in international relations

    OpenAIRE

    Buzan, Barry; Lawson, George

    2014-01-01

    International Relations (IR) has an ‘orthodox set’ of benchmark dates by which much of its research and teaching is organized: 1500, 1648, 1919, 1945 and 1989. This article argues that IR scholars need to question the ways in which these orthodox dates serve as internal and external points of reference, think more critically about how benchmark dates are established, and generate a revised set of benchmark dates that better reflects macro-historical international dynamics. The first part of t...

  5. WIPP benchmark II results using SANCHO

    International Nuclear Information System (INIS)

    Results of the second Benchmark problem in the WIPP code evaluation series using the finite element dynamic relaxation code SANCHO are presented. A description of SANCHO and its model for sliding interfaces is given, along with a discussion of the various small routines used for generating stress plot data. Conclusions and a discussion of this benchmark problem, as well as recommendations for a possible third benchmark problem are presented

  6. Benchmarking for Excellence and the Nursing Process

    Science.gov (United States)

    Sleboda, Claire

    1999-01-01

    Nursing is a service profession. The services provided are essential to life and welfare. Therefore, setting the benchmark for high quality care is fundamental. Exploring the definition of a benchmark value will help to determine a best practice approach. A benchmark is the descriptive statement of a desired level of performance against which quality can be judged. It must be sufficiently well understood by managers and personnel in order that it may serve as a standard against which to measure value.

  7. Computational Chemistry Comparison and Benchmark Database

    Science.gov (United States)

    SRD 101 NIST Computational Chemistry Comparison and Benchmark Database (Web, free access)   The NIST Computational Chemistry Comparison and Benchmark Database is a collection of experimental and ab initio thermochemical properties for a selected set of molecules. The goals are to provide a benchmark set of molecules for the evaluation of ab initio computational methods and allow the comparison between different ab initio computational methods for the prediction of thermochemical properties.

  8. Pynamic: the Python Dynamic Benchmark

    Energy Technology Data Exchange (ETDEWEB)

    Lee, G L; Ahn, D H; de Supinksi, B R; Gyllenhaal, J C; Miller, P J

    2007-07-10

    Python is widely used in scientific computing to facilitate application development and to support features such as computational steering. Making full use of some of Python's popular features, which improve programmer productivity, leads to applications that access extremely high numbers of dynamically linked libraries (DLLs). As a result, some important Python-based applications severely stress a system's dynamic linking and loading capabilities and also cause significant difficulties for most development environment tools, such as debuggers. Furthermore, using the Python paradigm for large scale MPI-based applications can create significant file IO and further stress tools and operating systems. In this paper, we present Pynamic, the first benchmark program to support configurable emulation of a wide-range of the DLL usage of Python-based applications for large scale systems. Pynamic has already accurately reproduced system software and tool issues encountered by important large Python-based scientific applications on our supercomputers. Pynamic provided insight for our system software and tool vendors, and our application developers, into the impact of several design decisions. As we describe the Pynamic benchmark, we will highlight some of the issues discovered in our large scale system software and tools using Pynamic.

  9. Method and system for benchmarking computers

    Science.gov (United States)

    Gustafson, John L.

    1993-09-14

    A testing system and method for benchmarking computer systems. The system includes a store containing a scalable set of tasks to be performed to produce a solution in ever-increasing degrees of resolution as a larger number of the tasks are performed. A timing and control module allots to each computer a fixed benchmarking interval in which to perform the stored tasks. Means are provided for determining, after completion of the benchmarking interval, the degree of progress through the scalable set of tasks and for producing a benchmarking rating relating to the degree of progress for each computer.

  10. Characterizing universal gate sets via dihedral benchmarking

    Science.gov (United States)

    Carignan-Dugas, Arnaud; Wallman, Joel J.; Emerson, Joseph

    2015-12-01

    We describe a practical experimental protocol for robustly characterizing the error rates of non-Clifford gates associated with dihedral groups, including small single-qubit rotations. Our dihedral benchmarking protocol is a generalization of randomized benchmarking that relaxes the usual unitary 2-design condition. Combining this protocol with existing randomized benchmarking schemes enables practical universal gate sets for quantum information processing to be characterized in a way that is robust against state-preparation and measurement errors. In particular, our protocol enables direct benchmarking of the π /8 gate even under the gate-dependent error model that is expected in leading approaches to fault-tolerant quantum computation.

  11. The Zoo, Benchmarks & You: How To Reach the Oregon State Benchmarks with Zoo Resources.

    Science.gov (United States)

    2002

    This document aligns Oregon state educational benchmarks and standards with Oregon Zoo resources. Benchmark areas examined include English, mathematics, science, social studies, and career and life roles. Brief descriptions of the programs offered by the zoo are presented. (SOE)

  12. Benchmarking Implementations of Functional Languages with ``Pseudoknot'', a Float-Intensive Benchmark

    NARCIS (Netherlands)

    Hartel, P.H.; Feeley, M.; Alt, M.; Augustsson, L.

    1996-01-01

    Over 25 implementations of different functional languages are benchmarked using the same program, a floatingpoint intensive application taken from molecular biology. The principal aspects studied are compile time and execution time for the various implementations that were benchmarked. An important

  13. Shielding benchmark tests of JENDL-3

    International Nuclear Information System (INIS)

    The integral test of neutron cross sections for major shielding materials in JENDL-3 has been performed by analyzing various shielding benchmark experiments. For the fission-like neutron source problem, the following experiments are analyzed: (1) ORNL Broomstick experiments for oxygen, iron and sodium, (2) ASPIS deep penetration experiments for iron, (3) ORNL neutron transmission experiments for iron, stainless steel, sodium and graphite, (4) KfK leakage spectrum measurements from iron spheres, (5) RPI angular neutron spectrum measurements in a graphite block. For D-T neutron source problem, the following two experiments are analyzed: (6) LLNL leakage spectrum measurements from spheres of iron and graphite, and (7) JAERI-FNS angular neutron spectrum measurements on beryllium and graphite slabs. Analyses have been performed using the radiation transport codes: ANISN(1D Sn), DIAC(1D Sn), DOT3.5(2D Sn) and MCNP(3D point Monte Carlo). The group cross sections for Sn transport calculations are generated with the code systems PROF-GROUCH-G/B and RADHEAT-V4. The point-wise cross sections for MCNP are produced with NJOY. For comparison, the analyses with JENDL-2 and ENDF/B-IV have been also carried out. The calculations using JENDL-3 show overall agreement with the experimental data as well as those with ENDF/B-IV. Particularly, JENDL-3 gives better results than JENDL-2 and ENDF/B-IV for sodium. It has been concluded that JENDL-3 is very applicable for fission and fusion reactor shielding analyses. (author)

  14. Shielding benchmark tests of JENDL-3

    Energy Technology Data Exchange (ETDEWEB)

    Kawai, Masayoshi [Toshiba Corp., Kawasaki, Kanagawa (Japan); Hasegawa, Akira; Ueki, Kohtaro; Yamano, Naoki; Sasaki, Kenji; Matsumoto, Yoshihiro; Takemura, Morio; Ohtani, Nobuo; Sakurai, Kiyoshi

    1994-03-01

    The integral test of neutron cross sections for major shielding materials in JENDL-3 has been performed by analyzing various shielding benchmark experiments. For the fission-like neutron source problem, the following experiments are analyzed: (1) ORNL Broomstick experiments for oxygen, iron and sodium, (2) ASPIS deep penetration experiments for iron, (3) ORNL neutron transmission experiments for iron, stainless steel, sodium and graphite, (4) KfK leakage spectrum measurements from iron spheres, (5) RPI angular neutron spectrum measurements in a graphite block. For D-T neutron source problem, the following two experiments are analyzed: (6) LLNL leakage spectrum measurements from spheres of iron and graphite, and (7) JAERI-FNS angular neutron spectrum measurements on beryllium and graphite slabs. Analyses have been performed using the radiation transport codes: ANISN(1D Sn), DIAC(1D Sn), DOT3.5(2D Sn) and MCNP(3D point Monte Carlo). The group cross sections for Sn transport calculations are generated with the code systems PROF-GROUCH-G/B and RADHEAT-V4. The point-wise cross sections for MCNP are produced with NJOY. For comparison, the analyses with JENDL-2 and ENDF/B-IV have been also carried out. The calculations using JENDL-3 show overall agreement with the experimental data as well as those with ENDF/B-IV. Particularly, JENDL-3 gives better results than JENDL-2 and ENDF/B-IV for sodium. It has been concluded that JENDL-3 is very applicable for fission and fusion reactor shielding analyses. (author).

  15. Benchmarking: A tool to enhance performance

    Energy Technology Data Exchange (ETDEWEB)

    Munro, J.F. [Oak Ridge National Lab., TN (United States); Kristal, J. [USDOE Assistant Secretary for Environmental Management, Washington, DC (United States); Thompson, G.; Johnson, T. [Los Alamos National Lab., NM (United States)

    1996-12-31

    The Office of Environmental Management is bringing Headquarters and the Field together to implement process improvements throughout the Complex through a systematic process of organizational learning called benchmarking. Simply stated, benchmarking is a process of continuously comparing and measuring practices, processes, or methodologies with those of other private and public organizations. The EM benchmarking program, which began as the result of a recommendation from Xerox Corporation, is building trust and removing barriers to performance enhancement across the DOE organization. The EM benchmarking program is designed to be field-centered with Headquarters providing facilitatory and integrative functions on an ``as needed`` basis. One of the main goals of the program is to assist Field Offices and their associated M&O/M&I contractors develop the capabilities to do benchmarking for themselves. In this regard, a central precept is that in order to realize tangible performance benefits, program managers and staff -- the ones closest to the work - must take ownership of the studies. This avoids the ``check the box`` mentality associated with some third party studies. This workshop will provide participants with a basic level of understanding why the EM benchmarking team was developed and the nature and scope of its mission. Participants will also begin to understand the types of study levels and the particular methodology the EM benchmarking team is using to conduct studies. The EM benchmarking team will also encourage discussion on ways that DOE (both Headquarters and the Field) can team with its M&O/M&I contractors to conduct additional benchmarking studies. This ``introduction to benchmarking`` is intended to create a desire to know more and a greater appreciation of how benchmarking processes could be creatively employed to enhance performance.

  16. Overlapping reliable control for a cable-stayed bridge benchmark

    OpenAIRE

    Bakule, Lubomir; Paulet-Crainiceanu, Fideliu; Rodellar Benedé, José; Rossell Garriga, Josep Maria

    2005-01-01

    The brief presents a reliable 1-out-2 reduced order control design strategy for a cable-stayed bridge benchmark using two overlapping subsystems and the linear quadratic Gaussian (LQG) design. Reliability with regard to controller failures is considered. Local controllers are designed for reduced order subsystems of expanded system. They are implemented and evaluated on the original overall system model. Two different sets of numerical experiments of reliable control design within 1-out-2 con...

  17. Benchmarking of tourism products and implementation in Galicia

    OpenAIRE

    Castro, Belén; Iglesias, Montserrat; Piñeira, María José; Paül, Valerià

    2011-01-01

    Benchmarking could be defined as the systematic process of identifying, comparing and learning from the best practices or successful experiences of a particular sector in diverse geographical areas by disaggregating and analysing in an orderly manner the set of factors that condition their success. The outcome of this process helps improve key processes of a sector or adapt those practices that after a long search have been proven successful, to other areas. This study has gath...

  18. Research Reactor Benchmarking Database: Facility Specification and Experimental Data

    International Nuclear Information System (INIS)

    This web publication contains the facility specifications, experiment descriptions, and corresponding experimental data for nine different research reactors covering a wide range of research reactor types, power levels and experimental configurations. Each data set was prepared in order to serve as a stand-alone resource of well documented experimental data, which can subsequently be used in benchmarking and validation of the neutronic and thermal-hydraulic computational methods and tools employed for improved utilization, operation and safety analysis of research reactors

  19. (Invited) Microreactors for Characterization and Benchmarking of Photocatalysts

    DEFF Research Database (Denmark)

    Vesborg, Peter Christian Kjærgaard; Dionigi, Fabio; Trimarco, Daniel Bøndergaard;

    2015-01-01

    In the field of photocatalysis the batch-nature of the typical benchmarking experiment makes it very laborious to obtain good kinetic data as a function of parameters such as illumination wavelength, irradiance, catalyst temperature, reactant composition, etc. Microreactors with on-line mass spec......] Dionigi et al. Rev. Sci. Instr., 84, p. 103910 (2013) [6] Bøndergaard et al. "Fast and sensitive method for detecting volatile species in liquids", submitted...

  20. HPC Analytics Support. Requirements for Uncertainty Quantification Benchmarks

    Energy Technology Data Exchange (ETDEWEB)

    Paulson, Patrick R. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Purohit, Sumit [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Rodriguez, Luke R. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States)

    2015-05-01

    This report outlines techniques for extending benchmark generation products so they support uncertainty quantification by benchmarked systems. We describe how uncertainty quantification requirements can be presented to candidate analytical tools supporting SPARQL. We describe benchmark data sets for evaluating uncertainty quantification, as well as an approach for using our benchmark generator to produce data sets for generating benchmark data sets.

  1. General benchmarks for quantum repeaters

    CERN Document Server

    Pirandola, Stefano

    2015-01-01

    Using a technique based on quantum teleportation, we simplify the most general adaptive protocols for key distribution, entanglement distillation and quantum communication over a wide class of quantum channels in arbitrary dimension. Thanks to this method, we bound the ultimate rates for secret key generation and quantum communication through single-mode Gaussian channels and several discrete-variable channels. In particular, we derive exact formulas for the two-way assisted capacities of the bosonic quantum-limited amplifier and the dephasing channel in arbitrary dimension, as well as the secret key capacity of the qubit erasure channel. Our results establish the limits of quantum communication with arbitrary systems and set the most general and precise benchmarks for testing quantum repeaters in both discrete- and continuous-variable settings.

  2. Human factors reliability Benchmark exercise

    International Nuclear Information System (INIS)

    The Joint Research Centre of the European Commission has organized a Human Factors Reliability Benchmark Exercise (HF-RBE) with the aim of assessing the state of the art in human reliability modelling and assessment. Fifteen teams from eleven countries, representing industry, utilities, licensing organisations and research institutes, participated in the HF-RBE. The HF-RBE was organized around two study cases: (1) analysis of routine functional Test and Maintenance (T and M) procedures: with the aim of assessing the probability of test induced failures, the probability of failures to remain unrevealed and the potential to initiate transients because of errors performed in the test; (2) analysis of human actions during an operational transient: with the aim of assessing the probability that the operators will correctly diagnose the malfunctions and take proper corrective action. This report contains the final summary reports produced by the participants in the exercise

  3. Experimental and computational benchmark tests

    International Nuclear Information System (INIS)

    A program involving principally NIST, LANL, and ORNL has been in progress for about four years now to establish a series of benchmark measurements and calculations related to the moderation and leakage of 252Cf neutrons from a source surrounded by spherical aqueous moderators of various thicknesses and compositions. The motivation for these studies comes from problems in criticality calculations concerning arrays of multiplying components, where the leakage from one component acts as a source for the other components. This talk compares experimental and calculated values for the fission rates of four nuclides - 235U, 239Pu, 238U, and 237Np - in the leakage spectrum from moderator spheres of diameters 76.2 mm, 101.6 mm, and 127.0 mm, with either pure water or enriched B-10 solutions as the moderator. Very detailed Monte Carlo calculations were done with the MCNP code, using a open-quotes light waterclose quotes S(α,β) scattering kernel

  4. Benchmark scenarios for the NMSSM

    CERN Document Server

    Djouadi, A; Ellwanger, U; Godbole, R; Hugonie, C; King, S F; Lehti, S; Moretti, S; Nikitenko, A; Rottlander, I; Schumacher, M; Teixeira, A

    2008-01-01

    We discuss constrained and semi--constrained versions of the next--to--minimal supersymmetric extension of the Standard Model (NMSSM) in which a singlet Higgs superfield is added to the two doublet superfields that are present in the minimal extension (MSSM). This leads to a richer Higgs and neutralino spectrum and allows for many interesting phenomena that are not present in the MSSM. In particular, light Higgs particles are still allowed by current constraints and could appear as decay products of the heavier Higgs states, rendering their search rather difficult at the LHC. We propose benchmark scenarios which address the new phenomenological features, consistent with present constraints from colliders and with the dark matter relic density, and with (semi--)universal soft terms at the GUT scale. We present the corresponding spectra for the Higgs particles, their couplings to gauge bosons and fermions and their most important decay branching ratios. A brief survey of the search strategies for these states a...

  5. Benchmarking Learning and Teaching: Developing a Method

    Science.gov (United States)

    Henderson-Smart, Cheryl; Winning, Tracey; Gerzina, Tania; King, Shalinie; Hyde, Sarah

    2006-01-01

    Purpose: To develop a method for benchmarking teaching and learning in response to an institutional need to validate a new program in Dentistry at the University of Sydney, Australia. Design/methodology/approach: After a collaborative partner, University of Adelaide, was identified, the areas of teaching and learning to be benchmarked, PBL…

  6. Beyond Benchmarking: Value-Adding Metrics

    Science.gov (United States)

    Fitz-enz, Jac

    2007-01-01

    HR metrics has grown up a bit over the past two decades, moving away from simple benchmarking practices and toward a more inclusive approach to measuring institutional performance and progress. In this article, the acknowledged "father" of human capital performance benchmarking provides an overview of several aspects of today's HR metrics…

  7. Evaluating software verification systems: benchmarks and competitions

    NARCIS (Netherlands)

    Beyer, Dirk; Huisman, Marieke; Klebanov, Vladimir; Monahan, Rosemary

    2014-01-01

    This report documents the program and the outcomes of Dagstuhl Seminar 14171 “Evaluating Software Verification Systems: Benchmarks and Competitions”. The seminar brought together a large group of current and future competition organizers and participants, benchmark maintainers, as well as practition

  8. Benchmark Two-Good Utility Functions

    NARCIS (Netherlands)

    de Jaegher, K.

    2007-01-01

    Benchmark two-good utility functions involving a good with zero income elasticity and unit income elasticity are well known. This paper derives utility functions for the additional benchmark cases where one good has zero cross-price elasticity, unit own-price elasticity, and zero own price elasticit

  9. Benchmarking for controllere: Metoder, teknikker og muligheder

    DEFF Research Database (Denmark)

    Bukh, Per Nikolaj; Sandalgaard, Niels; Dietrichson, Lars

    2008-01-01

    Der vil i artiklen blive stillet skarpt på begrebet benchmarking ved at præsentere og diskutere forskellige facetter af det. Der vil blive redegjort for fire forskellige anvendelser af benchmarking for at vise begrebets bredde og væsentligheden af at klarlægge formålet med et benchmarkingprojekt...

  10. The Linked Data Benchmark Council Project

    NARCIS (Netherlands)

    Boncz, P.A.; Fundulaki, I.; Gubichev, A.; Larriba-Pey, J.; Neumann, T.

    2013-01-01

    Despite the fast growth and increasing popularity, the broad field of RDF and Graph database systems lacks an independent authority for developing benchmarks, and for neutrally assessing benchmark results through industry-strength auditing which would allow to quantify and compare the performance of

  11. An Effective Approach for Benchmarking Implementation

    Directory of Open Access Journals (Sweden)

    B. M. Deros

    2011-01-01

    Full Text Available Problem statement: The purpose of this study is to present a benchmarking guideline, conceptual framework and computerized mini program to assists companies achieve better performance in terms of quality, cost, delivery, supply chain and eventually increase their competitiveness in the market. The study begins with literature review on benchmarking definition, barriers and advantages from the implementation and the study of benchmarking framework. Approach: Thirty respondents were involved in the case study. They comprise of industrial practitioners, which had assessed usability and practicability of the guideline, conceptual framework and computerized mini program. Results: A guideline and template were proposed to simplify the adoption of benchmarking techniques. A conceptual framework was proposed by integrating the Deming’s PDCA and Six Sigma DMAIC theory. It was provided a step-by-step method to simplify the implementation and to optimize the benchmarking results. A computerized mini program was suggested to assist the users in adopting the technique as part of improvement project. As the result from the assessment test, the respondents found that the implementation method provided an idea for company to initiate benchmarking implementation and it guides them to achieve the desired goal as set in a benchmarking project. Conclusion: The result obtained and discussed in this study can be applied in implementing benchmarking in a more systematic way for ensuring its success.

  12. Repeated Results Analysis for Middleware Regression Benchmarking

    Czech Academy of Sciences Publication Activity Database

    Bulej, Lubomír; Kalibera, T.; Tůma, P.

    2005-01-01

    Roč. 60, - (2005), s. 345-358. ISSN 0166-5316 R&D Projects: GA ČR GA102/03/0672 Institutional research plan: CEZ:AV0Z10300504 Keywords : middleware benchmarking * regression benchmarking * regression testing Subject RIV: JD - Computer Applications, Robotics Impact factor: 0.756, year: 2005

  13. FZK proposal on transient benchmark in accelerator driven transmuter

    International Nuclear Information System (INIS)

    The benchmark should help to understand the behavior of a subcritical system with external source under transient conditions. It should allow to test and compare available methods and codes; include the knowledge base gained by experiments. The benchmark should be largely flexible, starting from simple configurations and transients up to comprehensive dynamics problems. It should help improving the description of transient and accident related phenomena in an ADS. General proposals on transient benchmark in accelerator driven transmuter are described. Select one or more systems to be analysed in details: solid and movable (molten salt) fuel systems; oxide, nitride, and other fuels for the solid option; Pb/I, Pb, Na coolants for the solid option; Pu+MA, LWR Pu, introduction of Th, other fuels; ADS and critical reactors; small (k-eff>0.97), very small (>0.99) or large deviations from k=1; intermediate, small, large reactor. Select types type of analyses: transient, static (reactivity coefficients, power peak factors), MA incineration rate calculations. In the transient case select: unprotected (no Control Rod, constant beam power) reactivity transients: benchmark exercise, investigation of the influence of the subcriticality level, other types of analyses. The investigation of dynamics of the ADS could be structured as short time-scale and long time-scale. Study of the ADT with dedicated fuels and cores is proposed including the safety aspects. Benchmark cores and fuels are defined. Spec-time kinetics of subcritical ADS should be recalculated. ADS/critical reactor behaviour and point kinetics should be included

  14. A performance benchmark test for geodynamo simulations

    Science.gov (United States)

    Matsui, H.; Heien, E. M.

    2013-12-01

    In the last ten years, a number of numerical dynamo models have successfully represented basic characteristics of the geomagnetic field. As new models and numerical methods continue to be developed, it is important to update and extend benchmarks for testing these models. The first dynamo benchmark of Christensen et al. (2001) was applied to models based on spherical harmonic expansion methods. However, only a few groups have reported results of the dynamo benchmark using local methods (Harder and Hansen, 2005; Matsui and Okuda, 2005; Chan et al., 2007) because of the difficulty treating magnetic boundary conditions based on the local methods. On the other hand, spherical harmonics expansion methods perform poorly on massively parallel computers because global data communications are required for the spherical harmonics expansions to evaluate nonlinear terms. We perform benchmark tests to asses various numerical methods for the next generation of geodynamo simulations. The purpose of this benchmark test is to assess numerical geodynamo models on a massively parallel computational platform. To compare among many numerical methods as possible, we consider the model with the insulated magnetic boundary by Christensen et al. (2001) and with the pseudo vacuum magnetic boundary, because the pseudo vacuum boundaries are implemented easier by using the local method than the magnetic insulated boundaries. In the present study, we consider two kinds of benchmarks, so-called accuracy benchmark and performance benchmark. In the accuracy benchmark, we compare the dynamo models by using modest Ekman and Rayleigh numbers proposed by Christensen et. al. (2001). We investigate a required spatial resolution for each dynamo code to obtain less than 1% difference from the suggested solution of the benchmark test using the two magnetic boundary conditions. In the performance benchmark, we investigate computational performance under the same computational environment. We perform these

  15. KENO-IV code benchmark calculation, (4)

    International Nuclear Information System (INIS)

    A series of benchmark tests has been undertaken in JAERI in order to examine the capability of JAERI's criticality safety evaluation system consisting of the Monte Carlo calculation code KENO-IV and the newly developed multi-group constants library MGCL. The present paper describes the results of a test using criticality experiments about slab-cylinder system of uranium nitrate solution. In all, 128 cases of experiments have been calculated for the slab-cylinder configuration with and without plexiglass reflector, having the various critical parameters such as the number of cylinders and height of the uranium nitrate solution. It is shown among several important results that the code and library gives a fairly good multiplication factor, that is, k sub(eff) -- 1.0 for heavily reflected cases, whereas k sub(eff) -- 0.91 for the unreflected ones. This suggests the necessity of more advanced treatment of the criticality calculation for the system where neutrons can easily leak out during slowing down process. (author)

  16. Validation of neutron-transport calculations in benchmark facilities for improved damage-fluence predictions

    International Nuclear Information System (INIS)

    An accurate determination of damage fluence accumulated by reactor pressure vessels (RPV) as a function of time is essential in order to evaluate the vessel integrity for both pressurized thermal shock (PTS) transients and end-of-life considerations. The desired accuracy for neutron exposure parameters such as displacements per atom or fluence (E > 1 MeV) is of the order of 20 to 30%. However, these types of accuracies can only be obtained realistically by validation of nuclear data and calculational methods in benchmark facilities. The purposes of this paper are to review the needs and requirements for benchmark experiments, to discuss the status of current benchmark experiments, to summarize results and conclusions obtained so far, and to suggest areas where further benchmarking is needed

  17. The inverted Pendulum: A fundamental Benchmark in Control Theory and Robotics

    OpenAIRE

    Boubaker, Olfa

    2014-01-01

    For at least fifty years, the inverted pendulum has been the most popular benchmark, among others, for teaching and researches in control theory and robotics. This paper presents the key motivations for the use of that system and explains, in details, the main reflections on how the inverted pendulum benchmark gives an effective and efficient application. Several real experiences, virtual models and web-based remote control laboratories will be presented with emphasis on the practical design ...

  18. Non-exponential Fidelity Decay in Randomized Benchmarking with Low-Frequency Noise

    OpenAIRE

    Fogarty, M. A.; Veldhorst, M.; R. Harper; Yang, C.H.; Bartlett, S. D.; Flammia, S. T.; A. S. Dzurak

    2015-01-01

    We show that non-exponential fidelity decays in randomized benchmarking experiments on quantum dot qubits are consistent with numerical simulations that incorporate low-frequency noise. By expanding standard randomized benchmarking analysis to this experimental regime, we find that such non-exponential decays are better modeled by multiple exponential decay rates, leading to an instantaneous control fidelity for isotopically-purified-silicon MOS quantum dot qubits which can be as high as 99.9...

  19. National benchmarking against GLOBALGAP : Case studies of Good Agricultural Practices in Kenya, Malaysia, Mexico and Chile

    OpenAIRE

    Valk, van der, L.J.M.; Roest, De, Kees

    2009-01-01

    This desk study examines the experiences and lessons learned from four case studies of countries aiming at the GLOBALGAP benchmarking procedure for national Good Agricultural Practices, namely Chile, Kenya, Malaysia, and Mexico. Aspects that determine the origin and character of the benchmarking process are the current and future export markets of each country; the ownership of the National GAP Schemes, that is, government or private sector. Government led GAP initiatives are part of strategi...

  20. Benchmarking--Measuring and Comparing for Continuous Improvement.

    Science.gov (United States)

    Henczel, Sue

    2002-01-01

    Discussion of benchmarking focuses on the use of internal and external benchmarking by special librarians. Highlights include defining types of benchmarking; historical development; benefits, including efficiency, improved performance, increased competitiveness, and better decision making; problems, including inappropriate adaptation; developing a…

  1. Parallelization of NAS Benchmarks for Shared Memory Multiprocessors

    Science.gov (United States)

    Waheed, Abdul; Yan, Jerry C.; Saini, Subhash (Technical Monitor)

    1998-01-01

    This paper presents our experiences of parallelizing the sequential implementation of NAS benchmarks using compiler directives on SGI Origin2000 distributed shared memory (DSM) system. Porting existing applications to new high performance parallel and distributed computing platforms is a challenging task. Ideally, a user develops a sequential version of the application, leaving the task of porting to new generations of high performance computing systems to parallelization tools and compilers. Due to the simplicity of programming shared-memory multiprocessors, compiler developers have provided various facilities to allow the users to exploit parallelism. Native compilers on SGI Origin2000 support multiprocessing directives to allow users to exploit loop-level parallelism in their programs. Additionally, supporting tools can accomplish this process automatically and present the results of parallelization to the users. We experimented with these compiler directives and supporting tools by parallelizing sequential implementation of NAS benchmarks. Results reported in this paper indicate that with minimal effort, the performance gain is comparable with the hand-parallelized, carefully optimized, message-passing implementations of the same benchmarks.

  2. Toxicological benchmarks for screening potential contaminants of concern for effects on terrestrial plants: 1994 revision

    Energy Technology Data Exchange (ETDEWEB)

    Will, M.E.; Suter, G.W. II

    1994-09-01

    One of the initial stages in ecological risk assessment for hazardous waste sites is screening contaminants to determine which of them are worthy of further consideration as contaminants of potential concern. This process is termed contaminant screening. It is performed by comparing measured ambient concentrations of chemicals to benchmark concentrations. Currently, no standard benchmark concentrations exist for assessing contaminants in soil with respect to their toxicity to plants. This report presents a standard method for deriving benchmarks for this purpose (phytotoxicity benchmarks), a set of data concerning effects of chemicals in soil or soil solution on plants, and a set of phytotoxicity benchmarks for 38 chemicals potentially associated with United States Department of Energy (DOE) sites. In addition, background information on the phytotoxicity and occurrence of the chemicals in soils is presented, and literature describing the experiments from which data were drawn for benchmark derivation is reviewed. Chemicals that are found in soil at concentrations exceeding both the phytotoxicity benchmark and the background concentration for the soil type should be considered contaminants of potential concern.

  3. Toxicological benchmarks for screening potential contaminants of concern for effects on terrestrial plants: 1994 revision

    International Nuclear Information System (INIS)

    One of the initial stages in ecological risk assessment for hazardous waste sites is screening contaminants to determine which of them are worthy of further consideration as contaminants of potential concern. This process is termed contaminant screening. It is performed by comparing measured ambient concentrations of chemicals to benchmark concentrations. Currently, no standard benchmark concentrations exist for assessing contaminants in soil with respect to their toxicity to plants. This report presents a standard method for deriving benchmarks for this purpose (phytotoxicity benchmarks), a set of data concerning effects of chemicals in soil or soil solution on plants, and a set of phytotoxicity benchmarks for 38 chemicals potentially associated with United States Department of Energy (DOE) sites. In addition, background information on the phytotoxicity and occurrence of the chemicals in soils is presented, and literature describing the experiments from which data were drawn for benchmark derivation is reviewed. Chemicals that are found in soil at concentrations exceeding both the phytotoxicity benchmark and the background concentration for the soil type should be considered contaminants of potential concern

  4. Toxicological Benchmarks for Screening Potential Contaminants of Concern for Effects on Terrestrial Plants

    Energy Technology Data Exchange (ETDEWEB)

    Suter, G.W. II

    1993-01-01

    One of the initial stages in ecological risk assessment for hazardous waste sites is screening contaminants to determine which of them are worthy of further consideration as contaminants of potential concern. This process is termed contaminant screening. It is performed by comparing measured ambient concentrations of chemicals to benchmark concentrations. Currently, no standard benchmark concentrations exist for assessing contaminants in soil with respect to their toxicity to plants. This report presents a standard method for deriving benchmarks for this purpose (phytotoxicity benchmarks), a set of data concerning effects of chemicals in soil or soil solution on plants, and a set of phytotoxicity benchmarks for 38 chemicals potentially associated with United States Department of Energy (DOE) sites. In addition, background information on the phytotoxicity and occurrence of the chemicals in soils is presented, and literature describing the experiments from which data were drawn for benchmark derivation is reviewed. Chemicals that are found in soil at concentrations exceeding both the phytotoxicity benchmark and the background concentration for the soil type should be considered contaminants of potential concern.

  5. Computational benchmarking of fast neutron transport throughout large water thicknesses

    International Nuclear Information System (INIS)

    Neutron dosimetry experiments seem to point our difficulties in the treatment of large water thickness like those encountered between the core baffle and the pressure vessel. This paper describes the theoretical benchmark undertaken by EDF, SCK/CEN and TRACTEBEL ENERGY ENGINEERING, concerning the transport of fast neutrons throughout a one meter cube of water, located after a U-235 fission sources plate. The results showed no major discrepancies between the calculations up to 50 cm from the source, accepting that a P3 development of the Legendre polynomials is necessary for the Sn calculations. The main differences occurred after 50 cm, reaching 20 % at the end of the water cube. This results lead us to consider an experimental benchmark, dedicated to the problem of fast neutron deep penetration in water, which has been launched at SCK/CEN. (authors)

  6. Benchmarking of neutron production of heavy-ion transport codes

    International Nuclear Information System (INIS)

    Document available in abstract form only, full text of document follows: Accurate prediction of radiation fields generated by heavy ion interactions is important in medical applications, space missions, and in design and operation of rare isotope research facilities. In recent years, several well-established computer codes in widespread use for particle and radiation transport calculations have been equipped with the capability to simulate heavy ion transport and interactions. To assess and validate these capabilities, we performed simulations of a series of benchmark-quality heavy ion experiments with the computer codes FLUKA, MARS15, MCNPX, and PHITS. We focus on the comparisons of secondary neutron production. Results are encouraging; however, further improvements in models and codes and additional benchmarking are required. (authors)

  7. TRX and UO2 criticality benchmarks with SAM-CE

    International Nuclear Information System (INIS)

    A set of thermal reactor benchmark calculations with SAM-CE which have been conducted at both MAGI and at BNL are described. Their purpose was both validation of the SAM-CE reactor eigenvalue capability developed by MAGI and a substantial contribution to the data testing of both ENDF/B-IV and ENDF/B-V libraries. This experience also resulted in increased calculational efficiency of the code and an example is given. The benchmark analysis included the TRX-1 infinite cell using both ENDF/B-IV and ENDF/B-V cross section sets and calculations using ENDF/B-IV of the TRX-1 full core and TRX-2 cell. BAPL-UO2-1 calculations were conducted for the cell using both ENDF/B-IV and ENDF/B-V and for the full core with ENDF/B-V

  8. ES-RBE Event sequence reliability Benchmark exercise

    International Nuclear Information System (INIS)

    The event Sequence Reliability Benchmark Exercise (ES-RBE) can be considered as a logical extension of the other three Reliability Benchmark Exercices : the RBE on Systems Analysis, the RBE on Common Cause Failures and the RBE on Human Factors. The latter, constituting Activity No. 1, was concluded by the end of 1987. The ES-RBE covered the techniques that are currently used for analysing and quantifying sequences of events starting from an initiating event to various plant damage states, including analysis of various system failures and/or successes, human intervention failure and/or success and dependencies between systems. By this way, one of the scopes of the ES-RBE was to integrate the experiences gained in the previous exercises

  9. Developing integrated benchmarks for DOE performance measurement

    Energy Technology Data Exchange (ETDEWEB)

    Barancik, J.I.; Kramer, C.F.; Thode, Jr. H.C.

    1992-09-30

    The objectives of this task were to describe and evaluate selected existing sources of information on occupational safety and health with emphasis on hazard and exposure assessment, abatement, training, reporting, and control identifying for exposure and outcome in preparation for developing DOE performance benchmarks. Existing resources and methodologies were assessed for their potential use as practical performance benchmarks. Strengths and limitations of current data resources were identified. Guidelines were outlined for developing new or improved performance factors, which then could become the basis for selecting performance benchmarks. Data bases for non-DOE comparison populations were identified so that DOE performance could be assessed relative to non-DOE occupational and industrial groups. Systems approaches were described which can be used to link hazards and exposure, event occurrence, and adverse outcome factors, as needed to generate valid, reliable, and predictive performance benchmarks. Data bases were identified which contain information relevant to one or more performance assessment categories . A list of 72 potential performance benchmarks was prepared to illustrate the kinds of information that can be produced through a benchmark development program. Current information resources which may be used to develop potential performance benchmarks are limited. There is need to develop an occupational safety and health information and data system in DOE, which is capable of incorporating demonstrated and documented performance benchmarks prior to, or concurrent with the development of hardware and software. A key to the success of this systems approach is rigorous development and demonstration of performance benchmark equivalents to users of such data before system hardware and software commitments are institutionalized.

  10. Human factors reliability benchmark exercise

    International Nuclear Information System (INIS)

    The Joint Research Centre of the European Commission has organised a Human Factors Reliability Benchmark Exercise (HF-RBE) with the aim of assessing the state of the art in human reliability modelling and assessment. Fifteen teams from eleven countries, representing industry, utilities, licensing organisations and research institutes, participated in the HF-RBE. The HF-RBE was organised around two study cases: (1) analysis of routine functional Test and Maintenance (TPM) procedures: with the aim of assessing the probability of test induced failures, the probability of failures to remain unrevealed and the potential to initiate transients because of errors performed in the test; (2) analysis of human actions during an operational transient: with the aim of assessing the probability that the operators will correctly diagnose the malfunctions and take proper corrective action. This report summarises the contributions received from the participants and analyses these contributions on a comparative basis. The aim of this analysis was to compare the procedures, modelling techniques and quantification methods used, to obtain insight in the causes and magnitude of the variability observed in the results, to try to identify preferred human reliability assessment approaches and to get an understanding of the current state of the art in the field identifying the limitations that are still inherent to the different approaches

  11. Benchmarking in healthcare using aggregated indicators

    DEFF Research Database (Denmark)

    Traberg, Andreas; Jacobsen, Peter

    2010-01-01

    databases, the model is constructed as a comprehensive hierarchy of indicators. By aggregating the outcome of each indicator, the model is able to benchmark healthcare providing units. By assessing performance deeper in the hierarchy, a more detailed view of performance is obtained. The validity test of the...... model is performed at a Danish non-profit hospital, where four radiological sites are benchmarked against each other. Because of the multifaceted perspective on performance, the model proved valuable both as a benchmarking tool and as an internal decision support system....

  12. LAPUR-K BWR stability benchmark

    International Nuclear Information System (INIS)

    This paper documents the stability benchmark of the LAPUR-K code using the measurements taken at the Ringhals Unit 1 plant over four cycles of operation. This benchmark was undertaken to demonstrate the ability of LAPUR-K to calculate the decay ratios for both core-wide and regional mode oscillations. This benchmark contributes significantly to assuring that LAPUR-K can be used to define the exclusion region for the Monticello Plant in response to recent US Nuclear Regulatory Commission notices concerning oscillation observed at Boiling Water Reactor plants. Stability is part of Northern States Power Reload Safety Evaluation of the Monticello Plant

  13. Geant4 Computing Performance Benchmarking and Monitoring

    Science.gov (United States)

    Dotti, Andrea; Elvira, V. Daniel; Folger, Gunter; Genser, Krzysztof; Jun, Soon Yung; Kowalkowski, James B.; Paterno, Marc

    2015-12-01

    Performance evaluation and analysis of large scale computing applications is essential for optimal use of resources. As detector simulation is one of the most compute intensive tasks and Geant4 is the simulation toolkit most widely used in contemporary high energy physics (HEP) experiments, it is important to monitor Geant4 through its development cycle for changes in computing performance and to identify problems and opportunities for code improvements. All Geant4 development and public releases are being profiled with a set of applications that utilize different input event samples, physics parameters, and detector configurations. Results from multiple benchmarking runs are compared to previous public and development reference releases to monitor CPU and memory usage. Observed changes are evaluated and correlated with code modifications. Besides the full summary of call stack and memory footprint, a detailed call graph analysis is available to Geant4 developers for further analysis. The set of software tools used in the performance evaluation procedure, both in sequential and multi-threaded modes, include FAST, IgProf and Open|Speedshop. The scalability of the CPU time and memory performance in multi-threaded application is evaluated by measuring event throughput and memory gain as a function of the number of threads for selected event samples.

  14. Cost benchmarking of railway projects in Europe – can it help to reduce costs?

    DEFF Research Database (Denmark)

    Trabo, Inara; Landex, Alex; Nielsen, Otto Anker;

    This paper highlights the methodology of construction cost benchmarking of railway projects in the EU and its preliminary results. Benchmarking helps project managers learn from others, improve particular project areas, and reduce project costs. For railway projects, benchmarking is essential for...... the comparison of unit costs for major cost drivers (e.g. tunnels, bridges, etc.). This methodology was applied to the case study described in this paper, the first high-speed railway project in Denmark, “The New Line Copenhagen-Ringsted”. The aim was to avoid cost overruns and even reduce final...... budget outcomes by looking for the best practices in the construction and implementation of other high-speed lines in Europe and learning from their experience. The paper presents benchmarking from nine railway projects that are comparable with the Copenhagen-Ringsted project. The results of this...

  15. A BENCHMARK FOR DESIGNING USABLE AND SECURE TEXT-BASED CAPTCHAS

    Directory of Open Access Journals (Sweden)

    Suliman A. Alsuhibany

    2016-07-01

    Full Text Available An automated public Turing test to distinguish between computers and humans known as CAPTCHA is a widely used technique on many websites to protect their online services from malicious users. Two fundamental aspects of captcha considered in various studies in the literature are robustness and usability. A widely accepted standard benchmark, to guide the text-based captcha developers is not yet available. So this paper proposes a benchmark for designing usable-secure text-based captchas based on a community driven evaluation of the usability and security aspects. Based on this benchmark, we develop four new textbased captcha schemes, and conduct two separate experiments to evaluate both the security and usability perspectives of the developed schemes. The result of this evaluation indicates that the proposed benchmark provides a basis for designing usable-secure text-based captchas.

  16. DOE Commercial Building Benchmark Models: Preprint

    Energy Technology Data Exchange (ETDEWEB)

    Torcelini, P.; Deru, M.; Griffith, B.; Benne, K.; Halverson, M.; Winiarski, D.; Crawley, D. B.

    2008-07-01

    To provide a consistent baseline of comparison and save time conducting such simulations, the U.S. Department of Energy (DOE) has developed a set of standard benchmark building models. This paper will provide an executive summary overview of these benchmark buildings, and how they can save building analysts valuable time. Fully documented and implemented to use with the EnergyPlus energy simulation program, the benchmark models are publicly available and new versions will be created to maintain compatibility with new releases of EnergyPlus. The benchmark buildings will form the basis for research on specific building technologies, energy code development, appliance standards, and measurement of progress toward DOE energy goals. Having a common starting point allows us to better share and compare research results and move forward to make more energy efficient buildings.

  17. Medicare Contracting - Redacted Benchmark Metric Reports

    Data.gov (United States)

    U.S. Department of Health & Human Services — The Centers for Medicare and Medicaid Services has compiled aggregate national benchmark cost and workload metrics using data submitted to CMS by the AB MACs and...

  18. Benchmarking Optimization Software with Performance Profiles

    OpenAIRE

    Dolan, Elizabeth D.; Moré, Jorge J.

    2001-01-01

    We propose performance profiles-distribution functions for a performance metric-as a tool for benchmarking and comparing optimization software. We show that performance profiles combine the best features of other tools for performance evaluation.

  19. Neutronic Benchmarks for the Utilization of Mixed-Oxide Fuel: Joint U.S./Russian Progress Report for Fiscal Year 1997 - Volume 4, Part 2--Saxton Plutonium Program Critical Experiments

    International Nuclear Information System (INIS)

    Critical experiments with water-moderated, single-region PuO2-UO2 or UO2, and multiple-region PuO2-UO2- and UO2-fueled cores were performed at the CRX reactor critical facility at the Westinghouse Reactor Evaluation Center (WREC) at Waltz Mill, Pennsylvania in 1965 [1]. These critical experiments were part of the Saxton Plutonium Program. The mixed oxide (MOX) fuel used in these critical experiments and then loaded in the Saxton reactor contained 6.6 wt% PuO2 in a mixture of PuO2 and natural UO2. The Pu metal had the following isotopic mass percentages: 90.50% 239Pu; 8.57% 239Pu; 0.89% 240Pu; and 0.04% 241Pu. The purpose of these critical experiments was to verify the nuclear design of Saxton partial plutonium cores while obtaining parameters of fundamental significance such as buckling, control rod worth, soluble poison worth, flux, power peaking, relative pin power, and power sharing factors of MOX and UO2 lattices. For comparison purposes, the core was also loaded with uranium dioxide fuel rods only. This series is covered by experiments beginning with the designation SX

  20. Benchmarking carbon emissions performance in supply chains

    OpenAIRE

    Acquaye, Adolf; Genovese, Andrea; Barrett, John W.; Koh, Lenny

    2014-01-01

    Purpose – The paper aims to develop a benchmarking framework to address issues such as supply chain complexity and visibility, geographical differences and non-standardized data, ensuring that the entire supply chain environmental impact (in terms of carbon) and resource use for all tiers, including domestic and import flows, are evaluated. Benchmarking has become an important issue in supply chain management practice. However, challenges such as supply chain complexity and visibility, geogra...

  1. A framework for benchmarking land models

    Directory of Open Access Journals (Sweden)

    Y. Q. Luo

    2012-10-01

    Full Text Available Land models, which have been developed by the modeling community in the past few decades to predict future states of ecosystems and climate, have to be critically evaluated for their performance skills of simulating ecosystem responses and feedback to climate change. Benchmarking is an emerging procedure to measure performance of models against a set of defined standards. This paper proposes a benchmarking framework for evaluation of land model performances and, meanwhile, highlights major challenges at this infant stage of benchmark analysis. The framework includes (1 targeted aspects of model performance to be evaluated, (2 a set of benchmarks as defined references to test model performance, (3 metrics to measure and compare performance skills among models so as to identify model strengths and deficiencies, and (4 model improvement. Land models are required to simulate exchange of water, energy, carbon and sometimes other trace gases between the atmosphere and land surface, and should be evaluated for their simulations of biophysical processes, biogeochemical cycles, and vegetation dynamics in response to climate change across broad temporal and spatial scales. Thus, one major challenge is to select and define a limited number of benchmarks to effectively evaluate land model performance. The second challenge is to develop metrics of measuring mismatches between models and benchmarks. The metrics may include (1 a priori thresholds of acceptable model performance and (2 a scoring system to combine data–model mismatches for various processes at different temporal and spatial scales. The benchmark analyses should identify clues of weak model performance to guide future development, thus enabling improved predictions of future states of ecosystems and climate. The near-future research effort should be on development of a set of widely acceptable benchmarks that can be used to objectively, effectively, and reliably evaluate fundamental properties

  2. Benchmark Two-Good Utility Functions

    OpenAIRE

    de Jaegher, K.

    2007-01-01

    Benchmark two-good utility functions involving a good with zero income elasticity and unit income elasticity are well known. This paper derives utility functions for the additional benchmark cases where one good has zero cross-price elasticity, unit own-price elasticity, and zero own price elasticity. It is shown how each of these utility functions arises from a simple graphical construction based on a single given indifference curve. Also, it is shown that possessors of such utility function...

  3. Bundesländer-Benchmarking 2002

    OpenAIRE

    Blancke, Susanne; Hedrich, Horst; Schmid, Josef

    2002-01-01

    Das Bundesländer Benchmarking 2002 basiert auf einer Untersuchung ausgewählter Arbeitsmarkt- und Wirtschaftsindikatoren in den deutschen Bundesländern. Hierfür wurden drei Benchmarkings nach der Radar-Chart Methode vorgenommen: Eines welches nur Arbeitsmarktindikatoren betrachtet; eines, welches nur Wirtschaftsindikatoren betrachtet; und eines welches gemischte Arbeitsmarkt- und Wirtschaftsindikatoren beleuchtet. Verglichen wurden die Länder untereinander im Querschnitt zu zwei Zeitpunkten –...

  4. Benchmarking Deep Reinforcement Learning for Continuous Control

    OpenAIRE

    Duan, Yan; Chen, Xi; Houthooft, Rein; Schulman, John; Abbeel, Pieter

    2016-01-01

    Recently, researchers have made significant progress combining the advances in deep learning for learning feature representations with reinforcement learning. Some notable examples include training agents to play Atari games based on raw pixel data and to acquire advanced manipulation skills using raw sensory inputs. However, it has been difficult to quantify progress in the domain of continuous control due to the lack of a commonly adopted benchmark. In this work, we present a benchmark suit...

  5. Distributional benchmarking in tax policy evaluations

    OpenAIRE

    Thor O. Thoresen; Zhiyang Jia; Peter J. Lambert

    2013-01-01

    Given an objective to exploit cross-sectional micro data to evaluate the distributional effects of tax policies over a time period, the practitioner of public economics will find that the relevant literature offers a wide variety of empirical approaches. For example, studies vary with respect to the definition of individual well-being and to what extent explicit benchmarking techniques are utilized to describe policy effects. The present paper shows how the concept of distributional benchmark...

  6. Simple Benchmark Specifications for Space Radiation Protection

    Science.gov (United States)

    Singleterry, Robert C. Jr.; Aghara, Sukesh K.

    2013-01-01

    This report defines space radiation benchmark specifications. This specification starts with simple, monoenergetic, mono-directional particles on slabs and progresses to human models in spacecraft. This report specifies the models and sources needed to what the team performing the benchmark needs to produce in a report. Also included are brief descriptions of how OLTARIS, the NASA Langley website for space radiation analysis, performs its analysis.

  7. Features and technology of enterprise internal benchmarking

    OpenAIRE

    A. V. Dubodelova; Yurynets, O. V.

    2013-01-01

    The aim of the article. The aim of the article is to generalize characteristics, objectives, advantages of internal benchmarking. The stages sequence of internal benchmarking technology is formed. It is focused on continuous improvement of process of the enterprise by implementing existing best practices.The results of the analysis. Business activity of domestic enterprises in crisis business environment has to focus on the best success factors of their structural units by using standard rese...

  8. Overview of CSEWG shielding benchmark problems

    Energy Technology Data Exchange (ETDEWEB)

    Maerker, R.E.

    1979-01-01

    The fundamental philosophy behind the choosing of CSEWG shielding benchmarks is that the accuracy of a certain range of cross section data be adequately tested. The benchmarks, therefore, consist of measurements and calculations of these measurements. Calculations for which there are no measurements provide little information on the adequacy of the data, although they can perhaps indicate the sensitivity of results to variations in data.

  9. Dukovany NPP fuel cycle benchmark definition

    International Nuclear Information System (INIS)

    The new benchmark based on Dukovany NPP Unit-2 history is defined. The main goal of this benchmark is to compare results obtained by different codes used for neutron-physics calculation in organisations which are interested in this task. All needed are described in this paper or there are given references, where it is possible to obtain this information. Input data are presented in tables, requested output data format for automatic processing is described (Authors)

  10. Benchmarking Danish Vocational Education and Training Programmes

    DEFF Research Database (Denmark)

    Bogetoft, Peter; Wittrup, Jesper

    This study paper discusses methods whereby Danish vocational education and training colleges can be benchmarked, and presents results from a number of models. It is conceptually complicated to benchmark vocational colleges, as the various colleges in Denmark offer a wide range of course programmes......-related achievement. We attempt to summarise the various effects that the colleges have in two relevant figures, namely retention rates of students and employment rates among students who have completed training programmes....

  11. Under Pressure Benchmark for DDBMS Availability

    OpenAIRE

    Fior, Alessandro Gustavo; Meira, Jorge Augusto; Cunha De Almeida, Eduardo; Coelho, Ricardo Gonçalves; Didonet Del Fabro, Marcos; Le Traon, Yves

    2013-01-01

    The availability of Distributed Database Management Systems (DDBMS) is related to the probability of being up and running at a given point in time, and managing failures. One well-known and widely used mechanism to ensure availability is replication, which includes performance impact on maintaining data replicas across the DDBMS's machine nodes. Benchmarking can be used to measure such impact. In this article, we present a benchmark that evaluates the performance of DDBMS, considering availab...

  12. DWEB: A Data Warehouse Engineering Benchmark

    OpenAIRE

    Darmont, Jérôme; Bentayeb, Fadila; Boussaïd, Omar

    2005-01-01

    Data warehouse architectural choices and optimization techniques are critical to decision support query performance. To facilitate these choices, the performance of the designed data warehouse must be assessed. This is usually done with the help of benchmarks, which can either help system users comparing the performances of different systems, or help system engineers testing the effect of various design choices. While the TPC standard decision support benchmarks address the first point, they ...

  13. MPI Benchmarking Revisited: Experimental Design and Reproducibility

    OpenAIRE

    Hunold, Sascha; Carpen-Amarie, Alexandra

    2015-01-01

    The Message Passing Interface (MPI) is the prevalent programming model used on today's supercomputers. Therefore, MPI library developers are looking for the best possible performance (shortest run-time) of individual MPI functions across many different supercomputer architectures. Several MPI benchmark suites have been developed to assess the performance of MPI implementations. Unfortunately, the outcome of these benchmarks is often neither reproducible nor statistically sound. To overcome th...

  14. Benchmarking for Cost Improvement. Final report

    Energy Technology Data Exchange (ETDEWEB)

    1993-09-01

    The US Department of Energy`s (DOE) Office of Environmental Restoration and Waste Management (EM) conducted the Benchmarking for Cost Improvement initiative with three objectives: Pilot test benchmarking as an EM cost improvement tool; identify areas for cost improvement and recommend actions to address these areas; provide a framework for future cost improvement. The benchmarking initiative featured the use of four principal methods (program classification, nationwide cost improvement survey, paired cost comparison and component benchmarking). Interested parties contributed during both the design and execution phases. The benchmarking initiative was conducted on an accelerated basis. Of necessity, it considered only a limited set of data that may not be fully representative of the diverse and complex conditions found at the many DOE installations. The initiative generated preliminary data about cost differences and it found a high degree of convergence on several issues. Based on this convergence, the report recommends cost improvement strategies and actions. This report describes the steps taken as part of the benchmarking initiative and discusses the findings and recommended actions for achieving cost improvement. The results and summary recommendations, reported below, are organized by the study objectives.

  15. Clinically meaningful performance benchmarks in MS

    Science.gov (United States)

    Motl, Robert W.; Scagnelli, John; Pula, John H.; Sosnoff, Jacob J.; Cadavid, Diego

    2013-01-01

    Objective: Identify and validate clinically meaningful Timed 25-Foot Walk (T25FW) performance benchmarks in individuals living with multiple sclerosis (MS). Methods: Cross-sectional study of 159 MS patients first identified candidate T25FW benchmarks. To characterize the clinical meaningfulness of T25FW benchmarks, we ascertained their relationships to real-life anchors, functional independence, and physiologic measurements of gait and disease progression. Candidate T25FW benchmarks were then prospectively validated in 95 subjects using 13 measures of ambulation and cognition, patient-reported outcomes, and optical coherence tomography. Results: T25FW of 6 to 7.99 seconds was associated with a change in occupation due to MS, occupational disability, walking with a cane, and needing “some help” with instrumental activities of daily living; T25FW ≥8 seconds was associated with collecting Supplemental Security Income and government health care, walking with a walker, and inability to do instrumental activities of daily living. During prospective benchmark validation, we trichotomized data by T25FW benchmarks (10 seconds) ranges of performance. PMID:24174581

  16. ''FULL-CORE'' VVER-440 calculation benchmark

    International Nuclear Information System (INIS)

    Because of the difficulties with experimental validation of power distribution predicted by macro-code on the pin by pin level we decided to prepare a calculation benchmark named ''FULL-CORE'' VVER-440. This benchmark is a two-dimensional (2D) calculation benchmark based on the VVER-440 reactor core cold state geometry with taking into account the geometry of explicit radial reflector. The main task of this benchmark is to test the pin by pin power distribution in fuel assemblies predicted by macro-codes that are used for neutron-physics calculations especially for VVER-440 reactors. The proposal of this benchmark was presented at the 21st Symposium of AER in 2011. The reference solution has been calculated by MCNP code using Monte Carlo method and the results have been published in the AER community. The results of reference calculation were presented at the 22nd Symposium of AER in 2012. In this paper we will compare the available macro-codes results of this calculation benchmark.

  17. Action-Oriented Benchmarking: Concepts and Tools

    Energy Technology Data Exchange (ETDEWEB)

    California Energy Commission; Mathew, Paul; Mills, Evan; Mathew, Paul; Piette, Mary Ann; Bourassa, Norman; Brook, Martha

    2008-02-13

    Most energy benchmarking tools provide static feedback on how one building compares to a larger set of loosely similar buildings, without providing information at the end-use level or on what can be done to reduce consumption, cost, or emissions. In this article--Part 1 of a two-part series--we describe an 'action-oriented benchmarking' approach, which extends whole-building energy benchmarking to include analysis of system and component energy use metrics and features. Action-oriented benchmarking thereby allows users to generate more meaningful metrics and to identify, screen and prioritize potential efficiency improvements. This opportunity assessment process can then be used to inform and optimize a full-scale audit or commissioning process. We introduce a new web-based action-oriented benchmarking system and associated software tool-EnergyIQ. The benchmarking methods, visualizations, and user interface design are informed by an end-user needs assessment survey and best-practice guidelines from ASHRAE.

  18. Compilation report of VHTRC temperature coefficient benchmark calculations

    Energy Technology Data Exchange (ETDEWEB)

    Yasuda, Hideshi; Yamane, Tsuyoshi [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment

    1995-11-01

    A calculational benchmark problem has been proposed by JAERI to an IAEA Coordinated Research Program, `Verification of Safety Related Neutronic Calculation for Low-enriched Gas-cooled Reactors` to investigate the accuracy of calculation results obtained by using codes of the participating countries. This benchmark is made on the basis of assembly heating experiments at a pin-in block type critical assembly, VHTRC. Requested calculation items are the cell parameters, effective multiplication factor, temperature coefficient of reactivity, reaction rates, fission rate distribution, etc. Seven institutions from five countries have joined the benchmark works. Calculation results are summarized in this report with some remarks by the authors. Each institute analyzed the problem by applying the calculation code system which was prepared for the HTGR development of individual country. The values of the most important parameter, k{sub eff}, by all institutes showed good agreement with each other and with the experimental ones within 1%. The temperature coefficient agreed within 13%. The values of several cell parameters calculated by several institutes did not agree with the other`s ones. It will be necessary to check the calculation conditions again for getting better agreement. (J.P.N.).

  19. Criticality Benchmark Analysis of the HTTR Annular Startup Core Configurations

    Energy Technology Data Exchange (ETDEWEB)

    John D. Bess

    2009-11-01

    One of the high priority benchmarking activities for corroborating the Next Generation Nuclear Plant (NGNP) Project and Very High Temperature Reactor (VHTR) Program is evaluation of Japan's existing High Temperature Engineering Test Reactor (HTTR). The HTTR is a 30 MWt engineering test reactor utilizing graphite moderation, helium coolant, and prismatic TRISO fuel. A large amount of critical reactor physics data is available for validation efforts of High Temperature Gas-cooled Reactors (HTGRs). Previous international reactor physics benchmarking activities provided a collation of mixed results that inaccurately predicted actual experimental performance.1 Reevaluations were performed by the Japanese to reduce the discrepancy between actual and computationally-determined critical configurations.2-3 Current efforts at the Idaho National Laboratory (INL) involve development of reactor physics benchmark models in conjunction with the International Reactor Physics Experiment Evaluation Project (IRPhEP) for use with verification and validation methods in the VHTR Program. Annular cores demonstrate inherent safety characteristics that are of interest in developing future HTGRs.

  20. Criticality Benchmark Analysis of the HTTR Annular Startup Core Configurations

    International Nuclear Information System (INIS)

    One of the high priority benchmarking activities for corroborating the Next Generation Nuclear Plant (NGNP) Project and Very High Temperature Reactor (VHTR) Program is evaluation of Japan's existing High Temperature Engineering Test Reactor (HTTR). The HTTR is a 30 MWt engineering test reactor utilizing graphite moderation, helium coolant, and prismatic TRISO fuel. A large amount of critical reactor physics data is available for validation efforts of High Temperature Gas-cooled Reactors (HTGRs). Previous international reactor physics benchmarking activities provided a collation of mixed results that inaccurately predicted actual experimental performance.1 Reevaluations were performed by the Japanese to reduce the discrepancy between actual and computationally-determined critical configurations.2-3 Current efforts at the Idaho National Laboratory (INL) involve development of reactor physics benchmark models in conjunction with the International Reactor Physics Experiment Evaluation Project (IRPhEP) for use with verification and validation methods in the VHTR Program. Annular cores demonstrate inherent safety characteristics that are of interest in developing future HTGRs.

  1. Benchmarking Gas Path Diagnostic Methods: A Public Approach

    Science.gov (United States)

    Simon, Donald L.; Bird, Jeff; Davison, Craig; Volponi, Al; Iverson, R. Eugene

    2008-01-01

    Recent technology reviews have identified the need for objective assessments of engine health management (EHM) technology. The need is two-fold: technology developers require relevant data and problems to design and validate new algorithms and techniques while engine system integrators and operators need practical tools to direct development and then evaluate the effectiveness of proposed solutions. This paper presents a publicly available gas path diagnostic benchmark problem that has been developed by the Propulsion and Power Systems Panel of The Technical Cooperation Program (TTCP) to help address these needs. The problem is coded in MATLAB (The MathWorks, Inc.) and coupled with a non-linear turbofan engine simulation to produce "snap-shot" measurements, with relevant noise levels, as if collected from a fleet of engines over their lifetime of use. Each engine within the fleet will experience unique operating and deterioration profiles, and may encounter randomly occurring relevant gas path faults including sensor, actuator and component faults. The challenge to the EHM community is to develop gas path diagnostic algorithms to reliably perform fault detection and isolation. An example solution to the benchmark problem is provided along with associated evaluation metrics. A plan is presented to disseminate this benchmark problem to the engine health management technical community and invite technology solutions.

  2. BENCHMARKING WORKSHOPS AS A TOOL TO RAISE BUSINESS EXCELLENCE

    Directory of Open Access Journals (Sweden)

    Milos Jelic

    2011-03-01

    Full Text Available Annual competition for national award for business excellence appears to be a good opportunity for participating organizations to demonstrate their practices particularly those ones which enable them to excel. National quality award competition in Serbia (and Montenegro, namely "OSKAR KVALITETA" started in 1995 but was limited to competition cycle only. However, upon establishing Fund for Quality Culture and Excellence - FQCE in 2002, which took over OSKAR KVALITETA model, several changes took place. OSKAR KVALITETA turned to be annual competition in business excellence, but at the same time FQCE started to offer much wider portfolio of its services including levels of excellence programs, assessment and self-assessment training courses and benchmarking workshops. These benchmarking events have hosted by Award winners or other laureates in OSKAR KVALITETA competition who demonstrated excellence in regard of some particular criteria thus being in position to share their practice with other organizations. For six years experience in organizing benchmarking workshops FQCE scored 31 workshops covering major part of model issues. Increasing level of participation on the workshops and distinct positive trends of participants expressed satisfaction may serve as a reliable indicator that the workshops have been effective in actuating people to think and move in business excellence direction.

  3. Benchmarking of the FENDL-3 Neutron Cross-Section Data Library for Fusion Applications

    International Nuclear Information System (INIS)

    This report summarizes the benchmark analyses performed in a joint effort of ENEA (Italy), JAEA (Japan), KIT (Germany), and the University of Wisconsin (USA) with the objective to test and qualify the neutron induced general purpose FENDL-3.0 data library for fusion applications. The benchmark approach consisted of two major steps including the analysis of a simple ITER-like computational benchmark, and a series of analyses of benchmark experiments conducted previously at the 14 MeV neutron generator facilities at ENEA Frascati, Italy (FNG) and JAEA, Tokai-mura, Japan (FNS). The computational benchmark revealed a modest increase of the neutron flux levels in the deep penetration regions and a substantial increase of the gas production in steel components. The comparison to experimental results showed good agreement with no substantial differences between FENDL-3.0 and FENDL-2.1 for most of the responses analysed. There is a slight trend, however, for an increase of the fast neutron flux in the shielding experiment and a decrease in the breeder mock-up experiments. The photon flux spectra measured in the bulk shield and the tungsten experiments are significantly better reproduced with FENDL-3.0 data. In general, FENDL-3, as compared to FENDL-2.1, shows an improved performance for fusion neutronics applications. It is thus recommended to ITER to replace FENDL-2.1 as reference data library for neutronics calculation by FENDL-3.0. (author)

  4. Benchmarking von Krankenhausinformationssystemen – eine vergleichende Analyse deutschsprachiger Benchmarkingcluster

    Directory of Open Access Journals (Sweden)

    Jahn, Franziska

    2015-08-01

    Full Text Available Benchmarking is a method of strategic information management used by many hospitals today. During the last years, several benchmarking clusters have been established within the German-speaking countries. They support hospitals in comparing and positioning their information system’s and information management’s costs, performance and efficiency against other hospitals. In order to differentiate between these benchmarking clusters and to provide decision support in selecting an appropriate benchmarking cluster, a classification scheme is developed. The classification scheme observes both general conditions and examined contents of the benchmarking clusters. It is applied to seven benchmarking clusters which have been active in the German-speaking countries within the last years. Currently, performance benchmarking is the most frequent benchmarking type, whereas the observed benchmarking clusters differ in the number of benchmarking partners and their cooperation forms. The benchmarking clusters also deal with different benchmarking subjects. Assessing costs and quality application systems, physical data processing systems, organizational structures of information management and IT services processes are the most frequent benchmarking subjects. There is still potential for further activities within the benchmarking clusters to measure strategic and tactical information management, IT governance and quality of data and data-processing processes. Based on the classification scheme and the comparison of the benchmarking clusters, we derive general recommendations for benchmarking of hospital information systems.

  5. Results of VVER-440 fuel assembly head benchmark

    International Nuclear Information System (INIS)

    In the WWER-440/213 type reactors, the core outlet temperature field is monitored with in-core thermocouples, which are installed above 210 fuel assemblies. These measured temperatures are used in determination of the fuel assembly powers and they have important role in the reactor power limitation. For these reasons, correct interpretation of the thermocouple signals is an important question. In order to interpret the signals in correct way, knowledge of the coolant mixing in the assembly heads is necessary. Computational Fluid Dynamics codes and experiments can help to understand better these mixing processes and they can provide information which can support the more adequate interpretation of the thermocouple signals. This benchmark deals with the 3D Computational Fluid Dynamics modeling of the coolant mixing in the heads of the profiled fuel assemblies with 12,2 mm rod pitch. Two assemblies of the twenty third cycle of the Paks NPPs Unit 3 are investigated. One of them has symmetrical pin power profile and another possesses inclined profile. In this benchmark, the same fuel assemblies are investigated by the participants thus the results calculated with different codes and models can be compared with each other. Aims of benchmark was comparison of participants results with each other and with in-core measurement data of the Paks NPP in order to test the different Computational Fluid Dynamics codes and applied Computational Fluid Dynamics models. This paper contains OKB 'GIDROPRESSs' results of Computational Fluid Dynamics calculations this benchmark. Results are:-In-core thermocouple signals above the selected assemblies;-Deviations between the in- ore thermocouple signals and the outlet average coolant temperatures of the assemblies;-Axial velocity and temperature profiles along three diameters at the level of the thermocouple;- Axial velocity and temperature distributions in the cross section at the level of the thermocouple;-Axial velocity and temperature

  6. Shielding experimental benchmark storage, retrieval, and display system

    International Nuclear Information System (INIS)

    The complete description of an integral shielding benchmark experiment includes the radiation source, materials, physical geometry, and measurement data. This information is not usually contained in a single document, but must be gathered from several sources, including personal contact with the experimentalists. A comprehensive database of the experimental details is extremely useful and cost-effective in present day computations. Further, experimental data are vulnerable to being lost or destroyed as a result of facility closures, retirement of experimental personnel, and ignorance. A standard set of experiments, used globally, establishes a framework to validate and verify models in computer codes and guarantee comparative analyses between different computational systems. SINBAD is a database that was conceived in 1992 to store, retrieve, and display the measurements from international experiments for the past 50 years in nuclear shielding. Based at Oak Ridge National Laboratory's Radiation Safety Information and Computational Center (RSICC) SINBAD has a collection of integral benchmark experiments from around the world. SINBAD is shared with the Office of Economic and Cooperative Development Nuclear Energy Agency Data Bank, which provides contributions from Europe, Russia, and Japan. (author)

  7. Shielding Experimental Benchmark Storage, Retrieval, and Display System

    Energy Technology Data Exchange (ETDEWEB)

    Hunter, H.; Kodeli, I.; Marshall, W.J.; Parson, J.; Sartor, E.

    1999-10-17

    The complete description of an integral shielding benchmark experiment includes the radiation source, materials, physical geometry, and measurement data. This information is not usually contained in a single document, but must be gathered from several sources, including personal contact with the experimentalists. A comprehensive database of the experimental details is extremely useful and cost-effective in present day computations. Further, experimental data are vulnerable to being lost or destroyed as a result of facility closures, retirement of experimental personnel, and ignorance. A standard set of experiments, used globally, establishes a framework to validate and verify models in computer codes and guarantee comparative analyses between different computational systems. SINBAD is a database that was conceived in 1992 to store, retrieve, and display the measurements from international experiments for the past 50 years in nuclear shielding. Based at Oak Ridge National Laboratory's Radiation Safety Information and Computational Center (RSICC) SINBAD has a collection of integral benchmark experiments from around the world. SINBAD is shared with the Office of Economic and Cooperative Development/Nuclear Energy Agency Data Bank, which provides contributions from Europe, Russia, and Japan.

  8. Shielding experimental benchmark storage, retrieval, and display system

    Energy Technology Data Exchange (ETDEWEB)

    Hunter, H.T. [Oak Ridge National Laboratory, Oak Ridge, TN (United States); Parsons, J.L. [University of Tennessee, Castle Oak Place Knoxville, TN (United States); Marshall, W.J. [University of Missouri, Knoxville, TN (United States); Sartori, E. [OECD/Nuclear Energy Agency Data Bank, Issy les Moulineaux (France); Kodeli, I. [Institute Jozef Stefan, Ljubljana (Slovenia)

    2000-03-01

    The complete description of an integral shielding benchmark experiment includes the radiation source, materials, physical geometry, and measurement data. This information is not usually contained in a single document, but must be gathered from several sources, including personal contact with the experimentalists. A comprehensive database of the experimental details is extremely useful and cost-effective in present day computations. Further, experimental data are vulnerable to being lost or destroyed as a result of facility closures, retirement of experimental personnel, and ignorance. A standard set of experiments, used globally, establishes a framework to validate and verify models in computer codes and guarantee comparative analyses between different computational systems. SINBAD is a database that was conceived in 1992 to store, retrieve, and display the measurements from international experiments for the past 50 years in nuclear shielding. Based at Oak Ridge National Laboratory's Radiation Safety Information and Computational Center (RSICC) SINBAD has a collection of integral benchmark experiments from around the world. SINBAD is shared with the Office of Economic and Cooperative Development Nuclear Energy Agency Data Bank, which provides contributions from Europe, Russia, and Japan. (author)

  9. Benchmarking ENDF/B-VII.1, JENDL-4.0 and JEFF-3.1.1 with MCNP6

    Science.gov (United States)

    van der Marck, Steven C.

    2012-12-01

    Recent releases of three major world nuclear reaction data libraries, ENDF/B-VII.1, JENDL-4.0, and JEFF-3.1.1, have been tested extensively using benchmark calculations. The calculations were performed with the latest release of the continuous energy Monte Carlo neutronics code MCNP, i.e. MCNP6. Three types of benchmarks were used, viz. criticality safety benchmarks, (fusion) shielding benchmarks, and reference systems for which the effective delayed neutron fraction is reported. For criticality safety, more than 2000 benchmarks from the International Handbook of Criticality Safety Benchmark Experiments were used. Benchmarks from all categories were used, ranging from low-enriched uranium, compound fuel, thermal spectrum ones (LEU-COMP-THERM), to mixed uranium-plutonium, metallic fuel, fast spectrum ones (MIX-MET-FAST). For fusion shielding many benchmarks were based on IAEA specifications for the Oktavian experiments (for Al, Co, Cr, Cu, LiF, Mn, Mo, Si, Ti, W, Zr), Fusion Neutronics Source in Japan (for Be, C, N, O, Fe, Pb), and Pulsed Sphere experiments at Lawrence Livermore National Laboratory (for 6Li, 7Li, Be, C, N, O, Mg, Al, Ti, Fe, Pb, D2O, H2O, concrete, polyethylene and teflon). The new functionality in MCNP6 to calculate the effective delayed neutron fraction was tested by comparison with more than thirty measurements in widely varying systems. Among these were measurements in the Tank Critical Assembly (TCA in Japan) and IPEN/MB-01 (Brazil), both with a thermal spectrum, two cores in Masurca (France) and three cores in the Fast Critical Assembly (FCA, Japan), all with fast spectra. The performance of the three libraries, in combination with MCNP6, is shown to be good. The results for the LEU-COMP-THERM category are on average very close to the benchmark value. Also for most other categories the results are satisfactory. Deviations from the benchmark values do occur in certain benchmark series, or in isolated cases within benchmark series. Such

  10. MCNP calculations for Russian criticality-safety benchmarks

    International Nuclear Information System (INIS)

    The current edition of the International Handbook of Evaluated Criticality Safety Benchmark Experiments contains evaluations of 20 critical experiments performed and evaluated by the Institute for Experimental Physics of the Russian Federal Nuclear Center (VNIIEF) at Arzamas-16 and 16 critical experiments performed and evaluated by the Institute for Technical Physics of the Russian Federal Nuclear Center (VNIITF) at Chelyabinsk-70. These fast-spectrum experiments are of particular interest for data testing of ENDF/B-VI because they contain uranium metal systems of intermediate enrichment as well as uranium and plutonium metal systems with reflectors such as graphite, stainless steel, polyethylene, beryllium, and beryllium oxide. This paper presents the first published results for such systems using cross-section libraries based on ENDF/B-VI

  11. Full sphere hydrodynamic and dynamo benchmarks

    KAUST Repository

    Marti, P.

    2014-01-26

    Convection in planetary cores can generate fluid flow and magnetic fields, and a number of sophisticated codes exist to simulate the dynamic behaviour of such systems. We report on the first community activity to compare numerical results of computer codes designed to calculate fluid flow within a whole sphere. The flows are incompressible and rapidly rotating and the forcing of the flow is either due to thermal convection or due to moving boundaries. All problems defined have solutions that alloweasy comparison, since they are either steady, slowly drifting or perfectly periodic. The first two benchmarks are defined based on uniform internal heating within the sphere under the Boussinesq approximation with boundary conditions that are uniform in temperature and stress-free for the flow. Benchmark 1 is purely hydrodynamic, and has a drifting solution. Benchmark 2 is a magnetohydrodynamic benchmark that can generate oscillatory, purely periodic, flows and magnetic fields. In contrast, Benchmark 3 is a hydrodynamic rotating bubble benchmark using no slip boundary conditions that has a stationary solution. Results from a variety of types of code are reported, including codes that are fully spectral (based on spherical harmonic expansions in angular coordinates and polynomial expansions in radius), mixed spectral and finite difference, finite volume, finite element and also a mixed Fourier-finite element code. There is good agreement between codes. It is found that in Benchmarks 1 and 2, the approximation of a whole sphere problem by a domain that is a spherical shell (a sphere possessing an inner core) does not represent an adequate approximation to the system, since the results differ from whole sphere results. © The Authors 2014. Published by Oxford University Press on behalf of The Royal Astronomical Society.

  12. Higher education information technology management benchmarking in Europe

    OpenAIRE

    Juult, Janne

    2013-01-01

    Objectives of the Study: This study aims to facilitate the rapprochement of the European higher education benchmarking projects towards a unified European benchmarking project. Total of four higher education IT benchmarking projects are analysed by comparing their categorisation of benchmarking indicators and their data manipulation processes. Four select benchmarking projects are compared in this fashion for the first time. The focus is especially on the Finnish Bencheit project's point o...

  13. Revaluering benchmarking - A topical theme for the construction industry

    DEFF Research Database (Denmark)

    Rasmussen, Grane Mikael Gregaard

    2011-01-01

    Over the past decade, benchmarking has increasingly gained foothold in the construction industry. The predominant research, perceptions and uses of benchmarking are valued so strongly and uniformly, that what may seem valuable, is actually abstaining researchers and practitioners from studying and...... questioning the concept objectively. This paper addresses the underlying nature of benchmarking, and accounts for the importance of focusing attention on the sociological impacts benchmarking has in organizations. To understand these sociological impacts, benchmarking research needs to transcend the...

  14. Criteria of benchmark selection for efficient flexible multibody system formalisms

    Directory of Open Access Journals (Sweden)

    Valášek M.

    2007-10-01

    Full Text Available The paper deals with the selection process of benchmarks for testing and comparing efficient flexible multibody formalisms. The existing benchmarks are briefly summarized. The purposes for benchmark selection are investigated. The result of this analysis is the formulation of the criteria of benchmark selection for flexible multibody formalisms. Based on them the initial set of suitable benchmarks is described. Besides that the evaluation measures are revised and extended.

  15. Preliminary Benchmark Evaluation of Japan’s High Temperature Engineering Test Reactor

    Energy Technology Data Exchange (ETDEWEB)

    John Darrell Bess

    2009-05-01

    A benchmark model of the initial fully-loaded start-up core critical of Japan’s High Temperature Engineering Test Reactor (HTTR) was developed to provide data in support of ongoing validation efforts of the Very High Temperature Reactor Program using publicly available resources. The HTTR is a 30 MWt test reactor utilizing graphite moderation, helium coolant, and prismatic TRISO fuel. The benchmark was modeled using MCNP5 with various neutron cross-section libraries. An uncertainty evaluation was performed by perturbing the benchmark model and comparing the resultant eigenvalues. The calculated eigenvalues are approximately 2-3% greater than expected with an uncertainty of ±0.70%. The primary sources of uncertainty are the impurities in the core and reflector graphite. The release of additional HTTR data could effectively reduce the benchmark model uncertainties and bias. Sensitivity of the results to the graphite impurity content might imply that further evaluation of the graphite content could significantly improve calculated results. Proper characterization of graphite for future Next Generation Nuclear Power reactor designs will improve computational modeling capabilities. Current benchmarking activities include evaluation of the annular HTTR cores and assessment of the remaining start-up core physics experiments, including reactivity effects, reactivity coefficient, and reaction-rate distribution measurements. Long term benchmarking goals might include analyses of the hot zero-power critical, rise-to-power tests, and other irradiation, safety, and technical evaluations performed with the HTTR.

  16. MCNP (trademark) ENDF/B-VI iron benchmark calculations

    Science.gov (United States)

    Court, J. D.; Hendricks, J. S.

    Four iron shielding benchmarks have been calculated for, we believe the first time, with MCNP4A and its new ENDF/B-VI library. These calculations are part of the Hiroshima/Nagasaki dose re-evaluation for the National Academy of Sciences and the Defense Nuclear Agency. We believe these calculations are significant because they validate MCNP and the new ENDF/B-VI libraries. These calculations are compared to ENDF/B-V, experiment, and in some cases the recommended MCNP data library (a T-2 evaluation) and ENDF/IV.

  17. Benchmark assemblies of the Los Alamos Critical Assemblies Facility

    International Nuclear Information System (INIS)

    Several critical assemblies of precisely known materials composition and easily calculated and reproducible geometries have been constructed at the Los Alamos National Laboratory. Some of these machines, notably Jezebel, Flattop, Big Ten, and Godiva, have been used as benchmark assemblies for the comparison of the results of experimental measurements and computation of certain nuclear reaction parameters. These experiments are used to validate both the input nuclear data and the computational methods. The machines and the applications of these machines for integral nuclear data checks are described

  18. Toxicological benchmarks for wildlife: 1994 Revision

    International Nuclear Information System (INIS)

    The process by which ecological risks of environmental contaminants are evaluated is two-tiered. The first tier is a screening assessment where concentrations of contaminants in the environment are compared to toxicological benchmarks which represent concentrations of chemicals in environmental media (water, sediment, soil, food, etc.) that are presumed to be nonhazardous to the surrounding biota. The second tier is a baseline ecological risk assessment where toxicological benchmarks are one of several lines of evidence used to support or refute the presence of ecological effects. The report presents toxicological benchmarks for assessment of effects of 76 chemicals on 8 representative mammalian wildlife species and 31 chemicals on 9 avian wildlife species. The chemicals are some of those that occur at United States Department of Energy waste sites; the wildlife species were chosen because they are widely distributed and provide a representative range of body sizes and diets. Further descriptions of the chosen wildlife species and chemicals are provided in the report. The benchmarks presented in this report represent values believed to be nonhazardous for the listed wildlife species. These benchmarks only consider contaminant exposure through oral ingestion of contaminated media; exposure through inhalation or direct dermal exposure are not considered in this report

  19. Toxicological benchmarks for wildlife: 1994 Revision

    Energy Technology Data Exchange (ETDEWEB)

    Opresko, D.M.; Sample, B.E.; Suter, G.W. II

    1994-09-01

    The process by which ecological risks of environmental contaminants are evaluated is two-tiered. The first tier is a screening assessment where concentrations of contaminants in the environment are compared to toxicological benchmarks which represent concentrations of chemicals in environmental media (water, sediment, soil, food, etc.) that are presumed to be nonhazardous to the surrounding biota. The second tier is a baseline ecological risk assessment where toxicological benchmarks are one of several lines of evidence used to support or refute the presence of ecological effects. The report presents toxicological benchmarks for assessment of effects of 76 chemicals on 8 representative mammalian wildlife species and 31 chemicals on 9 avian wildlife species. The chemicals are some of those that occur at United States Department of Energy waste sites; the wildlife species were chosen because they are widely distributed and provide a representative range of body sizes and diets. Further descriptions of the chosen wildlife species and chemicals are provided in the report. The benchmarks presented in this report represent values believed to be nonhazardous for the listed wildlife species. These benchmarks only consider contaminant exposure through oral ingestion of contaminated media; exposure through inhalation or direct dermal exposure are not considered in this report.

  20. Air ingress benchmarking with computational fluid dynamics analysis

    International Nuclear Information System (INIS)

    The air ingress accident is a complicated accident scenario that may limit the deployment of high-temperature gas reactors. The complexity of this accident scenario is compounded by multiple physical phenomena that are involved in the air ingress event. These include diffusion, natural circulation, and complex chemical reactions with graphite and oxygen. In an attempt to better understand the phenomenon, the FLUENT-6 computational fluid dynamics code was used to assess two air ingress experiments. The first was the Japanese series of tests performed in the early 1990s by Takeda and Hishida. These separate effects tests were conducted to understand and model a multi-component experiment in which all three processes were included with the introduction of air in a heated graphite column. MIT used the FLUENT code to benchmark these series of tests with quite good results. These tests are generically applicable to prismatic reactors and the lower reflector regions of pebble-bed reactors. The second series of tests were performed at the NACOK facility for pebble bed reactors as reported by Kuhlmann [Kuhlmann, M.B., 1999. Experiments to investigate flow transfer and graphite corrosion in case of air ingress accidents in a high-temperature reactor]. These tests were aimed at understanding natural circulation of pebble bed reactors by simulating hot and cold legs of these reactors. The FLUENT code was also successfully used to simulate these tests. The results of these benchmarks and the findings will be presented