WorldWideScience

Sample records for consistency test pct

  1. Nuclear waste glass product consistency test (PCT), Version 5.0

    International Nuclear Information System (INIS)

    Jantzen, C.M.; Bibler, N.E.; Beam, D.C.; Ramsey, W.G.; Waters, B.J.

    1992-06-01

    Liquid high-level nuclear waste will be immobilized at the Savannah River Site (SRS) by vitrification in borosilicate glass. The glass will be produced in the Defense Waste Processing Facility (DWPF), poured into stainless steel canisters, and eventually disposed of in a geologic repository. In order to comply with the Waste Acceptance Preliminary Specifications (WAPS), the durability of the glass needs to be measured during production to assure its long term stability and radionuclide release properties. A durability test, designated the Produce Consistency Test (PCT), was developed for DWPF glass in order to meet the WAPS requirements. The response of the PCT procedure was based on extensive testing with glasses of widely different compositions. The PCT was determined to be very reproducible, to yield reliable results rapidly, and to be easily performed in shielded cell facilities with radioactive samples. Version 5.0 of the PCT procedure is attached

  2. Nuclear waste glass product consistency test (PCT): Version 7.0. Revision 3

    International Nuclear Information System (INIS)

    Jantzen, C.M.; Bibler, N.E.; Beam, D.C.; Ramsey, W.G.

    1994-06-01

    Liquid high-level nuclear waste will be immobilized at the Savannah River Site (SRS) by vitrification in borosilicate glass. The glass will be produced in the Defense Waste Processing Facility (DWPF), poured into stainless steel canisters, and eventually disposed of in a geologic repository. In order to comply with the Waste Acceptance Product Specifications (WAPS), the durability of the glass needs to be measured during production to assure its long term stability and radionuclide release properties. A durability test, designated the Product Consistency Test (PCT), was developed for DWPF glass in order to meet the WAPS requirements. The response of the PCT procedure was based on extensive testing with glasses of widely different compositions. The PCT was determined to be very reproducible, to yield reliable results rapidly, and to be easily performed in shielded cell facilities with radioactive samples. Version 7.0 of the PCT procedure is attached. This draft version has been submitted to ASTM for full committee (C26, Nuclear Fuel Cycle) ballot after being balloted successfully through subcommittee C26.13 on Repository Waste Package Materials Testing

  3. Development of an ASTM standard glass durability test, the Product Consistency Test (PCT), for high level radioactive waste glass

    International Nuclear Information System (INIS)

    Jantzen, C.M.; Bibler, N.E.; Beam, D.C.; Ramsey, W.G.

    1994-01-01

    The nation's first, and the world's largest, facility to immobilize high-level nuclear waste in durable borosilicate glass has started operation at the Savannah River Site (SRS) in Aiken, South Carolina. The product specifications on the glass wasteform produced in the Defense Waste Processing Facility (DWPF) required extensive characterization of the glass product before actual production began and for continued characterization during production. To aid in this characterization, a glass durability (leach) test was needed that was easily reproducible, could be performed remotely on highly radioactive samples, and could yield results rapidly. Several standard leach tests were examined with a variety of test configurations. Using existing tests as a starting point, the DWPF Product Consistency Test (PCT was developed in which crushed glass samples are exposed to 90 ± 2 degree C deionized water for seven days. Based on extensive testing, including a seven-laboratory round robin and confirmatory testing with radioactive samples, the PCT is very reproducible, yields reliable results rapidly, and can be performed in shielded cell facilities with radioactive samples

  4. Nonlinear relationship between the Product Consistency Test (PCT) response and the Al/B ratio in a soda-lime aluminoborosilicate glass

    Energy Technology Data Exchange (ETDEWEB)

    Farooqi, Rahmat Ullah, E-mail: rufarooqi@postech.ac.kr [Division of Advanced Nuclear Engineering, Pohang University of Science and Technology, 77 Cheongam-Ro, Nam-Gu, Pohang, Gyeongbuk 790-784 (Korea, Republic of); Hrma, Pavel [Division of Advanced Nuclear Engineering, Pohang University of Science and Technology, 77 Cheongam-Ro, Nam-Gu, Pohang, Gyeongbuk 790-784 (Korea, Republic of); Pacific Northwest National Laboratory, Richland, WA (United States)

    2016-06-15

    We have investigated the effect of Al/B ratio on the Product Consistency Test (PCT) response. In an aluminoborosilicate soda-lime glass based on a modified International Simple Glass, ISG-3, the Al/B ratio varied from 0 to 0.55 (in mole fractions). In agreement with various models of the PCT response as a function of glass composition, we observed a monotonic increase of B and Na releases with decreasing Al/B mole ratio, but only when the ratio was higher than 0.05. Below this value (Al/B < 0.05), we observed a sharp decrease that we attribute to B in tetrahedral coordination.

  5. Standard test methods for determining chemical durability of nuclear, hazardous, and mixed waste glasses and multiphase glass ceramics: The product consistency test (PCT)

    CERN Document Server

    American Society for Testing and Materials. Philadelphia

    2002-01-01

    1.1 These product consistency test methods A and B evaluate the chemical durability of homogeneous glasses, phase separated glasses, devitrified glasses, glass ceramics, and/or multiphase glass ceramic waste forms hereafter collectively referred to as “glass waste forms” by measuring the concentrations of the chemical species released to a test solution. 1.1.1 Test Method A is a seven-day chemical durability test performed at 90 ± 2°C in a leachant of ASTM-Type I water. The test method is static and conducted in stainless steel vessels. Test Method A can specifically be used to evaluate whether the chemical durability and elemental release characteristics of nuclear, hazardous, and mixed glass waste forms have been consistently controlled during production. This test method is applicable to radioactive and simulated glass waste forms as defined above. 1.1.2 Test Method B is a durability test that allows testing at various test durations, test temperatures, mesh size, mass of sample, leachant volume, a...

  6. Effects of container material on PCT leach test results for high-level nuclear waste glasses

    International Nuclear Information System (INIS)

    Xing, S.B.; Pegg, I.L.

    1994-01-01

    A glass-based waste form used for the immobilization of high-level nuclear wastes should exhibit good resistance to aqueous corrosion since typically this is the primary process by which radionucleides could be released into the environment upon failure of other barriers. In the USA, the Waste Acceptance Product Specifications (WAPS) provides a set of requirements to ensure the consistency of the waste forms produced and specifies the Product Consistency Test (PCT) as a measure of relative chemical durability. While the PCT procedure permits usage of both Teflon and stainless steel vessels for testing of simulated development glasses, Teflon is not permitted for testing of production glasses due to radiative degradation. The results presented in this paper indicate that there are very significant differences between tests conducted in the two types of vessels due to the well-known permeability of Teflon to atmospheric carbon dioxide which results in lowering of the solution pH and a consequent reduction in the leach rate of silicate glasses. A wide range of nuclear waste glass compositions was subjected to the PCT procedure using both Teflon and stainless steel vessels. The magnitude of the effect (up to a factor of four for B, Na, Li concentrations) depends strongly on glass composition, therefore the isolated checks performed previously were inconclusive. The permeability to CO, of two types of Teflon vessels specified in the PCT procedure was directly measured using buffer solutions: ingress of CO, is linear in time, strongly pH-dependent, and was as high as 100 ppm after 7 days. In actual PCT tests in Teflon vessels, the total CO, content was 560 ppm after 87 days and 1930 ppm after one year

  7. Should the post-coital test (PCT) be part of the routine fertility work-up?

    NARCIS (Netherlands)

    van der Steeg, Jan W.; Steures, Pieternel; Eijkemans, Marinus J. C.; Habbema, J. Dik; van der Veen, Fulco; Bossuyt, Patrick M. M.; Hompes, Peter G. A.; Mol, Ben W. J.

    2004-01-01

    BACKGROUND: This study aimed to determine whether medical history and semen analysis can predict the result of the post-coital test (PCT). METHODS: A previously reported data set of Dutch patients collected between 1985 and 1993 was used. Our study was limited to just patients with an ovulatory

  8. Short-term consistency testing vs. long-term behavior

    International Nuclear Information System (INIS)

    Ebert, W.L.; Bakel, A.J.; Wolf, S.F.; Strachan, D.M.

    1997-01-01

    The authors conducted Product Consistency Tests (PCTs) with a surrogate low-activity waste (LAW) glass to (1) evaluate the possible use of various test conditions in a specification test for LAW waste forms, (2) measure the reproducibility of the test at low temperatures, and (3) determine if the rates calculated from 7-day PCTs bound the rates measured in PCT conducted for longer durations, which represent more advanced corrosion. The effects of temperature and pH on the dissolution rate in PCTs are much less than the effects observed in dilute solutions due to the buildup of dissolved glass components in the PCTs. The precision of replicate 7-day tests at 20 and 40 degrees C was limited by the analytical uncertainty. The dissolution rates at all temperatures decreased with the test duration initially. However, the dissolution rates in tests at 70 and 90 degrees C increased when certain alteration phases formed after about 100 and 500 days, respectively; the rates in some tests exceeded that measured in a 7-day PCT. While the 7-day PCT does not provide a bounding rate for this glass at 70 or 90 degrees C, tests for longer durations are needed to determine if a 7-day test provides a bounding rate at lower temperatures

  9. Product consistency leach tests of Savannah River Site radioactive waste glasses

    International Nuclear Information System (INIS)

    Bibler, N.E.; Bates, J.K.

    1990-01-01

    The product consistency test (PCT) is a glass leach test developed at the Savannah River Site (SRS) to confirm the durability of radioactive nuclear waste glasses that will be produced in the Defense Waste Processing Facility. The PCT is a seven day, crushed glass leach test in deionized water at 90C. Final leachates are filtered and acidified prior to analysis. To demonstrate the reproducibility of the PCT when performed remotely, SRS and Argonne National Laboratory have performed the PCT on samples of two radioactive glasses. The tests were also performed to compare the releases of the radionuclides with the major nonradioactive glass components and to determine if radiation from the glass was affecting the results of the PCT. The test was performed in triplicate at each laboratory. For the major soluble elements, B, Li, Na, and Si, in the glass, each investigator obtained relative precisions in the range 2-5% in the triplicate tests. This range indicates good precision for the PCT when performed remotely with master slave manipulators in a shielded cell environment. When the results of the two laboratories were compared to each other, the agreement was within 20%. Normalized concentrations for the nonradioactive and radioactive elements in the PCT leachates measured at both facilities indicated that the radionuclides were released from the glass slower than the major soluble elements in the glass. For both laboratories, the normalized releases for both glasses were in the general order Li ∼ B ∼ Na > Si > Cs - 137 > Sb - 125 < Sr - 90. The normalized releases for the major soluble elements and the final pH values in the tests with radioactive glass are consistent with those for nonradioactive glasses with similar compositions. This indicates that there was no significant effect of radiation on the results of the PCT

  10. Product consistency leach tests of Savannah River Site radioactive waste glasses

    International Nuclear Information System (INIS)

    Bibler, N.E.; Bates, J.K.

    1989-01-01

    The Product Consistency Test (PCT) is a glass leach test that was developed at the Savannah River Site (SRS) to routinely confirm the durability of nuclear waste glasses that will be produced in the Defense Waste Processing Facility. The PCT is a 7 day, crushed glass leach test in deionized water at 90 degree C. Final leachates are filtered and acidified prior to analysis. To demonstrate the reproducibility of the PCT when performed remotely, SRS and Argonne National Laboratory have performed the PCT on samples of two radioactive glasses. The tests were also performed to compare the releases of the radionuclides with the major nonradioactive glass components and to determine if radiation from the glass was affecting the results of the PCT. The test was performed in triplicate at each laboratory. For the major soluble elements, B, Li, Na, and Si, in the glass, each investigator obtained relative precisions in the range 2--5% in the triplicate tests. This range indicates good precision for the PCT when performed remotely with master slave manipulators in a shielded cell environment

  11. Chemical Composition Analysis and Product Consistency Tests of the ORP Phase 5 Nepheline Study Glasses

    Energy Technology Data Exchange (ETDEWEB)

    Fox, K. M. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Edwards, T. B. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Caldwell, M. E. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Riley, W. T. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL)

    2018-02-01

    In this report, the Savannah River National Laboratory (SRNL) provides chemical analyses and Product Consistency Test (PCT) results for a series of simulated high-level waste glass compositions fabricated by the Pacific Northwest National Laboratory (PNNL). These data will be used in the development of improved models for the prediction of nepheline crystallization in support of the Hanford Tank Waste Treatment and Immobilization Plant (WTP).

  12. Porphyria Cutanea Tarda (PCT)

    Science.gov (United States)

    ... effective than phlebotomy or low dose hydroxychloroquine or chloroquine in treating individuals with PCT. However, these drugs may play a role in treating affected individuals in whom the use of the two front-line therapies is not ...

  13. Rapid and Sensitive Lateral Flow Immunoassay Method for Procalcitonin (PCT Based on Time-Resolved Immunochromatography

    Directory of Open Access Journals (Sweden)

    Xiang-Yang Shao

    2017-02-01

    Full Text Available Procalcitonin (PCT is a current, frequently-used marker for severe bacterial infection. The aim of this study was to develop a cost-effective detection kit for rapid quantitative and on-site detection of PCT. To develop the new PCT quantitative detecting kit, a double-antibody sandwich immunofluorescent assay was employed based on time-resolved immunofluorescent assay (TRFIA combined with lateral flow immunoassay (LFIA. The performance of the new developed kit was evaluated in the aspects of linearity, precision, accuracy, and specificity. Two-hundred thirty-four serum samples were enrolled to carry out the comparison test. The new PCT quantitative detecting kit exhibited a higher sensitivity (0.08 ng/mL. The inter-assay coefficient of variation (CV and the intra-assay CV were 5.4%–7.7% and 5.7%–13.4%, respectively. The recovery rates ranged from 93% to 105%. Furthermore, a high correlation (n = 234, r = 0.977, p < 0.0001 and consistency (Kappa = 0.875 were obtained when compared with the PCT kit from Roche Elecsys BRAHMS. Thus, the new quantitative method for detecting PCT has been successfully established. The results indicated that the newly-developed system based on TRFIA combined with LFIA was suitable for rapid and on-site detection for PCT, which might be a useful platform for other biomarkers in point-of-care tests.

  14. Modeling and Testing Legacy Data Consistency Requirements

    DEFF Research Database (Denmark)

    Nytun, J. P.; Jensen, Christian Søndergaard

    2003-01-01

    An increasing number of data sources are available on the Internet, many of which offer semantically overlapping data, but based on different schemas, or models. While it is often of interest to integrate such data sources, the lack of consistency among them makes this integration difficult....... This paper addresses the need for new techniques that enable the modeling and consistency checking for legacy data sources. Specifically, the paper contributes to the development of a framework that enables consistency testing of data coming from different types of data sources. The vehicle is UML and its...... accompanying XMI. The paper presents techniques for modeling consistency requirements using OCL and other UML modeling elements: it studies how models that describe the required consistencies among instances of legacy models can be designed in standard UML tools that support XMI. The paper also considers...

  15. Application of PCT to the EBR II ceramic waste form

    International Nuclear Information System (INIS)

    Ebert, W. L.; Lewis, M. A.; Johnson, S. G.

    2002-01-01

    We are evaluating the use of the Product Consistency Test (PCT) developed to monitor the consistency of borosilicate glass waste forms for application to the multiphase ceramic waste form (CWF) that will be used to immobilize waste salts generated during the electrometallurgical conditioning of spent sodium-bonded nuclear fuel from the Experimental Breeder Reactor No. 2 (EBR II). The CWF is a multiphase waste form comprised of about 70% sodalite, 25% borosilicate glass binder, and small amounts of halite and oxide inclusions. It must be qualified for disposal as a non-standard high-level waste (HLW) form. One of the requirements in the DOE Waste Acceptance System Requirements Document (WASRD) for HLW waste forms is that the consistency of the waste forms be monitored.[1] Use of the PCT is being considered for the CWF because of the similarities of the dissolution behaviors of both the sodalite and glass binder phases in the CWF to borosilicate HLW glasses. This paper provides (1) a summary of the approach taken in selecting a consistency test for CWF production and (2) results of tests conducted to measure the precision and sensitivity of the PCT conducted with simulated CWF

  16. Microstructure-strength relations in a hardenable stainless steel with 16 pct Cr, 1.5 pct Mo, and 5 pct Ni

    Science.gov (United States)

    Grobner, P. J.; Blšs, V.

    1984-07-01

    Metallographic studies have been conducted on a 0.024 pct C-16 pct Cr-1.5 pct Mo-5 pct Ni stainless steel to study the phase reactions associated with heat treatments and investigate the strengthening mechanisms of the steel. In the normalized condition, air cooled from 1010 °C, the microstructure consists of 20 pct ferrite and 80 pct martensite. Tempering in a temperature range between 500 and 600 °C results in a gradual transformation of martensite to a fine mixture of ferrite and austenite. At higher tempering temperatures, between 600 and 800 °C, progressively larger quantities of austenite form and are converted during cooling to proportionally increasing amounts of fresh martensite. The amount of retained austenite in the microstructure is reduced to zero at 800 °C, and the microstructure contains 65 pct re-formed martensite and 35 pct total ferrite. Chromium rich M23C6 carbides precipitate in the single tempered microstructures. The principal strengthening is produced by the presence of martensite in the microstructure. Additional strengthening is provided by a second tempering treatment at 400 °C due to the precipitation of ultrafine (Cr, Mo) (C,N) particles in the ferrite.

  17. Testing the visual consistency of web sites

    NARCIS (Netherlands)

    van der Geest, Thea; Loorbach, N.R.

    2005-01-01

    Consistency in the visual appearance of Web pages is often checked by experts, such as designers or reviewers. This article reports a card sort study conducted to determine whether users rather than experts could distinguish visual (in-)consistency in Web elements and pages. The users proved to

  18. Consistency test of the standard model

    International Nuclear Information System (INIS)

    Pawlowski, M.; Raczka, R.

    1997-01-01

    If the 'Higgs mass' is not the physical mass of a real particle but rather an effective ultraviolet cutoff then a process energy dependence of this cutoff must be admitted. Precision data from at least two energy scale experimental points are necessary to test this hypothesis. The first set of precision data is provided by the Z-boson peak experiments. We argue that the second set can be given by 10-20 GeV e + e - colliders. We pay attention to the special role of tau polarization experiments that can be sensitive to the 'Higgs mass' for a sample of ∼ 10 8 produced tau pairs. We argue that such a study may be regarded as a negative selfconsistency test of the Standard Model and of most of its extensions

  19. Chemical composition analysis and product consistency tests to support enhanced Hanford waste glass models: Results for the January, March, and April 2015 LAW glasses

    Energy Technology Data Exchange (ETDEWEB)

    Fox, K. M. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Edwards, T. B. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Riley, W. T. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Best, D. R. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL)

    2015-09-03

    In this report, the Savannah River National Laboratory provides chemical analyses and Product Consistency Test (PCT) results for several simulated low activity waste (LAW) glasses (designated as the January, March, and April 2015 LAW glasses) fabricated by the Pacific Northwest National Laboratory. The results of these analyses will be used as part of efforts to revise or extend the validation regions of the current Hanford Waste Treatment and Immobilization Plant glass property models to cover a broader span of waste compositions.

  20. Chemical composition analysis and product consistency tests to support Enhanced Hanford Waste Glass Models. Results for the Augusta and October 2014 LAW Glasses

    Energy Technology Data Exchange (ETDEWEB)

    Fox, K. M. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Edwards, T. B. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Best, D. R. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL)

    2015-07-07

    In this report, the Savannah River National Laboratory provides chemical analyses and Product Consistency Test (PCT) results for several simulated low activity waste (LAW) glasses (designated as the August and October 2014 LAW glasses) fabricated by the Pacific Northwest National Laboratory. The results of these analyses will be used as part of efforts to revise or extend the validation regions of the current Hanford Waste Treatment and Immobilization Plant glass property models to cover a broader span of waste compositions.

  1. Chemical composition analysis and product consistency tests supporting refinement of the Nepheline Model for the high aluminum Hanford glass composition region

    Energy Technology Data Exchange (ETDEWEB)

    Fox, K. M. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Edwards, T. B. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Mcclane, D. L. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL)

    2016-03-01

    In this report, Savannah River National Laboratory provides chemical analyses and Product Consistency Test (PCT) results for a series of simulated high level waste (HLW) glasses fabricated by Pacific Northwest National Laboratory (PNNL) as part of an ongoing nepheline crystallization study. The results of these analyses will be used to improve the ability to predict crystallization of nepheline as a function of composition and heat treatment for glasses formulated at high alumina concentrations.

  2. Chemical composition analysis and product consistency tests supporting refinement of the Nepheline model for the high aluminum Hanford Glass composition region

    Energy Technology Data Exchange (ETDEWEB)

    Fox, K. M. [Savannah River Site (SRS), Aiken, SC (United States); Edwards, T. B. [Savannah River Site (SRS), Aiken, SC (United States); Mcclane, D. L. [Savannah River Site (SRS), Aiken, SC (United States)

    2016-02-17

    In this report, SRNL provides chemical analyses and Product Consistency Test (PCT) results for a series of simulated HLW glasses fabricated by Pacific Northwest National Laboratory (PNNL) as part of an ongoing nepheline crystallization study. The results of these analyses will be used to improve the ability to predict crystallization of nepheline as a function of composition and heat treatment for glasses formulated at high alumina concentrations.

  3. Removal of radon by aeration: testing of various aeration techniques for small water works. For European Commission under Contract No FI4PCT960054 TENAWA project

    International Nuclear Information System (INIS)

    Salonen, L.; Mehtonen, J.; Turunen, H.; Mjoenes, L.; Hagberg, N.; Raff, O.

    2002-12-01

    Capability of various aeration techniques to remove radon from water in small waterworks was studied as a part of project (Treatment Techniques for Removing Natural Radionuclides from Drinking Water), which was carried out during 1997-1999 on a cost-shared basis (contract No. F14PCT960054) with The European Commission (CEC) under the supervision of the Directorate-General XII Radiation Protection Research Unit. In TENAWA project both laboratory and field experiments were performed in order to find reliable methods and equipment for removing natural radionuclides from ground water originating either from private wells or small waterworks. Because such techniques are more often needed in private households than at waterworks, the main emphasis of the research was aimed to solve the water treatment problems related to the private water supplies, especially bedrock wells. Radon was the most important radionuclide to be removed from water at waterworks whereas the removal of other radionuclides ( 234,238 U, 226,228 Ra, 210 Pb and 210 Po) was oft required from radonrich bedrock waters. The currently available methods and equipment were mainly tested during the field and laboratory experiments but the project was also aimed to find new materials, absorbents and membranes applicable for radionuclide removal from various types of ground waters (e.g. soft, hard, acidic). Because iron, manganese or organic occur in waters with radionuclides, their simultaneous removal was also studied. The project was divided into 13 work packages. In this report the results of the work package 2.2 are described. Elevated levels of radon and other natural radionuclides in European ground waters have been observed mainly in wide areas of the crystalline Scandinavian bedrock, especially in the granite rock areas of Finland and Sweden but also in more limited crystalline rock areas of Central and Southern Europe, Ukraine and Scotland. The radon removal efficiencies of different aeration methods

  4. Product consistency testing of three reference glasses in stainless steel and perfluoroalkoxy resin vessels

    International Nuclear Information System (INIS)

    Olson, K.M.; Smith, G.L.; Marschman, S.C.

    1995-03-01

    Because of their chemical durability, silicate glasses have been proposed and researched since the mid-1950s as a medium for incorporating high-level radioactive waste (HLW) generated from processing of nuclear materials. A number of different waste forms were evaluated and ranked in the early 1980s; durability (leach resistance) was the highest weighted factor. Borosilicate glass was rated the best waste form available for incorporation of HLW. Four different types of vessels and three different glasses were used to study the possible effect of vessel composition on durability test results from the Production Consistency Test (PCT). The vessels were 45-m 304 stainless steel vessels, 150-m 304 L stainless steel vessels, and 60-m perfluoroalkoxy (PFA) fluoropolymer resin vessels. The three glasses were the Environmental Assessment glass manufactured by Corning Incorporated and supplied by Westinghouse Savannah River company, and West Valley Nuclear Services reference glasses 5 and 6, manufactured and supplied by Catholic University of America. Within experimental error, no differences were found in durability test results using the 3 different glasses in the 304L stainless steel or PFA fluoropolymer resin vessels over the seven-day test period

  5. Removal of radon by aeration testing of various aeration techniques for small water works. For European Commission under Contract No FI4PCT960054 TENAWA project

    CERN Document Server

    Salonen, L; Mehtonen, J; Mjoenes, L; Raff, O; Turunen, H

    2002-01-01

    Capability of various aeration techniques to remove radon from water in small waterworks was studied as a part of project (Treatment Techniques for Removing Natural Radionuclides from Drinking Water), which was carried out during 1997-1999 on a cost-shared basis (contract No. F14PCT960054) with The European Commission (CEC) under the supervision of the Directorate-General XII Radiation Protection Research Unit. In TENAWA project both laboratory and field experiments were performed in order to find reliable methods and equipment for removing natural radionuclides from ground water originating either from private wells or small waterworks. Because such techniques are more often needed in private households than at waterworks, the main emphasis of the research was aimed to solve the water treatment problems related to the private water supplies, especially bedrock wells. Radon was the most important radionuclide to be removed from water at waterworks whereas the removal of other radionuclides ( sup 2 sup 3 sup 4...

  6. Cosmological consistency tests of gravity theory and cosmic acceleration

    Science.gov (United States)

    Ishak-Boushaki, Mustapha B.

    2017-01-01

    Testing general relativity at cosmological scales and probing the cause of cosmic acceleration are among the important objectives targeted by incoming and future astronomical surveys and experiments. I present our recent results on consistency tests that can provide insights about the underlying gravity theory and cosmic acceleration using cosmological data sets. We use statistical measures, the rate of cosmic expansion, the growth rate of large scale structure, and the physical consistency of these probes with one another.

  7. Potential application of the consistency approach for vaccine potency testing.

    Science.gov (United States)

    Arciniega, J; Sirota, L A

    2012-01-01

    The Consistency Approach offers the possibility of reducing the number of animals used for a potency test. However, it is critical to assess the effect that such reduction may have on assay performance. Consistency of production, sometimes referred to as consistency of manufacture or manufacturing, is an old concept implicit in regulation, which aims to ensure the uninterrupted release of safe and effective products. Consistency of manufacture can be described in terms of process capability, or the ability of a process to produce output within specification limits. For example, the standard method for potency testing of inactivated rabies vaccines is a multiple-dilution vaccination challenge test in mice that gives a quantitative, although highly variable estimate. On the other hand, a single-dilution test that does not give a quantitative estimate, but rather shows if the vaccine meets the specification has been proposed. This simplified test can lead to a considerable reduction in the number of animals used. However, traditional indices of process capability assume that the output population (potency values) is normally distributed, which clearly is not the case for the simplified approach. Appropriate computation of capability indices for the latter case will require special statistical considerations.

  8. Short-Cut Estimators of Criterion-Referenced Test Consistency.

    Science.gov (United States)

    Brown, James Dean

    1990-01-01

    Presents simplified methods for deriving estimates of the consistency of criterion-referenced, English-as-a-Second-Language tests, including (1) the threshold loss agreement approach using agreement or kappa coefficients, (2) the squared-error loss agreement approach using the phi(lambda) dependability approach, and (3) the domain score…

  9. Migraine patients consistently show abnormal vestibular bedside tests

    Directory of Open Access Journals (Sweden)

    Eliana Teixeira Maranhão

    2015-01-01

    Full Text Available Migraine and vertigo are common disorders, with lifetime prevalences of 16% and 7% respectively, and co-morbidity around 3.2%. Vestibular syndromes and dizziness occur more frequently in migraine patients. We investigated bedside clinical signs indicative of vestibular dysfunction in migraineurs.Objective To test the hypothesis that vestibulo-ocular reflex, vestibulo-spinal reflex and fall risk (FR responses as measured by 14 bedside tests are abnormal in migraineurs without vertigo, as compared with controls.Method Cross-sectional study including sixty individuals – thirty migraineurs, 25 women, 19-60 y-o; and 30 gender/age healthy paired controls.Results Migraineurs showed a tendency to perform worse in almost all tests, albeit only the Romberg tandem test was statistically different from controls. A combination of four abnormal tests better discriminated the two groups (93.3% specificity.Conclusion Migraine patients consistently showed abnormal vestibular bedside tests when compared with controls.

  10. Migraine patients consistently show abnormal vestibular bedside tests.

    Science.gov (United States)

    Maranhão, Eliana Teixeira; Maranhão-Filho, Péricles; Luiz, Ronir Raggio; Vincent, Maurice Borges

    2016-01-01

    Migraine and vertigo are common disorders, with lifetime prevalences of 16% and 7% respectively, and co-morbidity around 3.2%. Vestibular syndromes and dizziness occur more frequently in migraine patients. We investigated bedside clinical signs indicative of vestibular dysfunction in migraineurs. To test the hypothesis that vestibulo-ocular reflex, vestibulo-spinal reflex and fall risk (FR) responses as measured by 14 bedside tests are abnormal in migraineurs without vertigo, as compared with controls. Cross-sectional study including sixty individuals - thirty migraineurs, 25 women, 19-60 y-o; and 30 gender/age healthy paired controls. Migraineurs showed a tendency to perform worse in almost all tests, albeit only the Romberg tandem test was statistically different from controls. A combination of four abnormal tests better discriminated the two groups (93.3% specificity). Migraine patients consistently showed abnormal vestibular bedside tests when compared with controls.

  11. GLASS FABRICATION AND PRODUCT CONSISTENCY TESTING OF LANTHANIDE BOROSILICATE FRIT X COMPOSITION FOR PLUTONIUM DISPOSITION

    Energy Technology Data Exchange (ETDEWEB)

    Marra, J

    2006-11-15

    The Department of Energy Office of Environmental Management (DOE/EM) plans to conduct the Plutonium Disposition Project at the Savannah River Site (SRS) to disposition excess weapons-usable plutonium. A plutonium glass waste form is the preferred option for immobilization of the plutonium for subsequent disposition in a geologic repository. A reference glass composition (Lanthanide Borosilicate (LaBS) Frit B) was developed during the Plutonium Immobilization Program (PIP) to immobilize plutonium in the late 1990's. A limited amount of performance testing was performed on this baseline composition before efforts to further pursue Pu disposition via a glass waste form ceased. Recent FY05 studies have further investigated the LaBS Frit B formulation as well as development of a newer LaBS formulation denoted as LaBS Frit X. The objectives of this present task were to fabricate plutonium loaded LaBS Frit X glass and perform corrosion testing to provide near-term data that will increase confidence that LaBS glass product is suitable for disposal in the Yucca Mountain Repository. Specifically, testing was conducted in an effort to provide data to Yucca Mountain Project (YMP) personnel for use in performance assessment calculations. Plutonium containing LaBS glass with the Frit X composition with a 9.5 wt% PuO{sub 2} loading was prepared for testing. Glass was prepared to support Product Consistency Testing (PCT) at Savannah River National Laboratory (SRNL). The glass was thoroughly characterized using x-ray diffraction (XRD) and scanning electron microscopy coupled with energy dispersive spectroscopy (SEM/EDS) prior to performance testing. A series of PCTs were conducted at SRNL using quenched Pu Frit X glass with varying exposed surface areas. Effects of isothermal and can-in-canister heat treatments on the Pu Frit X glass were also investigated. Another series of PCTs were performed on these different heat-treated Pu Frit X glasses. Leachates from all these PCTs

  12. Clinical validity of panoramic radiographs with Digora PCT

    International Nuclear Information System (INIS)

    Matsuda, Yukiko; Araki, Kazuyuki; Hanazawa, Tomomi; Seki, Kenji; Okano, Tomohiro

    2006-01-01

    The quality of panoramic radiographs with a digital imaging system, Digora PCT, was compared with that of a Fuji Computed Radiography (FCR) system. The exposure settings that would produce the appropriate density for observation of the images were determined. The images of the Digora PCT and the FCR system were observed on a 17-inch CRT monitor and printed films respectively. Thirty images of patients were obtained with each system. Six oral radiologists observed the images and evaluated the visibility of pathologic conditions based on outline of mandible, mandibular canal and bone trabecula, shape of the tooth and periodontal hard tissue of the mandible, shape of the tooth and periodontal hard tissue of the maxilla, and outline of the bony structure of the maxillary sinus, using the following three categories: good, fair, and unacceptable. The images of each system were evaluated separately at an interval of 3 months. Mann-Whitney test with p=0.05 was used to analyze differences between the two systems. There were no images of either system rated as ''unacceptable''. The visibility of the shape of the tooth, periodontal hard tissue, and the outline of the bony structure of the maxillary sinus with the Digora PCT was inferior to that with the FCR system. However, it can be concluded that the image quality of the Digora PCT is adequate for dental purposes because there were no unacceptable images. (author)

  13. Prospective studies comparing PCT and CRP

    African Journals Online (AJOL)

    Yoshan Moodley

    2012-04-24

    Apr 24, 2012 ... bacterial cell wall components such as lipopolysac- charide, are known .... acquired pneumonia, PCT showed improved perfor- mance at predicting .... adrenomedullin to predict severity and outcome in community- acquired ...

  14. Long-term product consistency test of simulated 90-19/Nd HLW glass

    International Nuclear Information System (INIS)

    Gan, X.Y.; Zhang, Z.T.; Yuan, W.Y.; Wang, L.; Bai, Y.; Ma, H.

    2011-01-01

    Chemical durability of 90-19/Nd glass, a simulated high-level waste (HLW) glass in contact with the groundwater was investigated with a long-term product consistency test (PCT). Generally, it is difficult to observe the long term property of HLW glass due to the slow corrosion rate in a mild condition. In order to overcome this problem, increased contacting surface (S/V = 6000 m -1 ) and elevated temperature (150 o C) were employed to accelerate the glass corrosion evolution. The micro-morphological characteristics of the glass surface and the secondary minerals formed after the glass alteration were analyzed by SEM-EDS and XRD, and concentrations of elements in the leaching solution were determined by ICP-AES. In our experiments, two types of minerals, which have great impact on glass dissolution, were found to form on 90-19/Nd HLW glass surface when it was subjected to a long-term leaching in the groundwater. One is Mg-Fe-rich phyllosilicates with honeycomb structure; the other is aluminosilicates (zeolites). Mg and Fe in the leaching solution participated in the formation of phyllosilicates. The main components of phyllosilicates in alteration products of 90-19/Nd HLW glass are nontronite (Na 0.3 Fe 2 Si 4 O 10 (OH) 2 .4H 2 O) and montmorillonite (Ca 0.2 (Al,Mg) 2 Si 4 O 10 (OH) 2 .4H 2 O), and those of aluminosilicates are mordenite ((Na 2 ,K 2 ,Ca)Al 2 Si 10 O 24 .7H 2 O)) and clinoptilolite ((Na,K,Ca) 5 Al 6 Si 30 O 72 .18H 2 O). Minerals like Ca(Mg)SO 4 and CaCO 3 with low solubility limits are prone to form precipitant on the glass surface. Appearance of the phyllosilicates and aluminosilicates result in the dissolution rate of 90-19/Nd HLW glass resumed, which is increased by several times over the stable rate. As further dissolution of the glass, both B and Na in the glass were found to leach out in borax form.

  15. GLASS FABRICATION AND PRODUCT CONSISTENCY TESTING OF LANTHANIDE BOROSHILICATE FRIT X COMPOSITION FOR PLUTONIUM DISPOSITION

    Energy Technology Data Exchange (ETDEWEB)

    Marra, J

    2006-11-21

    The Department of Energy Office of Environmental Management (DOE/EM) plans to conduct the Plutonium Disposition Project at the Savannah River Site (SRS) to disposition excess weapons-usable plutonium. A plutonium glass waste form is the preferred option for immobilization of the plutonium for subsequent disposition in a geologic repository. A reference glass composition (Lanthanide Borosilicate (LaBS) Frit B) was developed during the Plutonium Immobilization Program (PIP) to immobilize plutonium in the late 1990's. A limited amount of performance testing was performed on this baseline composition before efforts to further pursue Pu disposition via a glass waste form ceased. Recent FY05 studies have further investigated the LaBS Frit B formulation as well as development of a newer LaBS formulation denoted as LaBS Frit X. The objectives of this present task were to fabricate plutonium loaded LaBS Frit X glass and perform corrosion testing to provide near-term data that will increase confidence that LaBS glass product is suitable for disposal in the Yucca Mountain Repository. Specifically, testing was conducted in an effort to provide data to Yucca Mountain Project (YMP) personnel for use in performance assessment calculations. Plutonium containing LaBS glass with the Frit X composition with a 9.5 wt% PuO{sub 2} loading was prepared for testing. Glass was prepared to support Product Consistency Testing (PCT) at Savannah River National Laboratory (SRNL). The glass was thoroughly characterized using x-ray diffraction (XRD) and scanning electron microscopy coupled with energy dispersive spectroscopy (SEM/EDS) prior to performance testing. A series of PCTs were conducted at SRNL using quenched Pu Frit X glass with varying exposed surface areas. Effects of isothermal and can-in-canister heat treatments on the Pu Frit X glass were also investigated. Another series of PCTs were performed on these different heat-treated Pu Frit X glasses. Leachates from all these PCTs

  16. Chemical composition analysis and product consistency tests to support enhanced Hanford waste glass models. Results for the third set of high alumina outer layer matrix glasses

    Energy Technology Data Exchange (ETDEWEB)

    Fox, K. M. [Savannah River Site (SRS), Aiken, SC (United States); Edwards, T. B. [Savannah River Site (SRS), Aiken, SC (United States)

    2015-12-01

    In this report, the Savannah River National Laboratory provides chemical analyses and Product Consistency Test (PCT) results for 14 simulated high level waste glasses fabricated by the Pacific Northwest National Laboratory. The results of these analyses will be used as part of efforts to revise or extend the validation regions of the current Hanford Waste Treatment and Immobilization Plant glass property models to cover a broader span of waste compositions. The measured chemical composition data are reported and compared with the targeted values for each component for each glass. All of the measured sums of oxides for the study glasses fell within the interval of 96.9 to 100.8 wt %, indicating recovery of all components. Comparisons of the targeted and measured chemical compositions showed that the measured values for the glasses met the targeted concentrations within 10% for those components present at more than 5 wt %. The PCT results were normalized to both the targeted and measured compositions of the study glasses. Several of the glasses exhibited increases in normalized concentrations (NCi) after the canister centerline cooled (CCC) heat treatment. Five of the glasses, after the CCC heat treatment, had NCB values that exceeded that of the Environmental Assessment (EA) benchmark glass. These results can be combined with additional characterization, including X-ray diffraction, to determine the cause of the higher release rates.

  17. Hot Ductility Behavior of an 8 Pct Cr Roller Steel

    Science.gov (United States)

    Wang, Zhenhua; Sun, Shuhua; Shi, Zhongping; Wang, Bo; Fu, Wantang

    2015-04-01

    The hot ductility of an 8 pct Cr roller steel was determined between 1173 K and 1473 K (900 °C and 1200 °C) at strain rates of 0.01 to 10 s-1 through tensile testing. The fracture morphology was observed using scanning electron microscopy, and the microstructure was examined through optical microscopy and transmission electron microscopy. The dependence of the hot ductility behavior on the deformation conditions, grain size, and precipitation was analyzed. The relationship between the reduction in area and the natural logarithm of the Zener-Hollomon parameter (ln Z) was found to be a second-order polynomial. When ln Z was greater than 40 s-1, the hot ductility was poor and fracture was mainly caused by incompatible deformation between the grains. When ln Z was between 32 and 40 s-1, the hot ductility was excellent and the main fracture mechanism was void linking. When ln Z was below 32 s-1, the hot ductility was poor and fracture was mainly caused by grain boundary sliding. A fine grain structure is beneficial for homogenous deformation and dynamic recrystallization, which induces better hot ductility. The effect of M7C3 carbide particles dispersed in the matrix on the hot ductility was small. The grain growth kinetics in the 8 pct Cr steel were obtained between 1373 K and 1473 K (1100 °C and 1200 °C). Finally, optimized preheating and forging procedures for 8 pct Cr steel rollers are provided.

  18. Selective Oxidation and Reactive Wetting During Hot-Dip Galvanizing of a 1.0 pct Al-0.5 pct Si TRIP-Assisted Steel

    Science.gov (United States)

    Bellhouse, E. M.; McDermid, J. R.

    2012-07-01

    Selective oxidation and reactive wetting during continuous galvanizing were studied for a low-alloy transformation induced plasticity (TRIP)-assisted steel with 0.2 pct C, 1.5 pct Mn, 1.0 pct Al and 0.5 pct Si. Three process atmospheres were tested during annealing prior to galvanizing: 220 K (-53 °C) dew point (dp) N2-20 pct H2, 243 K (-30 °C) dp N2-5 pct H2 and 278 K (+5 °C) dp N2-5 pct H2. The process atmosphere oxygen partial pressure affected the oxide chemistry, morphology and thickness. For the 220 K (-53 °C) dp and 243 K (-30 °C) dp process atmospheres, film and nodule-type manganese, silicon and aluminum containing oxides were observed at the surface. For the 278 K (+5 °C) dp atmosphere, MnO was observed at the grain boundaries and as thicker localized surface films. Oxide morphology, thickness and chemistry affected reactive wetting, with complete wetting being observed for the 220 K (-53 °C) dp and 243 K (-30 °C) dp process atmospheres and incomplete reactive wetting being observed for the 278 K (+5 °C) dp atmosphere. Complete reactive wetting for the 220 K (-53 °C) dp and 243 K (-30 °C) dp process atmospheres was attributed to a combination of zinc bridging of oxides, aluminothermic reduction of surface oxides and wetting of the oxides. Incomplete wetting for the 278 K (+5 °C) dp atmosphere was attributed to localized thick MnO films.

  19. Early diagnostic value of plasma PCT and BG assay for CRBSI after OLT.

    Science.gov (United States)

    Chen, J; Wang, Y; Shen, Z; Zhu, Z; Song, Y; Han, R

    2011-06-01

    The aim was to evaluate the role of procalcitonin (PCT) and (1-3)-β-D-glucan (BG) tests for early detection or exclusion of central venous catheter-related bloodstream infections (CRBSI) in patients after orthotopic liver transplantation (OLT). Fifty-five patients with clinically suspected CRBSI were assessed after OLT in this prospective study. On the day of clinical suspicion of CRBSI, blood samples were obtained from central venous catheters and a peripheral vein for blood cultures and from a peripheral vein for PCT and BG tests. Plasma PCT and BG values were measured by using an immunoluminometric assay and Fungitell BG assay, respectively. No prisoners or organs from prisoners were used in this study. Twenty-five patients (45%) were diagnosed with CRBIS. Among them, 13 (52%) displayed gram-positive bacteriemia, 11 (44%) gram-negative bacteriemia, and 1 (4%) fungemia. The PCT values were higher in CRBSI than in non-CRBSI patients (P = .003). CRBSI patients did not show significant increases in plasma BG values compared with non-CRBSI subjects (P = .051). PCT and BG area under receiver operating characteristic curves were 0.840 and 0.486, respectively. Sensitivity, specificity, and positive and negative predictive values of a PCT of ≥ 3.1 ng/mL for the diagnosis of CRBSI were 0.72, 0.87, 0.82, and 0.79, respectively. The figures for a BG of ≥ 83 pg/mL were 0.32, 0.90, 0.73, and 0.61, respectively. Among the 24 patients with bacteria infections, PCT was higher in patients with gram-negative than those with gram-positive bacterial infections (P = .022). We concluded that the PCT assay may be a useful rapid diagnostic adjunct for the diagnosis of suspected CRBSI in OLT patients. Copyright © 2011 Elsevier Inc. All rights reserved.

  20. Dark energy two decades after: observables, probes, consistency tests.

    Science.gov (United States)

    Huterer, Dragan; Shafer, Daniel L

    2018-01-01

    The discovery of the accelerating universe in the late 1990s was a watershed moment in modern cosmology, as it indicated the presence of a fundamentally new, dominant contribution to the energy budget of the universe. Evidence for dark energy, the new component that causes the acceleration, has since become extremely strong, owing to an impressive variety of increasingly precise measurements of the expansion history and the growth of structure in the universe. Still, one of the central challenges of modern cosmology is to shed light on the physical mechanism behind the accelerating universe. In this review, we briefly summarize the developments that led to the discovery of dark energy. Next, we discuss the parametric descriptions of dark energy and the cosmological tests that allow us to better understand its nature. We then review the cosmological probes of dark energy. For each probe, we briefly discuss the physics behind it and its prospects for measuring dark energy properties. We end with a summary of the current status of dark energy research.

  1. Porphyria Cutanea Tarda (PCT experience in Victoria, Australia: A case series and literature review

    Directory of Open Access Journals (Sweden)

    Quynh Le

    2018-01-01

    Full Text Available Background Porphyria Cutanea Tarda (PCT is a metabolic disorder resulting from a deficiency of hepatic enzyme uroporphyrinogen decarboxylase (UROD. UROD deficiency results in the accumulation of porphyrins, which are phototoxic and hepatotoxic. PCT patients are at increased risk of developing hepatocellular carcinoma. Aims We aim to describe a series of PCT patients presenting to a tertiary center over 35-year period from the 1980s to December 2015 and review current literature to date on PCT, with a focus on PCT management. Methods A search of the center’s dermatology department and biochemistry database were performed to identify patients diagnosed with PCT. Demographic data, underlying risk factors and management details were obtained. Statistical tests were performed to identify any possible association between the variables of interest. Results 34 patients were included in this study. Mean age of diagnosis was 48 years and there was no gender difference. 12 patients had Hepatitis C infection, 25 had excessive alcohol consumption, 13 had hereditary haemochromatosis. Eight patients developed oestrogen-associated or hormonal replacement therapy (HRT induced PCT. 33 patients (97 per cent responded to venesection. Six (18 per cent patients were prescribed hydroxychloroquine, either alone or concurrently with venesection. They all achieved remission. Average duration of follow up is 13 years. One patient developed hepatocellular carcinoma (HCC. Conclusion Our study has reinforced venesection as an effective treatment for PCT. Low dose hydroxychloroquine can be used in patients where venesection is contraindicated or not tolerated. General measures such as alcohol abstinence, visible violet light protection and trauma avoidance are recommended.

  2. Preliminary A ampersand PCT multiple detector design

    International Nuclear Information System (INIS)

    Roberson, G.P.; Martz, H.E.; Camp, D.C.; Decman, D.J.; Johansson, E.M.

    1997-01-01

    The next generation, multi-detector active and passive computed tomography (A ampersand PCT) scanner will be optimized for speed and accuracy. At the Lawrence Livermore National Lab (LLNL) we have demonstrated the trade-offs between different A ampersand PCT design parameters that affect the speed and quality of the assay results. These fundamental parameters govern the optimum system design. Although the multi-detector scanner design has priority put on speed to increase waste drum throughput, higher speed should not compromise assay accuracy. One way to increase the speed of the A ampersand PCT technology is to use multiple detectors. This yields a linear speedup by a factor approximately equal to the number of detectors used without a compromise in system accuracy. There are many different design scenarios that can be developed using multiple detectors. Here we describe four different scenarios and discuss the trade-offs between them. Also, some considerations are given in this design description for the implementation of a multiple detector technology in a field- deployable mobile trailer system

  3. Final Report - IHLW PCT, Spinel T1%, Electrical Conductivity, and Viscosity Model Development, VSL-07R1240-4

    Energy Technology Data Exchange (ETDEWEB)

    Kruger, Albert A.; Piepel, Gregory F.; Landmesser, S. M.; Pegg, I. L.; Heredia-Langner, Alejandro; Cooley, Scott K.; Gan, H.; Kot, W. K.

    2013-11-13

    This report is the last in a series of currently scheduled reports that presents the results from the High Level Waste (HLW) glass formulation development and testing work performed at the Vitreous State Laboratory (VSL) of the Catholic University of America (CUA) and the development of IHLW property-composition models performed jointly by Pacific Northwest National Laboratory (PNNL) and VSL for the River Protection Project-Waste Treatment and Immobilization Plant (RPP-WTP). Specifically, this report presents results of glass testing at VSL and model development at PNNL for Product Consistency Test (PCT), one-percent crystal fraction temperature (T1%), electrical conductivity (EC), and viscosity of HLW glasses. The models presented in this report may be augmented and additional validation work performed during any future immobilized HLW (IHLW) model development work. Completion of the test objectives is addressed.

  4. The use of a standardized PCT-algorithm reduces costs in intensive care in septic patients - a DRG-based simulation model

    Directory of Open Access Journals (Sweden)

    Wilke MH

    2011-12-01

    Full Text Available Abstract Introduction The management of bloodstream infections especially sepsis is a difficult task. An optimal antibiotic therapy (ABX is paramount for success. Procalcitonin (PCT is a well investigated biomarker that allows close monitoring of the infection and management of ABX. It has proven to be a cost-efficient diagnostic tool. In Diagnoses Related Groups (DRG based reimbursement systems, hospitals get only a fixed amount of money for certain treatments. Thus it's very important to obtain an optimal balance of clinical treatment and resource consumption namely the length of stay in hospital and especially in the Intensive Care Unit (ICU. We investigated which economic effects an optimized PCT-based algorithm for antibiotic management could have. Materials and methods We collected inpatient episode data from 16 hospitals. These data contain administrative and clinical information such as length of stay, days in the ICU or diagnoses and procedures. From various RCTs and reviews there are different algorithms for the use of PCT to manage ABX published. Moreover RCTs and meta-analyses have proven possible savings in days of ABX (ABD and length of stay in ICU (ICUD. As the meta-analyses use studies on different patient populations (pneumonia, sepsis, other bacterial infections, we undertook a short meta-analyses of 6 relevant studies investigating in sepsis or ventilator associated pneumonia (VAP. From this analyses we obtained savings in ABD and ICUD by calculating the weighted mean differences. Then we designed a new PCT-based algorithm using results from two very recent reviews. The algorithm contains evidence from several studies. From the patient data we calculated cost estimates using German National standard costing information for the German G-DRG system. We developed a simulation model where the possible savings and the extra costs for (in average 8 PCT tests due to our algorithm were brought into equation. Results We calculated ABD

  5. The Value of PCT, WBC and hs-CRP Tests in the Diagnosis and Prognosis of Children with Pneumonia%小儿肺炎中PCT、WBC、hs-CRP检测对疾病诊断及预后判断价值分析

    Institute of Scientific and Technical Information of China (English)

    夏兰兰; 李敬

    2017-01-01

    目的 分析小儿肺炎中PCT、WBC、hs-CRP检测对疾病诊断及预后的判断价值.方法 选取2015年5月至2016年4月在本院住院治疗的肺炎患儿120例作为观察组,健康儿童65例作为对照组.检测治疗前后患儿及健康儿童血清PCT、WBC、hs-CRP的变化.结果 治疗前,细菌感染组WBC数量(15.5±3.4)×109/L、hs-CRP水平(19.8±2.6) mg/L和PCT水平(21.9±2.1)ng/mL均明显高于非细菌感染组(6.8 ±2.3)×109/L、(4.5±2.3)mg/L、(4.7±1.8) ng/mL和健康对照组(8.1 ±1.4)×109/L、(3.5±1.2) mg/L、(1.8±0.5) ng/mL,且差异具有统计学意义(P<0.05).而非细菌感染组和健康对照组的WBC、hs-CRP和PCT结果差异无统计学意义(P>0.05);治疗后,细菌感染组WBC数量(7.3±1.4)×109/L、hs-CRP水平(5.1 ± 1.2) mg/L和PCT水平(2.3 ± 1.2) ng/mL均明显低于治疗前,差异具有统计学意义(P<0.05);WBC、hs-CRP、PCT联合检测的灵敏性95.2%、特异性93.0%、准确度93.3%显著高于WBC、hs-CRP、PCT单独检测,差异具有统计学意义(P<0.05).结论 在小儿肺炎治疗过程中,动态监测患儿血清的WBC、hs-CRP、PCT的变化,并结合临床,有助于诊疗和判断患儿疾病愈后的情况,具有显著的临床意义,值得推广应用.%Objective To estimate the value of PCT,WBC and hs-CRP tests in the diagnosis and prognosis of children with pneumonia.Methods 120 cases with pneumonia in children were selected from May 2015 to April 2016 in our hospital as the observation group,65 cases of healthy children as the control group.Changes of serum PCT,WBC and hs-CRP tests in children and healthy children before and after treatment were mearued.Results Before treatment,bacterial infection group have WBC count (15.5 ± 3.4) × 109/L,hs-CRP levels (19.8±2.6) mg/L and PCT level (21.9 ±2.1) ng/mL,which were significantly higher than those in non bacterial infection group (6.8 ±2.3) × 109/L,(4.5 ±2.3) mg/L,(4.7 ± 1.8) ng/mL and healthy control group (+ 8.1

  6. Value of serum PCT in early diagnosis of bacterial infection in patients with liver failure

    Directory of Open Access Journals (Sweden)

    WANG Chuanmin

    2017-06-01

    Full Text Available ObjectiveTo investigate the value of serum procalcitonin (PCT in early diagnosis of bacterial infection in patients with liver failure. MethodsA total of 463 patients with hepatitis B were selected from January to December, 2014, in the Department of Infectious Diseases, Taihe Hospital. According to the degree of liver injury, the patients were divided into four groups: mild liver injury group (n=120, moderate liver injury group (n=222, sever liver injury group (n=53, and liver failure group (n=68. Serum PCT was measured for all patients, and the white blood cell count (WBC and high-sensitivity C-reactive protein (hsCRP were measured for patients with liver failure. The clinical manifestations were observed and recorded. The t test was used for comparison of normally distributed continuous data, while the Kruskal-Wallis H test was used for non-normally distributed continuous data; the Mann-Whitney U test was used for pairwise comparison of continuous data. The chi-square test was used for comparison of categorical data. The receiver operating characteristic (ROC curve was used for the analysis of predictive value. ResultsThe liver failure group had a significantly higher PCT level than the severe liver injury group, moderate liver injury group, and mild liver injury group (0.81[0.34-2.15] vs 0.53[0.21-1.59], 0.35[010-1.18], and 0.17[0.10-0.60], χ2=25.091, P<0.05. The liver failure patients with PCT levels of <0.25 ng/ml (n=10, 0.25-0.5 ng/ml (n=10, and >0.5 ng/ml (n=48 had infection rates of 20%, 30%, and 66.7%, respectively, with a significant difference between the patients with a PCT level of >0.5 ng/ml and those with PCT levels of <0.25 ng/ml and 0.25-0.5 ng/ml (χ2=5631,4650,P=0018,0031. Among the liver failure patients, the infection cases had significantly higher PCT, WBC, and hsCRP than the non-infection cases (PCT: 3.72±1.33 ng/ml vs 0.34±0.12 ng/ml, t=-2.547, P=0.015; hsCRP: 16.70±7.03 mg

  7. A diagnostic test for apraxia in stroke patients: internal consistency and diagnostic value.

    NARCIS (Netherlands)

    Heugten, C.M. van; Dekker, J.; Deelman, B.G.; Stehmann-Saris, F.C.; Kinebanian, A.

    1999-01-01

    The internal consistency and the diagnostic value of a test for apraxia in patients having had a stroke are presented. Results indicate that the items of the test form a strong and consistent scale: Cronbach's alpha as well as the results of a Mokken scale analysis present good reliability and good

  8. Personal Hypothesis Testing: The Role of Consistency and Self-Schema.

    Science.gov (United States)

    Strohmer, Douglas C.; And Others

    1988-01-01

    Studied how individuals test hypotheses about themselves. Examined extent to which Snyder's bias toward confirmation persists when negative or nonconsistent personal hypothesis is tested. Found negativity or positivity did not affect hypothesis testing directly, though hypothesis consistency did. Found cognitive schematic variable (vulnerability…

  9. A Bayesian Decision-Theoretic Approach to Logically-Consistent Hypothesis Testing

    Directory of Open Access Journals (Sweden)

    Gustavo Miranda da Silva

    2015-09-01

    Full Text Available This work addresses an important issue regarding the performance of simultaneous test procedures: the construction of multiple tests that at the same time are optimal from a statistical perspective and that also yield logically-consistent results that are easy to communicate to practitioners of statistical methods. For instance, if hypothesis A implies hypothesis B, is it possible to create optimal testing procedures that reject A whenever they reject B? Unfortunately, several standard testing procedures fail in having such logical consistency. Although this has been deeply investigated under a frequentist perspective, the literature lacks analyses under a Bayesian paradigm. In this work, we contribute to the discussion by investigating three rational relationships under a Bayesian decision-theoretic standpoint: coherence, invertibility and union consonance. We characterize and illustrate through simple examples optimal Bayes tests that fulfill each of these requisites separately. We also explore how far one can go by putting these requirements together. We show that although fairly intuitive tests satisfy both coherence and invertibility, no Bayesian testing scheme meets the desiderata as a whole, strengthening the understanding that logical consistency cannot be combined with statistical optimality in general. Finally, we associate Bayesian hypothesis testing with Bayes point estimation procedures. We prove the performance of logically-consistent hypothesis testing by means of a Bayes point estimator to be optimal only under very restrictive conditions.

  10. How Well Does the Sum Score Summarize the Test? Summability as a Measure of Internal Consistency

    NARCIS (Netherlands)

    Goeman, J.J.; De, Jong N.H.

    2018-01-01

    Many researchers use Cronbach's alpha to demonstrate internal consistency, even though it has been shown numerous times that Cronbach's alpha is not suitable for this. Because the intention of questionnaire and test constructers is to summarize the test by its overall sum score, we advocate

  11. Standard Practice for Ensuring Test Consistency in Neutron-Induced Displacement Damage of Electronic Parts

    CERN Document Server

    American Society for Testing and Materials. Philadelphia

    2007-01-01

    1.1 This practice sets forth requirements to ensure consistency in neutron-induced displacement damage testing of silicon and gallium arsenide electronic piece parts. This requires controls on facility, dosimetry, tester, and communications processes that affect the accuracy and reproducibility of these tests. It provides background information on the technical basis for the requirements and additional recommendations on neutron testing. In addition to neutrons, reactors are used to provide gamma-ray pulses of intensities and durations that are not achievable elsewhere. This practice also provides background information and recommendations on gamma-ray testing of electronics using nuclear reactors. 1.2 Methods are presented for ensuring and validating consistency in neutron displacement damage testing of electronic parts such as integrated circuits, transistors, and diodes. The issues identified and the controls set forth in this practice address the characterization and suitability of the radiation environm...

  12. Requirements for UML and OWL Integration Tool for User Data Consistency Modeling and Testing

    DEFF Research Database (Denmark)

    Nytun, J. P.; Jensen, Christian Søndergaard; Oleshchuk, V. A.

    2003-01-01

    The amount of data available on the Internet is continuously increasing, consequentially there is a growing need for tools that help to analyse the data. Testing of consistency among data received from different sources is made difficult by the number of different languages and schemas being used....... In this paper we analyze requirements for a tool that support integration of UML models and ontologies written in languages like the W3C Web Ontology Language (OWL). The tool can be used in the following way: after loading two legacy models into the tool, the tool user connects them by inserting modeling......, an important part of this technique is attaching of OCL expressions to special boolean class attributes that we call consistency attributes. The resulting integration model can be used for automatic consistency testing of two instances of the legacy models by automatically instantiate the whole integration...

  13. Leaching characteristics of the metal waste form from the electrometallurgical treatment process: Product consistency testing

    International Nuclear Information System (INIS)

    Johnson, S. G.; Keiser, D. D.; Frank, S. M.; DiSanto, T.; Noy, M.

    1999-01-01

    Argonne National Laboratory is developing an electrometallurgical treatment for spent fuel from the experimental breeder reactor II. A product of this treatment process is a metal waste form that incorporates the stainless steel cladding hulls, zirconium from the fuel and the fission products that are noble to the process, i.e., Tc, Ru, Nb, Pd, Rh, Ag. The nominal composition of this waste form is stainless steel/15 wt% zirconium/1--4 wt% noble metal fission products/1--2 wt % U. Leaching results are presented from several tests and sample types: (1) 2 week monolithic immersion tests on actual metal waste forms produced from irradiated cladding hulls, (2) long term (>2 years) pulsed flow tests on samples containing technetium and uranium and (3) crushed sample immersion tests on cold simulated metal waste form samples. The test results will be compared and their relevance for waste form product consistency testing discussed

  14. Method used to test the imaging consistency of binocular camera's left-right optical system

    Science.gov (United States)

    Liu, Meiying; Wang, Hu; Liu, Jie; Xue, Yaoke; Yang, Shaodong; Zhao, Hui

    2016-09-01

    To binocular camera, the consistency of optical parameters of the left and the right optical system is an important factor that will influence the overall imaging consistency. In conventional testing procedure of optical system, there lacks specifications suitable for evaluating imaging consistency. In this paper, considering the special requirements of binocular optical imaging system, a method used to measure the imaging consistency of binocular camera is presented. Based on this method, a measurement system which is composed of an integrating sphere, a rotary table and a CMOS camera has been established. First, let the left and the right optical system capture images in normal exposure time under the same condition. Second, a contour image is obtained based on the multiple threshold segmentation result and the boundary is determined using the slope of contour lines near the pseudo-contour line. Third, the constraint of gray level based on the corresponding coordinates of left-right images is established and the imaging consistency could be evaluated through standard deviation σ of the imaging grayscale difference D (x, y) between the left and right optical system. The experiments demonstrate that the method is suitable for carrying out the imaging consistency testing for binocular camera. When the standard deviation 3σ distribution of imaging gray difference D (x, y) between the left and right optical system of the binocular camera does not exceed 5%, it is believed that the design requirements have been achieved. This method could be used effectively and paves the way for the imaging consistency testing of the binocular camera.

  15. Outcomes Definitions and Statistical Tests in Oncology Studies: A Systematic Review of the Reporting Consistency.

    Science.gov (United States)

    Rivoirard, Romain; Duplay, Vianney; Oriol, Mathieu; Tinquaut, Fabien; Chauvin, Franck; Magne, Nicolas; Bourmaud, Aurelie

    2016-01-01

    Quality of reporting for Randomized Clinical Trials (RCTs) in oncology was analyzed in several systematic reviews, but, in this setting, there is paucity of data for the outcomes definitions and consistency of reporting for statistical tests in RCTs and Observational Studies (OBS). The objective of this review was to describe those two reporting aspects, for OBS and RCTs in oncology. From a list of 19 medical journals, three were retained for analysis, after a random selection: British Medical Journal (BMJ), Annals of Oncology (AoO) and British Journal of Cancer (BJC). All original articles published between March 2009 and March 2014 were screened. Only studies whose main outcome was accompanied by a corresponding statistical test were included in the analysis. Studies based on censored data were excluded. Primary outcome was to assess quality of reporting for description of primary outcome measure in RCTs and of variables of interest in OBS. A logistic regression was performed to identify covariates of studies potentially associated with concordance of tests between Methods and Results parts. 826 studies were included in the review, and 698 were OBS. Variables were described in Methods section for all OBS studies and primary endpoint was clearly detailed in Methods section for 109 RCTs (85.2%). 295 OBS (42.2%) and 43 RCTs (33.6%) had perfect agreement for reported statistical test between Methods and Results parts. In multivariable analysis, variable "number of included patients in study" was associated with test consistency: aOR (adjusted Odds Ratio) for third group compared to first group was equal to: aOR Grp3 = 0.52 [0.31-0.89] (P value = 0.009). Variables in OBS and primary endpoint in RCTs are reported and described with a high frequency. However, statistical tests consistency between methods and Results sections of OBS is not always noted. Therefore, we encourage authors and peer reviewers to verify consistency of statistical tests in oncology studies.

  16. On the Consistency of Bootstrap Testing for a Parameter on the Boundary of the Parameter Space

    DEFF Research Database (Denmark)

    Cavaliere, Giuseppe; Nielsen, Heino Bohn; Rahbek, Anders

    2017-01-01

    It is well known that with a parameter on the boundary of the parameter space, such as in the classic cases of testing for a zero location parameter or no autoregressive conditional heteroskedasticity (ARCH) effects, the classic nonparametric bootstrap – based on unrestricted parameter estimates...... – leads to inconsistent testing. In contrast, we show here that for the two aforementioned cases, a nonparametric bootstrap test based on parameter estimates obtained under the null – referred to as ‘restricted bootstrap’ – is indeed consistent. While the restricted bootstrap is simple to implement...... in practice, novel theoretical arguments are required in order to establish consistency. In particular, since the bootstrap is analysed both under the null hypothesis and under the alternative, non-standard asymptotic expansions are required to deal with parameters on the boundary. Detailed proofs...

  17. Remaining useful life prediction based on variation coefficient consistency test of a Wiener process

    Directory of Open Access Journals (Sweden)

    Juan LI

    2018-01-01

    Full Text Available High-cost equipment is often reused after maintenance, and whether the information before the maintenance can be used for the Remaining Useful Life (RUL prediction after the maintenance is directly determined by the consistency of the degradation pattern before and after the maintenance. Aiming at this problem, an RUL prediction method based on the consistency test of a Wiener process is proposed. Firstly, the parameters of the Wiener process estimated by Maximum Likelihood Estimation (MLE are proved to be biased, and a modified unbiased estimation method is proposed and verified by derivation and simulations. Then, the h statistic is constructed according to the reciprocal of the variation coefficient of the Wiener process, and the sampling distribution is derived. Meanwhile, a universal method for the consistency test is proposed based on the sampling distribution theorem, which is verified by simulation data and classical crack degradation data. Finally, based on the consistency test of the degradation model, a weighted fusion RUL prediction method is presented for the fuel pump of an airplane, and the validity of the presented method is verified by accurate computation results of real data, which provides a theoretical and practical guidance for engineers to predict the RUL of equipment after maintenance.

  18. Vibrational multiconfiguration self-consistent field theory: implementation and test calculations.

    Science.gov (United States)

    Heislbetz, Sandra; Rauhut, Guntram

    2010-03-28

    A state-specific vibrational multiconfiguration self-consistent field (VMCSCF) approach based on a multimode expansion of the potential energy surface is presented for the accurate calculation of anharmonic vibrational spectra. As a special case of this general approach vibrational complete active space self-consistent field calculations will be discussed. The latter method shows better convergence than the general VMCSCF approach and must be considered the preferred choice within the multiconfigurational framework. Benchmark calculations are provided for a small set of test molecules.

  19. The internal consistency of the standard gamble: tests after adjusting for prospect theory.

    Science.gov (United States)

    Oliver, Adam

    2003-07-01

    This article reports a study that tests whether the internal consistency of the standard gamble can be improved upon by incorporating loss weighting and probability transformation parameters in the standard gamble valuation procedure. Five alternatives to the standard EU formulation are considered: (1) probability transformation within an EU framework; and, within a prospect theory framework, (2) loss weighting and full probability transformation, (3) no loss weighting and full probability transformation, (4) loss weighting and no probability transformation, and (5) loss weighting and partial probability transformation. Of the five alternatives, only the prospect theory formulation with loss weighting and no probability transformation offers an improvement in internal consistency over the standard EU valuation procedure.

  20. A non-parametric consistency test of the ΛCDM model with Planck CMB data

    Energy Technology Data Exchange (ETDEWEB)

    Aghamousa, Amir; Shafieloo, Arman [Korea Astronomy and Space Science Institute, Daejeon 305-348 (Korea, Republic of); Hamann, Jan, E-mail: amir@aghamousa.com, E-mail: jan.hamann@unsw.edu.au, E-mail: shafieloo@kasi.re.kr [School of Physics, The University of New South Wales, Sydney NSW 2052 (Australia)

    2017-09-01

    Non-parametric reconstruction methods, such as Gaussian process (GP) regression, provide a model-independent way of estimating an underlying function and its uncertainty from noisy data. We demonstrate how GP-reconstruction can be used as a consistency test between a given data set and a specific model by looking for structures in the residuals of the data with respect to the model's best-fit. Applying this formalism to the Planck temperature and polarisation power spectrum measurements, we test their global consistency with the predictions of the base ΛCDM model. Our results do not show any serious inconsistencies, lending further support to the interpretation of the base ΛCDM model as cosmology's gold standard.

  1. Test of Gross Motor Development : Expert Validity, confirmatory validity and internal consistence

    Directory of Open Access Journals (Sweden)

    Nadia Cristina Valentini

    2008-12-01

    Full Text Available The Test of Gross Motor Development (TGMD-2 is an instrument used to evaluate children’s level of motordevelopment. The objective of this study was to translate and verify the clarity and pertinence of the TGMD-2 items by expertsand the confirmatory factorial validity and the internal consistence by means of test-retest of the Portuguese TGMD-2. Across-cultural translation was used to construct the Portuguese version. The participants of this study were 7 professionalsand 587 children, from 27 schools (kindergarten and elementary from 3 to 10 years old (51.1% boys and 48.9% girls.Each child was videotaped performing the test twice. The videotaped tests were then scored. The results indicated thatthe Portuguese version of the TGMD-2 contains clear and pertinent motor items; demonstrated satisfactory indices ofconfirmatory factorial validity (χ2/gl = 3.38; Goodness-of-fit Index = 0.95; Adjusted Goodness-of-fit index = 0.92 and Tuckerand Lewis’s Index of Fit = 0.83 and test-retest internal consistency (locomotion r = 0.82; control of object: r = 0.88. ThePortuguese TGMD-2 demonstrated validity and reliability for the sample investigated.

  2. Test of Gross Motor Development: expert validity, confirmatory validity and internal consistence

    Directory of Open Access Journals (Sweden)

    Nadia Cristina Valentini

    2008-01-01

    The Test of Gross Motor Development (TGMD-2 is an instrument used to evaluate children’s level of motor development. The objective of this study was to translate and verify the clarity and pertinence of the TGMD-2 items by experts and the confirmatory factorial validity and the internal consistence by means of test-retest of the Portuguese TGMD-2. A cross-cultural translation was used to construct the Portuguese version. The participants of this study were 7 professionals and 587 children, from 27 schools (kindergarten and elementary from 3 to 10 years old (51.1% boys and 48.9% girls. Each child was videotaped performing the test twice. The videotaped tests were then scored. The results indicated that the Portuguese version of the TGMD-2 contains clear and pertinent motor items; demonstrated satisfactory indices of confirmatory factorial validity (÷2/gl = 3.38; Goodness-of-fit Index = 0.95; Adjusted Goodness-of-fit index = 0.92 and Tucker and Lewis’s Index of Fit = 0.83 and test-retest internal consistency (locomotion r = 0.82; control of object: r = 0.88. The Portuguese TGMD-2 demonstrated validity and reliability for the sample investigated.

  3. Force Concept Inventory-based multiple-choice test for investigating students’ representational consistency

    Directory of Open Access Journals (Sweden)

    Pasi Nieminen

    2010-08-01

    Full Text Available This study investigates students’ ability to interpret multiple representations consistently (i.e., representational consistency in the context of the force concept. For this purpose we developed the Representational Variant of the Force Concept Inventory (R-FCI, which makes use of nine items from the 1995 version of the Force Concept Inventory (FCI. These original FCI items were redesigned using various representations (such as motion map, vectorial and graphical, yielding 27 multiple-choice items concerning four central concepts underpinning the force concept: Newton’s first, second, and third laws, and gravitation. We provide some evidence for the validity and reliability of the R-FCI; this analysis is limited to the student population of one Finnish high school. The students took the R-FCI at the beginning and at the end of their first high school physics course. We found that students’ (n=168 representational consistency (whether scientifically correct or not varied considerably depending on the concept. On average, representational consistency and scientifically correct understanding increased during the instruction, although in the post-test only a few students performed consistently both in terms of representations and scientifically correct understanding. We also compared students’ (n=87 results of the R-FCI and the FCI, and found that they correlated quite well.

  4. Evaluation of the entropy consistent euler flux on 1D and 2D test problems

    Science.gov (United States)

    Roslan, Nur Khairunnisa Hanisah; Ismail, Farzad

    2012-06-01

    Perhaps most CFD simulations may yield good predictions of pressure and velocity when compared to experimental data. Unfortunately, these results will most likely not adhere to the second law of thermodynamics hence comprising the authenticity of predicted data. Currently, the test of a good CFD code is to check how much entropy is generated in a smooth flow and hope that the numerical entropy produced is of the correct sign when a shock is encountered. Herein, a shock capturing code written in C++ based on a recent entropy consistent Euler flux is developed to simulate 1D and 2D flows. Unlike other finite volume schemes in commercial CFD code, this entropy consistent flux (EC) function precisely satisfies the discrete second law of thermodynamics. This EC flux has an entropy-conserved part, preserving entropy for smooth flows and a numerical diffusion part that will accurately produce the proper amount of entropy, consistent with the second law. Several numerical simulations of the entropy consistent flux have been tested on two dimensional test cases. The first case is a Mach 3 flow over a forward facing step. The second case is a flow over a NACA 0012 airfoil while the third case is a hypersonic flow passing over a 2D cylinder. Local flow quantities such as velocity and pressure are analyzed and then compared with mainly the Roe flux. The results herein show that the EC flux does not capture the unphysical rarefaction shock unlike the Roe-flux and does not easily succumb to the carbuncle phenomenon. In addition, the EC flux maintains good performance in cases where the Roe flux is known to be superior.

  5. Thermodynamic consistency test procedure using orthogonal collocation and the Peng-Robinson equation of state

    International Nuclear Information System (INIS)

    Hamm, L.L.; Van Brunt, V.

    1982-08-01

    The Christiansen and Fredenslund programs for calculating vapor-liquid equilibria have been modified by replacing the Soave-Redlich-Kwong equation of state with the newly developed Peng-Robinson equation of state. This modification was shown to be a decided improvement for high pressure systems, especially in the critical and upper retrograde regions. Thermodynamic consistency tests were developed and used to evaluate and compare calculated values from both the modified and unmodified programs with reported experimental data for several vapor-liquid systems

  6. Fatigue crack growth rates and fracture toughness of rapidly solidified Al-8.5 pct Fe-1.2 pct V-1.7 pct Si alloys

    International Nuclear Information System (INIS)

    Hariprasad, S.; Sastry, S.M.L.; Jerina, K.L.

    1994-01-01

    The room-temperature fatigue crack growth rates (FCGR) and fracture toughness were evaluated for different crack plane Orientations of an Al-8.5 pct Fe-1.2 pct V-1.7 pct Si alloy produced by planar flow casting (PFC) and atomized melt deposition (AMD) processes. For the alloy produced by the PFC process, properties were determined in six different orientations, including the short transverse directions S-T and S-L. Diffusion bonding and adhesive bonding methods were used to prepare specimens for determining FCGR and fracture toughness in the short transverse direction. Interparticle boundaries control fracture properties in the alloy produced by PFC. Fracture toughness of the PFC alloy varies from 13.4 MPa√ bar m to 30.8 MPa√ bar m, depending on the orientation of the crack plane relative to the interparticle boundaries. Fatigue crack growth resistance and fracture toughness are greater in the L-T, L-S, and T-S directions than in the T-L, S-T, and S-L orientations. The alloy produced by AMD does not exhibit anisotropy in fracture toughness and fatigue crack growth resistance in the as-deposited condition or in the extruded condition. The fracture toughness varies from 17.2 MPa√ bar m to 18.5 MPa√ bar m for the as-deposited condition and from 19.8 MPa√ bar m to 21.0 MPa√ bar m for the extruded condition. Fracture properties are controlled by intrinsic factors in the alloy produced by AMD. Fatigue crack growth rates of the AMD alloy are comparable to those of the PFC alloy in the L-T orientation. The crack propagation modes were studied by optical metallographic examination of crack-microstructure interactions and scanning electron microscopy of the fracture surfaces

  7. General approach to the testing of binary solubility systems for thermodynamic consistency. Consolidated Fuel Reprocessing Program

    International Nuclear Information System (INIS)

    Hamm, L.L.; Van Brunt, V.

    1982-08-01

    A comparison of implicit Runge-Kutta and orthogonal collocation methods is made for the numerical solution to the ordinary differential equation which describes the high-pressure vapor-liquid equilibria of a binary system. The systems of interest are limited to binary solubility systems where one of the components is supercritical and exists as a noncondensable gas in the pure state. Of the two methods - implicit Runge-Kuta and orthogonal collocation - this paper attempts to present some preliminary but not necessarily conclusive results that the implicit Runge-Kutta method is superior for the solution to the ordinary differential equation utilized in the thermodynamic consistency testing of binary solubility systems. Due to the extreme nonlinearity of thermodynamic properties in the region near the critical locus, an extended cubic spline fitting technique is devised for correlating the P-x data. The least-squares criterion is employed in smoothing the experimental data. Even though the derivation is presented specifically for the correlation of P-x data, the technique could easily be applied to any thermodynamic data by changing the endpoint requirements. The volumetric behavior of the systems must be given or predicted in order to perform thermodynamic consistency tests. A general procedure is developed for predicting the volumetric behavior required and some indication as to the expected limit of accuracy is given

  8. The Work Softening by Deformation-Induced Disordering and Cold Rolling of 6.5 wt pct Si Steel Thin Sheets

    Science.gov (United States)

    Wang, Xianglong; Li, Haoze; Zhang, Weina; Liu, Zhenyu; Wang, Guodong; Luo, Zhonghan; Zhang, Fengquan

    2016-09-01

    As-cast strip of 6.5 wt pct Si steel was fabricated by twin-roll strip casting. After hot rolling at 1323 K (1050 °C), thin sheets with the thickness of 0.35 mm were produced by warm rolling at 373 K (100 °C) with rolling reductions of 15, 25, 35, 45, 55, and 65 pct. Influence of warm rolling reduction on ductility was investigated by room temperature bending test. The measurement of macro-hardness showed that "work softening" could begin when the warm rolling reduction exceeded 35 pct. The room temperature ductility of the thin sheets gradually increased with the increase of warm rolling reductions, and the plastic deformation during bending began to form when the warm rolling reduction was greater than 45 pct, the 65 pct rolled thin sheet exhibited the maximum plastic deformation of about 0.6 pct during bending at room temperature, with a few small dimples having been observed on the fracture surfaces. B2-ordered domains were formed in the 15, 25, 35, 45, and 55 pct rolled specimens, and their average size decreased with the increase of warm rolling reductions. By contrast, no B2-ordered domain could be found in the 65 pct rolled specimen. It had been observed that large-ordered domains could be split into several small parts by the slip of partial super-dislocations during warm rolling, which led to significant decrease of the order degree to cause the phenomenon of deformation-induced disordering. On the basis of these results, cold rolling schedule was developed to successfully fabricate 0.25-mm-thick sheets with good surface qualities and magnetic properties from warm rolled sheets.

  9. Nonparametric test of consistency between cosmological models and multiband CMB measurements

    Energy Technology Data Exchange (ETDEWEB)

    Aghamousa, Amir [Asia Pacific Center for Theoretical Physics, Pohang, Gyeongbuk 790-784 (Korea, Republic of); Shafieloo, Arman, E-mail: amir@apctp.org, E-mail: shafieloo@kasi.re.kr [Korea Astronomy and Space Science Institute, Daejeon 305-348 (Korea, Republic of)

    2015-06-01

    We present a novel approach to test the consistency of the cosmological models with multiband CMB data using a nonparametric approach. In our analysis we calibrate the REACT (Risk Estimation and Adaptation after Coordinate Transformation) confidence levels associated with distances in function space (confidence distances) based on the Monte Carlo simulations in order to test the consistency of an assumed cosmological model with observation. To show the applicability of our algorithm, we confront Planck 2013 temperature data with concordance model of cosmology considering two different Planck spectra combination. In order to have an accurate quantitative statistical measure to compare between the data and the theoretical expectations, we calibrate REACT confidence distances and perform a bias control using many realizations of the data. Our results in this work using Planck 2013 temperature data put the best fit ΛCDM model at 95% (∼ 2σ) confidence distance from the center of the nonparametric confidence set while repeating the analysis excluding the Planck 217 × 217 GHz spectrum data, the best fit ΛCDM model shifts to 70% (∼ 1σ) confidence distance. The most prominent features in the data deviating from the best fit ΛCDM model seems to be at low multipoles  18 < ℓ < 26 at greater than 2σ, ℓ ∼ 750 at ∼1 to 2σ and ℓ ∼ 1800 at greater than 2σ level. Excluding the 217×217 GHz spectrum the feature at ℓ ∼ 1800 becomes substantially less significance at ∼1 to 2σ confidence level. Results of our analysis based on the new approach we propose in this work are in agreement with other analysis done using alternative methods.

  10. Evaluating the Consistency of the FNA Test in Pathologically Proven Nodules of Thyroidectomy

    Directory of Open Access Journals (Sweden)

    Alireza Khazaei

    2018-02-01

    Full Text Available Fine Needle Aspiration (FNA is a selective diagnostic technique for the evaluation of non-toxic thyroid nodules. Thyroid FNA results are either undiagnosed or suspicious and indeterminate in 20-30% of cases. Therefore, this study seeks to determine the consistency of the FNA test in pathologically proven nodules of thyroidectomy. This is a descriptive cross-sectional study carried out on a total of 73 candidates for thyroidectomy who had been admitted to Imam Ali Hospital. A census sampling method has been used in this study. The FNA samples and pathology samples were evaluated and the consistency of the FNA test in pathologically proven nodules were compared. The SPSS software was used for data analysis. The mean age of the patients was 40.1 ± 12.9 years. 23.3% of the participants were male and 76.7% of them were female. The malignancy rate in the pathology was 65.8% (48 cases and 53.4% (39 cases in the FNA. Of the 48 positive cases, the FNA pathology diagnosed 35 cases (72.9% as positive and 13 cases (27.1% as negative. Of the 25 negative cases, the FNA pathology diagnosed 21 cases (84% as negative and 4 cases (16% as positive. Sensitivity, specificity, positive and negative predictive values of FNA in malignancy diagnosis were 72.92, 84, 89.74, and 61.76%, respectively. The results show that FNA does not have a high sensitivity in the diagnosis of malignancy, but has good specificity and the use of other diagnostic methods before the operation of thyroid nodules seems necessary.

  11. GEANT4 Tuning For pCT Development

    International Nuclear Information System (INIS)

    Yevseyeva, Olga; Assis, Joaquim T. de; Evseev, Ivan; Schelin, Hugo R.; Paschuk, Sergei A.; Milhoretto, Edney; Setti, Joao A. P.; Diaz, Katherin S.; Hormaza, Joel M.; Lopes, Ricardo T.

    2011-01-01

    Proton beams in medical applications deal with relatively thick targets like the human head or trunk. Thus, the fidelity of proton computed tomography (pCT) simulations as a tool for proton therapy planning depends in the general case on the accuracy of results obtained for the proton interaction with thick absorbers. GEANT4 simulations of proton energy spectra after passing thick absorbers do not agree well with existing experimental data, as showed previously. Moreover, the spectra simulated for the Bethe-Bloch domain showed an unexpected sensitivity to the choice of low-energy electromagnetic models during the code execution. These observations were done with the GEANT4 version 8.2 during our simulations for pCT. This work describes in more details the simulations of the proton passage through aluminum absorbers with varied thickness. The simulations were done by modifying only the geometry in the Hadrontherapy Example, and for all available choices of the Electromagnetic Physics Models. As the most probable reasons for these effects is some specific feature in the code, or some specific implicit parameters in the GEANT4 manual, we continued our study with version 9.2 of the code. Some improvements in comparison with our previous results were obtained. The simulations were performed considering further applications for pCT development.

  12. Unusual morphology of the omega phase in a Zr-1.75 At. pct Ni alloy

    International Nuclear Information System (INIS)

    Srivastava, D.; Mukhopadhyay, P.; Ramadasan, E.; Banerjee, S.

    1993-01-01

    The observations reported in the present communication were made in the course of a microstructural investigation on dilute Zr-Ni alloys. The alloys were prepared from nuclear-grade sponge zirconium and high-purity nickel by nonconsumable arc melting. Repeated melting was carried out to enhance homogeneity. A master alloy was used in the preparation of very dilute alloys. The observations reported here clearly indicate that the formation of the ω phase on β quenching, can occur in the binary Zr-Ni system at very low solute concentrations, suggesting that nickel is a strong ω stabilizer. Such a situation is known to obtain in the binary Zr-Mo system also. However, the unusual ω morphology observed in the present work has not been reported in the context of dilute Zr-Mo alloys. The appearance of sharp ω reflections and the absence of streaking and diffuse distributions in the selected area electron diffraction (SAD) patterns suggest that the Zr-1.75 at. pct Ni composition lies toward the solute lean end of the composition range associated with the formation of the athermal ω phase in the binary Zr-Ni system. In a more dilute Zr-Ni alloy (Zr-1.30 at. pct Ni), no athermal ω could be observed in β quenched specimens. The absence of streaking is also consistent with the fact that the ω phase in the β quenched Zr-1.75 at. pct Ni alloy did not occur in the form of very fine precipitates. The unusual ω morphology encountered in this work merits further investigation

  13. REFINEMENT OF THE REVERSE EXTRUSION TEST TO DETERMINE THE TWO CONSISTENCY LIMITS

    Directory of Open Access Journals (Sweden)

    Kamil KAYABALI

    2015-11-01

    coefficients (i.e., a and b were subjected to a multiple regression analysis along with the results of the conventional testing methods of fall-cone and rolling-device to determine the LL and PL as functions of the reverse extrusion parameters. It was shown that LL and PL can be predicted with a great degree of success using the reverse extrusion coefficients. While a great majority of the liquid limits found by using the fall-cone method were predicted with a ±10% error, almost all of the plastic limits found by the rolling device were predicted with a ±10% error. This refined investigation on the reverse extrusion test confirmed and proved that the reverse extrusion test is a simple, robust and inexpensive method capable of predicting both of two fundamental consistency limits using a single device.

  14. ILAW Glass Testing for Disposal at IDF: Phase 1 Testing

    Energy Technology Data Exchange (ETDEWEB)

    Papathanassiu, Adonia [The Catholic Univ. of America, Washington, DC (United States). Virteous State Lab.; Muller, Isabelle S. [The Catholic Univ. of America, Washington, DC (United States). Virteous State Lab.; Brandys, Marek [The Catholic Univ. of America, Washington, DC (United States). Virteous State Lab.; Gilbo, Konstantin [The Catholic Univ. of America, Washington, DC (United States). Virteous State Lab.; Barkatt, Aaron [The Catholic Univ. of America, Washington, DC (United States). Virteous State Lab.; Joseph, Innocent [EnergySolutions Federal EPC, Inc., Columbia, MD (United States); The Catholic Univ. of America, Washington, DC (United States). Virteous State Lab.; Pegg, Ian L. [The Catholic Univ. of America, Washington, DC (United States). Virteous State Lab.; Brown, Elvie E. [Washington River Protection Solutions, LLC, Richland, WA (United States); Swanberg, David J. [Washington River Protection Solutions, LLC, Richland, WA (United States)

    2011-04-11

    This document reports the results of the testing of phase 1 ORP LAW (low activity waste) glasses, also identified as enhanced LAW glasses. Testing involved are SPFT (Single Pass Flow Through), VHT (Vapor Hydration Test), and PCT (Product Consistency Test), along with the analytical tests (XRD and SEM-EDS). This report contains the data of the high waste loading ORP LAW glasses that will be used for the performance assessment of the IDF (Integrated Disposal Facility).

  15. ILAW Glass Testing for Disposal at IDF: Phase 1 Testing

    International Nuclear Information System (INIS)

    Papathanassiu, Adonia; Swanberg, David J.

    2011-01-01

    This document reports the results of the testing of phase 1 ORP LAW (low activity waste) glasses, also identified as enhanced LAW glasses. Testing involved are SPFT (Single Pass Flow Through), VHT (Vapor Hydration Test), and PCT (Product Consistency Test), along with the analytical tests (XRD and SEM-EDS). This report contains the data of the high waste loading ORP LAW glasses that will be used for the performance assessment of the IDF (Integrated Disposal Facility).

  16. Perfusion of prostate cancer: correlation between p-CT and whole-mount pathology - case report

    International Nuclear Information System (INIS)

    Luczynska, E.; Aniol, J.; Szczudlo, J.; Stelmach, A.; Jaszczynski, J.; Hartel, M.; Konopka, M.

    2006-01-01

    Prostate cancer is the second most common form of cancer among malignant neoplasm for men in Poland, next to lung cancer, as far as frequency is concerned. Incidence of this kind of neoplasm increases by about 3 % annually. In the last decade a growing number of this type of diseases has been observed and its detections are closely related to the development of biochemical (PSA - prostate-specific antigen) and diagnostic imaging technologies. A 60-year-old patient was diagnosed in the Oncology Institute because of an increasing level of PSA in his blood. The PSA level in March 2005 was 10,4 ng/ml. There was a slight increase of PSA during the year, up to 1,5 ng/ml (this is less than 25% / year). The patient came for the following check up to repeat the core-needle sextant biopsy, to exclude neoplasmatic process. Before operation the patient's prostate was tested by p-CT. The parametric maps revealed some disturbances of blood flow parameters. Blood flow - BF, blood volume - BV, mean transit time - MTT and permeability surface - PS were noted in the form of their asymmetry within peripheral zone in the right lobe. A pathological focus with increased BF, BV, PS and decreased MTT was revealed on the right side. This examination suggested that neoplasmatic process might be localized in this area. Core needle biopsy taken from the suspicious region revealed prostate cancer. That was also confirmed in histopathology examination after radical prostatectomy. P-CT examination can be performed during classical CT exam and it leads to obtaining morphological and functional data at the same time. P-CT examination allowed us to localize pathological process and helped to continue its verification by other diagnostic methods. (author)

  17. Culture, Cross-Role Consistency, and Adjustment: Testing Trait and Cultural Psychology Perspectives

    OpenAIRE

    Church, A. Timothy; Anderson-Harumi, Cheryl A.; del Prado, Alicia M.; Curtis, Guy J.; Tanaka-Matsumi, Junko; Valdez Medina, José L.; Mastor, Khairul A.; White, Fiona A.; Miramontes, Lilia A.; Katigbak, Marcia S.

    2008-01-01

    Trait and cultural psychology perspectives on cross-role consistency and its relation to adjustment were examined in two individualistic cultures, the United States (N = 231) and Australia (N = 195), and four collectivistic cultures, Mexico (N = 199), Philippines (N = 195), Malaysia (N = 217), and Japan (N = 180). Cross-role consistency in trait ratings was evident in all cultures, supporting trait perspectives. Cultural comparisons of mean consistency provided support for cultural psychology...

  18. PCT theorem for fields with arbitrary high-energy behavior

    International Nuclear Information System (INIS)

    Luecke, W.

    1986-01-01

    A neutral scalar field A(x) is considered that has to be smeared by Fourier transforms of C/sup infinity/ functions with compact support but otherwise fulfills all the Wightman axioms, except strict local commutativity. It is shown to fulfill the PCT symmetry condition (where Ω denotes the vacuum state vector) 1 ) xxx A(x/sub n/)Ω> = 1 )Ω> if and only if 1 ) xxx A(x/sub n/)Ω> - 1 )Ω> can be represented, in a sense, as an infinite sum of derivatives of measures with supports containing no Jost points

  19. Force Concept Inventory-Based Multiple-Choice Test for Investigating Students' Representational Consistency

    Science.gov (United States)

    Nieminen, Pasi; Savinainen, Antti; Viiri, Jouni

    2010-01-01

    This study investigates students' ability to interpret multiple representations consistently (i.e., representational consistency) in the context of the force concept. For this purpose we developed the Representational Variant of the Force Concept Inventory (R-FCI), which makes use of nine items from the 1995 version of the Force Concept Inventory…

  20. On Consistency Test Method of Expert Opinion in Ecological Security Assessment.

    Science.gov (United States)

    Gong, Zaiwu; Wang, Lihong

    2017-09-04

    To reflect the initiative design and initiative of human security management and safety warning, ecological safety assessment is of great value. In the comprehensive evaluation of regional ecological security with the participation of experts, the expert's individual judgment level, ability and the consistency of the expert's overall opinion will have a very important influence on the evaluation result. This paper studies the consistency measure and consensus measure based on the multiplicative and additive consistency property of fuzzy preference relation (FPR). We firstly propose the optimization methods to obtain the optimal multiplicative consistent and additively consistent FPRs of individual and group judgments, respectively. Then, we put forward a consistency measure by computing the distance between the original individual judgment and the optimal individual estimation, along with a consensus measure by computing the distance between the original collective judgment and the optimal collective estimation. In the end, we make a case study on ecological security for five cities. Result shows that the optimal FPRs are helpful in measuring the consistency degree of individual judgment and the consensus degree of collective judgment.

  1. Do Different Tests of Episodic Memory Produce Consistent Results in Human Adults?

    Science.gov (United States)

    Cheke, Lucy G.; Clayton, Nicola S.

    2013-01-01

    A number of different philosophical, theoretical, and empirical perspectives on episodic memory have led to the development of very different tests with which to assess it. Although these tests putatively assess the same psychological capacity, they have rarely been directly compared. Here, a sample of undergraduates was tested on three different…

  2. Phase Transformation Behavior of Medium Manganese Steels with 3 Wt Pct Aluminum and 3 Wt Pct Silicon During Intercritical Annealing

    Science.gov (United States)

    Sun, Binhan; Fazeli, Fateh; Scott, Colin; Yue, Stephen

    2016-10-01

    Medium manganese steels alloyed with sufficient aluminum and silicon amounts contain high fractions of retained austenite adjustable to various transformation-induced plasticity/twinning-induced plasticity effects, in addition to a reduced density suitable for lightweight vehicle body-in-white assemblies. Two hot rolled medium manganese steels containing 3 wt pct aluminum and 3 wt pct silicon were subjected to different annealing treatments in the present study. The evolution of the microstructure in terms of austenite transformation upon reheating and the subsequent austenite decomposition during quenching was investigated. Manganese content of the steels prevailed the microstructural response. The microstructure of the leaner alloy with 7 wt pct Mn (7Mn) was substantially influenced by the annealing temperature, including the variation of phase constituents, the morphology and composition of intercritical austenite, the Ms temperature and the retained austenite fraction. In contrast, the richer variant 10 wt pct Mn steel (10Mn) exhibited a substantially stable ferrite-austenite duplex phase microstructure containing a fixed amount of retained austenite which was found to be independent of the variations of intercritical annealing temperature. Austenite formation from hot band ferrite-pearlite/bainite mixtures was very rapid during annealing at 1273 K (1000 °C), regardless of Mn contents. Austenite growth was believed to be controlled at early stages by carbon diffusion following pearlite/bainite dissolution. The redistribution of Mn in ferrite and particularly in austenite at later stages was too subtle to result in a measureable change in austenite fraction. Further, the hot band microstructure of both steels contained a large fraction of coarse-grained δ-ferrite, which remained almost unchanged during intercritical annealing. A recently developed thermodynamic database was evaluated using the experimental data. The new database achieved a better agreement

  3. Quantum fields on manifolds: PCT and gravitationally induced thermal states

    International Nuclear Information System (INIS)

    Sewell, G.L.

    1982-01-01

    We formulate an axiomatic scheme, designed to provide a framework for a general, rigorous theory of relativistic quantum fields on a class of manifolds, that includes Kruskal's extension of Schwarzchild space-time, as well as Minkowski space-time. The scheme is an adaptation of Wightman's to this class of manifolds. We infer from it that, given an arbitrary field (in general, interacting) on a manifold X, the restriction of the field to a certain open submanifold X/sup( + ), whose boundaries are event horizons, satisfies the Kubo--Martin--Schwinger (KMS) thermal equilibrium conditions. This amounts to a rigorous, model-independent proof of a generalized Hawking--Unruh effect. Further, in cases where the field enjoys a certain PCT symmetry, the conjugation governing the KMS condition is just the PCT operator. The key to these results is an analogue, that we prove, of the Bisognano--Wichmann theorem [J. Math. Phys. 17, (1976), Theorem 1]. We also construct an alternative scheme by replacing a regularity condition at an event horizon by the assumption that the field in X/sup( + ) is in a ground, rather then a thermal, state. We show that, in this case, the observables in X/sup( + ) are uncorrelated to those in its causal complement, X/sup( - ), and thus that the event horizons act as physical barriers. Finally, we argue that the choice between the two schemes must be dictated by the prevailing conditions governing the state of the field

  4. Culture, cross-role consistency, and adjustment: testing trait and cultural psychology perspectives.

    Science.gov (United States)

    Church, A Timothy; Anderson-Harumi, Cheryl A; del Prado, Alicia M; Curtis, Guy J; Tanaka-Matsumi, Junko; Valdez Medina, José L; Mastor, Khairul A; White, Fiona A; Miramontes, Lilia A; Katigbak, Marcia S

    2008-09-01

    Trait and cultural psychology perspectives on cross-role consistency and its relation to adjustment were examined in 2 individualistic cultures, the United States (N=231) and Australia (N=195), and 4 collectivistic cultures, Mexico (N=199), the Philippines (N=195), Malaysia (N=217), and Japan (N=180). Cross-role consistency in trait ratings was evident in all cultures, supporting trait perspectives. Cultural comparisons of mean consistency provided support for cultural psychology perspectives as applied to East Asian cultures (i.e., Japan) but not collectivistic cultures more generally. Some but not all of the hypothesized predictors of consistency were supported across cultures. Cross-role consistency predicted aspects of adjustment in all cultures, but prediction was most reliable in the U.S. sample and weakest in the Japanese sample. Alternative constructs proposed by cultural psychologists--personality coherence, social appraisal, and relationship harmony--predicted adjustment in all cultures but were not, as hypothesized, better predictors of adjustment in collectivistic cultures than in individualistic cultures.

  5. Swift and Certain, Proportionate and Consistent: Key Values of Urine Drug Test Consequences for Probationers.

    Science.gov (United States)

    Cadwallader, Amy B

    2017-09-01

    Traditionally, urine drug testing (UDT) in the correctional population (both prison and community corrections) has been infrequent, is scheduled, and has a high possibility of delayed results. Of practical relevance is that scheduled testing is ineffective for identifying drug misuse. Of ethical relevance is that consequences of positive scheduled tests can be unpredictable-in the form of overly severe punishment or a lack of treatment options-and that the scheduled testing paradigm is a poor way to change behaviors. More innovative programs now use a UDT paradigm with more frequent, random testing providing rapid results and certain, swift consequences and addiction treatment when warranted or requested. Studies have shown these new programs-the foundation of which is frequent, random UDTs-to significantly reduce drug use, criminal recidivism, and incarceration. © 2017 American Medical Association. All Rights Reserved.

  6. Corrosion behavior of environmental assessment glass in product consistency tests of extended duration

    International Nuclear Information System (INIS)

    Bates, J.K.; Buck, E.C.; Ebert, W.L.; Luo, J.S.; Tam, S.W.

    1998-01-01

    We have conducted static dissolution tests to study the corrosion behavior of the Environmental Assessment (EA) glass, which is the benchmark glass for high-level waste glasses being produced at US Department of Energy facilities. These tests were conducted to evaluate the behavior of the EA glass under the same long-term and accelerated test conditions that are being used to evaluate the corrosion of waste glasses. Tests were conducted at 90 C in a tuff groundwater solution at glass surface area/solution volume (WV) ratios of about 2000 and 20,000 m -1 . The glass dissolved at three distinct dissolution rates in tests conducted at 2000 m -1 . Based on the release of boron, dissolution within the first seven days occurred at a rate of about 0.65 g/(m 2 · d). The rate between seven and 70 days decreased to 0.009 g/(m 2 · d). An increase in the dissolution rate occurred at longer times after the precipitation of zeolite phases analcime, gmelinite, and an aluminum silicate base. The dissolution rate after phase formation was about 0.18 g/(m 2 · d). The formation of the same zeolite alteration phases occurred after about 20 days in tests at 20,000 m - . The average dissolution rate over the first 20 days was 0.5 g/(m 2 · d) and the rate after phase formation was about 0.20 g/(m 2 · d). An intermediate stage with a lower rate was not observed in tests at 20,000 m -1 . The corrosion behavior of EA glass is similar to that observed for other high-level waste glasses reacted under the same test conditions. The dissolution rate of EA glass is higher than that of other high-level waste glasses both in 7-day tests and after alteration phases form

  7. Single-field consistency relations of large scale structure part III: test of the equivalence principle

    Energy Technology Data Exchange (ETDEWEB)

    Creminelli, Paolo [Abdus Salam International Centre for Theoretical Physics, Strada Costiera 11, Trieste, 34151 (Italy); Gleyzes, Jérôme; Vernizzi, Filippo [CEA, Institut de Physique Théorique, Gif-sur-Yvette cédex, F-91191 France (France); Hui, Lam [Physics Department and Institute for Strings, Cosmology and Astroparticle Physics, Columbia University, New York, NY, 10027 (United States); Simonović, Marko, E-mail: creminel@ictp.it, E-mail: jerome.gleyzes@cea.fr, E-mail: lhui@astro.columbia.edu, E-mail: msimonov@sissa.it, E-mail: filippo.vernizzi@cea.fr [SISSA, via Bonomea 265, Trieste, 34136 (Italy)

    2014-06-01

    The recently derived consistency relations for Large Scale Structure do not hold if the Equivalence Principle (EP) is violated. We show it explicitly in a toy model with two fluids, one of which is coupled to a fifth force. We explore the constraints that galaxy surveys can set on EP violation looking at the squeezed limit of the 3-point function involving two populations of objects. We find that one can explore EP violations of order 10{sup −3}÷10{sup −4} on cosmological scales. Chameleon models are already very constrained by the requirement of screening within the Solar System and only a very tiny region of the parameter space can be explored with this method. We show that no violation of the consistency relations is expected in Galileon models.

  8. Dilatometric and dielectric behaviour of Sm modified PCT ceramics

    International Nuclear Information System (INIS)

    Singh, Sarabjit; Thakur, O.P.; Prakash, Chandra; Raina, K.K.

    2005-01-01

    Samarium modified PCT ceramics with composition (Pb 0.76-x Sm x Ca 0.24 )(Ti 0.98 Mn 0.02 )O 3 ; x=0-0.08 in steps of 0.02 were prepared by conventional mixed-oxide method. Detailed dilatometric studies were carried out for green specimens in order to study sintering behaviour. Change in the dilatometric behaviour is correlated with the XRD results of powders calcined at different temperatures. Dielectric constant was observed to increase with increasing Sm concentration, which has been attributed to reduced tetragonality and better densification on Sm substitution. SEM micrographs have revealed the grain size of the samples. Ferroelectric hysteresis behaviour was studied for all the compositions

  9. Portable Computer Technology (PCT) Research and Development Program Phase 2

    Science.gov (United States)

    Castillo, Michael; McGuire, Kenyon; Sorgi, Alan

    1995-01-01

    The subject of this project report, focused on: (1) Design and development of two Advanced Portable Workstation 2 (APW 2) units. These units incorporate advanced technology features such as a low power Pentium processor, a high resolution color display, National Television Standards Committee (NTSC) video handling capabilities, a Personal Computer Memory Card International Association (PCMCIA) interface, and Small Computer System Interface (SCSI) and ethernet interfaces. (2) Use these units to integrate and demonstrate advanced wireless network and portable video capabilities. (3) Qualification of the APW 2 systems for use in specific experiments aboard the Mir Space Station. A major objective of the PCT Phase 2 program was to help guide future choices in computing platforms and techniques for meeting National Aeronautics and Space Administration (NASA) mission objectives. The focus being on the development of optimal configurations of computing hardware, software applications, and network technologies for use on NASA missions.

  10. Validity test and its consistency in the construction of patient loyalty model

    Science.gov (United States)

    Yanuar, Ferra

    2016-04-01

    The main objective of this present study is to demonstrate the estimation of validity values and its consistency based on structural equation model. The method of estimation was then implemented to an empirical data in case of the construction the patient loyalty model. In the hypothesis model, service quality, patient satisfaction and patient loyalty were determined simultaneously, each factor were measured by any indicator variables. The respondents involved in this study were the patients who ever got healthcare at Puskesmas in Padang, West Sumatera. All 394 respondents who had complete information were included in the analysis. This study found that each construct; service quality, patient satisfaction and patient loyalty were valid. It means that all hypothesized indicator variables were significant to measure their corresponding latent variable. Service quality is the most measured by tangible, patient satisfaction is the most mesured by satisfied on service and patient loyalty is the most measured by good service quality. Meanwhile in structural equation, this study found that patient loyalty was affected by patient satisfaction positively and directly. Service quality affected patient loyalty indirectly with patient satisfaction as mediator variable between both latent variables. Both structural equations were also valid. This study also proved that validity values which obtained here were also consistence based on simulation study using bootstrap approach.

  11. Consistency of test behaviour and individual difference in prescision of prediction

    NARCIS (Netherlands)

    Meijer, R.R.

    1998-01-01

    Ghiselli ((1956, 1960) argued that the precision of prediction on the basis of a test may vary for different individuals. To quantify the individual precision of prediction he compared the observed criterion scores with the expected criterion scores estimated on the basis of the total scores on a

  12. Individual behavioural differences in pigs: intra- and inter-test consistency

    NARCIS (Netherlands)

    Spoolder, H.A.M.; Burbidge, J.A.; Lawrence, A.B.; Simmins, P.H.

    1996-01-01

    Two studies have been carried out in pigs to determine the relation between escape behaviour and production parameters and between escape behaviour and other factors. In the first, 823 piglets were tested with the backtest at 10 and 17 days of age. Production parameters, such as average daily weight

  13. Factors associated with not testing for HIV and consistent condom use among men in Soweto, South Africa.

    Directory of Open Access Journals (Sweden)

    Sakhile Mhlongo

    Full Text Available Besides access to medical male circumcision, HIV testing, access to condoms and consistent condom use are additional strategies men can use to prevent HIV acquisition. We examine male behavior toward testing and condom use.To determine factors associated with never testing for HIV and consistent condom use among men who never test in Soweto.A cross-sectional survey in Soweto was conducted in 1539 men aged 18-32 years in 2007. Data were collected on socio-demographic and behavioral characteristics to determine factors associated with not testing and consistent condom use.Over two thirds (71% of men had not had an HIV test and the majority (55%, n = 602 were young (18-23. Of those not testing, condom use was poor (44%, n = 304. Men who were 18-23 years (aOR: 2.261, CI: 1.534-3.331, with primary (aOR: 2.096, CI: 1.058-4.153 or high school (aOR: 1.622, CI: 1.078-2.439 education, had sex in the last 6 months (aOR: 1.703, CI: 1.055-2.751, and had ≥1 sexual partner (aOR: 1.749, CI: 1.196-2.557 were more likely not to test. Of those reporting condom use (n = 1036, 67%, consistent condom use was 43% (n = 451. HIV testing did not correlate with condom use.Low rates of both condom use and HIV testing among men in a high HIV prevalence setting are worrisome and indicate an urgent need to develop innovative behavioral strategies to address this shortfall. Condom use is poor in this population whether tested or not tested for HIV, indicating no association between condom use and HIV testing.

  14. Operational quality control of daily precipitation using spatio-climatological consistency testing

    Science.gov (United States)

    Scherrer, S. C.; Croci-Maspoli, M.; van Geijtenbeek, D.; Naguel, C.; Appenzeller, C.

    2010-09-01

    Quality control (QC) of meteorological data is of utmost importance for climate related decisions. The search for an effective automated QC of precipitation data has proven difficult and many weather services still use mainly manual inspection of daily precipitation including MeteoSwiss. However, man power limitations force many weather services to move towards less labour intensive and more automated QC with the challenge to keeping data quality high. In the last decade, several approaches have been presented to objectify daily precipitation QC. Here we present a spatio-climatological approach that will be implemented operationally at MeteoSwiss. It combines the information from the event based spatial distribution of everyday's precipitation field and the historical information of the interpolation error using different precipitation intensity intervals. Expert judgement shows that the system is able to detect potential outliers very well (hardly any missed errors) without creating too many false alarms that need human inspection. 50-80% of all flagged values have been classified as real errors by the data editor. This is much better than the roughly 15-20% using standard spatial regression tests. Very helpful in the QC process is the automatic redistribution of accumulated several day sums. Manual inspection in operations can be reduced and the QC of precipitation objectified substantially.

  15. Determining the dissolution rates of actinide glasses: A time and temperature Product Consistency Test study

    International Nuclear Information System (INIS)

    Daniel, W.E.; Best, D.R.

    1995-01-01

    Vitrification has been identified as one potential option for the e materials such as Americium (Am), Curium (Cm), Neptunium (Np), and Plutonium (Pu). A process is being developed at the Savannah River Site to safely vitrify all of the highly radioactive Am/Cm material and a portion of the fissile (Pu) actinide materials stored on site. Vitrification of the Am/Cm will allow the material to be transported and easily stored at the Oak Ridge National Laboratory. The Am/Cm glass has been specifically designed to be (1) highly durable in aqueous environments and (2) selectively attacked by nitric acid to allow recovery of the valuable Am and Cm isotopes. A similar glass composition will allow for safe storage of surplus plutonium. This paper will address the composition, relative durability, and dissolution rate characteristics of the actinide glass, Loeffler Target, that will be used in the Americium/Curium Vitrification Project at Westinghouse Savannah River Company near Aiken, South Carolina. The first part discusses the tests performed on the Loeffler Target Glass concerning instantaneous dissolution rates. The second part presents information concerning pseudo-activation energy for the one week glass dissolution process

  16. Elevated Temperature Effects on the Plastic Anisotropy of an Extruded Mg-4 Wt Pct Li Alloy: Experiments and Polycrystal Modeling

    Science.gov (United States)

    Risse, Marcel; Lentz, Martin; Fahrenson, Christoph; Reimers, Walter; Knezevic, Marko; Beyerlein, Irene J.

    2017-01-01

    In this work, we study the deformation behavior of Mg-4 wt pct Li in uniaxial tension as a function of temperature and loading direction. Standard tensile tests were performed at temperatures in the range of 293 K (20 °C) ≤ T ≤ 473 K (200 °C) and in two in-plane directions: the extrusion and the transverse. We find that while the in-plane plastic anisotropy (PA) decreases with temperature, the anisotropy in failure strain and texture development increases. To uncover the temperature dependence in the critical stresses for slip and in the amounts of slip and twinning systems mediating deformation, we employ the elastic-plastic self-consistent polycrystal plasticity model with a thermally activated dislocation density based hardening law for activating slip with individual crystals. We demonstrate that the model, with a single set of intrinsic material parameters, achieves good agreement with the stress-strain curves, deformation textures, and intragranular misorientation axis analysis for all test directions and temperatures. With the model, we show that at all temperatures the in-plane tensile behavior is driven primarily by analysis explains that the in-plane PA decreases and failure strains increase with temperature as a result of a significant reduction in the activation stress for pyramidal multiple types of < a rangle and < {c + a} rangle slip. The results also show that because of the strong initial texture, in-plane texture development is anisotropic since prismatic slip dominates the deformation in one test, although it is not the easiest slip mode, and basal slip in the other. These findings reveal the relationship between the temperature-sensitive thresholds needed to activate crystallographic slip and the development of texture and macroscopic PA.

  17. Preferences for HIV test characteristics among young, Black Men Who Have Sex With Men (MSM) and transgender women: Implications for consistent HIV testing

    Science.gov (United States)

    Frye, Victoria; Hirshfield, Sabina; Chiasson, Mary Ann; Lucy, Debbie; Usher, DaShawn; McCrossin, Jermaine; Greene, Emily; Koblin, Beryl

    2018-01-01

    Background Promoting consistent HIV testing is critical among young, Black Men Who Have Sex With Men (MSM) and transgender women who are overrepresented among new HIV cases in the United States. New HIV test options are available, including mobile unit testing, one-minute testing, at home or self-testing and couples HIV testing and counseling (CHTC). In the context of these newer options, the objective of this study was to explore whether and how preferences for specific characteristics of the tests acted as barriers to and/or facilitators of testing in general and consistent testing specifically among young Black MSM and transgender women aged 16 to 29. Methods We conducted 30 qualitative, semi-structured, in-depth interviews with young, Black, gay, bisexual or MSM and transgender women in the New York City metropolitan area to identify preferences for specific HIV tests and aspects of HIV testing options. Participants were primarily recruited from online and mobile sites, followed by community-based, face-to-face recruitment strategies to specifically reach younger participants. Thematic coding was utilized to analyze the qualitative data based on a grounded theoretical approach. Results We identified how past experiences, perceived test characteristics (e.g., accuracy, cost, etc.) and beliefs about the “fit” between the individual, and the test relate to preferred testing methods and consistent testing. Three major themes emerged as important to preferences for HIV testing methods: the perceived accuracy of the test method, venue characteristics, and lack of knowledge or experience with the newer testing options, including self-testing and CHTC. Conclusions These findings suggest that increasing awareness of and access to newer HIV testing options (e.g., free or reduced price on home or self-tests or CHTC available at all testing venues) is critical if these new options are to facilitate increased levels of consistent testing among young, Black MSM and

  18. Preferences for HIV test characteristics among young, Black Men Who Have Sex With Men (MSM) and transgender women: Implications for consistent HIV testing.

    Science.gov (United States)

    Frye, Victoria; Wilton, Leo; Hirshfield, Sabina; Chiasson, Mary Ann; Lucy, Debbie; Usher, DaShawn; McCrossin, Jermaine; Greene, Emily; Koblin, Beryl

    2018-01-01

    Promoting consistent HIV testing is critical among young, Black Men Who Have Sex With Men (MSM) and transgender women who are overrepresented among new HIV cases in the United States. New HIV test options are available, including mobile unit testing, one-minute testing, at home or self-testing and couples HIV testing and counseling (CHTC). In the context of these newer options, the objective of this study was to explore whether and how preferences for specific characteristics of the tests acted as barriers to and/or facilitators of testing in general and consistent testing specifically among young Black MSM and transgender women aged 16 to 29. We conducted 30 qualitative, semi-structured, in-depth interviews with young, Black, gay, bisexual or MSM and transgender women in the New York City metropolitan area to identify preferences for specific HIV tests and aspects of HIV testing options. Participants were primarily recruited from online and mobile sites, followed by community-based, face-to-face recruitment strategies to specifically reach younger participants. Thematic coding was utilized to analyze the qualitative data based on a grounded theoretical approach. We identified how past experiences, perceived test characteristics (e.g., accuracy, cost, etc.) and beliefs about the "fit" between the individual, and the test relate to preferred testing methods and consistent testing. Three major themes emerged as important to preferences for HIV testing methods: the perceived accuracy of the test method, venue characteristics, and lack of knowledge or experience with the newer testing options, including self-testing and CHTC. These findings suggest that increasing awareness of and access to newer HIV testing options (e.g., free or reduced price on home or self-tests or CHTC available at all testing venues) is critical if these new options are to facilitate increased levels of consistent testing among young, Black MSM and transgender women. Addressing perceptions of

  19. Preferences for HIV test characteristics among young, Black Men Who Have Sex With Men (MSM and transgender women: Implications for consistent HIV testing.

    Directory of Open Access Journals (Sweden)

    Victoria Frye

    Full Text Available Promoting consistent HIV testing is critical among young, Black Men Who Have Sex With Men (MSM and transgender women who are overrepresented among new HIV cases in the United States. New HIV test options are available, including mobile unit testing, one-minute testing, at home or self-testing and couples HIV testing and counseling (CHTC. In the context of these newer options, the objective of this study was to explore whether and how preferences for specific characteristics of the tests acted as barriers to and/or facilitators of testing in general and consistent testing specifically among young Black MSM and transgender women aged 16 to 29.We conducted 30 qualitative, semi-structured, in-depth interviews with young, Black, gay, bisexual or MSM and transgender women in the New York City metropolitan area to identify preferences for specific HIV tests and aspects of HIV testing options. Participants were primarily recruited from online and mobile sites, followed by community-based, face-to-face recruitment strategies to specifically reach younger participants. Thematic coding was utilized to analyze the qualitative data based on a grounded theoretical approach.We identified how past experiences, perceived test characteristics (e.g., accuracy, cost, etc. and beliefs about the "fit" between the individual, and the test relate to preferred testing methods and consistent testing. Three major themes emerged as important to preferences for HIV testing methods: the perceived accuracy of the test method, venue characteristics, and lack of knowledge or experience with the newer testing options, including self-testing and CHTC.These findings suggest that increasing awareness of and access to newer HIV testing options (e.g., free or reduced price on home or self-tests or CHTC available at all testing venues is critical if these new options are to facilitate increased levels of consistent testing among young, Black MSM and transgender women. Addressing

  20. Resting functional imaging tools (MRS, SPECT, PET and PCT).

    Science.gov (United States)

    Van Der Naalt, J

    2015-01-01

    Functional imaging includes imaging techniques that provide information about the metabolic and hemodynamic status of the brain. Most commonly applied functional imaging techniques in patients with traumatic brain injury (TBI) include magnetic resonance spectroscopy (MRS), single photon emission computed tomography (SPECT), positron emission tomography (PET) and perfusion CT (PCT). These imaging modalities are used to determine the extent of injury, to provide information for the prediction of outcome, and to assess evidence of cerebral ischemia. In TBI, secondary brain damage mainly comprises ischemia and is present in more than 80% of fatal cases with traumatic brain injury (Graham et al., 1989; Bouma et al., 1991; Coles et al., 2004). In particular, while SPECT measures cerebral perfusion and MRS determines metabolism, PET is able to assess both perfusion and cerebral metabolism. This chapter will describe the application of these techniques in traumatic brain injury separately for the major groups of severity comprising the mild and moderate to severe group. The application in TBI and potential difficulties of each technique is described. The use of imaging techniques in children will be separately outlined. © 2015 Elsevier B.V. All rights reserved.

  1. The eye-complaint questionnaire in a visual display unit work environment: Internal consistency and test-retest reliability

    NARCIS (Netherlands)

    Steenstra, Ivan A.; Sluiter, Judith K.; Frings-Dresen, Monique H. W.

    2009-01-01

    The internal consistency and test-retest reliability of a 10-item eye-complaint questionnaire (ECQ) were examined within a sample of office workers. Repeated within-subjects measures were performed within a single day and over intervals of 1 and 7 d. Questionnaires were completed by 96 workers (70%

  2. Prediction and Cross-Situational Consistency of Daily Behavior across Cultures: Testing Trait and Cultural Psychology Perspectives

    Science.gov (United States)

    Church, A. Timothy; Katigbak, Marcia S.; Reyes, Jose Alberto S.; Salanga, Maria Guadalupe C.; Miramontes, Lilia A.; Adams, Nerissa B.

    2008-01-01

    Trait and cultural psychology perspectives on the cross-situational consistency of behavior, and the predictive validity of traits, were tested in a daily process study in the United States (N = 68), an individualistic culture, and the Philippines (N = 80), a collectivistic culture. Participants completed the Revised NEO Personality Inventory (Costa & McCrae, 1992) and a measure of self-monitoring, then reported their daily behaviors and associated situational contexts for approximately 30 days. Consistent with trait perspectives, the Big Five traits predicted daily behaviors in both cultures, and relative (interindividual) consistency was observed across many, although not all, situational contexts. The frequency of various Big Five behaviors varied across relevant situational contexts in both cultures and, consistent with cultural psychology perspectives, there was a tendency for Filipinos to exhibit greater situational variability than Americans. Self-monitoring showed some ability to account for individual differences in situational variability in the American sample, but not the Filipino sample. PMID:22146866

  3. Analytical evaluation of Lumipulse® BRAHMS PCT CLEIA assay and clinical performances in an unselected population as compared with central lab PCT assay.

    Science.gov (United States)

    Dupuy, Anne Marie; Né, Maxence; Bargnoux, Anne Sophie; Badiou, Stéphanie; Cristol, Jean Paul

    2017-03-01

    We report the analytical performances of the Lumipulse®G BRAHMS PCT assay (Fujirebio, Courteboeuf, France) and the concordance with BRAHMS PCT Kryptor CompactPlus© results from central laboratory. Lumipulse®G BRAHMS PCT immunoassay on Lumipulse®G600II instrument is a chemiluminescence enzyme immunoassay (CLEIA). Analytical performances included imprecision study, linearity, limit of detection and comparison study on 138 plasma specimen on Lumipulse®G600II vs plasma on Kryptor CompactPlus©. The intra and inter assay imprecision of Lumipulse®G BRAHMS PCT was between 2 and 5%. The LoD in our condition was 0.0029ng/mL in accordance with the LoD provided by the manufacturer (0.0048ng/mL). The linear equation of linearity was y=1,001×-0,052 with r 2 =0.99, with a mean recovery (SD) percentage of 1.8% (8%). Correlation studies showed a good correlation (r=0.99) between plasma on Kryptor and Lumipulse, with a bias of 0.02 in the range from 0.12 to 1ng/mL. The new adaptation developed from Fujirebio on quantification of PCT with CLEIA technology from monoclonal antibodies from ThermoFisher appears to be acceptable for clinical use. Copyright © 2016 The Canadian Society of Clinical Chemists. Published by Elsevier Inc. All rights reserved.

  4. Expanding the Reach of Participatory Risk Management: Testing an Online Decision-Aiding Framework for Informing Internally Consistent Choices.

    Science.gov (United States)

    Bessette, Douglas L; Campbell-Arvai, Victoria; Arvai, Joseph

    2016-05-01

    This article presents research aimed at developing and testing an online, multistakeholder decision-aiding framework for informing multiattribute risk management choices associated with energy development and climate change. The framework was designed to provide necessary background information and facilitate internally consistent choices, or choices that are in line with users' prioritized objectives. In order to test different components of the decision-aiding framework, a six-part, 2 × 2 × 2 factorial experiment was conducted, yielding eight treatment scenarios. The three factors included: (1) whether or not users could construct their own alternatives; (2) the level of detail regarding the composition of alternatives users would evaluate; and (3) the way in which a final choice between users' own constructed (or highest-ranked) portfolio and an internally consistent portfolio was presented. Participants' self-reports revealed the framework was easy to use and providing an opportunity to develop one's own risk-management alternatives (Factor 1) led to the highest knowledge gains. Empirical measures showed the internal consistency of users' decisions across all treatments to be lower than expected and confirmed that providing information about alternatives' composition (Factor 2) resulted in the least internally consistent choices. At the same time, those users who did not develop their own alternatives and were not shown detailed information about the composition of alternatives believed their choices to be the most internally consistent. These results raise concerns about how the amount of information provided and the ability to construct alternatives may inversely affect users' real and perceived internal consistency. © 2015 Society for Risk Analysis.

  5. VLE measurements using a static cell vapor phase manual sampling method accompanied with an empirical data consistency test

    International Nuclear Information System (INIS)

    Freitag, Joerg; Kosuge, Hitoshi; Schmelzer, Juergen P.; Kato, Satoru

    2015-01-01

    Highlights: • We use a new, simple static cell vapor phase manual sampling method (SCVMS) for VLE (x, y, T) measurement. • The method is applied to non-azeotropic, asymmetric and two-liquid phase forming azeotropic binaries. • The method is approved by a data consistency test, i.e., a plot of the polarity exclusion factor vs. pressure. • The consistency test reveals that with the new SCVMS method accurate VLE near ambient temperature can be measured. • Moreover, the consistency test approves that the effect of air in the SCVMS system is negligible. - Abstract: A new static cell vapor phase manual sampling (SCVMS) method is used for the simple measurement of constant temperature x, y (vapor + liquid) equilibria (VLE). The method was applied to the VLE measurements of the (methanol + water) binary at T/K = (283.2, 298.2, 308.2 and 322.9), asymmetric (acetone + 1-butanol) binary at T/K = (283.2, 295.2, 308.2 and 324.2) and two-liquid phase forming azeotropic (water + 1-butanol) binary at T/K = (283.2 and 298.2). The accuracy of the experimental data was approved by a data consistency test, that is, an empirical plot of the polarity exclusion factor, β, vs. the system pressure, P. The SCVMS data are accurate, because the VLE data converge to the same lnβ vs. lnP straight line determined from conventional distillation-still method and a headspace gas chromatography method

  6. Frequency and determinants of consistent STI/HIV testing among men who have sex with men testing at STI outpatient clinics in the Netherlands: a longitudinal study.

    Science.gov (United States)

    Visser, Maartje; Heijne, Janneke C M; Hogewoning, Arjan A; van Aar, Fleur

    2017-09-01

    Men who have sex with men (MSM) are at highest risk for STIs and HIV infections in the Netherlands. However, official guidelines on STI testing among MSM are lacking. They are advised to test for STIs at least every six months, but their testing behaviour is not well known. This study aimed to get insight into the proportion and determinants of consistent 6-monthly STI testing among MSM testing at STI outpatient clinics in the Netherlands. This study included longitudinal surveillance data of STI consultations among MSM from all 26 STI outpatient clinics in the Netherlands between 1 June 2014 and 31 December 2015. Multinomial logistic regression analysis was used to identify determinants of consistent 6-monthly testing compared with single testing and inconsistent testing. Determinants of time between consultations among men with multiple consultations were analysed using a Cox Prentice-Williams-Peterson gap-time model. A total of 34 605 STI consultations of 18 634 MSM were included. 8966 (48.1%) men had more than one consultation, and 3516 (18.9%) men tested consistently 6-monthly. Indicators of high sexual risk behaviour, including having a history of STI, being HIV positive and having more than 10 sex partners, were positively associated with both being a consistent tester and returning to the STI clinic sooner. Men who were notified by a partner or who reported STI symptoms were also more likely to return to the STI clinic sooner, but were less likely to be consistent testers, identifying a group of event-driven testers. The proportion of consistent 6-monthly testers among MSM visiting Dutch STI outpatient clinics was low. Testing behaviour was associated with sexual risk behaviour, but exact motives to test consistently remain unclear. Evidence-based testing guidelines are needed to achieve optimal reductions in STI transmission in the future. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence

  7. Changes of serum procalcitonin (PCT) and IL-6 levels in patients with sepsis

    International Nuclear Information System (INIS)

    Wang Jinjiang

    2007-01-01

    Objective: To investigate the importance of determination of changes of serum procalcitonin (PCT) and IL-6 levels in patients with sepsis. Methods: Serum PCT (with double-sandwich immunofluorescence assay) and IL-6 (with ELISA) levels were measured repeatedly in 130 patients with sepsis on d1, d3, d5, d7 after admission. Values in 130 healthy individuals were also measured as control. Results: The serum levels of PCT and IL-6 in the patients with sepsis of admission were significantly higher than those in controls. The levels dropped markedly in the survivors by d7. Among the septic patients, the levels in the succumbed patients were significantly higher those in the survivors (P<0.05). Conclusion: Serum PCT and IL-6 values appeared to be of prognostic value in patients with sepsis. (authors)

  8. PCT Uncertainty Analysis Using Unscented Transform with Random Orthogonal Matrix

    Energy Technology Data Exchange (ETDEWEB)

    Fynana, Douglas A.; Ahn, Kwang-Il [KAERI, Daejeon (Korea, Republic of); Lee, John C. [Univ. of Michigan, Michigan (United States)

    2015-05-15

    Most Best Estimate Plus Uncertainty (BEPU) methods employ nonparametric order statistics through Wilks' formula to quantify uncertainties of best estimate simulations of nuclear power plant (NPP) transients. 95%/95% limits, the 95''t{sup h} percentile at a 95% confidence level, are obtained by randomly sampling all uncertainty contributors through conventional Monte Carlo (MC). Advantages are simple implementation of MC sampling of input probability density functions (pdfs) and limited computational expense of 1''s{sup t}, 2''n{sup d}, and 3''r{sup d} order Wilks' formula requiring only 59, 93, or 124 simulations, respectively. A disadvantage of small sample size is large sample to sample variation of statistical estimators. This paper presents a new efficient sampling based algorithm for accurate estimation of mean and variance of the output parameter pdf. The algorithm combines a deterministic sampling method, the unscented transform (UT), with random sampling through the generation of a random orthogonal matrix (ROM). The UT guarantees the mean, covariance, and 3''r{sup d} order moments of the multivariate input parameter distributions are exactly preserved by the sampled input points and the orthogonal transformation of the points by a ROM guarantees the sample error of all 4''t{sup h} order and higher moments are unbiased. The UT with ROM algorithm is applied to the uncertainty quantification of the peak clad temperature (PCT) during a large break loss-of-coolant accident (LBLOCA) in an OPR1000 NPP to demonstrate the applicability of the new algorithm to BEPU. This paper presented a new algorithm combining the UT with ROM for efficient multivariate parameter sampling that ensures sample input covariance and 3''r{sup d} order moments are exactly preserved and 4''th moment errors are small and unbiased. The advantageous sample properties guarantee higher order accuracy and

  9. Assessment of test-retest reliability and internal consistency of the Wisconsin Gait Scale in hemiparetic post-stroke patients

    Directory of Open Access Journals (Sweden)

    Guzik Agnieszka

    2016-09-01

    Full Text Available Introduction: A proper assessment of gait pattern is a significant aspect in planning the process of teaching gait in hemiparetic post-stroke patients. The Wisconsin Gait Scale (WGS is an observational tool for assessing post-stroke patients’ gait. The aim of the study was to assess test-retest reliability and internal consistency of the WGS and examine correlations between gait assessment made with the WGS and gait speed, Brunnström scale, Ashworth’s scale and the Barthel Index.

  10. Doubling immunochemistry laboratory testing efficiency with the cobas e 801 module while maintaining consistency in analytical performance.

    Science.gov (United States)

    Findeisen, P; Zahn, I; Fiedler, G M; Leichtle, A B; Wang, S; Soria, G; Johnson, P; Henzell, J; Hegel, J K; Bendavid, C; Collet, N; McGovern, M; Klopprogge, K

    2018-06-04

    The new immunochemistry cobas e 801 module (Roche Diagnostics) was developed to meet increasing demands on routine laboratories to further improve testing efficiency, while maintaining high quality and reliable data. During a non-interventional multicenter evaluation study, the overall performance, functionality and reliability of the new module was investigated under routine-like conditions. It was tested as a dedicated immunochemistry system at four sites and as a consolidator combined with clinical chemistry at three sites. We report on testing efficiency and analytical performance of the new module. Evaluation of sample workloads with site-specific routine request patterns demonstrated increased speed and almost doubled throughput (maximal 300 tests per h), thus revealing that one cobas e 801 module can replace two cobas e 602 modules while saving up to 44% floor space. Result stability was demonstrated by QC analysis per assay throughout the study. Precision testing over 21 days yielded excellent results within and between labs, and, method comparison performed versus the cobas e 602 module routine results showed high consistency of results for all assays under study. In a practicability assessment related to performance and handling, 99% of graded features met (44%) or even exceeded (55%) laboratory expectations, with enhanced reagent management and loading during operation being highlighted. By nearly doubling immunochemistry testing efficiency on the same footprint as a cobas e 602 module, the new module has a great potential to further consolidate and enhance laboratory testing while maintaining high quality analytical performance with Roche platforms. Copyright © 2018 The Canadian Society of Clinical Chemists. Published by Elsevier Inc. All rights reserved.

  11. Full-scale locomotive dynamic crash testing and correlations : locomotive consist colliding with steel coil truck at grade crossing (test 3).

    Science.gov (United States)

    2011-09-01

    This report presents the test results and finite element correlations of a full-scale dynamic collision between a locomotive and a highway truck loaded with two heavy steel coils. The locomotive consist was moving at 58 miles per hour before it struc...

  12. PROCALCITONIN TESTING IN RHEUMATOLOGY

    Directory of Open Access Journals (Sweden)

    D. V. Bukhanova

    2017-01-01

    Full Text Available Currently, differential diagnosis of systemic bacterial infection and active rheumatic process remains a challenging problem in rheumatology. In the review, current data on the role of procalcitonin biomarker in diagnosis and differential diagnosis of rheumatic diseases (RD and infectious pathology are presented. In particular, some authors recommend procalcitonin (PCT test as a marker of bacterial infection in bones and joints at levels above 0.5 ng/ml; at PCT level below 0.3 ng/ml, infection can be ruled out. In patients with microcrystalline arthritis, data on the significance of PCT for differential diagnosis are contradictory. PCT level doesn’t correlate with systemic lupus erythematosus activity and is elevated only during bacterial infection proportionally to its systematicity. In some studies, elevated PCT level was observed in ANCA-associated vasculitis with high activity without bacterial infection. It was shown that in 80 % of adults with Still’s disease, PCT level was higher than the threshold value even without infection. For patients with RD hospitalized in intensive care units, PCT clearance is a more informative predictive characteristic than its level, regardless of the cause of PCT elevation (infection, injury, severe organ damage, etc.; slowdown of its decrease is a factor of poor prognosis and is associated with higher mortality. At the same time, PCT level positively correlates with the SOFA score in presence of bacterial infection. For some rheumatic diseases, the threshold PCT value at which the test has optimal sensitivity and specificity is yet to be established. Nonetheless, PCT should be evaluated in relation to the clinical picture and data of additional examinations. The effect of various therapy methods used in rheumatology on PCT level requires further research.

  13. Application of the methodology of surface of answer in the determination of the PCT in the simulation of a LOFT; Aplicacion de la metodologia de superficies de respuesta en la determinacion del PCT en la simulacion de un LOFT

    Energy Technology Data Exchange (ETDEWEB)

    Alva N, J. [IPN, Escuela Superior de Fisica y Matematicas, Departamento de Ingenieria Nuclear, Av. IPN s/n, Col. Lindavista, Mexico 07738 D.F. (Mexico); Ortiz V, J.; Amador G, R. [ININ, 52750 La Marquesa, Estado de Mexico (Mexico)]. e-mail: neriaesfm@gmail.com

    2008-07-01

    This article summarizes the main typical of the methodology of surfaces and answer (MSA) and its connections with the lineal regression analysis. Also, an example of the application of MSA in the prediction of the principle cladding temperature (PCT) of a combustible assembly of a nuclear reactor, whose used data were taken from the simulation of a LOFT (Loss Of Fluid Test) during a course of experts. The made prediction will be used like one first approach to predict the behavior of the PCT, this is made in order to diminish the time of calculation when realizing the executions of codes thermal hydraulics of better estimation. The present work comprises of the theoretical base of the project in charge to delineate a methodology of uncertainty analysis for codes of better estimation, employees in the thermal hydraulics analysis and safety of plants and nuclear reactors. The institutions that participate in such project are: ININ, CFE, IPN and CNSNS, is possible to mention that this project is sponsored by the IAEA. (Author)

  14. Metabolic borderzone in acutely ischemic canine myocardium demonstrated by positron-CT (PCT)

    International Nuclear Information System (INIS)

    Schwaiger, M.; Hansen, H.; Selin, C.; Wittmer, S.; Barrio, J.; Schelbert, H.R.

    1984-01-01

    Acute coronary ligation in dogs results in an area of myocardial dysfunction that exceeds the area of subsequent necrosis suggesting the existence of an ischemic ''borderzone'' of reversibly injured myocardium. The authors tested this hypothesis in 9 closed chest dogs with C-11 plamitate (CPA) and serial PCT imaging after an LAD occlusion. Using a blood flow (MBF) image obtained with iv N-13 ammonia prior to CPA, regions of interest were assigned on the serial CPA cross-sectional images to the center (IC) and border (IB) of the ischemic segment and to control myocardium (CO). CPA uptake was closely related to MBF (r=0.88) implicating flow as a major determinant of CPA uptake. Clearance helftimes (T 1/2) and relative sizes (RS) of the early rapid phase on the C-11 tissue time activity curves were determined for IC, IB and CO. In IC, MBF, RS and T 1/2 were markedly depressed indicating impaired CPA utilization and oxidation. In IB, MBF was less than in CO though only insignificantly, while RS and T 1/2 were highly abnormal. The authors conclude that FFA metabolism in areas adjacent to ischemic segments but without significant MBF decreases in abnormal, presenting evidence for a metabolic borderzone which now can be identified noninvasively with positron emission tomography

  15. Evaluation of uniformity and noise consistency test in computerized tomography; Avaliacao do teste de constancia de uniformidade e ruido em tomografia computadorizada

    Energy Technology Data Exchange (ETDEWEB)

    Garrafiel, Fabricio N.; Lykawka, Rochelle; Bacelar, Alexandre; Anes, Mauricio [Hospital de Clinicas de Porto Alegre (HCPA), Porto Alegre, RS (Brazil); Alva-Sanchez, Mirko S.; Pianoschi, Thatiane [Universidade Federal de Ciencias da Saude de Porto Alegre (UFCSPA), Porto Alegre, RS (Brazil)

    2016-07-01

    The quality control in computed tomography (CT) includes constancy tests concerning CT number measured in a water tank and its response as uniformity, noise and accuracy. Differences to It's performance limits and frequency established by the national legislation and the manufacturers are well known. The objective of this study was to evaluate these tests proceeding the periodicity of national legislation. The tests were conducted over a three equipment of different manufacturers. Each equipment has acquisition parameters and specific phantoms. The equipment presented adequate performance and consistency when compared their results to the limits determined by manufacturers and national legislation. Thus, it is important to evaluate these differences and tailor the testing protocols to meet the established quality requirements. (author)

  16. Application of the methodology of surface of answer in the determination of the PCT in the simulation of a LOFT

    International Nuclear Information System (INIS)

    Alva N, J.; Ortiz V, J.; Amador G, R.

    2008-01-01

    This article summarizes the main typical of the methodology of surfaces and answer (MSA) and its connections with the lineal regression analysis. Also, an example of the application of MSA in the prediction of the principle cladding temperature (PCT) of a combustible assembly of a nuclear reactor, whose used data were taken from the simulation of a LOFT (Loss Of Fluid Test) during a course of experts. The made prediction will be used like one first approach to predict the behavior of the PCT, this is made in order to diminish the time of calculation when realizing the executions of codes thermal hydraulics of better estimation. The present work comprises of the theoretical base of the project in charge to delineate a methodology of uncertainty analysis for codes of better estimation, employees in the thermal hydraulics analysis and safety of plants and nuclear reactors. The institutions that participate in such project are: ININ, CFE, IPN and CNSNS, is possible to mention that this project is sponsored by the IAEA. (Author)

  17. Integration and consistency testing of groundwater flow models with hydro-geochemistry in site investigations in Finland

    International Nuclear Information System (INIS)

    Pitkaenen, P.; Loefman, J.; Korkealaakso, J.; Koskinen, L.; Ruotsalainen, P.; Hautojaervi, A.; Aeikaes, T.

    1999-01-01

    In the assessment of the suitability and safety of a geological repository for radioactive waste the understanding of the fluid flow at a site is essential. In order to build confidence in the assessment of the hydrogeological performance of a site in various conditions, integration of hydrological and hydrogeochemical methods and studies provides the primary method for investigating the evolution that has taken place in the past, and for predicting future conditions at the potential disposal site. A systematic geochemical sampling campaign was started since the beginning of 1990's in the Finnish site investigation programme. This enabled the initiating of integration and evaluation of site scale hydrogeochemical and groundwater flow models. Hydrogeochemical information has been used to screen relevant external processes and variables for definition of the initial and boundary conditions in hydrological simulations. The results obtained from interpretation and modelling hydrogeochemical evolution have been employed in testing the hydrogeochemical consistency of conceptual flow models. Integration and testing of flow models with hydrogeochemical information are considered to improve significantly the hydrogeological understanding of a site and increases confidence in conceptual hydrogeological models. (author)

  18. A general approach to the testing of binary solubility systems for thermodynamic consistency. Consolidated Fuel Reprocessing Program

    Science.gov (United States)

    Hamm, L. L.; Vanbrunt, V.

    1982-08-01

    The numerical solution to the ordinary differential equation which describes the high-pressure vapor-liquid equilibria of a binary system where one of the components is supercritical and exists as a noncondensable gas in the pure state is considered with emphasis on the implicit Runge-Kuta and orthogonal collocation methods. Some preliminary results indicate that the implicit Runge-Kutta method is superior. Due to the extreme nonlinearity of thermodynamic properties in the region near the critical locus, and extended cubic spline fitting technique is devised for correlating the P-x data. The least-squares criterion is employed in smoothing the experimental data. The technique could easily be applied to any thermodynamic data by changing the endpoint requirements. The volumetric behavior of the systems must be given or predicted in order to perform thermodynamic consistency tests. A general procedure is developed for predicting the volumetric behavior required and some indication as to the expected limit of accuracy is given.

  19. How TEM Projection Artifacts Distort Microstructure Measurements: A Case Study in a 9 pct Cr-Mo-V Steel

    Science.gov (United States)

    Monsegue, Niven; Reynolds, William T.; Hawk, Jeffrey A.; Murayama, Mitsuhiro

    2014-08-01

    Morphological data obtained from two-dimensional (2D) and three-dimensional (3D) transmission electron microscopy (TEM) observations were compared to assess the effects of TEM projection errors for submicron-size precipitates. The microstructure consisted of M23C6 carbides in a 9 pct Cr-Mo-V heat resistant steel before and after exposure to creep conditions. Measurements obtained from about 800 carbides demonstrate that particle size and spacing estimates made from 2D observations overestimate the more accurate values obtained from 3D reconstructions. The 3D analysis also revealed the M23C6 precipitates lengthen anisotropically along lath boundary planes, suggesting that coarsening during the early stage of creep in this alloy system is governed by grain boundary diffusion.

  20. Corrosion and Mechanical Properties of Al-5 At. Pct Cr Produced by Cryomilling and Subsequent Consolidation at Various Temperatures

    Science.gov (United States)

    Esquivel, J.; Darling, K. A.; Murdoch, H. A.; Gupta, R. K.

    2018-04-01

    An Al-5 at. pct Cr alloy was produced by high-energy ball milling at liquid nitrogen temperature followed by consolidation using equal-channel axial extrusion at 200 °C, 300 °C and 450 °C. The microstructure and corrosion response were compared with a cast alloy of the same composition. Rather than the intermetallics expected by the phase diagram and seen in the cast alloy, consolidated HEBM alloys exhibited extended solid solubility of Cr in the aluminum matrix in addition to a finely dispersed Cr-rich phase. This led to improvement in the corrosion behavior as investigated via potentiodynamic polarization and constant immersion tests in NaCl solution. Hardness and tensile tests were performed to evaluate the mechanical properties. The highest consolidation temperature (450 °C) contributed to significant grain growth and Cr diffusion, lessening the beneficial effects of processing with HEBM.

  1. Microstructures and Surface Stabilities of {Ni-0.4C-6Ta- xCr, 0 ≤ x ≤ 50 Wt Pct} Cast Alloys at High Temperature

    Science.gov (United States)

    Berthod, Patrice

    2018-06-01

    Nickel-based cast alloys rich in chromium and reinforced by TaC carbides are potentially very interesting alloys for applications at elevated temperatures. Unfortunately, unlike cobalt-chromium and iron-chromium alloys, it is difficult to obtain exclusively TaC as primary carbides in Ni-Cr alloys. In alloys containing 30 wt pct Cr tantalum, carbides coexist with chromium carbides. The latter tend to weaken the alloy at elevated temperatures because they become rapidly spherical and then quickly lose their reinforcing effect. In this work, we attempted to stabilize TaC as a single carbide phase by testing different chromium contents in the [0, 50 wt pct] range. Six alloys containing 0.4C and 6Ta, weight contents corresponding to equivalent molar contents, were elaborated by foundry, and their as-cast microstructures were characterized. Samples of all alloys were exposed to 1127 °C and 1237 °C for 24 hours to characterize their stabilized microstructures. The surface fractions of chromium carbides and tantalum carbides were measured by image analysis, and their evolutions vs the chromium content were studied. For the chosen C and Ta contents, it appears that obtaining TaC only is possible by decreasing the chromium content to 10 wt pct. At the same time, TaC fractions are unfortunately too low because a large portion of tantalum integrates into the solid solution in the matrix. A second consequence is a critical decrease in oxidation resistance. Other possible methods to stabilize TaC as a single carbide are evocated, such as the simultaneous increase in Ta and decrease in chromium from 30 wt pct Cr.

  2. BRST quantization, IOSp(D,2vertical stroke2) invariance and the PCT theorem. Pt. 2

    International Nuclear Information System (INIS)

    Gozzi, E.; Reuter, M.

    1989-01-01

    In this paper we investigate the IOSp(D,2vertical stroke2) extension of the Poincare group in the BRST quantization of the spinning particle. We construct the corresponding field theory and prove its dimensional reduction a la Parisi-Sourlas. As for the scalar particle, we show that the PCT transformation is induced by a certain element of the SO(D,2) subgroup of the IOSp(D,2vertical stroke2). This gives us the tools for a simple interpretation of the PCT theorem for Dirac fields. We then continue by clarifying the role of the 'modular' transformation (i.e. world-line orientation-reversing diffeomorphism) and indicating how the PCT transformation could be regarded as a modular transformation in a 'SO(D,2)-rotated' frame. In this work we also consider BRST field theories leading to D-dimensional chiral theories. In this case, of the full IOSp(D,2vertical stroke2) symmetry, only the subgroup IO(D-1,1)xOSp(1,1vertical stroke2) survives together with the (isolated) SO(D,2) transformation that implements PCT. (orig.)

  3. Assessment of Effect on LBLOCA PCT for Change in Upper Head Nodalization

    International Nuclear Information System (INIS)

    Kang, Dong Gu; Huh, Byung Gil; Yoo, Seung Hun; Bang, Youngseok; Seul, Kwangwon; Cho, Daehyung

    2014-01-01

    In this study, the best estimate plus uncertainty (BEPU) analysis of LBLOCA for original and modified nodalizations was performed, and the effect on LBLOCA PCT for change in upper head nodalization was assessed. In this study, the best estimate plus uncertainty (BEPU) analysis of LBLOCA for original and modified nodalizations was performed, and the effect on LBLOCA PCT for change in upper head nodalization was assessed. It is confirmed that modification of upper head nodalization influences PCT behavior, especially in the reflood phase. In conclusions, the modification of nodalization to reflect design characteristic of upper head temperature should be done to predict PCT behavior accurately in LBLOCA analysis. In the best estimate (BE) method with the uncertainty evaluation, the system nodalization is determined by the comparative studies of the experimental data. Up to now, it was assumed that the temperature of the upper dome in OPR-1000 was close to that of the cold leg. However, it was found that the temperature of the upper head/dome might be a little lower than or similar to that of the hot leg through the evaluation of the detailed design data. Since the higher upper head temperature affects blowdown quenching and peak cladding temperature in the reflood phase, the nodalization for upper head should be modified

  4. Kinetics of Z-Phase Precipitation in 9 to 12 pct Cr Steels

    DEFF Research Database (Denmark)

    Danielsen, Hilmar Kjartansson; Nunzio, Paolo Emilio di; Hald, John

    2013-01-01

    precipitated Z-phase 20 to 50 times faster than the 9 pct Cr steel. Transmission electron microscopy (TEM) was applied to follow the Z-phase precipitation, using energy-dispersive X-ray spectroscopy (EDS) line scans and atomic resolution imaging. © The Minerals, Metals & Materials Society and ASM International...

  5. Dissolution rates of DWPF glasses from long-term PCT

    International Nuclear Information System (INIS)

    Ebert, W.L.; Tam, S.W.

    1996-01-01

    We have characterized the corrosion behavior of several Defense Waste Processing Facility (DWPF) reference waste glasses by conducting static dissolution tests with crushed glasses. Glass dissolution rates were calculated from measured B concentrations in tests conducted for up to five years. The dissolution rates of all glasses increased significantly after certain alteration phases precipitated. Calculation of the dissolution rates was complicated by the decrease in the available surface area as the glass dissolves. We took the loss of surface area into account by modeling the particles to be spheres, then extracting from the short-term test results the dissolution rate corresponding to a linear decrease in the radius of spherical particles. The measured extent of dissolution in tests conducted for longer times was less than predicted with this linear dissolution model. This indicates that advanced stages of corrosion are affected by another process besides dissolution, which we believe to be associated with a decrease in the precipitation rate of the alteration phases. These results show that the dissolution rate measured soon after the formation of certain alteration phases provides an upper limit for the long-term dissolution rate, and can be used to determine a bounding value for the source term for radionuclide release from waste glasses. The long-term dissolution rates measured in tests at 20,000 per m at 90 degrees C in tuff groundwater at pH values near 12 for the Environmental Assessment glass and glasses made with SRL 131 and SRL 202 frits, respectively

  6. Does computer-synthesized speech manifest personality? Experimental tests of recognition, similarity-attraction, and consistency-attraction.

    Science.gov (United States)

    Nass, C; Lee, K M

    2001-09-01

    Would people exhibit similarity-attraction and consistency-attraction toward unambiguously computer-generated speech even when personality is clearly not relevant? In Experiment 1, participants (extrovert or introvert) heard a synthesized voice (extrovert or introvert) on a book-buying Web site. Participants accurately recognized personality cues in text to speech and showed similarity-attraction in their evaluation of the computer voice, the book reviews, and the reviewer. Experiment 2, in a Web auction context, added personality of the text to the previous design. The results replicated Experiment 1 and demonstrated consistency (voice and text personality)-attraction. To maximize liking and trust, designers should set parameters, for example, words per minute or frequency range, that create a personality that is consistent with the user and the content being presented.

  7. Interface Consistency

    DEFF Research Database (Denmark)

    Staunstrup, Jørgen

    1998-01-01

    This paper proposes that Interface Consistency is an important issue for the development of modular designs. Byproviding a precise specification of component interfaces it becomes possible to check that separately developedcomponents use a common interface in a coherent matter thus avoiding a very...... significant source of design errors. Awide range of interface specifications are possible, the simplest form is a syntactical check of parameter types.However, today it is possible to do more sophisticated forms involving semantic checks....

  8. Contraction Twinning Dominated Tensile Deformation and Subsequent Fracture in Extruded Mg-1Mn (Wt Pct) at Ambient Temperature

    Science.gov (United States)

    Chakkedath, A.; Maiti, T.; Bohlen, J.; Yi, S.; Letzig, D.; Eisenlohr, P.; Boehlert, C. J.

    2018-03-01

    Due to their excellent strength-to-weight ratio, Mg alloys are attractive for applications where weight savings are critical. However, the limited cold formability of wrought Mg alloys severely restricts their widespread usage. In order to study the role that deformation twinning might play in limiting the elongation-to-failure ({ɛ} f ), in-situ tensile tests along the extrusion axis of Mg-1Mn (wt pct) were performed at 323 K, 423 K, and 523 K. The alloy exhibited a strong basal texture such that most of the grains experienced compression along their -axis during deformation. At 323 K, fracture occurred at about 10 pct strain. Although basal, prismatic, and pyramidal slip activity was observed along with extension twinning, contraction twinning significantly influenced the deformation, and such twins evolved into {10{\\bar{1}} 1}-{10{\\bar{1}} 2} double twins. Crystal plasticity simulation showed localized shear deformation within the contraction twins and double twins due to the enhanced activity of basal slip in the reoriented twin volume. Due to this, the twin-matrix interface was identified to be a potential crack initiation site. Thus, contraction twins were considered to have led to the failure of the material at a relatively low strain, suggesting that this deformation mode is detrimental to the cold formability of Mg and its alloys. With increasing temperature, there was a significant decrease in the activity of contraction twinning as well as extension twinning, along with a decrease in the tensile strength and an increase in the {ɛ} f value. A combination of basal, prismatic, and pyramidal slips accounted for a large percentage of the observed deformation activity at 423 K and 523 K. The lack of contraction twinning was explained by the expected decrease in the critical resolved shear stress values for pyramidal slip, and the improved {ɛ} f values at elevated temperatures were attributed to the vanishing activity of contraction twinning.

  9. Contraction Twinning Dominated Tensile Deformation and Subsequent Fracture in Extruded Mg-1Mn (Wt Pct) at Ambient Temperature

    Science.gov (United States)

    Chakkedath, A.; Maiti, T.; Bohlen, J.; Yi, S.; Letzig, D.; Eisenlohr, P.; Boehlert, C. J.

    2018-06-01

    Due to their excellent strength-to-weight ratio, Mg alloys are attractive for applications where weight savings are critical. However, the limited cold formability of wrought Mg alloys severely restricts their widespread usage. In order to study the role that deformation twinning might play in limiting the elongation-to-failure ({ɛ} _{ {f}}), in-situ tensile tests along the extrusion axis of Mg-1Mn (wt pct) were performed at 323 K, 423 K, and 523 K. The alloy exhibited a strong basal texture such that most of the grains experienced compression along their -axis during deformation. At 323 K, fracture occurred at about 10 pct strain. Although basal, prismatic, and pyramidal slip activity was observed along with extension twinning, contraction twinning significantly influenced the deformation, and such twins evolved into {10{\\bar{1}}1}-{10{\\bar{1}}2} double twins. Crystal plasticity simulation showed localized shear deformation within the contraction twins and double twins due to the enhanced activity of basal slip in the reoriented twin volume. Due to this, the twin-matrix interface was identified to be a potential crack initiation site. Thus, contraction twins were considered to have led to the failure of the material at a relatively low strain, suggesting that this deformation mode is detrimental to the cold formability of Mg and its alloys. With increasing temperature, there was a significant decrease in the activity of contraction twinning as well as extension twinning, along with a decrease in the tensile strength and an increase in the {ɛ} _{ {f}} value. A combination of basal, prismatic, and pyramidal slips accounted for a large percentage of the observed deformation activity at 423 K and 523 K. The lack of contraction twinning was explained by the expected decrease in the critical resolved shear stress values for pyramidal slip, and the improved {ɛ} _{ {f}} values at elevated temperatures were attributed to the vanishing activity of contraction twinning.

  10. Self-consistency and sum-rule tests in the Kramers-Kronig analysis of optical data: Applications to aluminum

    International Nuclear Information System (INIS)

    Shiles, E.; Sasaki, T.; Inokuti, M.; Smith, D.Y.

    1980-01-01

    An iterative, self-consistent procedure for the Kramers-Kronig analysis of data from reflectance, ellipsometric, transmission, and electron-energy-loss measurements is presented. This procedure has been developed for practical dispersion analysis since experimentally no single optical function can be readily measured over the entire range of frequencies as required by the Kramers-Kronig relations. The present technique is applied to metallic aluminum as an example. The results are then examined for internal consistency and for systematic errors by various optical sum rules. The present procedure affords a systematic means of preparing a self-consistent set of optical functions provided some optical or energy-loss data are available in all important spectral regions. The analysis of aluminum discloses that currently available data exhibit an excess oscillator strength, apparently in the vicinity of the L edge. A possible explanation is a systematic experimental error in the absorption-coefficient measurements resulting from surface layers: possibly oxides: present in thin-film transmission samples. A revised set of optical functions has been prepared by an ad hoc reduction of the reported absorption coefficient above the L edge by 14%. These revised data lead to a total oscillator strength consistent with the known electron density and are in agreement with dc-conductivity and stopping-power measurements as well as with absorption coefficients inferred from the cross sections of neighboring elements in the periodic table. The optical functions resulting from this study show evidence for both the redistribution of oscillator strength between energy levels and the effects on real transitions of the shielding of conduction electrons by virtual processes in the core states

  11. An Audit of Repeat Testing at an Academic Medical Center: Consistency of Order Patterns With Recommendations and Potential Cost Savings.

    Science.gov (United States)

    Hueth, Kyle D; Jackson, Brian R; Schmidt, Robert L

    2018-05-31

    To evaluate the prevalence of potentially unnecessary repeat testing (PURT) and the associated economic burden for an inpatient population at a large academic medical facility. We evaluated all inpatient test orders during 2016 for PURT by comparing the intertest times to published recommendations. Potential cost savings were estimated using the Centers for Medicare & Medicaid Services maximum allowable reimbursement rate. We evaluated result positivity as a determinant of PURT through logistic regression. Of the evaluated 4,242 repeated target tests, 1,849 (44%) were identified as PURT, representing an estimated cost-savings opportunity of $37,376. Collectively, the association of result positivity and PURT was statistically significant (relative risk, 1.2; 95% confidence interval, 1.1-1.3; P < .001). PURT contributes to unnecessary health care costs. We found that a small percentage of providers account for the majority of PURT, and PURT is positively associated with result positivity.

  12. Measuring the Consistency in Change in Hepatitis B Knowledge among Three Different Types of Tests: True/False, Multiple Choice, and Fill in the Blanks Tests.

    Science.gov (United States)

    Sahai, Vic; Demeyere, Petra; Poirier, Sheila; Piro, Felice

    1998-01-01

    The recall of information about Hepatitis B demonstrated by 180 seventh graders was tested with three test types: (1) short-answer; (2) true/false; and (3) multiple-choice. Short answer testing was the most reliable. Suggestions are made for the use of short-answer tests in evaluating student knowledge. (SLD)

  13. Basic investigation of the laminated alginate impression technique: Setting time, permanent deformation, elastic deformation, consistency, and tensile bond strength tests.

    Science.gov (United States)

    Kitamura, Aya; Kawai, Yasuhiko

    2015-01-01

    Laminated alginate impression for edentulous is simple and time efficient compared to border molding technique. The purpose of this study was to examine clinical applicability of the laminated alginate impression, by measuring the effects of different Water/Powder (W/P) and mixing methods, and different bonding methods in the secondary impression of alginate impression. Three W/P: manufacturer-designated mixing water amount (standard), 1.5-fold (1.5×) and 1.75-fold (1.75×) water amount were mixed by manual and automatic mixing methods. Initial and complete setting time, permanent and elastic deformation, and consistency of the secondary impression were investigated (n=10). Additionally, tensile bond strength between the primary and secondary impression were measured in the following surface treatment; air blow only (A), surface baking (B), and alginate impression material bonding agent (ALGI-BOND: AB) (n=12). Initial setting times significantly shortened with automatic mixing for all W/P (p<0.05). The permanent deformation decreased and elastic deformation increased as high W/P, regardless of the mixing method. Elastic deformation significantly reduced in 1.5× and 1.75× with automatic mixing (p<0.05). All of these properties resulted within JIS standards. For all W/P, AB showed a significantly high bonding strength as compared to A and B (p<0.01). The increase of mixing water, 1.5× and 1.75×, resulted within JIS standards in setting time, suggesting its applicability in clinical setting. The use of automatic mixing device decreased elastic strain and shortening of the curing time. For the secondary impression application of adhesives on the primary impression gives secure adhesion. Copyright © 2014 Japan Prosthodontic Society. Published by Elsevier Ltd. All rights reserved.

  14. Relationship between acute kidney injury and serum procalcitonin (PCT) concentration in critically ill patients with influenza infection.

    Science.gov (United States)

    Rodríguez, A; Reyes, L F; Monclou, J; Suberviola, B; Bodí, M; Sirgo, G; Solé-Violán, J; Guardiola, J; Barahona, D; Díaz, E; Martín-Loeches, I; Restrepo, M I

    2018-02-09

    Serum procalcitonin (PCT) concentration could be increased in patients with renal dysfunction in the absence of bacterial infection. To determine the interactions among serum renal biomarkers of acute kidney injury (AKI) and serum PCT concentration, in patients admitted to the intensive care unit (ICU) due to lung influenza infection. Secondary analysis of a prospective multicentre observational study. 148 Spanish ICUs. ICU patients admitted with influenza infection without bacterial co-infection. Clinical, laboratory and hemodynamic variables were recorded. AKI was classified as AKI I or II based on creatinine (Cr) concentrations (≥1.60-2.50mg/dL and Cr≥2.51-3.99mg/dL, respectively). Patients with chronic renal disease, receiving renal replacement treatment or with Cr>4mg/dL were excluded. Spearman's correlation, simple and multiple linear regression analysis were performed. None. Out of 663 patients included in the study, 52 (8.2%) and 10 (1.6%) developed AKI I and II, respectively. Patients with AKI were significantly older, had more comorbid conditions and were more severally ill. PCT concentrations were higher in patients with AKI (2.62 [0.60-10.0]ng/mL vs. 0.40 [0.13-1.20]ng/mL, p=0.002). Weak correlations between Cr/PCT (rho=0.18) and Urea (U)/PCT (rho=0.19) were identified. Simple linear regression showed poor interaction between Cr/U and PCT concentrations (Cr R 2 =0.03 and U R 2 =0.018). Similar results were observed during multiple linear regression analysis (Cr R 2 =0.046 and U R 2 =0.013). Although PCT concentrations were slightly higher in patients with AKI, high PCT concentrations are not explained by AKI and could be warning sign of a potential bacterial infection. Copyright © 2018 Elsevier España, S.L.U. y SEMICYUC. All rights reserved.

  15. Internal consistency, reliability, and temporal stability of the Oxford Happiness Questionnaire short-form: Test-retest data over two weeks

    OpenAIRE

    MCGUCKIN, CONOR

    2006-01-01

    PUBLISHED The Oxford Happiness Questionnaire short-form is a recently developed eight-item measure of happiness. This study evaluated the internal consistency reliability and test-retest reliability of the Oxford Happiness Questionnaire short-form among 55 Northern Irish undergraduate university students who completed the measure on two occasions separated by two weeks. Internal consistency of the measure on both occasions was satisfactory at both Time 1 (alpha = .62) and Time 2 (alpha = ....

  16. Consistency tests of Ampcalculator and chiral amplitudes in SU(3) Chiral Perturbation Theory: A tutorial-based approach

    International Nuclear Information System (INIS)

    Ananthanarayan, B.; Sentitemsu Imsong, I.; Das, Diganta

    2012-01-01

    Ampcalculator (AMPC) is a Mathematica copyright based program that was made publicly available some time ago by Unterdorfer and Ecker. It enables the user to compute several processes at one loop (upto O(p 4 )) in SU(3) chiral perturbation theory. They include computing matrix elements and form factors for strong and non-leptonic weak processes with at most six external states. It was used to compute some novel processes and was tested against well-known results by the original authors. Here we present the results of several thorough checks of the package. Exhaustive checks performed by the original authors are not publicly available, and hence the present effort. Some new results are obtained from the software especially in the kaon odd-intrinsic parity non-leptonic decay sector involving the coupling G 27 . Another illustrative set of amplitudes at tree level we provide is in the context of τ-decays with several mesons including quark mass effects, of use to the BELLE experiment. All eight meson-meson scattering amplitudes have been checked. The Kaon-Compton amplitude has been checked and a minor error in the published results has been pointed out. This exercise is a tutorial-based one, wherein several input and output notebooks are also being made available as ancillary files on the arXiv. Some of the additional notebooks we provide contain explicit expressions that we have used for comparison with established results. The purpose is to encourage users to apply the software to suit their specific needs. An automatic amplitude generator of this type can provide error-free outputs that could be used as inputs for further simplification, and in varied scenarios such as applications of chiral perturbation theory at finite temperature, density and volume. This can also be used by students as a learning aid in low-energy hadron dynamics. (orig.)

  17. Use of the SSF equations in the Kojima-Moon-Ochi thermodynamic consistency test of isothermal vapour-liquid equilibrium data

    Directory of Open Access Journals (Sweden)

    SLOBODAN P. SERBANOVIC

    2000-12-01

    Full Text Available The Kojima-Moon-Ochi (KMO thermodynamic consistency test of vapour–liquid equilibrium (VLE measurements for 32 isothermal data sets of binary systems of various complexity was applied using two fitting equations: the Redlich-Kister equation and the Sum of Symmetrical Functions. It was shown that the enhanced reliability of the fitting of the experimental data can change the conclusions drawn on their thermodynamic consistency in those cases of VLE data sets that are estimated to be near the border of consistency.

  18. Detection of PCT and urinary β2 -MG enhances the accuracy for localization diagnosing pediatric urinary tract infection.

    Science.gov (United States)

    Fang, Jian; Luan, Jiangwei; Zhu, Gaohong; Qi, Chang; Wang, Dandan

    2017-09-01

    The purpose of this article was to investigate whether the combination of urinary beta 2 microglobulin (urinary β 2 -MG) and procalcitonin (PCT) diagnosis could enhance the localization diagnostic precision of pediatric urinary tract infection comparing with single diagnosis. A study was conducted in the Nephrology Department of Wuhan women and children's health care centre. This study incorporated 85 participants, including 35 children who were diagnosed as upper urinary tract infection (UUTI) with the symptom of fever and 50 children who conducted lower urinary tract infection (LUTI). Levels of PCT and urinary β 2 -MG in both UUTI and LUTI patients were measured and compared. The level of PCT and β 2 -MG were both significantly higher in UUTI group compared with in LUTI group. AUC of urinary β 2 -MG ROC (sensitivity of 71.4%, specificity of 90.0%) was significantly smaller than that of PCT ROC (sensitivity of 77.1%, specificity of 96.0%) in the single diagnosis. Although in the combined diagnosis, the sensitivity and specificity increased to 88.6% and 98%, respectively. Both PCT and β 2 -MG could be used to localize the UTI. Introducing urinary β 2 -MG into PCT diagnosis could increase the sensitivity and specificity of UTI lesion diagnosis in clinical practice. © 2016 The Authors Journal of Clinical Laboratory Analysis Published by Wiley Periodicals, Inc.

  19. Accelerated Leach Testing of GLASS: ALTGLASS Version 3.0

    Energy Technology Data Exchange (ETDEWEB)

    Trivelpiece, Cory L. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Jantzen, Carol M. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Crawford, Charles L. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL)

    2016-12-31

    The Accelerated Leach Testing of GLASS (ALTGLASS) database is a collection of data from short- and long-term product consistency tests (PCT, ASTM C1285 A and B) on high level waste (HLW) as well as low activity waste (LAW) glasses. The database provides both U.S. and international researchers with an archive of experimental data for the purpose of studying, modeling, or validating existing models of nuclear waste glass corrosion. The ALTGLASS database is maintained and updated by researchers at the Savannah River National Laboratory (SRNL). This newest version, ALTGLASS Version 3.0, has been updated with an additional 503 rows of data representing PCT results from corrosion experiments conducted in the United States by the Savannah River National Laboratory, Pacific Northwest National Laboratory, Argonne National Laboratory, and the Vitreous State Laboratory (SRNL, PNNL, ANL, VSL, respectively) as well as the National Nuclear Laboratory (NNL) in the United Kingdom.

  20. Accelerated Leach Testing of GLASS: ALTGLASS Version 3.0

    International Nuclear Information System (INIS)

    Trivelpiece, Cory L.; Jantzen, Carol M.; Crawford, Charles L.

    2016-01-01

    The Accelerated Leach Testing of GLASS (ALTGLASS) database is a collection of data from short- and long-term product consistency tests (PCT, ASTM C1285 A and B) on high level waste (HLW) as well as low activity waste (LAW) glasses. The database provides both U.S. and international researchers with an archive of experimental data for the purpose of studying, modeling, or validating existing models of nuclear waste glass corrosion. The ALTGLASS database is maintained and updated by researchers at the Savannah River National Laboratory (SRNL). This newest version, ALTGLASS Version 3.0, has been updated with an additional 503 rows of data representing PCT results from corrosion experiments conducted in the United States by the Savannah River National Laboratory, Pacific Northwest National Laboratory, Argonne National Laboratory, and the Vitreous State Laboratory (SRNL, PNNL, ANL, VSL, respectively) as well as the National Nuclear Laboratory (NNL) in the United Kingdom.

  1. Microanalysis on the Hydrogen Ion Irradiated 50 wt pct TiC-C Films

    Institute of Scientific and Technical Information of China (English)

    Hui JIANG; Yaoguang LIU; Ningkang HUANG

    2007-01-01

    The 50 wt pct TiC-C films were prepared on stainless steel substrates by using a technique of ion beam mixing.These films were irradiated by hydrogen ion beam with a dose of 1×1018 ions/cm2 and an energy of 5 keV.Microanalysis of X-ray photoelectron spectroscopy (XPS) and secondary ion mass spectroscopy (SIMS) were used to analyze the films before and after hydrogen ion irradiation and to study the mechanism of hydrogen resistance.

  2. Microstructure and Interfacial Reactions During Vacuum Brazing of Stainless Steel to Titanium Using Ag-28 pct Cu Alloy

    Science.gov (United States)

    Laik, A.; Shirzadi, A. A.; Sharma, G.; Tewari, R.; Jayakumar, T.; Dey, G. K.

    2015-02-01

    Microstructural evolution and interfacial reactions during vacuum brazing of grade-2 Ti and 304L-type stainless steel (SS) using eutectic alloy Ag-28 wt pct Cu were investigated. A thin Ni-depleted zone of -Fe(Cr, Ni) solid solution formed on the SS-side of the braze zone (BZ). Cu from the braze alloy, in combination with the dissolved Fe and Ti from the base materials, formed a layer of ternary compound , adjacent to Ti in the BZ. In addition, four binary intermetallic compounds, CuTi, CuTi, CuTi and CuTi formed as parallel contiguous layers in the BZ. The unreacted Ag solidified as islands within the layers of CuTi and CuTi. Formation of an amorphous phase at certain locations in the BZ could be revealed. The -Ti(Cu) layer, formed due to diffusion of Cu into Ti-based material, transformed to an -Ti + CuTi eutectoid with lamellar morphology. Tensile test showed that the brazed joints had strength of 112 MPa and failed at the BZ. The possible sequence of events that led to the final microstructure and the mode of failure of these joints were delineated.

  3. Test-Retest Reliability, Convergent Validity, and Internal Consistency of the Persian Version of Fullerton Advanced Balance Scale in Iranian Community-Dwelling Older Adults

    OpenAIRE

    Azar Sabet; Akram Azad; Ghorban Taghizadeh

    2016-01-01

    Objectives: This study was performed to evaluate convergent validity, test-retest reliability and internal consistency of the Persian translation of the Fullerton advanced balance (FAB) for use in Iranian community- dwelling older adults and improve the quality of their functional balance assessment. Methods & Materials: The original scale was translated with forward-backward protocol. In the next step, using convenience sampling and inclusion criteria, 88 functionally indep...

  4. Laboratory testing of glasses for Lockheed Idaho Technology Company: Final report

    International Nuclear Information System (INIS)

    Ellison, A.J.G.; Buck, E.C.; Dietz, N.L.; Ebert, W.L.; Luo, J.S.; Wolf, S.F.; Bates, J.K.

    1997-06-01

    Tests have been conducted at Argonne National Laboratory (ANL) in support of the efforts of Lockheed Idaho Technology Company (LITCO) to vitrify high-level waste calcines. Tests were conducted with three classes of LITCO glass formulations: Formula 127 (fluorine-bearing), Formula 532 (fluorine-free), and 630 series (both single- and mixed-alkali) glasses. The test matrices included, as appropriate, the Product Consistency Test Method B (PCT-B), the Materials Characterization Center Test 1 (MCC-1), and the Argonne vapor hydration test (VHT). Test durations ranged from 7 to 183 d. In 7-d PCT-Bs, normalized mass losses of major glass-forming elements for the LITCO glasses are similar to, or lower than, normalized mass losses obtained for other domestic candidate waste glasses. Formula 532 glasses form zeolite alteration phases relatively early in their reaction with water. The formation of those phases increased the dissolution rate. In contrast, the Formula 127 glass is highly durable and forms alteration phases only after prolonged exposure to water in tests with very high surface area to volume ratios; these alteration phases have a relatively small effect on the rate of glass corrosion. No alteration phases formed within the maximum test duration of 183 d in PCT-Bs with the 630 series glasses. The corrosion behavior of the mixed-alkali 630 series glasses is similar to that of 630 series glasses containing sodium alone. In VHTs, both single- and mixed-alkali glasses form zeolite phases that increase the rate of glass reaction. The original 630 series glasses and those based on a revised surrogate calcine formulation react at the same rate in PCT-Bs and form the same major alteration phases in VHTs

  5. Effect of Aluminum Coating on the Surface Properties of Ti-(~49 at. pct) Ni Alloy

    Science.gov (United States)

    Sinha, Arijit; Khan, Gobinda Gopal; Mondal, Bholanath; Majumdar, Jyotsna Dutta; Chattopadhyay, Partha Protim

    2015-08-01

    Stable porous layer of mixed Al2O3 and TiO2 has been formed on the Ti-(~49 at. pct) Ni alloy surface with an aim to suppress leaching of Ni from the alloy surface in contact with bio-fluid and to enhance the process of osseointegration. Aluminum coating on the Ni-Ti alloy surface prior to the anodization treatment has resulted in enhancement of depth and uniformity of pores. Thermal oxidation of the anodized aluminum-coated Ni-Ti samples has exhibited the formation of Al2O3 and TiO2 phases with dense porous structure. The nanoindentation and nanoscratch measurements have indicated a remarkable improvement in the hardness, wear resistance, and adhesiveness of the porous aluminum-coated Ni-Ti sample after thermal oxidation.

  6. Hvornår har 1.000 landbrug 90 pct. af landbrugsjorden?

    DEFF Research Database (Denmark)

    Hansen, Henning Otte

    2017-01-01

    relativt lige i forhold til resten af EU. De 1.000 største landbrug driver ca. 22 pct. af landbrugsjorden, og andelen er stigende. Selv om strukturudviklingen ser ud til at fortsætte upåvirket af de økonomiske og markedsmæssige forhold, kan man identificere en række forhold, som er drivkræfter bag...... strukturudviklingen. De senere års liberalisering af landbrugsloven har sandsynligvis været med til fremme strukturudviklingen yderligere, og også fjernelsen af EU‘s mælkekvoter ser ud til at have genskabt den hidtidige udvikling. Om strukturudviklingen er en fordel eller en ulempe kan ikke afgøres entydigt, idet...

  7. Product consistency test and toxicity characteristic leaching procedure results of the ceramic waste form from the electrometallurgical treatment process for spent fuel

    International Nuclear Information System (INIS)

    Johnson, S. G.; Adamic, M. L.: DiSanto, T.; Warren, A. R.; Cummings, D. G.; Foulkrod, L.; Goff, K. M.

    1999-01-01

    The ceramic waste form produced from the electrometallurgical treatment of sodium bonded spent fuel from the Experimental Breeder Reactor-II was tested using two immersion tests with separate and distinct purposes. The product consistency test is used to assess the consistency of the waste forms produced and thus is an indicator of a well-controlled process. The toxicity characteristic leaching procedure is used to determine whether a substance is to be considered hazardous by the Environmental Protection Agency. The proposed high level waste repository will not be licensed to receive hazardous waste, thus any waste forms destined to be placed there cannot be of a hazardous nature as defined by the Resource Conservation and Recovery Act. Results are presented from the first four fully radioactive ceramic waste forms produced and from seven ceramic waste forms produced from cold surrogate materials. The fully radioactive waste forms are approximately 2 kg in weight and were produced with salt used to treat 100 driver subassemblies of spent fuel

  8. Risk-informed analysis of the large break loss of coolant accident and PCT margin evaluation with the RISMC methodology

    International Nuclear Information System (INIS)

    Liang, T.H.; Liang, K.S.; Cheng, C.K.; Pei, B.S.; Patelli, E.

    2016-01-01

    Highlights: • With RISMC methodology, both aleatory and epistemic uncertainties have been considered. • 14 probabilistically significant sequences have been identified and quantified. • A load spectrum for LBLOCA has been conducted with CPCT and SP of each dominant sequence. • Comparing to deterministic methodologies, the risk-informed PCT margin can be greater by 44–62 K. • The SP of the referred sequence to cover 99% in the load spectrum is only 5.07 * 10 −3 . • The occurrence probability of the deterministic licensing sequence is 5.46 * 10 −5 . - Abstract: For general design basis accidents, such as SBLOCA and LBLOCA, the traditional deterministic safety analysis methodologies are always applied to analyze events based on a so called surrogate or licensing sequence, without considering how low this sequence occurrence probability is. In the to-be-issued 10 CFR 50.46a, the LBLOCA will be categorized as accidents beyond design basis and the PCT margin shall be evaluated in a risk-informed manner. According to the risk-informed safety margin characterization (RISMC) methodology, a process has been suggested to evaluate the risk-informed PCT margin. Following the RISMC methodology, a load spectrum of PCT for LBLOCA has been generated for the Taiwan’s Maanshan Nuclear Power plant and 14 probabilistic significant sequences have been identified. It was observed in the load spectrum that the conditional PCT generally ascends with the descending sequence occurrence probability. With the load spectrum covering both aleatory and epistemic uncertainties, the risk-informed PCT margin can be evaluated by either expecting value estimation method or sequence probability coverage method. It was found that by comparing with the traditional deterministic methodology, the PCT margin evaluated by the RISMC methodology can be greater by 44–62 K. Besides, to have a cumulated occurrence probability over 99% in the load spectrum, the occurrence probability of the

  9. Risk-informed analysis of the large break loss of coolant accident and PCT margin evaluation with the RISMC methodology

    Energy Technology Data Exchange (ETDEWEB)

    Liang, T.H. [Institute of Nuclear Engineering and Science, National Tsing Hua University, 101 Sec. 2, Kuang-Fu Road, Hsinchu 30013, Taiwan (China); Liang, K.S., E-mail: ksliang@alum.mit.edu [Institute of Nuclear Engineering and Science, National Tsing Hua University, 101 Sec. 2, Kuang-Fu Road, Hsinchu 30013, Taiwan (China); Cheng, C.K.; Pei, B.S. [Institute of Nuclear Engineering and Science, National Tsing Hua University, 101 Sec. 2, Kuang-Fu Road, Hsinchu 30013, Taiwan (China); Patelli, E. [Institute of Risk and Uncertainty, University of Liverpool, Room 610, Brodie Tower, L69 3GQ (United Kingdom)

    2016-11-15

    Highlights: • With RISMC methodology, both aleatory and epistemic uncertainties have been considered. • 14 probabilistically significant sequences have been identified and quantified. • A load spectrum for LBLOCA has been conducted with CPCT and SP of each dominant sequence. • Comparing to deterministic methodologies, the risk-informed PCT margin can be greater by 44–62 K. • The SP of the referred sequence to cover 99% in the load spectrum is only 5.07 * 10{sup −3}. • The occurrence probability of the deterministic licensing sequence is 5.46 * 10{sup −5}. - Abstract: For general design basis accidents, such as SBLOCA and LBLOCA, the traditional deterministic safety analysis methodologies are always applied to analyze events based on a so called surrogate or licensing sequence, without considering how low this sequence occurrence probability is. In the to-be-issued 10 CFR 50.46a, the LBLOCA will be categorized as accidents beyond design basis and the PCT margin shall be evaluated in a risk-informed manner. According to the risk-informed safety margin characterization (RISMC) methodology, a process has been suggested to evaluate the risk-informed PCT margin. Following the RISMC methodology, a load spectrum of PCT for LBLOCA has been generated for the Taiwan’s Maanshan Nuclear Power plant and 14 probabilistic significant sequences have been identified. It was observed in the load spectrum that the conditional PCT generally ascends with the descending sequence occurrence probability. With the load spectrum covering both aleatory and epistemic uncertainties, the risk-informed PCT margin can be evaluated by either expecting value estimation method or sequence probability coverage method. It was found that by comparing with the traditional deterministic methodology, the PCT margin evaluated by the RISMC methodology can be greater by 44–62 K. Besides, to have a cumulated occurrence probability over 99% in the load spectrum, the occurrence probability

  10. Impact of Alzheimer's Disease on Caregiver Questionnaire: internal consistency, convergent validity, and test-retest reliability of a new measure for assessing caregiver burden.

    Science.gov (United States)

    Cole, Jason C; Ito, Diane; Chen, Yaozhu J; Cheng, Rebecca; Bolognese, Jennifer; Li-McLeod, Josephine

    2014-09-04

    There is a lack of validated instruments to measure the level of burden of Alzheimer's disease (AD) on caregivers. The Impact of Alzheimer's Disease on Caregiver Questionnaire (IADCQ) is a 12-item instrument with a seven-day recall period that measures AD caregiver's burden across emotional, physical, social, financial, sleep, and time aspects. Primary objectives of this study were to evaluate psychometric properties of IADCQ administered on the Web and to determine most appropriate scoring algorithm. A national sample of 200 unpaid AD caregivers participated in this study by completing the Web-based version of IADCQ and Short Form-12 Health Survey Version 2 (SF-12v2™). The SF-12v2 was used to measure convergent validity of IADCQ scores and to provide an understanding of the overall health-related quality of life of sampled AD caregivers. The IADCQ survey was also completed four weeks later by a randomly selected subgroup of 50 participants to assess test-retest reliability. Confirmatory factor analysis (CFA) was implemented to test the dimensionality of the IADCQ items. Classical item-level and scale-level psychometric analyses were conducted to estimate psychometric characteristics of the instrument. Test-retest reliability was performed to evaluate the instrument's stability and consistency over time. Virtually none (2%) of the respondents had either floor or ceiling effects, indicating the IADCQ covers an ideal range of burden. A single-factor model obtained appropriate goodness of fit and provided evidence that a simple sum score of the 12 items of IADCQ can be used to measure AD caregiver's burden. Scales-level reliability was supported with a coefficient alpha of 0.93 and an intra-class correlation coefficient (for test-retest reliability) of 0.68 (95% CI: 0.50-0.80). Low-moderate negative correlations were observed between the IADCQ and scales of the SF-12v2. The study findings suggest the IADCQ has appropriate psychometric characteristics as a

  11. Internal consistency, test-retest reliability and measurement error of the self-report version of the social skills rating system in a sample of Australian adolescents.

    Directory of Open Access Journals (Sweden)

    Sharmila Vaz

    Full Text Available The social skills rating system (SSRS is used to assess social skills and competence in children and adolescents. While its characteristics based on United States samples (US are published, corresponding Australian figures are unavailable. Using a 4-week retest design, we examined the internal consistency, retest reliability and measurement error (ME of the SSRS secondary student form (SSF in a sample of Year 7 students (N = 187, from five randomly selected public schools in Perth, western Australia. Internal consistency (IC of the total scale and most subscale scores (except empathy on the frequency rating scale was adequate to permit independent use. On the importance rating scale, most IC estimates for girls fell below the benchmark. Test-retest estimates of the total scale and subscales were insufficient to permit reliable use. ME of the total scale score (frequency rating for boys was equivalent to the US estimate, while that for girls was lower than the US error. ME of the total scale score (importance rating was larger than the error using the frequency rating scale. The study finding supports the idea of using multiple informants (e.g. teacher and parent reports, not just student as recommended in the manual. Future research needs to substantiate the clinical meaningfulness of the MEs calculated in this study by corroborating them against the respective Minimum Clinically Important Difference (MCID.

  12. Internal consistency, test-retest reliability and measurement error of the self-report version of the social skills rating system in a sample of Australian adolescents.

    Science.gov (United States)

    Vaz, Sharmila; Parsons, Richard; Passmore, Anne Elizabeth; Andreou, Pantelis; Falkmer, Torbjörn

    2013-01-01

    The social skills rating system (SSRS) is used to assess social skills and competence in children and adolescents. While its characteristics based on United States samples (US) are published, corresponding Australian figures are unavailable. Using a 4-week retest design, we examined the internal consistency, retest reliability and measurement error (ME) of the SSRS secondary student form (SSF) in a sample of Year 7 students (N = 187), from five randomly selected public schools in Perth, western Australia. Internal consistency (IC) of the total scale and most subscale scores (except empathy) on the frequency rating scale was adequate to permit independent use. On the importance rating scale, most IC estimates for girls fell below the benchmark. Test-retest estimates of the total scale and subscales were insufficient to permit reliable use. ME of the total scale score (frequency rating) for boys was equivalent to the US estimate, while that for girls was lower than the US error. ME of the total scale score (importance rating) was larger than the error using the frequency rating scale. The study finding supports the idea of using multiple informants (e.g. teacher and parent reports), not just student as recommended in the manual. Future research needs to substantiate the clinical meaningfulness of the MEs calculated in this study by corroborating them against the respective Minimum Clinically Important Difference (MCID).

  13. A review of culturally adapted versions of the Oswestry Disability Index: the adaptation process, construct validity, test-retest reliability and internal consistency.

    Science.gov (United States)

    Sheahan, Peter J; Nelson-Wong, Erika J; Fischer, Steven L

    2015-01-01

    The Oswestry Disability Index (ODI) is a self-report-based outcome measure used to quantify the extent of disability related to low back pain (LBP), a substantial contributor to workplace absenteeism. The ODI tool has been adapted for use by patients in several non-English speaking nations. It is unclear, however, if these adapted versions of the ODI are as credible as the original ODI developed for English-speaking nations. The objective of this study was to conduct a review of the literature to identify culturally adapted versions of the ODI and to report on the adaptation process, construct validity, test-retest reliability and internal consistency of these ODIs. Following a pragmatic review process, data were extracted from each study with regard to these four outcomes. While most studies applied adaptation processes in accordance with best-practice guidelines, there were some deviations. However, all studies reported high-quality psychometric properties: group mean construct validity was 0.734 ± 0.094 (indicated via a correlation coefficient), test-retest reliability was 0.937 ± 0.032 (indicated via an intraclass correlation coefficient) and internal consistency was 0.876 ± 0.047 (indicated via Cronbach's alpha). Researchers can be confident when using any of these culturally adapted ODIs, or when comparing and contrasting results between cultures where these versions were employed. Implications for Rehabilitation Low back pain is the second leading cause of disability in the world, behind only cancer. The Oswestry Disability Index (ODI) has been developed as a self-report outcome measure of low back pain for administration to patients. An understanding of the various cross-cultural adaptations of the ODI is important for more concerted multi-national research efforts. This review examines 16 cross-cultural adaptations of the ODI and should inform the work of health care and rehabilitation professionals.

  14. Characterisation of the hydrodynamic properties within THE EDZ around drifts at level - 490 m of the Bure URL: a methodology for consistent interpretation of hydraulic tests

    International Nuclear Information System (INIS)

    Baechler, S.; Lavanchy, J.M.; Armand, G.; Cruchaudet, M.

    2010-01-01

    realistic manner) the combined effects of the EDZ around the drift and the mechanical disturbance due to drilling. Independent information from other investigation methods were used to critically revise the conceptual model and formation parameters. In particular, the absence of break-out in the borehole section close to the drift suggests that there is no additional mechanical disturbance induced by drilling in the pre-existing EDZ around a drift. Furthermore, preliminary information from the diffusion experiment (DIR) within undisturbed rock sections could confirm that the mechanically disturbed zone around a borehole may not exceed about 0.1 m (approximately one borehole diameter). Therefore several critical tests series from previous investigations were reanalysed using a revised conceptual model, constraining the extent of the mechanical disturbed zone around the borehole (conceptualized as a circular sheath around the test interval with a higher permeability than the formation further away from the borehole). The results of the reanalyses, performed with the numerical borehole simulator MULTISIM, are presented. The conclusions of the reanalyses in this borehole can be summarized as follows: - The test data could be reasonably fitted with the new constraints on the conceptual model. - In some cases the good quality of the simulations was, however, directly associated with higher storativity values of the EDZ (realistic considering unsaturated conditions). - Overall the new simulation results obtained from selected test series performed at various times are more consistent, without unexplained variations of the EDZ properties over time (i.e. permeability, storage coefficient). - The new conductivity profile indicates lower values between 0.8 and 2.0 m, which is consistent with drill core observations. - The significant decrease in permeability over 2.5 years between 0.3 and 0.5 m is confirmed. There has been no significant change of permeability since 2005 beyond 0.8 m

  15. A PCT-wide collaborative clinical audit selecting recall intervals for patients according to risk.

    Science.gov (United States)

    Cannell, P J

    2011-03-26

    This audit was carried out to assess the level to which recall intervals were individually and appropriately selected for patients attending dental practices across a primary care trust (PCT) area in Essex. A retrospective audit was carried out by reference to patient records to assess various criteria, including whether patients were categorised according to risk of oral disease, whether an appropriate recall had been selected and whether a discussion regarding a recall interval had been undertaken. An educational event highlighting the issue of recall intervals was held. Subsequent to this a prospective audit was undertaken to assess relevant criteria. Prospective audit data showed a marked increase in the use of patient risk assessments for caries, periodontal disease, oral cancer and non-carious tooth surface loss (NCTSL). Recall intervals were also more often selected based on a patient's risk status and discussed with the patient compared to that observed in the retrospective audit data. This audit was successful as a tool to bring about change in the behaviour of dentists regarding their determination of appropriate recall intervals for patients. Whether that change in behaviour is long-term or transient requires further investigation.

  16. Structural Consistency, Consistency, and Sequential Rationality.

    OpenAIRE

    Kreps, David M; Ramey, Garey

    1987-01-01

    Sequential equilibria comprise consistent beliefs and a sequentially ra tional strategy profile. Consistent beliefs are limits of Bayes ratio nal beliefs for sequences of strategies that approach the equilibrium strategy. Beliefs are structurally consistent if they are rationaliz ed by some single conjecture concerning opponents' strategies. Consis tent beliefs are not necessarily structurally consistent, notwithstan ding a claim by Kreps and Robert Wilson (1982). Moreover, the spirit of stru...

  17. Test of a non-physical barrier consisting of light, sound, and bubble screen to block upstream movement of sea lamprey in an experimental raceway

    Science.gov (United States)

    Miehls, Scott M.; Johnson, Nicholas S.; Hrodey, Pete J.

    2017-01-01

    Control of the invasive Sea Lamprey Petromyzon marinus is critical for management of commercial and recreational fisheries in the Laurentian Great Lakes. Use of physical barriers to block Sea Lampreys from spawning habitat is a major component of the control program. However, the resulting interruption of natural streamflow and blockage of nontarget species present substantial challenges. Development of an effective nonphysical barrier would aid the control of Sea Lampreys by eliminating their access to spawning locations while maintaining natural streamflow. We tested the effect of a nonphysical barrier consisting of strobe lights, low-frequency sound, and a bubble screen on the movement of Sea Lampreys in an experimental raceway designed as a two-choice maze with a single main channel fed by two identical inflow channels (one control and one blocked). Sea Lampreys were more likely to move upstream during trials when the strobe light and low-frequency sound were active compared with control trials and trials using the bubble screen alone. For those Sea Lampreys that did move upstream to the confluence of inflow channels, no combination of stimuli or any individual stimulus significantly influenced the likelihood that Sea Lampreys would enter the blocked inflow channel, enter the control channel, or return downstream.

  18. The Impact of Couple HIV Testing and Counseling on Consistent Condom Use Among Pregnant Women and Their Male Partners: An Observational Study.

    Science.gov (United States)

    Rosenberg, Nora E; Graybill, Lauren A; Wesevich, Austin; McGrath, Nuala; Golin, Carol E; Maman, Suzanne; Bhushan, Nivedita; Tsidya, Mercy; Chimndozi, Limbikani; Hoffman, Irving F; Hosseinipour, Mina C; Miller, William C

    2017-08-01

    In sub-Saharan Africa couple HIV testing and counseling (CHTC) has been associated with substantial increases in safe sex, especially when at least one partner is HIV infected. However, this relationship has not been characterized in an Option B+ context. The study was conducted at the antenatal clinic at Bwaila District Hospital in Lilongwe, Malawi in 2016 under an Option B+ program. Ninety heterosexual couples with an HIV-infected pregnant woman (female-positive couples) and 47 couples with an HIV-uninfected pregnant woman (female-negative couples) were enrolled in an observational study. Each couple member was assessed immediately before and 1 month after CHTC for safe sex (abstinence or consistent condom use in the last month). Generalized estimating equations were used to model change in safe sex before and after CHTC and to compare safe sex between female-positive and female-negative couples. Mean age was 26 years among women and 32 years among men. Before CHTC, safe sex was comparable among female-positive couples (8%) and female-negative couples (2%) [risk ratio (RR): 3.7, 95% confidence interval (CI): 0.5 to 29.8]. One month after CHTC, safe sex was higher among female-positive couples (75%) than among female-negative couples (3%) (RR: 30.0, 95% CI: 4.3 to 207.7). Safe sex increased substantially after CTHC for female-positive couples (RR 9.6, 95% CI: 4.6 to 20.0), but not for female-negative couples (RR: 1.2, 95% CI: 0.1 to 18.7). Engaging pregnant couples in CHTC can have prevention benefits for couples with an HIV-infected pregnant woman, but additional prevention approaches may be needed for couples with an HIV-uninfected pregnant woman.

  19. Tradução, adaptação e avaliação da consistência interna do Eating Behaviours and Body Image Test para uso com crianças do sexo feminino Translation, adaptation and internal consistency evaluation of the Eating Behaviours and Body Image Test for female children

    Directory of Open Access Journals (Sweden)

    Elizângela Moreira Careta Galindo

    2007-02-01

    Full Text Available Este trabalho tem por objetivo traduzir, adaptar e validar o Eating Behaviours and Body Image Test, para uso com crianças de uma cidade do interior do estado de São Paulo. Foram sujeitos do estudo 261 escolares do sexo feminino, na faixa etária de 9 a 12 anos. Por meio da análise fatorial, com rotação varimax avaliou-se a consistência interna do instrumento. Esta análise, realizada com o auxílio do programa Statistical Package for Social Sciences, versão 10.0, revelou dois fatores. Para o instrumento total a consistência interna foi adequada (coeficiente a de Cronbach: 0,89 e para os dois fatores (1 e 2 os valores de a também foram considerados satisfatórios (alfa=0,90 e alfa=0,80, respectivamente, mostrando, assim, que o Eating Behaviours and Body Image Test é útil para uma avaliação precoce, rastreando atitudes indicadoras de possíveis distúrbios no comportamento alimentar. Foram mantidas as características psicométricas do instrumento original.This study aimed to translate, adapt and validate the Eating Bahaviours and Body Image Test, to be used with children in a city in upstate São Paulo. Study subjects were 261 female students aging from 9 to 12 years. The internal consistency of the instrument was evaluated by means of factorial analysis with varimax rotation. This analysis was accomplished through Statistical Package for Social Sciences, version 10.0, revealing two factors. The internal consistency was adequate for the total instrument (Cronbach's alpha=0.89 and a values were also considered satisfactory for the two factors (1 and 2 (alpha=0.90 and alpha=0.80, respectively, which demonstrated that the Eating Bahaviours and Body Image Test is useful for an initial evaluation, tracing symptoms that indicate possible eating behavior disorders. The psychometric characteristics of the original instrument were maintained.

  20. Hybrid method for consistent model of the Pacific absolute plate motion and a test for inter-hotspot motion since 70Ma

    Science.gov (United States)

    Harada, Y.; Wessel, P.; Sterling, A.; Kroenke, L.

    2002-12-01

    Inter-hotspot motion within the Pacific plate is one of the most controversial issues in recent geophysical studies. However, it is a fact that many geophysical and geological data including ages and positions of seamount chains in the Pacific plate can largely be explained by a simple model of absolute motion derived from assumptions of rigid plates and fixed hotspots. Therefore we take the stand that if a model of plate motion can explain the ages and positions of Pacific hotspot tracks, inter-hotspot motion would not be justified. On the other hand, if any discrepancies between the model and observations are found, the inter-hotspot motion may then be estimated from these discrepancies. To make an accurate model of the absolute motion of the Pacific plate, we combined two different approaches: the polygonal finite rotation method (PFRM) by Harada and Hamano (2000) and the hot-spotting technique developed by Wessel and Kroenke (1997). The PFRM can determine accurate positions of finite rotation poles for the Pacific plate if the present positions of hotspots are known. On the other hand, the hot-spotting technique can predict present positions of hotspots if the absolute plate motion is given. Therefore we can undertake iterative calculations using the two methods. This hybrid method enables us to determine accurate finite rotation poles for the Pacific plate solely from geometry of Hawaii, Louisville and Easter(Crough)-Line hotspot tracks from around 70 Ma to present. Information of ages can be independently assigned to the model after the poles and rotation angles are determined. We did not detect any inter-hotspot motion from the geometry of these Pacific hotspot tracks using this method. The Ar-Ar ages of Pacific seamounts including new age data of ODP Leg 197 are used to test the newly determined model of the Pacific plate motion. The ages of Hawaii, Louisville, Easter(Crough)-Line, and Cobb hotspot tracks are quite consistent with each other from 70 Ma to

  1. Effect of Boron Addition on Microstructural Evolution and Room-Temperature Mechanical Properties of Novel Fe66- x CrNiB x Si ( x = 0, 0.25, 0.50 and 0.75 Wt Pct) Advanced High-Strength Steels

    Science.gov (United States)

    Askari-Paykani, Mohsen; Shahverdi, Hamid Reza; Miresmaeili, Reza

    2016-11-01

    In this study, the Vickers hardnesses and room-temperature uniaxial tensile behaviors of four Fe66- x CrNiB x Si ( x = 0 (0B), 0.25 (25B), 0.50 (50B), and 0.75 (75B) wt pct) advanced high-strength steels (AHSSs) in the as-hot-rolled and heat-treated (1373 K (1100 °C)/2 h + 973 K (700 °C)/20 min) conditions were investigated. Microstructural evolution after solidification, hot rolling, heat treatment, and uniaxial tensile tests of 0B, 25B, 50B, and 75B AHSSs was also characterized using field emission gun scanning electron microscopy and X-ray diffraction. The tensile behaviors of the 0B, 25B, 50B, and 75B AHSSs were manifested by an excellent combination of strength and ductility over 34.7 and 47.1 GPa pct, 36.9 and 42.3 GPa pct, 45.9 and 46.4 GPa pct, and 11.9 and 47.8 GPa pct, respectively, arising from microband-induced plasticity in the 0B, 50B, and 75B AHSSs and transformation-induced plasticity in the 25B specimens. All specimens in the as-hot-rolled and heat-treated states showed an austenitic matrix grain. Adding boron to the base alloy (0B) resulted in grain refinement, M2B dispersion, precipitation hardening, and solid solution strengthening, which led to an increase in strength. The results of the present work show promise for automotive applications that require excellent properties and reduced specific weight.

  2. Is cosmology consistent?

    International Nuclear Information System (INIS)

    Wang Xiaomin; Tegmark, Max; Zaldarriaga, Matias

    2002-01-01

    We perform a detailed analysis of the latest cosmic microwave background (CMB) measurements (including BOOMERaNG, DASI, Maxima and CBI), both alone and jointly with other cosmological data sets involving, e.g., galaxy clustering and the Lyman Alpha Forest. We first address the question of whether the CMB data are internally consistent once calibration and beam uncertainties are taken into account, performing a series of statistical tests. With a few minor caveats, our answer is yes, and we compress all data into a single set of 24 bandpowers with associated covariance matrix and window functions. We then compute joint constraints on the 11 parameters of the 'standard' adiabatic inflationary cosmological model. Our best fit model passes a series of physical consistency checks and agrees with essentially all currently available cosmological data. In addition to sharp constraints on the cosmic matter budget in good agreement with those of the BOOMERaNG, DASI and Maxima teams, we obtain a heaviest neutrino mass range 0.04-4.2 eV and the sharpest constraints to date on gravity waves which (together with preference for a slight red-tilt) favor 'small-field' inflation models

  3. The test of variables of attention (TOVA): Internal consistency (Q1 vs. Q2 and Q3 vs. Q4) in children with Attention Deficit/Hyperactivity Disorder (ADHD)

    Science.gov (United States)

    The internal consistency of the Test of Variables of Attention (TOVA) was examined in a cohort of 6- to 12-year-old children (N = 63) strictly diagnosed with ADHD. The internal consistency of errors of omission (OMM), errors of commission (COM), response time (RT), and response time variability (RTV...

  4. Post-test analysis of the ROSA/LSTF and PKL counterpart test

    Energy Technology Data Exchange (ETDEWEB)

    Carlos, S., E-mail: scarlos@iqn.upv.es [Departament d’Enginyeria Química i Nuclear, Universitat Politècnica de València, Camí de Vera, 14, València Spain (Spain); Querol, A., E-mail: anquevi@iqn.upv.es [Departament d’Enginyeria Química i Nuclear, Universitat Politècnica de València, Camí de Vera, 14, València Spain (Spain); Instituto de Seguridad Industrial, Radiofísica y Medioambiental, Universitat Politècnica de València, Camí de Vera, 14, València (Spain); Gallardo, S., E-mail: sergalbe@iqn.upv.es [Departament d’Enginyeria Química i Nuclear, Universitat Politècnica de València, Camí de Vera, 14, València Spain (Spain); Instituto de Seguridad Industrial, Radiofísica y Medioambiental, Universitat Politècnica de València, Camí de Vera, 14, València (Spain); Sanchez-Saez, F., E-mail: frasansa@etsii.upv.es [Departament d’Enginyeria Química i Nuclear, Universitat Politècnica de València, Camí de Vera, 14, València Spain (Spain); and others

    2016-02-15

    Highlights: • TRACE modelization for PKL and ROSA/LSTF installations. • Secondary-side depressurization as accident management action. • CET vs PCT relation. • Analysis of differences in the vessel models. - Abstract: Experimental facilities are scaled models of commercial nuclear power plants, and are of great importance to improve nuclear power plants safety. Thus, the results obtained in the experiments undertaken in such facilities are essential to develop and improve the models implemented in the thermal-hydraulic codes, which are used in safety analysis. The experiments and inter-comparisons of the simulated results are usually performed in the frame of international programmes in which different groups of several countries simulate the behaviour of the plant under the accidental conditions established, using different codes and models. The results obtained are compared and studied to improve the knowledge on codes performance and nuclear safety. Thus, the Nuclear Energy Agency (NEA), in the nuclear safety work area, auspices several programmes which involve experiments in different experimental facilities. Among the experiments proposed in NEA programmes, one on them consisted of performing a counterpart test between ROSA/LSTF and PKL facilities, with the main objective of determining the effectiveness of late accident management actions in a small break loss of coolant accident (SBLOCA). This study was proposed as a result of the conclusion obtained by the NEA Working Group on the Analysis and Management of Accidents, which analyzed different installations and observed differences in the measurements of core exit temperature (CET) and maximum peak cladding temperature (PCT). In particular, the transient consists of a small break loss of coolant accident (SBLOCA) in a hot leg with additional failure of safety systems but with accident management measures (AM), consisting of a fast secondary-side depressurization, activated by the CET. The paper

  5. Post-test analysis of the ROSA/LSTF and PKL counterpart test

    International Nuclear Information System (INIS)

    Carlos, S.; Querol, A.; Gallardo, S.; Sanchez-Saez, F.

    2016-01-01

    Highlights: • TRACE modelization for PKL and ROSA/LSTF installations. • Secondary-side depressurization as accident management action. • CET vs PCT relation. • Analysis of differences in the vessel models. - Abstract: Experimental facilities are scaled models of commercial nuclear power plants, and are of great importance to improve nuclear power plants safety. Thus, the results obtained in the experiments undertaken in such facilities are essential to develop and improve the models implemented in the thermal-hydraulic codes, which are used in safety analysis. The experiments and inter-comparisons of the simulated results are usually performed in the frame of international programmes in which different groups of several countries simulate the behaviour of the plant under the accidental conditions established, using different codes and models. The results obtained are compared and studied to improve the knowledge on codes performance and nuclear safety. Thus, the Nuclear Energy Agency (NEA), in the nuclear safety work area, auspices several programmes which involve experiments in different experimental facilities. Among the experiments proposed in NEA programmes, one on them consisted of performing a counterpart test between ROSA/LSTF and PKL facilities, with the main objective of determining the effectiveness of late accident management actions in a small break loss of coolant accident (SBLOCA). This study was proposed as a result of the conclusion obtained by the NEA Working Group on the Analysis and Management of Accidents, which analyzed different installations and observed differences in the measurements of core exit temperature (CET) and maximum peak cladding temperature (PCT). In particular, the transient consists of a small break loss of coolant accident (SBLOCA) in a hot leg with additional failure of safety systems but with accident management measures (AM), consisting of a fast secondary-side depressurization, activated by the CET. The paper

  6. An Improved Cognitive Model of the Iowa and Soochow Gambling Tasks With Regard to Model Fitting Performance and Tests of Parameter Consistency

    Directory of Open Access Journals (Sweden)

    Junyi eDai

    2015-03-01

    Full Text Available The Iowa Gambling Task (IGT and the Soochow Gambling Task (SGT are two experience-based risky decision-making tasks for examining decision-making deficits in clinical populations. Several cognitive models, including the expectancy-valence learning model (EVL and the prospect valence learning model (PVL, have been developed to disentangle the motivational, cognitive, and response processes underlying the explicit choices in these tasks. The purpose of the current study was to develop an improved model that can fit empirical data better than the EVL and PVL models and, in addition, produce more consistent parameter estimates across the IGT and SGT. Twenty-six opiate users (mean age 34.23; SD 8.79 and 27 control participants (mean age 35; SD 10.44 completed both tasks. Eighteen cognitive models varying in evaluation, updating, and choice rules were fit to individual data and their performances were compared to that of a statistical baseline model to find a best fitting model. The results showed that the model combining the prospect utility function treating gains and losses separately, the decay-reinforcement updating rule, and the trial-independent choice rule performed the best in both tasks. Furthermore, the winning model produced more consistent individual parameter estimates across the two tasks than any of the other models.

  7. Choice consistency and preference stability in test-retests of discrete choice experiment and open-ended willingness to pay elicitation formats

    NARCIS (Netherlands)

    Brouwer, R.; Logar, I.; Sheremet, O.I.

    2017-01-01

    This study tests the temporal stability of preferences, choices and willingness to pay (WTP) values using both discrete choice experiment (DCE) and open-ended (OE) WTP elicitation formats. The same sample is surveyed three times over the course of two years using each time the same choice sets.

  8. An examination of the interparticle contact area during sintering of W-0.3 wt pct Co

    International Nuclear Information System (INIS)

    Mitlin, D.; German, R.M.

    1998-01-01

    As a powder compact sinters, its microstructure evolves. One way to quantify the scale of the microstructure is to consider the interparticle contact area. This study examines two known models for calculating the interparticle contact area: the classic two-sphere model and the Voronoi cell model. Both models have particular assumptions about the microstructure that make them not applicable for treating densification to near full density with concurrent grain growth. The classic two-sphere model assumes a regular packing of particles and a perfectly spherical particle geometry and neglects an increasing particle coordination number with sintering. The Voronoi cell model assumes that the scale of the microstructure remains constant; i.e., as long as the compact is densifying, grain growth does not occur. The authors propose a modified Voronoi cell that accounts for an increasing grain size, making it applicable to a general case where grain growth occurs during sintering. The three models are compared to the interparticle contact area data, obtained by stereology techniques, for W-0.3 wt pct Co sintered from green state to near full density. The original Voronoi cell model fits the data only at low temperatures, before the onset of grain growth. Below approximately 90 pct relative density, the two-sphere model with an assumed coordination number of six (coordination number in a green compact) and the modified Voronoi cell model provide a good fit to the data. At higher densities, both models overestimate the interparticle contact area

  9. Consistent Condom Use by Female Sex Workers in Kolkata, India: Testing Theories of Economic Insecurity, Behavior Change, Life Course Vulnerability and Empowerment.

    Science.gov (United States)

    Fehrenbacher, Anne E; Chowdhury, Debasish; Ghose, Toorjo; Swendeman, Dallas

    2016-10-01

    Consistent condom use (CCU) is the primary HIV/STI prevention option available to sex workers globally but may be undermined by economic insecurity, life-course vulnerabilities, behavioral factors, disempowerment, or lack of effective interventions. This study examines predictors of CCU in a random household survey of brothel-based female sex workers (n = 200) in two neighborhoods served by Durbar (the Sonagachi Project) in Kolkata, India. Multivariate logistic regression analyses indicated that CCU was significantly associated with perceived HIV risk, community mobilization participation, working more days in sex work, and higher proportion of occasional clients to regular clients. Exploratory analyses stratifying by economic insecurity indicators (i.e., debt, savings, income, housing security) indicate that perceived HIV risk and community mobilization were only associated with CCU for economically secure FSW. Interventions with FSW must prioritize economic security and access to social protections as economic insecurity may undermine the efficacy of more direct condom use intervention strategies.

  10. Are There Consistent Grazing Indicators in Drylands? Testing Plant Functional Types of Various Complexity in South Africa’s Grassland and Savanna Biomes

    Science.gov (United States)

    Linstädter, Anja; Schellberg, Jürgen; Brüser, Katharina; Moreno García, Cristian A.; Oomen, Roelof J.; du Preez, Chris C.; Ruppert, Jan C.; Ewert, Frank

    2014-01-01

    Despite our growing knowledge on plants’ functional responses to grazing, there is no consensus if an optimum level of functional aggregation exists for detecting grazing effects in drylands. With a comparative approach we searched for plant functional types (PFTs) with a consistent response to grazing across two areas differing in climatic aridity, situated in South Africa’s grassland and savanna biomes. We aggregated herbaceous species into PFTs, using hierarchical combinations of traits (from single- to three-trait PFTs). Traits relate to life history, growth form and leaf width. We first confirmed that soil and grazing gradients were largely independent from each other, and then searched in each biome for PFTs with a sensitive response to grazing, avoiding confounding with soil conditions. We found no response consistency, but biome-specific optimum aggregation levels. Three-trait PFTs (e.g. broad-leaved perennial grasses) and two-trait PFTs (e.g. perennial grasses) performed best as indicators of grazing effects in the semi-arid grassland and in the arid savanna biome, respectively. Some PFTs increased with grazing pressure in the grassland, but decreased in the savanna. We applied biome-specific grazing indicators to evaluate if differences in grazing management related to land tenure (communal versus freehold) had effects on vegetation. Tenure effects were small, which we mainly attributed to large variability in grazing pressure across farms. We conclude that the striking lack of generalizable PFT responses to grazing is due to a convergence of aridity and grazing effects, and unlikely to be overcome by more refined classification approaches. Hence, PFTs with an opposite response to grazing in the two biomes rather have a unimodal response along a gradient of additive forces of aridity and grazing. The study advocates for hierarchical trait combinations to identify localized indicator sets for grazing effects. Its methodological approach may also be

  11. Are there consistent grazing indicators in Drylands? Testing plant functional types of various complexity in South Africa's Grassland and Savanna Biomes.

    Directory of Open Access Journals (Sweden)

    Anja Linstädter

    Full Text Available Despite our growing knowledge on plants' functional responses to grazing, there is no consensus if an optimum level of functional aggregation exists for detecting grazing effects in drylands. With a comparative approach we searched for plant functional types (PFTs with a consistent response to grazing across two areas differing in climatic aridity, situated in South Africa's grassland and savanna biomes. We aggregated herbaceous species into PFTs, using hierarchical combinations of traits (from single- to three-trait PFTs. Traits relate to life history, growth form and leaf width. We first confirmed that soil and grazing gradients were largely independent from each other, and then searched in each biome for PFTs with a sensitive response to grazing, avoiding confounding with soil conditions. We found no response consistency, but biome-specific optimum aggregation levels. Three-trait PFTs (e.g. broad-leaved perennial grasses and two-trait PFTs (e.g. perennial grasses performed best as indicators of grazing effects in the semi-arid grassland and in the arid savanna biome, respectively. Some PFTs increased with grazing pressure in the grassland, but decreased in the savanna. We applied biome-specific grazing indicators to evaluate if differences in grazing management related to land tenure (communal versus freehold had effects on vegetation. Tenure effects were small, which we mainly attributed to large variability in grazing pressure across farms. We conclude that the striking lack of generalizable PFT responses to grazing is due to a convergence of aridity and grazing effects, and unlikely to be overcome by more refined classification approaches. Hence, PFTs with an opposite response to grazing in the two biomes rather have a unimodal response along a gradient of additive forces of aridity and grazing. The study advocates for hierarchical trait combinations to identify localized indicator sets for grazing effects. Its methodological approach may

  12. Are there consistent grazing indicators in Drylands? Testing plant functional types of various complexity in South Africa's Grassland and Savanna Biomes.

    Science.gov (United States)

    Linstädter, Anja; Schellberg, Jürgen; Brüser, Katharina; Moreno García, Cristian A; Oomen, Roelof J; du Preez, Chris C; Ruppert, Jan C; Ewert, Frank

    2014-01-01

    Despite our growing knowledge on plants' functional responses to grazing, there is no consensus if an optimum level of functional aggregation exists for detecting grazing effects in drylands. With a comparative approach we searched for plant functional types (PFTs) with a consistent response to grazing across two areas differing in climatic aridity, situated in South Africa's grassland and savanna biomes. We aggregated herbaceous species into PFTs, using hierarchical combinations of traits (from single- to three-trait PFTs). Traits relate to life history, growth form and leaf width. We first confirmed that soil and grazing gradients were largely independent from each other, and then searched in each biome for PFTs with a sensitive response to grazing, avoiding confounding with soil conditions. We found no response consistency, but biome-specific optimum aggregation levels. Three-trait PFTs (e.g. broad-leaved perennial grasses) and two-trait PFTs (e.g. perennial grasses) performed best as indicators of grazing effects in the semi-arid grassland and in the arid savanna biome, respectively. Some PFTs increased with grazing pressure in the grassland, but decreased in the savanna. We applied biome-specific grazing indicators to evaluate if differences in grazing management related to land tenure (communal versus freehold) had effects on vegetation. Tenure effects were small, which we mainly attributed to large variability in grazing pressure across farms. We conclude that the striking lack of generalizable PFT responses to grazing is due to a convergence of aridity and grazing effects, and unlikely to be overcome by more refined classification approaches. Hence, PFTs with an opposite response to grazing in the two biomes rather have a unimodal response along a gradient of additive forces of aridity and grazing. The study advocates for hierarchical trait combinations to identify localized indicator sets for grazing effects. Its methodological approach may also be useful

  13. Low Temperature Waste Immobilization Testing Vol. I

    Energy Technology Data Exchange (ETDEWEB)

    Russell, Renee L.; Schweiger, Michael J.; Westsik, Joseph H.; Hrma, Pavel R.; Smith, D. E.; Gallegos, Autumn B.; Telander, Monty R.; Pitman, Stan G.

    2006-09-14

    The Pacific Northwest National Laboratory (PNNL) is evaluating low-temperature technologies to immobilize mixed radioactive and hazardous waste. Three waste forms—alkali-aluminosilicate hydroceramic cement, “Ceramicrete” phosphate-bonded ceramic, and “DuraLith” alkali-aluminosilicate geopolymer—were selected through a competitive solicitation for fabrication and characterization of waste-form properties. The three contractors prepared their respective waste forms using simulants of a Hanford secondary waste and Idaho sodium bearing waste provided by PNNL and characterized their waste forms with respect to the Toxicity Characteristic Leaching Procedure (TCLP) and compressive strength. The contractors sent specimens to PNNL, and PNNL then conducted durability (American National Standards Institute/American Nuclear Society [ANSI/ANS] 16.1 Leachability Index [LI] and modified Product Consistency Test [PCT]) and compressive strength testing (both irradiated and as-received samples). This report presents the results of these characterization tests.

  14. Consistent model driven architecture

    Science.gov (United States)

    Niepostyn, Stanisław J.

    2015-09-01

    The goal of the MDA is to produce software systems from abstract models in a way where human interaction is restricted to a minimum. These abstract models are based on the UML language. However, the semantics of UML models is defined in a natural language. Subsequently the verification of consistency of these diagrams is needed in order to identify errors in requirements at the early stage of the development process. The verification of consistency is difficult due to a semi-formal nature of UML diagrams. We propose automatic verification of consistency of the series of UML diagrams originating from abstract models implemented with our consistency rules. This Consistent Model Driven Architecture approach enables us to generate automatically complete workflow applications from consistent and complete models developed from abstract models (e.g. Business Context Diagram). Therefore, our method can be used to check practicability (feasibility) of software architecture models.

  15. Bitcoin Meets Strong Consistency

    OpenAIRE

    Decker, Christian; Seidel, Jochen; Wattenhofer, Roger

    2014-01-01

    The Bitcoin system only provides eventual consistency. For everyday life, the time to confirm a Bitcoin transaction is prohibitively slow. In this paper we propose a new system, built on the Bitcoin blockchain, which enables strong consistency. Our system, PeerCensus, acts as a certification authority, manages peer identities in a peer-to-peer network, and ultimately enhances Bitcoin and similar systems with strong consistency. Our extensive analysis shows that PeerCensus is in a secure state...

  16. Consistent classical supergravity theories

    International Nuclear Information System (INIS)

    Muller, M.

    1989-01-01

    This book offers a presentation of both conformal and Poincare supergravity. The consistent four-dimensional supergravity theories are classified. The formulae needed for further modelling are included

  17. Sulfur Solubility Testing and Characterization of Hanford LAW Phase 2, Inner Layer Matrix Glasses

    Energy Technology Data Exchange (ETDEWEB)

    Fox, K. M. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Edwards, T. B. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Caldwell, M. E. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Riley, W. T. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL)

    2017-11-27

    In this report, the Savannah River National Laboratory (SRNL) provides chemical analyses and Product Consistency Test (PCT) results for a series of simulated low activity waste (LAW) glass compositions. A procedure developed at the Pacific Northwest National Laboratory (PNNL) for producing sulfur saturated melts (SSMs) was carried out at both SRNL and PNNL to fabricate the glasses characterized in this report. This method includes triplicate melting steps with excess sodium sulfate, followed by grinding and washing to remove unincorporated sulfur salts. The wash solutions were also analyzed as part of this study. These data will be used in the development of improved sulfur solubility models for LAW glass.

  18. Assessment of the Effects on PCT Evaluation of Enhanced Fuel Model Facilitated by Coupling the MARS Code with the FRAPTRAN Code

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Hyong Chol; Lee, Young Jin; Han, Sam Hee [NSE Technology Inc., Daejeon (Korea, Republic of)

    2016-10-15

    The principal objectives of the two safety criteria, peak cladding temperature (PCT) and total oxidation limits, are to ensure that the fuel rod claddings remain sufficiently ductile so that they do not crack and fragment during a LOCA. Another important purpose of the PCT limit is to ensure that the fuel cladding does not enter the regime of runaway oxidation and uncontrollable heat-up. However, even when the PCT limit is satisfied, it is known that cladding failures may still occur in a certain percentage of the fuel rods during a LOCA. This is largely because a 100% fuel failure is assumed for the radiological consequence analysis in the US regulatory practices. In this study, we analyze the effects of cladding failure and other fuel model features on PCT during a LOCA using the MARS-FRAPTRAN coupled code. MARS code has been coupled with FRAPTRAN code to extend fuel modeling capability. The coupling allows feedback of FRAPTRAN results in real time. Because of the significant impact of fuel models on key safety parameters such as PCT, detailed and accurate fuel models should be employed when evaluating PCT in LOCA analysis. It is noteworthy that the ECCS evaluation models laid out in the Appendix K to 10CFR50 require a provision for predicting cladding swelling and rupture and require to assume that the inside of the cladding react with steam after the rupture. The metal-water reaction energy can have significantly large effect on the reflood PCT, especially when fuel failure occurs. Effects of applying an advanced fuel model on the PCT evaluation can be clearly seen when comparing the MARS and the FRAPTRAN results in both the one-way calculation and the feedback calculation. As long as MARS and FRAPTRAN are used respectively in the ranges where they have been validated, the coupled calculation results are expected to be valid and to reveal various aspects of phenomena which have not been discovered in previous uncoupled calculations by MARS or FRAPTRAN.

  19. Consistency of orthodox gravity

    Energy Technology Data Exchange (ETDEWEB)

    Bellucci, S. [INFN, Frascati (Italy). Laboratori Nazionali di Frascati; Shiekh, A. [International Centre for Theoretical Physics, Trieste (Italy)

    1997-01-01

    A recent proposal for quantizing gravity is investigated for self consistency. The existence of a fixed-point all-order solution is found, corresponding to a consistent quantum gravity. A criterion to unify couplings is suggested, by invoking an application of their argument to more complex systems.

  20. Quasiparticles and thermodynamical consistency

    International Nuclear Information System (INIS)

    Shanenko, A.A.; Biro, T.S.; Toneev, V.D.

    2003-01-01

    A brief and simple introduction into the problem of the thermodynamical consistency is given. The thermodynamical consistency relations, which should be taken into account under constructing a quasiparticle model, are found in a general manner from the finite-temperature extension of the Hellmann-Feynman theorem. Restrictions following from these relations are illustrated by simple physical examples. (author)

  1. Testing of Large-Scale ICV Glasses with Hanford LAW Simulant

    Energy Technology Data Exchange (ETDEWEB)

    Hrma, Pavel R.; Kim, Dong-Sang; Vienna, John D.; Matyas, Josef; Smith, Donald E.; Schweiger, Michael J.; Yeager, John D.

    2005-03-01

    Preliminary glass compositions for immobilizing Hanford low-activity waste (LAW) by the in-container vitrification (ICV) process were initially fabricated at crucible- and engineering-scale, including simulants and actual (radioactive) LAW. Glasses were characterized for vapor hydration test (VHT) and product consistency test (PCT) responses and crystallinity (both quenched and slow-cooled samples). Selected glasses were tested for toxicity characteristic leach procedure (TCLP) responses, viscosity, and electrical conductivity. This testing showed that glasses with LAW loading of 20 mass% can be made readily and meet all product constraints by a far margin. Glasses with over 22 mass% Na2O can be made to meet all other product quality and process constraints. Large-scale testing was performed at the AMEC, Geomelt Division facility in Richland. Three tests were conducted using simulated LAW with increasing loadings of 12, 17, and 20 mass% Na2O. Glass samples were taken from the test products in a manner to represent the full expected range of product performance. These samples were characterized for composition, density, crystalline and non-crystalline phase assemblage, and durability using the VHT, PCT, and TCLP tests. The results, presented in this report, show that the AMEC ICV product with meets all waste form requirements with a large margin. These results provide strong evidence that the Hanford LAW can be successfully vitrified by the ICV technology and can meet all the constraints related to product quality. The economic feasibility of the ICV technology can be further enhanced by subsequent optimization.

  2. Evaluation and validation of a single-dilution potency assay based upon serology of vaccines containing diphtheria toxoid: analysis for consistency in production and testing at the laboratory of the Control of Biological Products of the RIVM

    NARCIS (Netherlands)

    Akkermans AM; Hendriksen CFM; Marsman FR; de Jong WH; van de Donk HJM

    1993-01-01

    A single-dilution assay can be a valid procedure to demonstrate that a product exceeds the minimal requirement given for potency provided that consistency in production and testing has been proven. Information is presented justifying the use of a single dilution assay based upon quantitative

  3. The diagnostic value of CRP, IL-8, PCT, and sTREM-1 in the detection of bacterial infections in pediatric oncology patients with febrile neutropenia

    NARCIS (Netherlands)

    Miedema, Karin G. E.; de Bont, Eveline S. J. M.; Elferink, Rob F. M. Oude; van Vliet, Michel J.; Nijhuis, Claudi S. M. Oude; Kamps, Willem A.; Tissing, Wim J. E.

    2011-01-01

    In this study, we evaluated C-reactive protein (CRP), interleukin (IL)-8, procalcitonin (PCT), and soluble triggering receptor expressed on myeloid cells-1 (sTREM-1) as predictors for bacterial infection in febrile neutropenia, plus their usefulness in febrile neutropenia during chemotherapy-induced

  4. A PCT algorithm for discontinuation of antibiotic therapy is a cost-effective way to reduce antibiotic exposure in adult intensive care patients with sepsis

    NARCIS (Netherlands)

    Kip, Michelle Maria Aleida; Kusters, Ron; IJzerman, Maarten Joost; Steuten, Lotte Maria Gertruda

    2015-01-01

    Objective: Procalcitonin (PCT) is a specific marker for differentiating bacterial from non-infective causes of inflammation. It can be used to guide initiation and duration of antibiotic therapy in intensive care unit (ICU) patients with suspected sepsis, and might reduce the duration of hospital

  5. Expression of MMP-2 and TIMP-1 in cerebrospinal fluid and the correlation with dynamic changes of serum PCT in neonatal purulent meningitis

    Science.gov (United States)

    Chen, Huilan; Wu, Fei; Fu, Rong; Feng, Xiangchun

    2018-01-01

    Matrix metalloproteinase 2 (MMP-2) and tissue inhibitor of metalloproteinase-1 (TIMP-1) levels in cerebrospinal fluid of pediatric patients with neonatal purulent meningitis were observed to analyze changes in serum procalcitonin (PCT) and the correlation among the three factors (MMP-2, TIMP-1 and PCT). Sixty pediatric patients with neonatal purulent meningitis from April 2015 to December 2016 were enrolled as the purulent meningitis group and 60 pediatric patients with viral encephalitis treated during the same period were enrolled as the viral encephalitis group. Additionally, 60 healthy newborns who underwent physical examinations in our hospital during the same period were enrolled as the control group. The levels of MMP-2 were 136.73±25.42 ng/ml in the purulent meningitis group, 45.32±6.57 ng/ml in the viral encephalitis group and 1.32±0.51 ng/ml in the control group and the differences between the three groups were statistically significant (F=15.052, pfluid were 374.55±36.04 ng/ml in the purulent meningitis group, 176.61±21.06 ng/ml in the viral encephalitis group and 7.72±2.44 ng/ml in the control group. The serum levels of PCT were 14.56±2.21 ng/ml in the purulent meningitis group, 9.04±1.17 ng/ml in the viral encephalitis group and 0.38±0.14 ng/ml in the control group. The level of MMP-2 in cerebrospinal fluid of pediatric patients in the purulent meningitis group was positively correlated with the level of serum PCT (r=0.582, pfluid of pediatric patients in the viral encephalitis group was positively correlated with the level of serum PCT (r=0.635, p<0.05). In conclusion, MMP-2 and TIMP-1 were positively correlated with the levels of serum PCT, suggesting that MMP-2, TIMP-1 and PCT were involved in the occurrence and development of neonatal purulent meningitis. PMID:29399119

  6. Heat treatment for improvement in lower temperature mechanical properties of 0.40 pct C-Cr-Mo ultrahigh strength steel

    Science.gov (United States)

    Tomita, Yoshiyuki; Okabayashi, Kunio

    1983-11-01

    In the previous paper, it was reported that isothermal heat treatment of a commercial Japanese 0.40 pct C-Ni-Cr-Mo ultrahigh strength steel (AISI 4340 type) at 593 K for a short time followed by water quenching, in which a mixed structure of 25 vol pct lower bainite and 75 vol pct martensite is produced, results in the improvement of low temperature mechanical properties (287 to 123 K). The purpose of this paper is to study whether above new heat treatment will still be effective in commercial practice for improving low temperature mechanical properties of the ultrahigh strength steel when applied to a commercial Japanese 0.40 pct C-Cr-Mo ultrahigh strength steel which is economical because it lacks the expensive nickel component (AISI 4140 type). At and above 203 K this new heat treatment, as compared with the conventional 1133 K direct water quenching treatment, significantly improved the strength, tensile ductility, and notch toughness of the 0.40 pct C-Cr-Mo ultrahigh strength steel. At and above 203 K the new heat treatment also produced superior fracture ductility and notch toughness results at similar strength levels as compared to those obtained by using γ α' repetitive heat treatment for the same steel. However, the new heat treatment remarkably decreased fracture ductility and notch toughness of the 0.40 pct C-Cr-Mo ultrahigh strength steel below 203 K, and thus no significant improvement in the mechanical properties was noticeable as compared with the properties produced by the conventional 1133 K direct water quenching treatment and the γ α' repetitive heat treatment. This contrasts with the fact that the new heat treatment, as compared with the conventional 1133 K direct water quenching treatment and the γ α' repetitive heat treatment, dramatically improved the notch toughness of the 0.40 pct C-Ni-Cr-Mo ultrahigh strength steel, providing a better combination of strength and ductility throughout the 287 to 123 K temperature range. The difference

  7. Stress Corrosion Cracking Behavior of Multipass TIG-Welded AA2219 Aluminum Alloy in 3.5 wt pct NaCl Solution

    Science.gov (United States)

    Venugopal, A.; Sreekumar, K.; Raja, V. S.

    2012-09-01

    The stress corrosion cracking (SCC) behavior of the AA2219 aluminum alloy in the single-pass (SP) and multipass (MP) welded conditions was examined and compared with that of the base metal (BM) in 3.5 wt pct NaCl solution using a slow-strain-rate technique (SSRT). The reduction in ductility was used as a parameter to evaluate the SCC susceptibility of both the BM and welded joints. The results showed that the ductility ratio ( ɛ NaCl/( ɛ air) was 0.97 and 0.96, respectively, for the BM and MP welded joint, and the same was marginally reduced to 0.9 for the SP welded joint. The fractographic examination of the failed samples revealed a typical ductile cracking morphology for all the base and welded joints, indicating the good environmental cracking resistance of this alloy under all welded conditions. To understand the decrease in the ductility of the SP welded joint, preexposure SSRT followed by microstructural observations were made, which showed that the decrease in ductility ratio of the SP welded joint was caused by the electrochemical pitting that assisted the nucleation of cracks in the form of corrosion induced mechanical cracking rather than true SCC failure of the alloy. The microstructural examination and polarization tests demonstrated a clear grain boundary (GB) sensitization of the PMZ, resulting in severe galvanic corrosion of the SP weld joint, which initiated the necessary conditions for the localized corrosion and cracking along the PMZ. The absence of PMZ and a refined fusion zone (FZ) structure because of the lesser heat input and postweld heating effect improved the galvanic corrosion resistance of the MP welded joint greatly, and thus, failure occurred along the FZ.

  8. Consistency in PERT problems

    OpenAIRE

    Bergantiños, Gustavo; Valencia-Toledo, Alfredo; Vidal-Puga, Juan

    2016-01-01

    The program evaluation review technique (PERT) is a tool used to schedule and coordinate activities in a complex project. In assigning the cost of a potential delay, we characterize the Shapley rule as the only rule that satisfies consistency and other desirable properties.

  9. Consistency argued students of fluid

    Science.gov (United States)

    Viyanti; Cari; Suparmi; Winarti; Slamet Budiarti, Indah; Handika, Jeffry; Widyastuti, Fatma

    2017-01-01

    Problem solving for physics concepts through consistency arguments can improve thinking skills of students and it is an important thing in science. The study aims to assess the consistency of the material Fluid student argmentation. The population of this study are College students PGRI Madiun, UIN Sunan Kalijaga Yogyakarta and Lampung University. Samples using cluster random sampling, 145 samples obtained by the number of students. The study used a descriptive survey method. Data obtained through multiple-choice test and interview reasoned. Problem fluid modified from [9] and [1]. The results of the study gained an average consistency argmentation for the right consistency, consistency is wrong, and inconsistent respectively 4.85%; 29.93%; and 65.23%. Data from the study have an impact on the lack of understanding of the fluid material which is ideally in full consistency argued affect the expansion of understanding of the concept. The results of the study as a reference in making improvements in future studies is to obtain a positive change in the consistency of argumentations.

  10. Effect of Repair Welding on Electrochemical Corrosion and Stress Corrosion Cracking Behavior of TIG Welded AA2219 Aluminum Alloy in 3.5 Wt Pct NaCl Solution

    Science.gov (United States)

    Venugopal, A.; Sreekumar, K.; Raja, V. S.

    2010-12-01

    The stress corrosion cracking (SCC) behavior of AA2219 aluminum alloy in the as-welded (AW) and repair-welded (RW) conditions was examined and compared with that of the base metal (BM) in 3.5 wt pct NaCl solution using the slow strain rate technique (SSRT). The reduction in ductility was used as a parameter to evaluate the SCC susceptibility of both BM and welded joints. The results show that the ductility ratio ( ɛ NaCl/( ɛ air)) of the BM was close to one (0.97) and reduced to 0.9 for the AW joint. This value further reduced to 0.77 after carrying out one repair welding operation. However, the RW specimen exhibited higher ductility than the single-weld specimens even in 3.5 wt pct NaCl solution. SSRT results obtained using pre-exposed samples followed by post-test metallographic observations clearly showed localized pitting corrosion along the partially melted zone (PMZ), signifying that the reduction in ductility ratio of both the AW and RW joints was more due to mechanical overload failure, caused by the localized corrosion and a consequent reduction in specimen thickness, than due to SCC. Also, the RW joint exhibited higher ductility than the AW joint both in air and the environment, although SCC index (SI) for the former is lower than that of the latter. Fractographic examination of the failed samples, in general, revealed a typical ductile cracking morphology for all the base and welded joints, indicating the good environmental cracking resistance of this alloy. Microstructural examination and polarization tests further demonstrate grain boundary melting along the PMZ, and that provided the necessary electrochemical condition for the preferential cracking on that zone of the weldment.

  11. The effect of compositional parameters on the TCLP and PCT durability of environmental glasses

    International Nuclear Information System (INIS)

    Resce, J.L.; Overcamp, T.J.

    1995-01-01

    The relationship between glass composition and the chemical durability of environmental waste glass is very important for both the development of glass formulations and the prediction of glass durability for process control. The development of such a model is extremely difficult for several reasons. Firstly, chemical durability is dependent upon the type of leach test employed; the leach tests themselves being only crude approximations of actual environmental conditions or long term behavior. Secondly, devitrification or crystallinity can also play a major role in durability, but is much more difficult to quantify. Lastly, the development of any one model for all glass types is impractical because of the wide variety of wastestreams, the heterogeneity of the wastestreams, and the large variety of components within each wastestream. Several ongoing efforts have been directed toward this goal, but as yet, no model has been proven acceptable

  12. Reporting consistently on CSR

    DEFF Research Database (Denmark)

    Thomsen, Christa; Nielsen, Anne Ellerup

    2006-01-01

    This chapter first outlines theory and literature on CSR and Stakeholder Relations focusing on the different perspectives and the contextual and dynamic character of the CSR concept. CSR reporting challenges are discussed and a model of analysis is proposed. Next, our paper presents the results...... of a case study showing that companies use different and not necessarily consistent strategies for reporting on CSR. Finally, the implications for managerial practice are discussed. The chapter concludes by highlighting the value and awareness of the discourse and the discourse types adopted...... in the reporting material. By implementing consistent discourse strategies that interact according to a well-defined pattern or order, it is possible to communicate a strong social commitment on the one hand, and to take into consideration the expectations of the shareholders and the other stakeholders...

  13. Geometrically Consistent Mesh Modification

    KAUST Repository

    Bonito, A.

    2010-01-01

    A new paradigm of adaptivity is to execute refinement, coarsening, and smoothing of meshes on manifolds with incomplete information about their geometry and yet preserve position and curvature accuracy. We refer to this collectively as geometrically consistent (GC) mesh modification. We discuss the concept of discrete GC, show the failure of naive approaches, and propose and analyze a simple algorithm that is GC and accuracy preserving. © 2010 Society for Industrial and Applied Mathematics.

  14. Consistent force fields for saccharides

    DEFF Research Database (Denmark)

    Rasmussen, Kjeld

    1999-01-01

    Consistent force fields for carbohydrates were hitherto developed by extensive optimization ofpotential energy function parameters on experimental data and on ab initio results. A wide range of experimental data is used: internal structures obtained from gas phase electron diffraction and from x......-anomeric effects are accounted for without addition of specific terms. The work is done in the framework of the Consistent Force Field which originatedin Israel and was further developed in Denmark. The actual methods and strategies employed havebeen described previously. Extensive testing of the force field...

  15. Glass consistency and glass performance

    International Nuclear Information System (INIS)

    Plodinec, M.J.; Ramsey, W.G.

    1994-01-01

    Glass produced by the Defense Waste Processing Facility (DWPF) will have to consistently be more durable than a benchmark glass (evaluated using a short-term leach test), with high confidence. The DWPF has developed a Glass Product Control Program to comply with this specification. However, it is not clear what relevance product consistency has on long-term glass performance. In this report, the authors show that DWPF glass, produced in compliance with this specification, can be expected to effectively limit the release of soluble radionuclides to natural environments. However, the release of insoluble radionuclides to the environment will be limited by their solubility, and not glass durability

  16. The Rucio Consistency Service

    CERN Document Server

    Serfon, Cedric; The ATLAS collaboration

    2016-01-01

    One of the biggest challenge with Large scale data management system is to ensure the consistency between the global file catalog and what is physically on all storage elements. To tackle this issue, the Rucio software which is used by the ATLAS Distributed Data Management system has been extended to automatically handle lost or unregistered files (aka Dark Data). This system automatically detects these inconsistencies and take actions like recovery or deletion of unneeded files in a central manner. In this talk, we will present this system, explain the internals and give some results.

  17. Consistent Quantum Theory

    Science.gov (United States)

    Griffiths, Robert B.

    2001-11-01

    Quantum mechanics is one of the most fundamental yet difficult subjects in physics. Nonrelativistic quantum theory is presented here in a clear and systematic fashion, integrating Born's probabilistic interpretation with Schrödinger dynamics. Basic quantum principles are illustrated with simple examples requiring no mathematics beyond linear algebra and elementary probability theory. The quantum measurement process is consistently analyzed using fundamental quantum principles without referring to measurement. These same principles are used to resolve several of the paradoxes that have long perplexed physicists, including the double slit and Schrödinger's cat. The consistent histories formalism used here was first introduced by the author, and extended by M. Gell-Mann, J. Hartle and R. Omnès. Essential for researchers yet accessible to advanced undergraduate students in physics, chemistry, mathematics, and computer science, this book is supplementary to standard textbooks. It will also be of interest to physicists and philosophers working on the foundations of quantum mechanics. Comprehensive account Written by one of the main figures in the field Paperback edition of successful work on philosophy of quantum mechanics

  18. Characterisation of the hydraulic properties within the EDZ around drifts at level -490 m of the Meuse/Haute-Marne URL: A methodology for consistent interpretation of hydraulic tests

    Science.gov (United States)

    Baechler, S.; Lavanchy, J. M.; Armand, G.; Cruchaudet, M.

    In order to investigate potential changes of the hydraulic properties of the EDZ over time in the Meuse/Haute-Marne underground research laboratory (URL), several hydraulic tests campaigns have been carried out between 2005 and 2008 in dedicated boreholes. After several test series, inconsistencies were noticed in the results, indicating, in particular cases, erratic, inexplicable property changes over time and spatial contrasts. It was therefore difficult to determine reliably potential trends of the EDZ hydraulic properties. It appeared necessary to re-evaluate both the interpretation concepts and assumptions applied to the numerical analyses of test data on the EDZ, trying to better constrain the flow model and the parameter variables. In order to improve the understanding of the geometrical, geomechanical and hydraulic properties of the EDZ, independent information from other investigation methods has been used to critically revise the conceptual model and formation parameters. In particular, results from a diffusion experiment and ultrasonic measurements allowed constraining the extent of the mechanical damaged zone around the borehole (BDZ). Storativity parameters were fitted due to their expected variability. Indeed, high storativity values can be presumed under the unsaturated conditions of the EDZ. The results of the reanalyses, performed with the numerical borehole simulator MULTISIM, demonstrated the good quality and consistency of the revised conceptual model with constrained BDZ and variable storativity. Overall the new simulation results obtained from selected test series are now very consistent. The revised conceptual model demonstrated its capacity to better represent the evolution and extension of the EDZ around a drift in Meuse/Haute-Marne URL. Further consistency checks are proposed to confirm the new model assumptions and the estimates of the single phase flow model in the EDZ.

  19. Characterisation of the hydraulic properties within the EDZ around drifts at level -490 m of the Meuse/Haute-Marne URL: A methodology for consistent interpretation of hydraulic tests

    International Nuclear Information System (INIS)

    Baechler, S.; Lavanchy, J. M.; Armand, G.; Cruchaudet, M.

    2011-01-01

    In order to investigate potential changes of the hydraulic properties of the EDZ over time in the Meuse/Haute-Marne underground research laboratory (URL), several hydraulic tests campaigns have been carried out between 2005 and 2008 in dedicated boreholes. After several test series, inconsistencies were noticed in the results, indicating, in particular cases, erratic, inexplicable property changes over time and spatial contrasts. It was therefore difficult to determine reliably potential trends of the EDZ hydraulic properties. It appeared necessary to re-evaluate both the interpretation concepts and assumptions applied to the numerical analyses of test data on the EDZ, trying to better constrain the flow model and the parameter variables. In order to improve the understanding of the geometrical, geomechanical and hydraulic properties of the EDZ, independent information from other investigation methods has been used to critically revise the conceptual model and formation parameters. In particular, results from a diffusion experiment and ultrasonic measurements allowed constraining the extent of the mechanical damaged zone around the borehole (BDZ). Storativity parameters were fitted due to their expected variability. Indeed, high storativity values can be presumed under the unsaturated conditions of the EDZ. The results of the reanalyses, performed with the numerical borehole simulator MULTISIM, demonstrated the good quality and consistency of the revised conceptual model with constrained BDZ and variable storativity. Overall the new simulation results obtained from selected test series are now very consistent. The revised conceptual model demonstrated its capacity to better represent the evolution and extension of the EDZ around a drift in Meuse/Haute-Marne URL. Further consistency checks are proposed to confirm the new model assumptions and the estimates of the single phase flow model in the EDZ. (authors)

  20. Construct validity, test-retest reliability and internal consistency of the Thai version of the disabilities of the arm, shoulder and hand questionnaire (DASH-TH) in patients with carpal tunnel syndrome.

    Science.gov (United States)

    Buntragulpoontawee, Montana; Phutrit, Suphatha; Tongprasert, Siam; Wongpakaran, Tinakon; Khunachiva, Jeeranan

    2018-03-27

    This study evaluated additional psychometric properties of the Thai version of the disabilities of the arm, shoulder and hand questionnaire (DASH-TH) which included, test-retest reliability, construct validity, internal consistency of in patients with carpal tunnel syndrome. As for determining construct validity, the Thai EuroQOL questionnaire (EQ-5D-5L) was also administered in order to examine convergent and divergent validity. Fifty patients completed both questionnaires. The DASH-TH showed excellent test-retest reliability (intraclass correlation coefficient = 0.811) and internal consistency (Cronbach's alpha = 0.911). The exploratory factor analysis yielded a six-factor solution while the confirmatory factor analysis denoted that the hypothesized model adequately fit the data with a comparative fit index of 0.967 and a Tucker-Lewis index of 0.964. The related subscales between the DASH-TH and the Thai EQ-5D-5L were significantly correlated, indicating the DASH-TH's convergent and discriminant validity. The DASH-TH demonstrated good reliability, internal consistency construct validity, and multidimensionality, in assessing the upper extremity function in carpal tunnel syndrome patients.

  1. The Principle of Energetic Consistency

    Science.gov (United States)

    Cohn, Stephen E.

    2009-01-01

    A basic result in estimation theory is that the minimum variance estimate of the dynamical state, given the observations, is the conditional mean estimate. This result holds independently of the specifics of any dynamical or observation nonlinearity or stochasticity, requiring only that the probability density function of the state, conditioned on the observations, has two moments. For nonlinear dynamics that conserve a total energy, this general result implies the principle of energetic consistency: if the dynamical variables are taken to be the natural energy variables, then the sum of the total energy of the conditional mean and the trace of the conditional covariance matrix (the total variance) is constant between observations. Ensemble Kalman filtering methods are designed to approximate the evolution of the conditional mean and covariance matrix. For them the principle of energetic consistency holds independently of ensemble size, even with covariance localization. However, full Kalman filter experiments with advection dynamics have shown that a small amount of numerical dissipation can cause a large, state-dependent loss of total variance, to the detriment of filter performance. The principle of energetic consistency offers a simple way to test whether this spurious loss of variance limits ensemble filter performance in full-blown applications. The classical second-moment closure (third-moment discard) equations also satisfy the principle of energetic consistency, independently of the rank of the conditional covariance matrix. Low-rank approximation of these equations offers an energetically consistent, computationally viable alternative to ensemble filtering. Current formulations of long-window, weak-constraint, four-dimensional variational methods are designed to approximate the conditional mode rather than the conditional mean. Thus they neglect the nonlinear bias term in the second-moment closure equation for the conditional mean. The principle of

  2. Interpretations of referral appropriateness by senior health managers in five PCT areas in England: a qualitative investigation.

    Science.gov (United States)

    Blundell, N; Clarke, Aileen; Mays, N

    2010-06-01

    To explore interpretations of "appropriate" and "inappropriate" elective referral from primary to secondary surgical care among senior clinical and non-clinical managers in five purposively sampled primary care trusts (PCTs) and their main associated acute hospitals in the English National Health Service (NHS). Semi-structured face-to-face interviews were undertaken with senior managerial staff from clinical and non-clinical backgrounds. Interviews were tape-recorded, transcribed and analysed according to the Framework approach developed at the National Centre for Social Research using N6 (NUD*IST6) qualitative data analysis software. Twenty-two people of 23 approached were interviewed (between three and five respondents per PCT and associated acute hospital). Three attributes relating to appropriateness of referral were identified: necessity: whether a patient with given characteristics was believed suitable for referral; destination or level: where or to whom a patient should be referred; and quality (or process): how a referral was carried out, including (eg, investigations undertaken before referral, information contained in the referral and extent of patient involvement in the referral decision. Attributes were hierarchical. "Necessity" was viewed as the most fundamental attribute, followed by "destination" and, finally, "quality". In general, but not always, all three attributes were perceived as necessary for a referral to be defined as appropriate. For senior clinical and non-clinical managers at the local level in the English NHS, three hierarchical attributes (necessity, appropriateness of destination and quality of referral process) contributed to the overall concept of appropriateness of referral from primary to secondary surgical care.

  3. Estructura factorial y consistencia interna de una versión española del Purpose-In-Life Test / Factorial Structure and Internal Consistency of a Spanish Version of the Purpose-In-Life Test

    Directory of Open Access Journals (Sweden)

    Joaquín García-Alandete

    2013-06-01

    Full Text Available El objetivo de este trabajo fue examinar la estructura factorial y la consistencia interna de la versión española del Purpose-In-Life Test, instrumento que mide el logro de sentido de la vida desde los supuestos de la logoterapia. En el estudio participaron 457 universitarios españoles (320 mujeres, 137 hombres de 18 a 55 años de edad, M = 21.80, DE = 4.56. Se realizaron análisis descriptivos y correlaciones entre los ítems y el total de la escala inicial, un análisis factorial exploratorio, la estimación de la consistencia interna de los factores y de la escala obtenida, el análisis factorial confirmatorio de la misma, la prueba t para comparación de medias entre mujeres y hombres y la prueba de Kruskal-Wallis para el efecto de la edad. Los resultados mostraron una estructura de dos factores correlacionados con aceptable consistencia interna de la escala y de los factores, diferencias significativas entre mujeres y hombres en la puntuación total y uno de los factores, y no significativas en función de la edad. El análisis factorial confirmatorio muestra un adecuado ajuste, apoyando el modelo propuesto.

  4. Calibration and GEANT4 Simulations of the Phase II Proton Compute Tomography (pCT) Range Stack Detector

    Energy Technology Data Exchange (ETDEWEB)

    Uzunyan, S. A. [Northern Illinois Univ., DeKalb, IL (United States); Blazey, G. [Northern Illinois Univ., DeKalb, IL (United States); Boi, S. [Northern Illinois Univ., DeKalb, IL (United States); Coutrakon, G. [Northern Illinois Univ., DeKalb, IL (United States); Dyshkant, A. [Northern Illinois Univ., DeKalb, IL (United States); Francis, K. [Northern Illinois Univ., DeKalb, IL (United States); Hedin, D. [Northern Illinois Univ., DeKalb, IL (United States); Johnson, E. [Northern Illinois Univ., DeKalb, IL (United States); Kalnins, J. [Northern Illinois Univ., DeKalb, IL (United States); Zutshi, V. [Northern Illinois Univ., DeKalb, IL (United States); Ford, R. [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Rauch, J. E. [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Rubinov, P. [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Sellberg, G. [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Wilson, P. [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Naimuddin, M. [Univ. of Delhi, New Delhi (India)

    2015-12-29

    Northern Illinois University in collaboration with Fermi National Accelerator Laboratory (FNAL) and Delhi University has been designing and building a proton CT scanner for applications in proton treatment planning. The Phase II proton CT scanner consists of eight planes of tracking detectors with two X and two Y coordinate measurements both before and after the patient. In addition, a range stack detector consisting of a stack of thin scintillator tiles, arranged in twelve eight-tile frames, is used to determine the water equivalent path length (WEPL) of each track through the patient. The X-Y coordinates and WEPL are required input for image reconstruction software to find the relative (proton) stopping powers (RSP) value of each voxel in the patient and generate a corresponding 3D image. In this Note we describe tests conducted in 2015 at the proton beam at the Central DuPage Hospital in Warrenville, IL, focusing on the range stack calibration procedure and comparisons with the GEANT~4 range stack simulation.

  5. Final Report for Crucible -Scale Radioactive Vitrification and Product Test of Waste Envelope B (AZ-102) Low-Activity Waste Glass

    International Nuclear Information System (INIS)

    CRAWFORD, CHARLES

    2004-01-01

    A proof-of-technology demonstration for the Hanford River Protection Project (RPP) Waste treatment and Immobilization Plant (WTP) was performed by the Savannah River Technology Center (SRTC). As part of this demonstration, treated AZ-102 Low-Activity Waste supernate was vitrified using a crucible-scale furnace. Initial glass samples were quench-cooled and characterized for metals and radionuclides. The glass was also durability tested using the American Society for Testing and Materials (ASTM) Product Consistency Test (PCT) protocol. These tests used the AZ-102 glass formulation Low Activity Waste (LAW) B88 that targeted AZ-102 waste loading at 5 wt% Na2O. After these initial results were obtained with the quench-cooled LAWB88 glass, a prototypical container centerline cooling (CCC) program was supplied to SRTC by WTP. A portion of the quench-cooled LAWB88 glass was remelted and centerline cooled. Samples from the CCC low-activity AZ-102 glass waste form were durability tested using the PCT and characterized for crystalline phase identification.This final report documents the characterization and durability of this AZ-102 glass

  6. Atom-Probe Tomographic Investigation of Austenite Stability and Carbide Precipitation in a TRIP-Assisted 10 Wt Pct Ni Steel and Its Weld Heat-Affected Zones

    Science.gov (United States)

    Jain, Divya; Seidman, David N.; Barrick, Erin J.; DuPont, John N.

    2018-04-01

    Newly developed low-carbon 10 wt pct Ni-Mo-Cr-V martensitic steels rely on the Ni-enriched, thermally stable austenite [formed via multistep intercritical Quench-Lamellarization-Tempering ( QLT)-treatment] for their superior mechanical properties, specifically ballistic resistance. Critical to the thermal stability of austenite is its composition, which can be severely affected in the weld heat-affected zones (HAZs) and thus needs investigations. This article represents the first study of the nanoscale redistributions of C, Ni, and Mn in single-pass HAZ microstructures of QLT-treated 10 wt pct Ni steels. Local compositions of Ni-rich regions (representative of austenite compositions) in the HAZs are determined using site-specific 3-D atom-probe tomography (APT). Martensite-start temperatures are then calculated for these compositions, employing the Ghosh-Olson thermodynamic and kinetics approach. These calculations predict that austenite (present at high temperatures) in the HAZs is susceptible to a martensitic transformation upon cooling to room temperature, unlike the austenite in the QLT-treated base-metal. While C in the QLT-treated base-metal is consumed primarily in MC and M2C-type carbide precipitates (M is Mo, Cr, V), its higher concentration in the Ni-rich regions in the HAZs indicates the dissolution of carbide precipitates, particularly M2C carbide precipitates. The role of M2C carbide precipitates and austenite stability is discussed in relation to the increase in microhardness values observed in the HAZs, relative to the QLT-treated base-metal. Insights gained from this research on austenite stability and carbide precipitation in the single-pass HAZ microstructures will assist in designing multiple weld cycles for these novel 10 wt pct Ni steels.

  7. Effects of the Cr2O3 Content on the Viscosity of CaO-SiO2-10 Pct Al2O3-Cr2O3 Quaternary Slag

    Science.gov (United States)

    Wu, Tuo; Zhang, Yanling; Yuan, Fang; An, Zhuoqing

    2018-04-01

    The present study experimentally investigates the effect of Cr2O3 on the viscosity of molten slags. The viscosities of CaO-SiO2-10 pct Al2O3-Cr2O3 quaternary slags with two different binary basicities (R, basic slag with R = 1.2 and acidic slag with R = 0.8) were measured by the rotating cylindrical method from 1813 K to 1953 K (1540 °C to 1680 °C). The results showed that the viscosity of both types of slag decreased as the Cr2O3 content increased, but the viscosity of acidic slags exhibited a greater decrease. The slags showed good Newtonian behavior at such high temperatures. Cr2O3 could act as a network modifier to simplify the Si-O-Si tetrahedral structure, as verified by the Raman spectral analysis, which was consistent with the decreasing trend of viscosity. The activation energy of viscous flow decreased slightly with increasing Cr2O3, but increasing the basicity seemed to be more effective in decreasing the viscosity than adding Cr2O3.

  8. Bridge deck surface temperature monitoring by infrared thermography and inner structure identification using PPT and PCT analysis methods

    Science.gov (United States)

    Dumoulin, Jean

    2013-04-01

    One of the objectives of ISTIMES project was to evaluate the potentialities offered by the integration of different electromagnetic techniques able to perform non-invasive diagnostics for surveillance and monitoring of transport infrastructures. Among the EM methods investigated, we focused our research and development efforts on uncooled infrared camera techniques due to their promising potential level of dissemination linked to their relative low cost on the market. On the other hand, works were also carried out to identify well adapted implementation protocols and key limits of Pulse Phase Thermography (PPT) and Principal Component Thermography (PCT) processing methods to analyse thermal image sequence and retrieve information about the inner structure. So the first part of this research works addresses infrared thermography measurement when it is used in quantitative mode (not in laboratory conditions) and not in qualitative mode (vision applied to survey). In such context, it requires to process in real time thermal radiative corrections on raw data acquired to take into account influences of natural environment evolution with time, thanks to additional measurements. But, camera sensor has to be enough smart to apply in real time calibration law and radiometric corrections in a varying atmosphere. So, a complete measurement system was studied and developed [1] with low cost infrared cameras available on the market. In the system developed, infrared camera is coupled with other sensors to feed simplified radiative models running, in real time, on GPU available on small PC. The whole measurement system was implemented on the "Musmeci" bridge located in Potenza (Italy). No traffic interruption was required during the mounting of our measurement system. The infrared camera was fixed on top of a mast at 6 m elevation from the surface of the bridge deck. A small weather station was added on the same mast at 1 m under the camera. A GPS antenna was also fixed at the

  9. Utility of Procalcitonin (PCT and Mid regional pro-Adrenomedullin (MR-proADM in risk stratification of critically ill febrile patients in Emergency Department (ED. A comparison with APACHE II score

    Directory of Open Access Journals (Sweden)

    Travaglino Francesco

    2012-08-01

    Full Text Available Abstract Background The aim of our study was to evaluate the prognostic value of MR-proADM and PCT levels in febrile patients in the ED in comparison with a disease severity index score, the APACHE II score. We also evaluated the ability of MR-proADM and PCT to predict hospitalization. Methods This was an observational, multicentric study. We enrolled 128 patients referred to the ED with high fever and a suspicion of severe infection such as sepsis, lower respiratory tract infections, urinary tract infections, gastrointestinal infections, soft tissue infections, central nervous system infections, or osteomyelitis. The APACHE II score was calculated for each patient. Results MR-proADM median values in controls were 0.5 nmol/l as compared with 0.85 nmol/l in patients (P P . MR-proADM and PCT levels were significantly increased in accordance with the Apache II quartiles (P  respectively. In the respiratory infections, urinary infections, and sepsis-septic shock groups we found a correlation between the Apache II and MR-proADM respectively and MR-proADM and PCT respectively. We evaluated the ability of MR-proADM and PCT to predict hospitalization in patients admitted to our emergency departments complaining of fever. MR-proADM alone had an AUC of 0.694, while PCT alone had an AUC of 0.763. The combined use of PCT and MR-proADM instead showed an AUC of 0.79. Conclusions The present study highlights the way in which MR-proADM and PCT may be helpful to the febrile patient’s care in the ED. Our data support the prognostic role of MR-proADM and PCT in that setting, as demonstrated by the correlation with the APACHE II score. The combined use of the two biomarkers can predict a subsequent hospitalization of febrile patients. The rational use of these two molecules could lead to several advantages, such as faster diagnosis, more accurate risk stratification, and optimization of the treatment, with consequent benefit to the patient and

  10. Measuring process and knowledge consistency

    DEFF Research Database (Denmark)

    Edwards, Kasper; Jensen, Klaes Ladeby; Haug, Anders

    2007-01-01

    When implementing configuration systems, knowledge about products and processes are documented and replicated in the configuration system. This practice assumes that products are specified consistently i.e. on the same rule base and likewise for processes. However, consistency cannot be taken...... for granted; rather the contrary, and attempting to implement a configuration system may easily ignite a political battle. This is because stakes are high in the sense that the rules and processes chosen may only reflect one part of the practice, ignoring a majority of the employees. To avoid this situation......, this paper presents a methodology for measuring product and process consistency prior to implementing a configuration system. The methodology consists of two parts: 1) measuring knowledge consistency and 2) measuring process consistency. Knowledge consistency is measured by developing a questionnaire...

  11. Corrosion Resistance Behavior of Single-Layer Cathodic Arc PVD Nitride-Base Coatings in 1M HCl and 3.5 pct NaCl Solutions

    Science.gov (United States)

    Adesina, Akeem Yusuf; Gasem, Zuhair M.; Madhan Kumar, Arumugam

    2017-04-01

    The electrochemical behavior of single-layer TiN, CrN, CrAlN, and TiAlN coatings on 304 stainless steel substrate, deposited using state-of-the-art and industrial size cathodic arc PVD machine, were evaluated in 1M HCl and 3.5 pct NaCl solutions. The corrosion behavior of the blank and coated substrates was analyzed by electrochemical impedance spectroscopy (EIS), linear polarization resistance, and potentiodynamic polarization. Bond-coat layers of pure-Ti, pure-Cr, alloyed-CrAl, and alloyed-TiAl for TiN, CrN, CrAlN, and TiAlN coatings were, respectively, first deposited for improved coating adhesion before the actual coating. The average coating thickness was about 1.80 µm. Results showed that the corrosion potentials ( E corr) of the coated substrates were shifted to more noble values which indicated improvement of the coated substrate resistance to corrosion susceptibility. The corrosion current densities were lower for all coated substrates as compared to the blank substrate. Similarly, EIS parameters showed that these coatings possessed improved resistance to defects and pores in similar solution compared to the same nitride coatings developed by magnetron sputtering. The charge transfer resistance ( R ct) can be ranked in the following order: TiAlN > CrN > TiN > CrAlN in both media except in NaCl solution where R ct of TiN is lowest. While the pore resistance ( R po) followed the order: CrAlN > CrN > TiAlN > TiN in HCl solution and TiAlN > CrN > CrAlN > TiN in NaCl solution. It is found that TiAlN coating has the highest protective efficiencies of 79 and 99 pct in 1M HCl and 3.5 pct NaCl, respectively. SEM analysis of the corroded substrates in both media was also presented.

  12. Final Report - ILAW PCT, VHT, Viscosity, and Electrical Conductivity Model Development, VSL-07R1230-1

    Energy Technology Data Exchange (ETDEWEB)

    Kruger, Albert A.; Cooley, S. K.; Joseph, I.; Pegg, I. L.; Piepel, G. F.; Gan, H.; Muller, I.

    2013-12-17

    This report describes the results of work and testing specified by the Test Specifications (24590-LAW-TSP-RT-01-013 Rev.1 and 24590-WTP-TSP-RT-02-001 Rev.0), Test Plans (VSL-02T4800-1 Rev.1 & TP-RPP-WTP-179 Rev.1), and Text Exception (24590-WTP-TEF-RT-03-040). The work and any associated testing followed established quality assurance requirements and conducted as authorized. The descriptions provided in this test report are an accurate account of both the conduct of the work and the data collected. Results required by the Test Plans are reported. Also reported are any unusual or anomalous occurrences that are different from the starting hypotheses. The test results and this report have been reviewed and verified.

  13. Effects of aging temperature on electrical conductivity and hardness of Cu-3 at. pct Ti alloy aged in a hydrogen atmosphere

    KAUST Repository

    Semboshi, S.

    2011-08-01

    To improve the balance of the electrical conductivity and mechanical strength for dilute Cu-Ti alloys by aging in a hydrogen atmosphere, the influence of aging temperature ranging from 673 K to 773 K (400 °C to 500 °C) on the properties of Cu-3 at. pct Ti alloy was studied. The Vickers hardness increases steadily with aging time and starts to fall at 3 hours at 773 K (500 °C), 10 hours at 723 K (450 °C), or over 620 hours at 673 K (400 °C), which is the same as the case of conventional aging in vacuum. The maximum hardness increases from 220 to 236 with the decrease of aging temperature, which is slightly lower than aging at the same temperature in vacuum. The electrical conductivity at the maximum hardness also increases from 18 to 32 pct of pure copper with the decrease of the temperature, which is enhanced by a factor of 1.3 to 1.5 in comparison to aging in vacuum. Thus, aging at 673 K (400 °C) in a hydrogen atmosphere renders fairly good balance of strength and conductivity, although it takes nearly a month to achieve. The microstructural changes during aging were examined by transmission electron microscopy (TEM) and atom-probe tomography (APT), it was confirmed that precipitation of the Cu4Ti phase occurs first and then particles of TiH2 form as the third phase, thereby efficiently removing the Ti solutes in the matrix.

  14. Coordinating user interfaces for consistency

    CERN Document Server

    Nielsen, Jakob

    2001-01-01

    In the years since Jakob Nielsen's classic collection on interface consistency first appeared, much has changed, and much has stayed the same. On the one hand, there's been exponential growth in the opportunities for following or disregarding the principles of interface consistency-more computers, more applications, more users, and of course the vast expanse of the Web. On the other, there are the principles themselves, as persistent and as valuable as ever. In these contributed chapters, you'll find details on many methods for seeking and enforcing consistency, along with bottom-line analys

  15. Choice, internal consistency, and rationality

    OpenAIRE

    Aditi Bhattacharyya; Prasanta K. Pattanaik; Yongsheng Xu

    2010-01-01

    The classical theory of rational choice is built on several important internal consistency conditions. In recent years, the reasonableness of those internal consistency conditions has been questioned and criticized, and several responses to accommodate such criticisms have been proposed in the literature. This paper develops a general framework to accommodate the issues raised by the criticisms of classical rational choice theory, and examines the broad impact of these criticisms from both no...

  16. Self-consistent quark bags

    International Nuclear Information System (INIS)

    Rafelski, J.

    1979-01-01

    After an introductory overview of the bag model the author uses the self-consistent solution of the coupled Dirac-meson fields to represent a bound state of strongly ineteracting fermions. In this framework he discusses the vivial approach to classical field equations. After a short description of the used numerical methods the properties of bound states of scalar self-consistent Fields and the solutions of a self-coupled Dirac field are considered. (HSI) [de

  17. Time-consistent and market-consistent evaluations

    NARCIS (Netherlands)

    Pelsser, A.; Stadje, M.A.

    2014-01-01

    We consider evaluation methods for payoffs with an inherent financial risk as encountered for instance for portfolios held by pension funds and insurance companies. Pricing such payoffs in a way consistent to market prices typically involves combining actuarial techniques with methods from

  18. Usefulness of a semi-quantitative procalcitonin test and the A-DROP Japanese prognostic scale for predicting mortality among adults hospitalized with community-acquired pneumonia.

    Science.gov (United States)

    Kasamatsu, Yu; Yamaguchi, Toshimasa; Kawaguchi, Takashi; Tanaka, Nagaaki; Oka, Hiroko; Nakamura, Tomoyuki; Yamagami, Keiko; Yoshioka, Katsunobu; Imanishi, Masahito

    2012-02-01

    The solid-phase immunoassay, semi-quantitative procalcitonin (PCT) test (B R A H M S PCT-Q) can be used to rapidly categorize PCT levels into four grades. However, the usefulness of this kit for determining the prognosis of adult patients with community-acquired pneumonia (CAP) is unclear. A prospective study was conducted in two Japanese hospitals to evaluate the usefulness of this PCT test in determining the prognosis of adult patients with CAP. The accuracy of the age, dehydration, respiratory failure, orientation disturbance, pressure (A-DROP) scale proposed by the Japanese Respiratory Society for prediction of mortality due to CAP was also investigated. Hospitalized CAP patients (n = 226) were enrolled in the study. Comprehensive examinations were performed to determine PCT and CRP concentrations, disease severity based on the A-DROP, pneumonia severity index (PSI) and confusion, urea, respiratory rate, blood pressure, age ≥65 (CURB-65) scales and the causative pathogens. The usefulness of the biomarkers and prognostic scales for predicting each outcome were then examined. Twenty of the 170 eligible patients died. PCT levels were strongly positively correlated with PSI (ρ = 0.56, P scale were found to be useful for predicting mortality in adult patients with CAP. © 2011 The Authors. Respirology © 2011 Asian Pacific Society of Respirology.

  19. Student Effort, Consistency and Online Performance

    Directory of Open Access Journals (Sweden)

    Hilde Patron

    2011-07-01

    Full Text Available This paper examines how student effort, consistency, motivation, and marginal learning, influence student grades in an online course. We use data from eleven Microeconomics courses taught online for a total of 212 students. Our findings show that consistency, or less time variation, is a statistically significant explanatory variable, whereas effort, or total minutes spent online, is not. Other independent variables include GPA and the difference between a pre-test and a post-test. The GPA is used as a measure of motivation, and the difference between a post-test and pre-test as marginal learning. As expected, the level of motivation is found statistically significant at a 99% confidence level, and marginal learning is also significant at a 95% level.

  20. Consistent thermodynamic properties of lipids systems

    DEFF Research Database (Denmark)

    Cunico, Larissa; Ceriani, Roberta; Sarup, Bent

    different pressures, with azeotrope behavior observed. Available thermodynamic consistency tests for TPx data were applied before performing parameter regressions for Wilson, NRTL, UNIQUAC and original UNIFAC models. The relevance of enlarging experimental databank of lipids systems data in order to improve......Physical and thermodynamic properties of pure components and their mixtures are the basic requirement for process design, simulation, and optimization. In the case of lipids, our previous works[1-3] have indicated a lack of experimental data for pure components and also for their mixtures...... the performance of predictive thermodynamic models was confirmed in this work by analyzing the calculated values of original UNIFAC model. For solid-liquid equilibrium (SLE) data, new consistency tests have been developed [2]. Some of the developed tests were based in the quality tests proposed for VLE data...

  1. Market-consistent actuarial valuation

    CERN Document Server

    Wüthrich, Mario V

    2016-01-01

    This is the third edition of this well-received textbook, presenting powerful methods for measuring insurance liabilities and assets in a consistent way, with detailed mathematical frameworks that lead to market-consistent values for liabilities. Topics covered are stochastic discounting with deflators, valuation portfolio in life and non-life insurance, probability distortions, asset and liability management, financial risks, insurance technical risks, and solvency. Including updates on recent developments and regulatory changes under Solvency II, this new edition of Market-Consistent Actuarial Valuation also elaborates on different risk measures, providing a revised definition of solvency based on industry practice, and presents an adapted valuation framework which takes a dynamic view of non-life insurance reserving risk.

  2. Consistent guiding center drift theories

    International Nuclear Information System (INIS)

    Wimmel, H.K.

    1982-04-01

    Various guiding-center drift theories are presented that are optimized in respect of consistency. They satisfy exact energy conservation theorems (in time-independent fields), Liouville's theorems, and appropriate power balance equations. A theoretical framework is given that allows direct and exact derivation of associated drift-kinetic equations from the respective guiding-center drift-orbit theories. These drift-kinetic equations are listed. Northrop's non-optimized theory is discussed for reference, and internal consistency relations of G.C. drift theories are presented. (orig.)

  3. Weak consistency and strong paraconsistency

    Directory of Open Access Journals (Sweden)

    Gemma Robles

    2009-11-01

    Full Text Available In a standard sense, consistency and paraconsistency are understood as, respectively, the absence of any contradiction and as the absence of the ECQ (“E contradictione quodlibet” rule that allows us to conclude any well formed formula from any contradiction. The aim of this paper is to explain the concepts of weak consistency alternative to the standard one, the concepts of paraconsistency related to them and the concept of strong paraconsistency, all of which have been defined by the author together with José M. Méndez.

  4. Time-consistent actuarial valuations

    NARCIS (Netherlands)

    Pelsser, A.A.J.; Salahnejhad Ghalehjooghi, A.

    2016-01-01

    Time-consistent valuations (i.e. pricing operators) can be created by backward iteration of one-period valuations. In this paper we investigate the continuous-time limits of well-known actuarial premium principles when such backward iteration procedures are applied. This method is applied to an

  5. Dynamically consistent oil import tariffs

    International Nuclear Information System (INIS)

    Karp, L.; Newbery, D.M.

    1992-01-01

    The standard theory of optimal tariffs considers tariffs on perishable goods produced abroad under static conditions, in which tariffs affect prices only in that period. Oil and other exhaustable resources do not fit this model, for current tariffs affect the amount of oil imported, which will affect the remaining stock and hence its future price. The problem of choosing a dynamically consistent oil import tariff when suppliers are competitive but importers have market power is considered. The open-loop Nash tariff is solved for the standard competitive case in which the oil price is arbitraged, and it was found that the resulting tariff rises at the rate of interest. This tariff was found to have an equilibrium that in general is dynamically inconsistent. Nevertheless, it is shown that necessary and sufficient conditions exist under which the tariff satisfies the weaker condition of time consistency. A dynamically consistent tariff is obtained by assuming that all agents condition their current decisions on the remaining stock of the resource, in contrast to open-loop strategies. For the natural case in which all agents choose their actions simultaneously in each period, the dynamically consistent tariff was characterized, and found to differ markedly from the time-inconsistent open-loop tariff. It was shown that if importers do not have overwhelming market power, then the time path of the world price is insensitive to the ability to commit, as is the level of wealth achieved by the importer. 26 refs., 4 figs

  6. Consistently violating the non-Gaussian consistency relation

    International Nuclear Information System (INIS)

    Mooij, Sander; Palma, Gonzalo A.

    2015-01-01

    Non-attractor models of inflation are characterized by the super-horizon evolution of curvature perturbations, introducing a violation of the non-Gaussian consistency relation between the bispectrum's squeezed limit and the power spectrum's spectral index. In this work we show that the bispectrum's squeezed limit of non-attractor models continues to respect a relation dictated by the evolution of the background. We show how to derive this relation using only symmetry arguments, without ever needing to solve the equations of motion for the perturbations

  7. Consistence of Network Filtering Rules

    Institute of Scientific and Technical Information of China (English)

    SHE Kun; WU Yuancheng; HUANG Juncai; ZHOU Mingtian

    2004-01-01

    The inconsistence of firewall/VPN(Virtual Private Network) rule makes a huge maintainable cost.With development of Multinational Company,SOHO office,E-government the number of firewalls/VPN will increase rapidly.Rule table in stand-alone or network will be increased in geometric series accordingly.Checking the consistence of rule table manually is inadequate.A formal approach can define semantic consistence,make a theoretic foundation of intelligent management about rule tables.In this paper,a kind of formalization of host rules and network ones for auto rule-validation based on SET theory were proporsed and a rule validation scheme was defined.The analysis results show the superior performance of the methods and demonstrate its potential for the intelligent management based on rule tables.

  8. Self-consistent radial sheath

    International Nuclear Information System (INIS)

    Hazeltine, R.D.

    1988-12-01

    The boundary layer arising in the radial vicinity of a tokamak limiter is examined, with special reference to the TEXT tokamak. It is shown that sheath structure depends upon the self-consistent effects of ion guiding-center orbit modification, as well as the radial variation of E /times/ B-induced toroidal rotation. Reasonable agreement with experiment is obtained from an idealized model which, however simplified, preserves such self-consistent effects. It is argued that the radial sheath, which occurs whenever confining magnetic field-lines lie in the plasma boundary surface, is an object of some intrinsic interest. It differs from the more familiar axial sheath because magnetized charges respond very differently to parallel and perpendicular electric fields. 11 refs., 1 fig

  9. The effect of quench rate on the microstructure, mechanical properties, and corrosion behavior of U-6 Wt Pct Nb

    International Nuclear Information System (INIS)

    Eckelmeyer, K.H.; Romiy, A.D.; Weirick, L.J.

    1984-01-01

    The effect of cooling rate on microstructure, mechanical behavior, corrosion resistance, and subsequent age hardenability is discussed. Cooling rates in excess of 20 Ks -1 cause the parent γ-phase to transform martensitically to a niobium supersaturated variant of the α-phase. This phase exhibits low hardness and strength, high ductility, good corrosion resistance, and age hardenability. As cooling rate decreases from 10 Ks -1 to 0.2 Ks -1 , microstructural changes (consistent with spinodal decomposition) occur to an increasing extent. These changes produce increases in hardness and strength and decreases in ductility, corrosion resistance, and age hardenability. At cooling rates less than 0.2 Ks -1 the parent phase undergoes cellular decomposition to a coarse two-phase lamellar microstructure which exhibits intermediate strength and ductility, reduced corrosion resistance, and no age hardenability. An analysis of the cooling rates indicates that fully martensitic microstructures can be obtained in plates as thick as 50 mm

  10. Lagrangian multiforms and multidimensional consistency

    Energy Technology Data Exchange (ETDEWEB)

    Lobb, Sarah; Nijhoff, Frank [Department of Applied Mathematics, University of Leeds, Leeds LS2 9JT (United Kingdom)

    2009-10-30

    We show that well-chosen Lagrangians for a class of two-dimensional integrable lattice equations obey a closure relation when embedded in a higher dimensional lattice. On the basis of this property we formulate a Lagrangian description for such systems in terms of Lagrangian multiforms. We discuss the connection of this formalism with the notion of multidimensional consistency, and the role of the lattice from the point of view of the relevant variational principle.

  11. Consistency and Communication in Committees

    OpenAIRE

    Inga Deimen; Felix Ketelaar; Mark T. Le Quement

    2013-01-01

    This paper analyzes truthtelling incentives in pre-vote communication in heterogeneous committees. We generalize the classical Condorcet jury model by introducing a new informational structure that captures consistency of information. In contrast to the impossibility result shown by Coughlan (2000) for the classical model, full pooling of information followed by sincere voting is an equilibrium outcome of our model for a large set of parameter values implying the possibility of ex post confli...

  12. Deep Feature Consistent Variational Autoencoder

    OpenAIRE

    Hou, Xianxu; Shen, Linlin; Sun, Ke; Qiu, Guoping

    2016-01-01

    We present a novel method for constructing Variational Autoencoder (VAE). Instead of using pixel-by-pixel loss, we enforce deep feature consistency between the input and the output of a VAE, which ensures the VAE's output to preserve the spatial correlation characteristics of the input, thus leading the output to have a more natural visual appearance and better perceptual quality. Based on recent deep learning works such as style transfer, we employ a pre-trained deep convolutional neural net...

  13. ELIMINATION OF THE CHARACTERIZATION OF DWPF POUR STREAM SAMPLE AND THE GLASS FABRICATION AND TESTING OF THE DWPF SLUDGE BATCH QUALIFICATION SAMPLE

    Energy Technology Data Exchange (ETDEWEB)

    Amoroso, J.; Peeler, D.; Edwards, T.

    2012-05-11

    A recommendation to eliminate all characterization of pour stream glass samples and the glass fabrication and Product Consistency Test (PCT) of the sludge batch qualification sample was made by a Six-Sigma team chartered to eliminate non-value-added activities for the Defense Waste Processing Facility (DWPF) sludge batch qualification program and is documented in the report SS-PIP-2006-00030. That recommendation was supported through a technical data review by the Savannah River National Laboratory (SRNL) and is documented in the memorandums SRNL-PSE-2007-00079 and SRNL-PSE-2007-00080. At the time of writing those memorandums, the DWPF was processing sludge-only waste but, has since transitioned to a coupled operation (sludge and salt). The SRNL was recently tasked to perform a similar data review relevant to coupled operations and re-evaluate the previous recommendations. This report evaluates the validity of eliminating the characterization of pour stream glass samples and the glass fabrication and Product Consistency Test (PCT) of the sludge batch qualification samples based on sludge-only and coupled operations. The pour stream sample has confirmed the DWPF's ability to produce an acceptable waste form from Slurry Mix Evaporator (SME) blending and product composition/durability predictions for the previous sixteen years but, ultimately the pour stream analysis has added minimal value to the DWPF's waste qualification strategy. Similarly, the information gained from the glass fabrication and PCT of the sludge batch qualification sample was determined to add minimal value to the waste qualification strategy since that sample is routinely not representative of the waste composition ultimately processed at the DWPF due to blending and salt processing considerations. Moreover, the qualification process has repeatedly confirmed minimal differences in glass behavior from actual radioactive waste to glasses fabricated from simulants or batch chemicals. In

  14. Elimination Of The Characterization Of DWPF Pour Stream Sample And The Glass Fabrication And Testing Of The DWPF Sludge Batch Qualification Sample

    International Nuclear Information System (INIS)

    Amoroso, J.; Peeler, D.; Edwards, T.

    2012-01-01

    A recommendation to eliminate all characterization of pour stream glass samples and the glass fabrication and Product Consistency Test (PCT) of the sludge batch qualification sample was made by a Six-Sigma team chartered to eliminate non-value-added activities for the Defense Waste Processing Facility (DWPF) sludge batch qualification program and is documented in the report SS-PIP-2006-00030. That recommendation was supported through a technical data review by the Savannah River National Laboratory (SRNL) and is documented in the memorandums SRNL-PSE-2007-00079 and SRNL-PSE-2007-00080. At the time of writing those memorandums, the DWPF was processing sludge-only waste but, has since transitioned to a coupled operation (sludge and salt). The SRNL was recently tasked to perform a similar data review relevant to coupled operations and re-evaluate the previous recommendations. This report evaluates the validity of eliminating the characterization of pour stream glass samples and the glass fabrication and Product Consistency Test (PCT) of the sludge batch qualification samples based on sludge-only and coupled operations. The pour stream sample has confirmed the DWPF's ability to produce an acceptable waste form from Slurry Mix Evaporator (SME) blending and product composition/durability predictions for the previous sixteen years but, ultimately the pour stream analysis has added minimal value to the DWPF's waste qualification strategy. Similarly, the information gained from the glass fabrication and PCT of the sludge batch qualification sample was determined to add minimal value to the waste qualification strategy since that sample is routinely not representative of the waste composition ultimately processed at the DWPF due to blending and salt processing considerations. Moreover, the qualification process has repeatedly confirmed minimal differences in glass behavior from actual radioactive waste to glasses fabricated from simulants or batch chemicals. In contrast, the

  15. Maintaining consistency in distributed systems

    Science.gov (United States)

    Birman, Kenneth P.

    1991-01-01

    In systems designed as assemblies of independently developed components, concurrent access to data or data structures normally arises within individual programs, and is controlled using mutual exclusion constructs, such as semaphores and monitors. Where data is persistent and/or sets of operation are related to one another, transactions or linearizability may be more appropriate. Systems that incorporate cooperative styles of distributed execution often replicate or distribute data within groups of components. In these cases, group oriented consistency properties must be maintained, and tools based on the virtual synchrony execution model greatly simplify the task confronting an application developer. All three styles of distributed computing are likely to be seen in future systems - often, within the same application. This leads us to propose an integrated approach that permits applications that use virtual synchrony with concurrent objects that respect a linearizability constraint, and vice versa. Transactional subsystems are treated as a special case of linearizability.

  16. Image recognition and consistency of response

    Science.gov (United States)

    Haygood, Tamara M.; Ryan, John; Liu, Qing Mary A.; Bassett, Roland; Brennan, Patrick C.

    2012-02-01

    Purpose: To investigate the connection between conscious recognition of an image previously encountered in an experimental setting and consistency of response to the experimental question. Materials and Methods: Twenty-four radiologists viewed 40 frontal chest radiographs and gave their opinion as to the position of a central venous catheter. One-to-three days later they again viewed 40 frontal chest radiographs and again gave their opinion as to the position of the central venous catheter. Half of the radiographs in the second set were repeated images from the first set and half were new. The radiologists were asked of each image whether it had been included in the first set. For this study, we are evaluating only the 20 repeated images. We used the Kruskal-Wallis test and Fisher's exact test to determine the relationship between conscious recognition of a previously interpreted image and consistency in interpretation of the image. Results. There was no significant correlation between recognition of the image and consistency in response regarding the position of the central venous catheter. In fact, there was a trend in the opposite direction, with radiologists being slightly more likely to give a consistent response with respect to images they did not recognize than with respect to those they did recognize. Conclusion: Radiologists' recognition of previously-encountered images in an observer-performance study does not noticeably color their interpretation on the second encounter.

  17. Decentralized Consistent Updates in SDN

    KAUST Repository

    Nguyen, Thanh Dang

    2017-04-10

    We present ez-Segway, a decentralized mechanism to consistently and quickly update the network state while preventing forwarding anomalies (loops and blackholes) and avoiding link congestion. In our design, the centralized SDN controller only pre-computes information needed by the switches during the update execution. This information is distributed to the switches, which use partial knowledge and direct message passing to efficiently realize the update. This separation of concerns has the key benefit of improving update performance as the communication and computation bottlenecks at the controller are removed. Our evaluations via network emulations and large-scale simulations demonstrate the efficiency of ez-Segway, which compared to a centralized approach, improves network update times by up to 45% and 57% at the median and the 99th percentile, respectively. A deployment of a system prototype in a real OpenFlow switch and an implementation in P4 demonstrate the feasibility and low overhead of implementing simple network update functionality within switches.

  18. BWR 1 % main recirculation line break LOCA tests, RUNs 917 and 918, without HPCS at ROSA-III program

    International Nuclear Information System (INIS)

    Suzuki, Mitsuhiro; Okazaki, Motoaki; Anoda, Yoshinari; Kumamaru, Hiroshige; Nakamura, Hideo; Yonomoto, Taisuke; Koizumi, Yasuo; Tasaka, Kanji

    1988-07-01

    In a case of small break loss-of-coolant accident (LOCA) at a boiling water reactor (BWR) system, it is important to lower the system pressure to cool down the reactor system by using either the high pressure core spray (HPCS) or the automatic depressurization system (ADS). The report presents characteristic test results of RUNs 918 and 917, which were performed at the rig-of-safety assessment (ROSA)-III program simulating a 1 % break BWR LOCA with an assumption of HPCS failure, and clarifies effects of the ADS delay time on a small break LOCA. The ROSA-III test facility simulates principal components of a BWR/6 system with volumetric scaling factor of 1/424. It is experimentally concluded that the ADS delay time shorter than 4 minutes results in a similar PCT as that in a standard case, in which the PCT is observed after actuation of the low pressure core spray (LPCS). And the ADS delay time longer than 4 minutes results in higher PCT than in the standard case. In the latter, the PCT depends on the ADS time, a 220 K higher PCT, for example, in a case of 10 minutes ADS delay compared with the standard case. (author) 52 refs. 299 figs

  19. Test

    DEFF Research Database (Denmark)

    Bendixen, Carsten

    2014-01-01

    Bidrag med en kortfattet, introducerende, perspektiverende og begrebsafklarende fremstilling af begrebet test i det pædagogiske univers.......Bidrag med en kortfattet, introducerende, perspektiverende og begrebsafklarende fremstilling af begrebet test i det pædagogiske univers....

  20. Integrated Disposal Facility FY 2012 Glass Testing Summary Report

    Energy Technology Data Exchange (ETDEWEB)

    Pierce, Eric M. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Kerisit, Sebastien N. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Krogstad, Eirik J. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Burton, Sarah D. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Bjornstad, Bruce N. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Freedman, Vicky L. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Cantrell, Kirk J. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Snyder, Michelle MV [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Crum, Jarrod V. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Westsik, Joseph H. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States)

    2013-03-29

    PNNL is conducting work to provide the technical basis for estimating radionuclide release from the engineered portion of the disposal facility for Hanford immobilized low-activity waste (ILAW). Before the ILAW can be disposed, DOE must conduct a performance assessment (PA) for the Integrated Disposal Facility (IDF) that describes the long-term impacts of the disposal facility on public health and environmental resources. As part of the ILAW glass testing program, PNNL is implementing a strategy, consisting of experimentation and modeling, to provide the technical basis for estimating radionuclide release from the glass waste form in support of future IDF PAs. Key activities in FY12 include upgrading the STOMP/eSTOMP codes to do near-field modeling, geochemical modeling of PCT tests to determine the reaction network to be used in the STOMP codes, conducting PUF tests on selected glasses to simulate and accelerate glass weathering, developing a Monte Carlo simulation tool to predict the characteristics of the weathered glass reaction layer as a function of glass composition, and characterizing glasses and soil samples exhumed from an 8-year lysimeter test. The purpose of this report is to summarize the progress made in fiscal year (FY) 2012 and the first quarter of FY 2013 toward implementing the strategy with the goal of developing an understanding of the long-term corrosion behavior of LAW glasses.

  1. Consistency relations in effective field theory

    Energy Technology Data Exchange (ETDEWEB)

    Munshi, Dipak; Regan, Donough, E-mail: D.Munshi@sussex.ac.uk, E-mail: D.Regan@sussex.ac.uk [Astronomy Centre, School of Mathematical and Physical Sciences, University of Sussex, Brighton BN1 9QH (United Kingdom)

    2017-06-01

    The consistency relations in large scale structure relate the lower-order correlation functions with their higher-order counterparts. They are direct outcome of the underlying symmetries of a dynamical system and can be tested using data from future surveys such as Euclid. Using techniques from standard perturbation theory (SPT), previous studies of consistency relation have concentrated on continuity-momentum (Euler)-Poisson system of an ideal fluid. We investigate the consistency relations in effective field theory (EFT) which adjusts the SPT predictions to account for the departure from the ideal fluid description on small scales. We provide detailed results for the 3D density contrast δ as well as the scaled divergence of velocity θ-bar . Assuming a ΛCDM background cosmology, we find the correction to SPT results becomes important at k ∼> 0.05 h/Mpc and that the suppression from EFT to SPT results that scales as square of the wave number k , can reach 40% of the total at k ≈ 0.25 h/Mpc at z = 0. We have also investigated whether effective field theory corrections to models of primordial non-Gaussianity can alter the squeezed limit behaviour, finding the results to be rather insensitive to these counterterms. In addition, we present the EFT corrections to the squeezed limit of the bispectrum in redshift space which may be of interest for tests of theories of modified gravity.

  2. Effect of Temperature and Fluid Flow on Dendrite Growth During Solidification of Al-3 Wt Pct Cu Alloy by the Two-Dimensional Cellular Automaton Method

    Science.gov (United States)

    Gu, Cheng; Wei, Yanhong; Liu, Renpei; Yu, Fengyi

    2017-12-01

    A two-dimensional cellular automaton-finite volume model was developed to simulate dendrite growth of Al-3 wt pct Cu alloy during solidification to investigate the effect of temperature and fluid flow on dendrite morphology, solute concentration distribution, and dendrite growth velocity. Different calculation conditions that may influence the results of the simulation, including temperature and flow, were considered. The model was also employed to study the effect of different undercoolings, applied temperature fields, and forced flow velocities on solute segregation and dendrite growth. The initial temperature and fluid flow have a significant impact on the dendrite morphologies and solute profiles during solidification. The release of energy is operated with solidification and results in the increase of temperature. A larger undercooling leads to larger solute concentration near the solid/liquid interface and solute concentration gradient at the same time-step. Solute concentration in the solid region tends to increase with the increase of undercooling. Four vortexes appear under the condition when natural flow exists: the two on the right of the dendrite rotate clockwise, and those on the left of the dendrite rotate counterclockwise. With the increase of forced flow velocity, the rejected solute in the upstream region becomes easier to be washed away and enriched in the downstream region, resulting in acceleration of the growth of the dendrite in the upstream and inhibiting the downstream dendrite growth. The dendrite perpendicular to fluid flow shows a coarser morphology in the upstream region than that of the downstream. Almost no secondary dendrite appears during the calculation process.

  3. Thermally Stable Ni-rich Austenite Formed Utilizing Multistep Intercritical Heat Treatment in a Low-Carbon 10 Wt Pct Ni Martensitic Steel

    Science.gov (United States)

    Jain, Divya; Isheim, Dieter; Zhang, Xian J.; Ghosh, Gautam; Seidman, David N.

    2017-08-01

    Austenite reversion and its thermal stability attained during the transformation is key to enhanced toughness and blast resistance in transformation-induced-plasticity martensitic steels. We demonstrate that the thermal stability of Ni-stabilized austenite and kinetics of the transformation can be controlled by forming Ni-rich regions in proximity of pre-existing (retained) austenite. Atom probe tomography (APT) in conjunction with thermodynamic and kinetic modeling elucidates the role of Ni-rich regions in enhancing growth kinetics of thermally stable austenite, formed utilizing a multistep intercritical ( Quench- Lamellarization- Tempering (QLT)-type) heat treatment for a low-carbon 10 wt pct Ni steel. Direct evidence of austenite formation is provided by dilatometry, and the volume fraction is quantified by synchrotron X-ray diffraction. The results indicate the growth of nm-thick austenite layers during the second intercritical tempering treatment (T-step) at 863 K (590 °C), with austenite retained from first intercritical treatment (L-step) at 923 K (650 °C) acting as a nucleation template. For the first time, the thermal stability of austenite is quantified with respect to its compositional evolution during the multistep intercritical treatment of these steels. Austenite compositions measured by APT are used in combination with the thermodynamic and kinetic approach formulated by Ghosh and Olson to assess thermal stability and predict the martensite-start temperature. This approach is particularly useful as empirical relations cannot be extrapolated for the highly Ni-enriched austenite investigated in the present study.

  4. Effect of Heat Treatment on Morphology of Fe-Rich Intermetallics in Hypereutectic Al-Si-Cu-Ni Alloy with 1.26 pct Fe

    Science.gov (United States)

    Sha, Meng; Wu, Shusen; Wan, Li; Lü, Shulin

    2013-12-01

    Cobalt is generally considered as the element that can neutralize the negative effects of iron in Al alloys, such as inducing fracture and failure for stress concentration. Nevertheless, Fe-rich intermetallics would be inclined to form coarse plate-like δ-Al4(Fe, Co, Ni)Si2 particles when the content of Fe was high, which could also cause inferior mechanical properties. The dissolution and transformation of δ-Al4(Fe, Co, Ni)Si2 phase in solution heat-treated samples of Al-20Si-1.85Cu-1.05Ni-1.26Fe-1.35Co alloy were studied using optical microscopy, image analysis, and scanning electron microscopy. The effects of solution heat treatment time ranging from 0 to 9 hours at 783.15 K (510 °C) on mechanical properties were also investigated. The coarse plate-like δ-Al4(Fe, Co, Ni)Si2 particles varied slowly through concurrent dissolution along widths and at the plate tips as solution treatment time increased, which could be explained from diffusion-induced grain boundary migration. Solution heat treatment also has an important influence on mechanical properties. The maximum ultimate tensile strength and yield strength after T6 treatment were 258 and 132 MPa, respectively, while the maximum hardness was 131 HB. Compared with those of the samples in the as-cast state, they increased by 53, 42, and 28 pct, respectively. Moreover, δ-Al4(Fe, Co, Ni)Si2 phase, which appears as a coarse plate-like particle in two dimensions, is actually a cuboid in three dimensions. The length of this cuboid is close to the width, while the height is much smaller.

  5. Consistency relation for cosmic magnetic fields

    DEFF Research Database (Denmark)

    Jain, R. K.; Sloth, M. S.

    2012-01-01

    If cosmic magnetic fields are indeed produced during inflation, they are likely to be correlated with the scalar metric perturbations that are responsible for the cosmic microwave background anisotropies and large scale structure. Within an archetypical model of inflationary magnetogenesis, we show...... that there exists a new simple consistency relation for the non-Gaussian cross correlation function of the scalar metric perturbation with two powers of the magnetic field in the squeezed limit where the momentum of the metric perturbation vanishes. We emphasize that such a consistency relation turns out...... to be extremely useful to test some recent calculations in the literature. Apart from primordial non-Gaussianity induced by the curvature perturbations, such a cross correlation might provide a new observational probe of inflation and can in principle reveal the primordial nature of cosmic magnetic fields. DOI...

  6. Consistency of color representation in smart phones.

    Science.gov (United States)

    Dain, Stephen J; Kwan, Benjamin; Wong, Leslie

    2016-03-01

    One of the barriers to the construction of consistent computer-based color vision tests has been the variety of monitors and computers. Consistency of color on a variety of screens has necessitated calibration of each setup individually. Color vision examination with a carefully controlled display has, as a consequence, been a laboratory rather than a clinical activity. Inevitably, smart phones have become a vehicle for color vision tests. They have the advantage that the processor and screen are associated and there are fewer models of smart phones than permutations of computers and monitors. Colorimetric consistency of display within a model may be a given. It may extend across models from the same manufacturer but is unlikely to extend between manufacturers especially where technologies vary. In this study, we measured the same set of colors in a JPEG file displayed on 11 samples of each of four models of smart phone (iPhone 4s, iPhone5, Samsung Galaxy S3, and Samsung Galaxy S4) using a Photo Research PR-730. The iPhones are white LED backlit LCD and the Samsung are OLEDs. The color gamut varies between models and comparison with sRGB space shows 61%, 85%, 117%, and 110%, respectively. The iPhones differ markedly from the Samsungs and from one another. This indicates that model-specific color lookup tables will be needed. Within each model, the primaries were quite consistent (despite the age of phone varying within each sample). The worst case in each model was the blue primary; the 95th percentile limits in the v' coordinate were ±0.008 for the iPhone 4 and ±0.004 for the other three models. The u'v' variation in white points was ±0.004 for the iPhone4 and ±0.002 for the others, although the spread of white points between models was u'v'±0.007. The differences are essentially the same for primaries at low luminance. The variation of colors intermediate between the primaries (e.g., red-purple, orange) mirror the variation in the primaries. The variation in

  7. Sensitivity Study of the Peak Cladding Temperature for the Pipe Break Accidents of the 3-Pin Fuel Test Loop

    International Nuclear Information System (INIS)

    Park, S. K.; Chi, D. Y.; Sim, B. S.; Park, K. N.; Ahn, S. H.; Lee, J. M.; Lee, C. Y.; Kim, H. R.

    2005-12-01

    The effect of the thermal hydraulic operation parameters, the stroke times of safety-related valves, the node number of test fuel for MARS modeling, and the axial power distribution on the peak cladding temperature (PCT) has been investigated for the loss of coolant accident of the 3-pin fuel test loop. The thermal hydraulic operation parameters investigated are the thermal power of the fuel test loop and the flow rate, temperature, and pressure of the main cooling water. The effect of the thermal power and the coolant temperature on the peak cladding temperature is dominant as compared with that of the coolant flow rate and pressure. The maximum PCT increases up to about 34.3K for the room 1 LOCA when the thermal power increase by 5% of the normal operation power and decreases up to about 38.9K for the room 1 LOCA when the coolant temperature decrease by 2% of the normal operation temperature. The effect of the stroke time of the loop isolation valves on the PCT is also dominant. However the effect of the stroke time of the safety injection valves and depressurization vent valves are negligible. Especially the maximum PCT increases up to 25.7K with the increase of the design stroke time of the cold leg loop isolation valve by 13% and decreases up to 25.1K with the decrease of the design stroke time by 13%. The maximum PCT increases by 3.3K as the number of nodes increases from 7 to 14 for the MARS model of test fuel. Three different axial power distributions are also investigated. The maximum PCT occurs for the room 1 LOCA in case the peak power is shifted to the downstream by 20cm

  8. A data base and a standard material for use in acceptance testing of low-activity waste products

    International Nuclear Information System (INIS)

    Wolf, S.F.; Ebert, W.L.; Luo, J.S.; Strachan, D.M.

    1998-04-01

    The authors have conducted replicate dissolution tests following the product consistency test (PCT) procedure to measure the mean and standard deviation of the solution concentrations of B, Na, and Si at various combinations of temperature, duration, and glass/water mass ratio. Tests were conducted with a glass formulated to be compositionally similar to low-activity waste products anticipated for Hanford to evaluate the adequacy of test methods that have been designated in privatization contracts for use in product acceptance. An important finding from this set of tests is that the solution concentrations generated in tests at 20 C will likely be too low to measure the dissolution rates of waste products reliably. Based on these results, the authors recommend that the acceptance test be conducted at 40 C. Tests at 40 C generated higher solution concentrations, were more easily conducted, and the measured rates were easily related to those at 20 C. Replicate measurements of other glass properties were made to evaluate the possible use of LRM-1 as a standard material. These include its composition, homogeneity, density, compressive strength, the Na leachability index with the ANSI/ANS 16.1 leach test, and if the glass is characteristically hazardous with the toxicity characteristic leach procedure. The values of these properties were within the acceptable limits identified for Hanford low-activity waste products. The reproducibility of replicate tests and analyses indicates that the glass would be a suitable standard material

  9. Tests with ceramic waste form materials made by pressureless consolidation

    International Nuclear Information System (INIS)

    Lewis, M. A.; Hash, M. C.; Hebden, A. S.; Ebert, W. L.

    2002-01-01

    A multiphase waste form referred to as the ceramic waste form (CWF) will be used to immobilize radioactively contaminated salt wastes recovered after the electrometallurgical treatment of spent sodium-bonded nuclear fuel. The CWF is made by first occluding salt in zeolite and then encapsulating the zeolite in a borosilicate binder glass. A variety of surrogate CWF materials were made using pressureless consolidation (PC) methods for comparison with CWF consolidated using a hot isostatic press (HIP) method and to study the effects of glass/zeolite batching ratio and processing conditions on the physical and chemical properties of the resulting materials. The data summarized in this report will also be used to support qualification of the PC CWF for disposal in the proposed federal high-level radioactive waste repository at Yucca Mountain. The phase composition and microstructure of HIP CWF and PC CWF are essentially identical: both are composed of about 70% sodalite, 25% binder glass, and a 5% total of inclusion phases (halite, nepheline, and various oxides and silicates). The primary difference is that PC CWF materials have higher porosities than HIP CWFs. The product consistency test (PCT) that was initially developed to monitor homogeneous glass waste forms was used to measure the chemical durabilities of the CWF materials. Series of replicate tests with several PC CWF materials indicate that the PCT can be conducted with the same precision with CWF materials as with borosilicate glasses. Short-term (7-day) PCTs were used to evaluate the repeatability of making the PC CWF and the effects of the glass/zeolite mass ratio, process temperature, and processing time on the chemical durability. Long-term (up to 1 year) PCTs were used to compare the durabilities of HIP and PC CWFs and to estimate the apparent solubility limit for the PC CWF that is needed for modeling. The PC and HIP CWF materials had similar disabilities, based on the release of silicon in long

  10. The Consistency Between Clinical and Electrophysiological Diagnoses

    Directory of Open Access Journals (Sweden)

    Esra E. Okuyucu

    2009-09-01

    Full Text Available OBJECTIVE: The aim of this study was to provide information concerning the impact of electrophysiological tests in the clinical management and diagnosis of patients, and to evaluate the consistency between referring clinical diagnoses and electrophysiological diagnoses. METHODS: The study included 957 patients referred to the electroneuromyography (ENMG laboratory from different clinics with different clinical diagnoses in 2008. Demographic data, referring clinical diagnoses, the clinics where the requests wanted, and diagnoses after ENMG testing were recorded and statistically evaluated. RESULTS: In all, 957 patients [644 (67.3% female and 313 (32.7% male] were included in the study. Mean age of the patients was 45.40 ± 14.54 years. ENMG requests were made by different specialists; 578 (60.4% patients were referred by neurologists, 122 (12.8% by orthopedics, 140 (14.6% by neurosurgeons, and 117 (12.2% by physical treatment and rehabilitation departments. According to the results of ENMG testing, 513 (53.6% patients’ referrals were related to their referral diagnosis, whereas 397 (41.5% patients had normal ENMG test results, and 47 (4.9% patients had a diagnosis that differed from the referring diagnosis. Among the relation between the referral diagnosis and electrophysiological diagnosis according to the clinics where the requests were made, there was no statistical difference (p= 0.794, but there were statistically significant differences between the support of different clinical diagnoses, such as carpal tunnel syndrome, polyneuropathy, radiculopathy-plexopathy, entrapment neuropathy, and myopathy based on ENMG test results (p< 0.001. CONCLUSION: ENMG is a frequently used neurological examination. As such, referrals for ENMG can be made to either support the referring diagnosis or to exclude other diagnoses. This may explain the inconsistency between clinical referring diagnoses and diagnoses following ENMG

  11. [Consistent Declarative Memory with Depressive Symptomatology].

    Science.gov (United States)

    Botelho de Oliveira, Silvia; Flórez, Ruth Natalia Suárez; Caballero, Diego Andrés Vásquez

    2012-12-01

    Some studies have suggested that potentiated remembrance of negative events on people with depressive disorders seems to be an important factor in the etiology, course and maintenance of depression. Evaluate the emotional memory in people with and without depressive symptomatology by means of an audio-visual test. 73 university students were evaluated, male and female, between 18 and 40 years old, distributed in two groups: with depressive symptomatology (32) and without depressive symptomatology (40), using the Scale from the Center of Epidemiologic Studies for Depression (CES-D, English Abbreviation) and a cutting point of 20. There were not meaningful differences between free and voluntary recalls, with and without depressive symptomatology, in spite of the fact that both groups had granted a higher emotional value to the audio-visual test and that they had associated it with emotional sadness. People with depressive symptomatology did not exhibit the effect of mnemonic potentiation generally associated to the content of the emotional version of the test; therefore, the hypothesis of emotional consistency was not validated. Copyright © 2012 Asociación Colombiana de Psiquiatría. Publicado por Elsevier España. All rights reserved.

  12. US bureau of mines small-scale arc melter tests

    International Nuclear Information System (INIS)

    O'Connor, W.K.; Oden, L.L.; Turner, P.C.; Davis, D.L.

    1993-01-01

    The US Bureau of Mines, in cooperation with the Idaho National Engineering Laboratory (INEL), conducted over 30 hours of melting tests to vitrify simulated low-level radioactive wastes from the INEL. Radioactive Waste Management Complex (RWMC). Five separate waste compositions were investigated, each consisting of noncontaminated soil from the RWMC and surrogate materials used to simulate the actual buried wastes. The RWMC soil and five waste compositions were melted in a 50-lb, single-phase electric arc furnace with a water-cooled shell. These tests were conducted to determine melting parameters in preparation for a large-scale melting campaign to be conducted in the Bureau's 1-metric ton (mt), water-cooled-wall, 3-phase electric arc furnace. Bulk chemical composition was determined for each of the feed materials and for the slag, metal, fume solids, and offgas furnace products, and distributions were calculated for the key elements. The material balance for the furnace operation indicates that from 63 to 84 pct of the feed reported to the slag. Cerium, used as the surrogate for the radionuclides in the wastes, demonstrated an extremely strong affinity for the slag product. Although slag temperatures as low as 1,250 C were recorded when melting the RWMC soil, temperatures in excess of 1,600 C were necessary to achieve the fluidity required for a successful slag tap

  13. Evaluating Temporal Consistency in Marine Biodiversity Hotspots.

    Science.gov (United States)

    Piacenza, Susan E; Thurman, Lindsey L; Barner, Allison K; Benkwitt, Cassandra E; Boersma, Kate S; Cerny-Chipman, Elizabeth B; Ingeman, Kurt E; Kindinger, Tye L; Lindsley, Amy J; Nelson, Jake; Reimer, Jessica N; Rowe, Jennifer C; Shen, Chenchen; Thompson, Kevin A; Heppell, Selina S

    2015-01-01

    With the ongoing crisis of biodiversity loss and limited resources for conservation, the concept of biodiversity hotspots has been useful in determining conservation priority areas. However, there has been limited research into how temporal variability in biodiversity may influence conservation area prioritization. To address this information gap, we present an approach to evaluate the temporal consistency of biodiversity hotspots in large marine ecosystems. Using a large scale, public monitoring dataset collected over an eight year period off the US Pacific Coast, we developed a methodological approach for avoiding biases associated with hotspot delineation. We aggregated benthic fish species data from research trawls and calculated mean hotspot thresholds for fish species richness and Shannon's diversity indices over the eight year dataset. We used a spatial frequency distribution method to assign hotspot designations to the grid cells annually. We found no areas containing consistently high biodiversity through the entire study period based on the mean thresholds, and no grid cell was designated as a hotspot for greater than 50% of the time-series. To test if our approach was sensitive to sampling effort and the geographic extent of the survey, we followed a similar routine for the northern region of the survey area. Our finding of low consistency in benthic fish biodiversity hotspots over time was upheld, regardless of biodiversity metric used, whether thresholds were calculated per year or across all years, or the spatial extent for which we calculated thresholds and identified hotspots. Our results suggest that static measures of benthic fish biodiversity off the US West Coast are insufficient for identification of hotspots and that long-term data are required to appropriately identify patterns of high temporal variability in biodiversity for these highly mobile taxa. Given that ecological communities are responding to a changing climate and other

  14. Sludge characterization: the role of physical consistency

    Energy Technology Data Exchange (ETDEWEB)

    Spinosa, Ludovico; Wichmann, Knut

    2003-07-01

    The physical consistency is an important parameter in sewage sludge characterization as it strongly affects almost all treatment, utilization and disposal operations. In addition, in many european Directives a reference to the physical consistency is reported as a characteristic to be evaluated for fulfilling the regulations requirements. Further, in many analytical methods for sludge different procedures are indicated depending on whether a sample is liquid or not, is solid or not. Three physical behaviours (liquid, paste-like and solid) can be observed with sludges, so the development of analytical procedures to define the boundary limit between liquid and paste-like behaviours (flowability) and that between solid and paste-like ones (solidity) is of growing interest. Several devices can be used for evaluating the flowability and solidity properties, but often they are costly and difficult to be operated in the field. Tests have been carried out to evaluate the possibility to adopt a simple extrusion procedure for flowability measurements, and a Vicat needle for solidity ones. (author)

  15. Consistency of extreme flood estimation approaches

    Science.gov (United States)

    Felder, Guido; Paquet, Emmanuel; Penot, David; Zischg, Andreas; Weingartner, Rolf

    2017-04-01

    Estimations of low-probability flood events are frequently used for the planning of infrastructure as well as for determining the dimensions of flood protection measures. There are several well-established methodical procedures to estimate low-probability floods. However, a global assessment of the consistency of these methods is difficult to achieve, the "true value" of an extreme flood being not observable. Anyway, a detailed comparison performed on a given case study brings useful information about the statistical and hydrological processes involved in different methods. In this study, the following three different approaches for estimating low-probability floods are compared: a purely statistical approach (ordinary extreme value statistics), a statistical approach based on stochastic rainfall-runoff simulation (SCHADEX method), and a deterministic approach (physically based PMF estimation). These methods are tested for two different Swiss catchments. The results and some intermediate variables are used for assessing potential strengths and weaknesses of each method, as well as for evaluating the consistency of these methods.

  16. Personality consistency analysis in cloned quarantine dog candidates

    Directory of Open Access Journals (Sweden)

    Jin Choi

    2017-01-01

    Full Text Available In recent research, personality consistency has become an important characteristic. Diverse traits and human-animal interactions, in particular, are studied in the field of personality consistency in dogs. Here, we investigated the consistency of dominant behaviours in cloned and control groups followed by the modified Puppy Aptitude Test, which consists of ten subtests to ascertain the influence of genetic identity. In this test, puppies are exposed to stranger, restraint, prey-like object, noise, startling object, etc. Six cloned and four control puppies participated and the consistency of responses at ages 7–10 and 16 weeks in the two groups was compared. The two groups showed different consistencies in the subtests. While the average scores of the cloned group were consistent (P = 0.7991, those of the control group were not (P = 0.0089. Scores of Pack Drive and Fight or Flight Drive were consistent in the cloned group, however, those of the control group were not. Scores of Prey Drive were not consistent in either the cloned or the control group. Therefore, it is suggested that consistency of dominant behaviour is affected by genetic identity and some behaviours can be influenced more than others. Our results suggest that cloned dogs could show more consistent traits than non-cloned. This study implies that personality consistency could be one of the ways to analyse traits of puppies.

  17. Hanford Immobilized LAW Product Acceptance: Tanks Focus Area Testing Data Package II

    International Nuclear Information System (INIS)

    Schulz, Rebecca L.; Lorier, Troy H.; Peeler, David K.; Brown, Kevin G.; Reamer, Irene A.; Vienna, John D.; Jiricka, Antonin; Jorgensen, Benaiah M.; Smith, Donald E.

    2001-01-01

    This report is a continuation of the Hanford Immobilized Low Activity Waste (LAW) Product Acceptance (HLP): Initial Tanks Focus Area Testing Data Package (Vienna (and others) 2000). In addition to new 5000-h product consistency test (PCT), vapor hydration test (VHT), and alteration products data, some previously reported data together with relevant background information are included for an easily accessible source of reference when comparing the response of the various glasses to different test conditions. A matrix of 55 glasses was developed and tested to identify the impact of glass composition on long-term corrosion behavior and to develop an acceptable composition region for Hanford LAW glasses. Of the 55 glasses, 45 were designed to systematically vary the glass composition, and 10 were selected because large and growing databases on their corrosion characteristics had accumulated. The targeted and measured compositions of these glasses are found in the Appendix A. All glasses were fabricated according to standard procedures and heat treated to simulate the slow cooling that will occur in a portion of the waste glass after vitrification in the planned treatment facility at Hanford

  18. Speed Consistency in the Smart Tachograph.

    Science.gov (United States)

    Borio, Daniele; Cano, Eduardo; Baldini, Gianmarco

    2018-05-16

    In the transportation sector, safety risks can be significantly reduced by monitoring the behaviour of drivers and by discouraging possible misconducts that entail fatigue and can increase the possibility of accidents. The Smart Tachograph (ST), the new revision of the Digital Tachograph (DT), has been designed with this purpose: to verify that speed limits and compulsory rest periods are respected by drivers. In order to operate properly, the ST periodically checks the consistency of data from different sensors, which can be potentially manipulated to avoid the monitoring of the driver behaviour. In this respect, the ST regulation specifies a test procedure to detect motion conflicts originating from inconsistencies between Global Navigation Satellite System (GNSS) and odometry data. This paper provides an experimental evaluation of the speed verification procedure specified by the ST regulation. Several hours of data were collected using three vehicles and considering light urban and highway environments. The vehicles were equipped with an On-Board Diagnostics (OBD) data reader and a GPS/Galileo receiver. The tests prescribed by the regulation were implemented with specific focus on synchronization aspects. The experimental analysis also considered aspects such as the impact of tunnels and the presence of data gaps. The analysis shows that the metrics selected for the tests are resilient to data gaps, latencies between GNSS and odometry data and simplistic manipulations such as data scaling. The new ST forces an attacker to falsify data from both sensors at the same time and in a coherent way. This makes more difficult the implementation of frauds in comparison to the current version of the DT.

  19. Procalcitonin as a biomarker for severe Plasmodium falciparum disease: a critical appraisal of a semi-quantitative point-of-care test in a cohort of travellers with imported malaria.

    Science.gov (United States)

    Hesselink, Dennis A; Burgerhart, Jan-Steven; Bosmans-Timmerarends, Hanna; Petit, Pieter; van Genderen, Perry J J

    2009-09-01

    Imported malaria occurs as a relatively rare event in developed countries. As a consequence, most clinicians have little experience in making clinical assessments of disease severity and decisions regarding the need for parenteral therapy or high-level monitoring. In this study, the diagnostic accuracy of procalcitonin (PCT) for severe Plasmodium falciparum disease was assessed in a cohort of 100 consecutive travellers with various species of imported malaria. In all patients, PCT was measured on admission with a semi-quantitative 'point-of-care' test. Patients with severe P. falciparum malaria had significantly higher median PCT levels on admission as compared with patients with uncomplicated P. falciparum disease. In addition, PCT levels in patients with non-falciparum malaria were also higher compared with patients with non-severe falciparum malaria but lower compared with severe P. falciparum malaria. At a cut-off point of 10 ng/mL, PCT had a sensitivity of 0,67 and a specificity of 0,94 for severe falciparum disease. However, at lower cut-off points the specificity and positive predictive value were rather poor although the sensitivity and negative predictive value remained high. Potential drawbacks in the interpretation of elevated PCT levels on admission may be caused by infections with non-falciparum species and by concomitant bacterial infections. Semi-quantitative determination of PCT on admission is of limited use in the initial clinical assessment of disease severity in travellers with imported malaria, especially in settings with limited experience with the treatment of malaria.

  20. Procalcitonin as a biomarker for severe Plasmodium falciparum disease: a critical appraisal of a semi-quantitative point-of-care test in a cohort of travellers with imported malaria

    Directory of Open Access Journals (Sweden)

    Petit Pieter

    2009-09-01

    Full Text Available Abstract Background Imported malaria occurs as a relatively rare event in developed countries. As a consequence, most clinicians have little experience in making clinical assessments of disease severity and decisions regarding the need for parenteral therapy or high-level monitoring. In this study, the diagnostic accuracy of procalcitonin (PCT for severe Plasmodium falciparum disease was assessed in a cohort of 100 consecutive travellers with various species of imported malaria. Methods and results In all patients, PCT was measured on admission with a semi-quantitative 'point-of-care' test. Patients with severe P. falciparum malaria had significantly higher median PCT levels on admission as compared with patients with uncomplicated P. falciparum disease. In addition, PCT levels in patients with non-falciparum malaria were also higher compared with patients with non-severe falciparum malaria but lower compared with severe P. falciparum malaria. At a cut-off point of 10 ng/mL, PCT had a sensitivity of 0,67 and a specificity of 0,94 for severe falciparum disease. However, at lower cut-off points the specificity and positive predictive value were rather poor although the sensitivity and negative predictive value remained high. Discussion Potential drawbacks in the interpretation of elevated PCT levels on admission may be caused by infections with non-falciparum species and by concomitant bacterial infections. Conclusion Semi-quantitative determination of PCT on admission is of limited use in the initial clinical assessment of disease severity in travellers with imported malaria, especially in settings with limited experience with the treatment of malaria.

  1. A consistent thermodynamic database for cement minerals

    International Nuclear Information System (INIS)

    Blanc, P.; Claret, F.; Burnol, A.; Marty, N.; Gaboreau, S.; Tournassat, C.; Gaucher, E.C.; Giffault, E.; Bourbon, X.

    2010-01-01

    work - the formation enthalpy and the Cp(T) function are taken from the literature or estimated - finally, the Log K(T) function is calculated, based on the selected dataset and it is compared to experimental data gathered at different temperatures. Each experimental point is extracted from solution compositions by using PHREEQC with a selection of aqueous complexes, consistent with the Thermochimie database. The selection was tested namely by drawing activity diagrams, allowing to assess phases relations. An example of such a diagram, drawn in the CaO-Al 2 O 3 -SiO 2 -H 2 O system is displayed. It can be seen that low pH concrete alteration proceeds essentially in decreasing the C/S ratio in C-S-H phases to the point where C-S-H are no longer stable and replaced by zeolite, then clay minerals. This evolution corresponds to a decrease in silica activity, which is consistent with the pH decrease, as silica concentration depends essentially on pH. Some rather consistent phase relations have been obtained for the SO 3 -Al 2 O 3 -CaO-CO 2 -H 2 O system. Addition of iron III enlarges the AFm-SO 4 stability field to the low temperature domain, whereas it decreases the pH domain where ettringite is stable. On the other hand, the stability field of katoite remains largely ambiguous, namely with respect to a hydro-garnet/grossular solid solution. With respect to other databases this work was made in consistency with a larger mineral selection, so that it can be used for modelling works in the cement clay interaction context

  2. Comparison of PCT, CRP, D-Dimer, Lactate, TNF-α, IL-1β, IL-6 and lL-10 in Development of Systemic Inflammatory Response Syndrome and Sepsis on Patients with Isolated Head Trauma and Polytrauma

    Directory of Open Access Journals (Sweden)

    Emine Dağlı

    2012-12-01

    Full Text Available Objective: In this study, it was aimed to compare the relationship between Glasgow Coma Scale (GKS, ISS values, PCT, CRP, D-Dimer, laktat, TNF-α, IL-1β, IL-6, IL-10 in patients with polytrauma and isolated head trauma in conjunction with SIRS and sepsis. Material and Method: Total of 68 patients (32 polytrauma, 36 isolated head trauma aged between 18-65 years were enrolled in the study. For 7 days of follow up, the biochemical parameters were analysed on the days 0, 1, 3, 5, 7 and the ISS, GCS score and growth rates of SIRS and sepsis were recorded. Results: It was demonstrated that for patients with isolated head trauma, SIRS (80.6%, sepsis (38.9% and mortality values (71.4% were higher and there is a statistically important linear and inverse relationship between ISS and GCS values (p<0.05. During sepsis and SIRS phases, CRP elevated in both groups, whereas PCT only in the polytrauma group. D-Dimer values were investigated high in both groups, but atypically decreased on day 3 in isolated head trauma group. Lactate, TNF-α and, IL-1β were within reference values, IL-6 and IL-10 values were elevated in both groups but were higher in the polytrauma group. Conclusion: Although multiple variations were detected in serum markers of pro-inflamatory and acute phase proteins, we thought that these are inadequate in predicting mortality and complications such as SIRS and sepsis.

  3. Consistency Anchor Formalization and Correctness Proofs

    OpenAIRE

    Miguel, Correia; Bessani, Alysson

    2014-01-01

    This is report contains the formal proofs for the techniques for increasing the consistency of cloud storage as presented in "Bessani et al. SCFS: A Cloud-backed File System. Proc. of the 2014 USENIX Annual Technical Conference. June 2014." The consistency anchor technique allows one to increase the consistency provided by eventually consistent cloud storage services like Amazon S3. This technique has been used in the SCFS (Shared Cloud File System) cloud-backed file system for solving rea...

  4. Personality consistency in dogs: a meta-analysis.

    Science.gov (United States)

    Fratkin, Jamie L; Sinn, David L; Patall, Erika A; Gosling, Samuel D

    2013-01-01

    Personality, or consistent individual differences in behavior, is well established in studies of dogs. Such consistency implies predictability of behavior, but some recent research suggests that predictability cannot be assumed. In addition, anecdotally, many dog experts believe that 'puppy tests' measuring behavior during the first year of a dog's life are not accurate indicators of subsequent adult behavior. Personality consistency in dogs is an important aspect of human-dog relationships (e.g., when selecting dogs suitable for substance-detection work or placement in a family). Here we perform the first comprehensive meta-analysis of studies reporting estimates of temporal consistency of dog personality. A thorough literature search identified 31 studies suitable for inclusion in our meta-analysis. Overall, we found evidence to suggest substantial consistency (r = 0.43). Furthermore, personality consistency was higher in older dogs, when behavioral assessment intervals were shorter, and when the measurement tool was exactly the same in both assessments. In puppies, aggression and submissiveness were the most consistent dimensions, while responsiveness to training, fearfulness, and sociability were the least consistent dimensions. In adult dogs, there were no dimension-based differences in consistency. There was no difference in personality consistency in dogs tested first as puppies and later as adults (e.g., 'puppy tests') versus dogs tested first as puppies and later again as puppies. Finally, there were no differences in consistency between working versus non-working dogs, between behavioral codings versus behavioral ratings, and between aggregate versus single measures. Implications for theory, practice, and future research are discussed.

  5. Personality consistency in dogs: a meta-analysis.

    Directory of Open Access Journals (Sweden)

    Jamie L Fratkin

    Full Text Available Personality, or consistent individual differences in behavior, is well established in studies of dogs. Such consistency implies predictability of behavior, but some recent research suggests that predictability cannot be assumed. In addition, anecdotally, many dog experts believe that 'puppy tests' measuring behavior during the first year of a dog's life are not accurate indicators of subsequent adult behavior. Personality consistency in dogs is an important aspect of human-dog relationships (e.g., when selecting dogs suitable for substance-detection work or placement in a family. Here we perform the first comprehensive meta-analysis of studies reporting estimates of temporal consistency of dog personality. A thorough literature search identified 31 studies suitable for inclusion in our meta-analysis. Overall, we found evidence to suggest substantial consistency (r = 0.43. Furthermore, personality consistency was higher in older dogs, when behavioral assessment intervals were shorter, and when the measurement tool was exactly the same in both assessments. In puppies, aggression and submissiveness were the most consistent dimensions, while responsiveness to training, fearfulness, and sociability were the least consistent dimensions. In adult dogs, there were no dimension-based differences in consistency. There was no difference in personality consistency in dogs tested first as puppies and later as adults (e.g., 'puppy tests' versus dogs tested first as puppies and later again as puppies. Finally, there were no differences in consistency between working versus non-working dogs, between behavioral codings versus behavioral ratings, and between aggregate versus single measures. Implications for theory, practice, and future research are discussed.

  6. Consistency of hand preference: predictions to intelligence and school achievement.

    Science.gov (United States)

    Kee, D W; Gottfried, A; Bathurst, K

    1991-05-01

    Gottfried and Bathurst (1983) reported that hand preference consistency measured over time during infancy and early childhood predicts intellectual precocity for females, but not for males. In the present study longitudinal assessments of children previously classified by Gottfried and Bathurst as consistent or nonconsistent in cross-time hand preference were conducted during middle childhood (ages 5 to 9). Findings show that (a) early measurement of hand preference consistency for females predicts school-age intellectual precocity, (b) the locus of the difference between consistent vs. nonconsistent females is in verbal intelligence, and (c) the precocity of the consistent females was also revealed on tests of school achievement, particularly tests of reading and mathematics.

  7. A new approach to hull consistency

    Directory of Open Access Journals (Sweden)

    Kolev Lubomir

    2016-06-01

    Full Text Available Hull consistency is a known technique to improve the efficiency of iterative interval methods for solving nonlinear systems describing steady-states in various circuits. Presently, hull consistency is checked in a scalar manner, i.e. successively for each equation of the nonlinear system with respect to a single variable. In the present poster, a new more general approach to implementing hull consistency is suggested which consists in treating simultaneously several equations with respect to the same number of variables.

  8. Replica consistency in a Data Grid

    International Nuclear Information System (INIS)

    Domenici, Andrea; Donno, Flavia; Pucciani, Gianni; Stockinger, Heinz; Stockinger, Kurt

    2004-01-01

    A Data Grid is a wide area computing infrastructure that employs Grid technologies to provide storage capacity and processing power to applications that handle very large quantities of data. Data Grids rely on data replication to achieve better performance and reliability by storing copies of data sets on different Grid nodes. When a data set can be modified by applications, the problem of maintaining consistency among existing copies arises. The consistency problem also concerns metadata, i.e., additional information about application data sets such as indices, directories, or catalogues. This kind of metadata is used both by the applications and by the Grid middleware to manage the data. For instance, the Replica Management Service (the Grid middleware component that controls data replication) uses catalogues to find the replicas of each data set. Such catalogues can also be replicated and their consistency is crucial to the correct operation of the Grid. Therefore, metadata consistency generally poses stricter requirements than data consistency. In this paper we report on the development of a Replica Consistency Service based on the middleware mainly developed by the European Data Grid Project. The paper summarises the main issues in the replica consistency problem, and lays out a high-level architectural design for a Replica Consistency Service. Finally, results from simulations of different consistency models are presented

  9. Field test results for radioactive waste drum characterization with Waste Inspection Tomography (WIT)

    Energy Technology Data Exchange (ETDEWEB)

    Bernardi, R.T. [Bio-Imaging Research, Inc., Lincolnshire, IL (United States)

    1997-11-01

    This paper summarizes the design, fabrication, factory testing, evaluation and demonstration of waste inspection tomography (WIT). WIT consists of a self-sufficient, mobile semi-trailer for Non-Destructive Evaluation and Non-Destructive Assay (NDE/NDA) characterization of nuclear waste drums using X-ray and gamma-ray tomographic techniques. The 23-month WIT Phase I initial test results include 2 MeV Digital Radiography (DR), Computed Tomography (CT), Anger camera imaging, Single Photon Emission Computed Tomography (SPECT), Gamma-Ray Spectroscopy, Collimated Gamma Scanning (CGS), and Active and Passive Computed Tomography (A&PCT) using a 1.4 mCi source of {sup 166}Ho. These techniques were initially demonstrated on a 55-gallon phantom drum with three simulated waste matrices of combustibles, heterogeneous metals, and cement using check sources of gamma active isotopes. Waste matrix identification, isotopic identification, and attenuation-corrected gamma activity determination were all demonstrated nondestructively and noninvasively. Preliminary field tests results with nuclear waste drums are summarized. WIT has inspected drums with 0 to 20 grams plutonium 239. The minimum measured was 0.131 gram plutonium 239 in cement. 8 figs.

  10. Student Effort, Consistency, and Online Performance

    Science.gov (United States)

    Patron, Hilde; Lopez, Salvador

    2011-01-01

    This paper examines how student effort, consistency, motivation, and marginal learning, influence student grades in an online course. We use data from eleven Microeconomics courses taught online for a total of 212 students. Our findings show that consistency, or less time variation, is a statistically significant explanatory variable, whereas…

  11. Translationally invariant self-consistent field theories

    International Nuclear Information System (INIS)

    Shakin, C.M.; Weiss, M.S.

    1977-01-01

    We present a self-consistent field theory which is translationally invariant. The equations obtained go over to the usual Hartree-Fock equations in the limit of large particle number. In addition to deriving the dynamic equations for the self-consistent amplitudes we discuss the calculation of form factors and various other observables

  12. Sticky continuous processes have consistent price systems

    DEFF Research Database (Denmark)

    Bender, Christian; Pakkanen, Mikko; Sayit, Hasanjan

    Under proportional transaction costs, a price process is said to have a consistent price system, if there is a semimartingale with an equivalent martingale measure that evolves within the bid-ask spread. We show that a continuous, multi-asset price process has a consistent price system, under...

  13. Consistent-handed individuals are more authoritarian.

    Science.gov (United States)

    Lyle, Keith B; Grillo, Michael C

    2014-01-01

    Individuals differ in the consistency with which they use one hand over the other to perform everyday activities. Some individuals are very consistent, habitually using a single hand to perform most tasks. Others are relatively inconsistent, and hence make greater use of both hands. More- versus less-consistent individuals have been shown to differ in numerous aspects of personality and cognition. In several respects consistent-handed individuals resemble authoritarian individuals. For example, both consistent-handedness and authoritarianism have been linked to cognitive inflexibility. Therefore we hypothesised that consistent-handedness is an external marker for authoritarianism. Confirming our hypothesis, we found that consistent-handers scored higher than inconsistent-handers on a measure of submission to authority, were more likely to identify with a conservative political party (Republican), and expressed less-positive attitudes towards out-groups. We propose that authoritarianism may be influenced by the degree of interaction between the left and right brain hemispheres, which has been found to differ between consistent- and inconsistent-handed individuals.

  14. Consistent spectroscopy for a extended gauge model

    International Nuclear Information System (INIS)

    Oliveira Neto, G. de.

    1990-11-01

    The consistent spectroscopy was obtained with a Lagrangian constructed with vector fields with a U(1) group extended symmetry. As consistent spectroscopy is understood the determination of quantum physical properties described by the model in an manner independent from the possible parametrizations adopted in their description. (L.C.J.A.)

  15. Personality Consistency in Dogs: A Meta-Analysis

    Science.gov (United States)

    Fratkin, Jamie L.; Sinn, David L.; Patall, Erika A.; Gosling, Samuel D.

    2013-01-01

    Personality, or consistent individual differences in behavior, is well established in studies of dogs. Such consistency implies predictability of behavior, but some recent research suggests that predictability cannot be assumed. In addition, anecdotally, many dog experts believe that ‘puppy tests’ measuring behavior during the first year of a dog's life are not accurate indicators of subsequent adult behavior. Personality consistency in dogs is an important aspect of human-dog relationships (e.g., when selecting dogs suitable for substance-detection work or placement in a family). Here we perform the first comprehensive meta-analysis of studies reporting estimates of temporal consistency of dog personality. A thorough literature search identified 31 studies suitable for inclusion in our meta-analysis. Overall, we found evidence to suggest substantial consistency (r = 0.43). Furthermore, personality consistency was higher in older dogs, when behavioral assessment intervals were shorter, and when the measurement tool was exactly the same in both assessments. In puppies, aggression and submissiveness were the most consistent dimensions, while responsiveness to training, fearfulness, and sociability were the least consistent dimensions. In adult dogs, there were no dimension-based differences in consistency. There was no difference in personality consistency in dogs tested first as puppies and later as adults (e.g., ‘puppy tests’) versus dogs tested first as puppies and later again as puppies. Finally, there were no differences in consistency between working versus non-working dogs, between behavioral codings versus behavioral ratings, and between aggregate versus single measures. Implications for theory, practice, and future research are discussed. PMID:23372787

  16. Self-consistency corrections in effective-interaction calculations

    International Nuclear Information System (INIS)

    Starkand, Y.; Kirson, M.W.

    1975-01-01

    Large-matrix extended-shell-model calculations are used to compute self-consistency corrections to the effective interaction and to the linked-cluster effective interaction. The corrections are found to be numerically significant and to affect the rate of convergence of the corresponding perturbation series. The influence of various partial corrections is tested. It is concluded that self-consistency is an important effect in determining the effective interaction and improving the rate of convergence. (author)

  17. Consistency in the World Wide Web

    DEFF Research Database (Denmark)

    Thomsen, Jakob Grauenkjær

    Tim Berners-Lee envisioned that computers will behave as agents of humans on the World Wide Web, where they will retrieve, extract, and interact with information from the World Wide Web. A step towards this vision is to make computers capable of extracting this information in a reliable...... and consistent way. In this dissertation we study steps towards this vision by showing techniques for the specication, the verication and the evaluation of the consistency of information in the World Wide Web. We show how to detect certain classes of errors in a specication of information, and we show how...... the World Wide Web, in order to help perform consistent evaluations of web extraction techniques. These contributions are steps towards having computers reliable and consistently extract information from the World Wide Web, which in turn are steps towards achieving Tim Berners-Lee's vision. ii...

  18. Consistent histories and operational quantum theory

    International Nuclear Information System (INIS)

    Rudolph, O.

    1996-01-01

    In this work a generalization of the consistent histories approach to quantum mechanics is presented. We first critically review the consistent histories approach to nonrelativistic quantum mechanics in a mathematically rigorous way and give some general comments about it. We investigate to what extent the consistent histories scheme is compatible with the results of the operational formulation of quantum mechanics. According to the operational approach, nonrelativistic quantum mechanics is most generally formulated in terms of effects, states, and operations. We formulate a generalized consistent histories theory using the concepts and the terminology which have proven useful in the operational formulation of quantum mechanics. The logical rule of the logical interpretation of quantum mechanics is generalized to the present context. The algebraic structure of the generalized theory is studied in detail

  19. Self-consistent areas law in QCD

    International Nuclear Information System (INIS)

    Makeenko, Yu.M.; Migdal, A.A.

    1980-01-01

    The problem of obtaining the self-consistent areas law in quantum chromodynamics (QCD) is considered from the point of view of the quark confinement. The exact equation for the loop average in multicolor QCD is reduced to a bootstrap form. Its iterations yield new manifestly gauge invariant perturbation theory in the loop space, reproducing asymptotic freedom. For large loops, the areas law apprears to be a self-consistent solution

  20. Consistency of the MLE under mixture models

    OpenAIRE

    Chen, Jiahua

    2016-01-01

    The large-sample properties of likelihood-based statistical inference under mixture models have received much attention from statisticians. Although the consistency of the nonparametric MLE is regarded as a standard conclusion, many researchers ignore the precise conditions required on the mixture model. An incorrect claim of consistency can lead to false conclusions even if the mixture model under investigation seems well behaved. Under a finite normal mixture model, for instance, the consis...

  1. Self-consistent asset pricing models

    Science.gov (United States)

    Malevergne, Y.; Sornette, D.

    2007-08-01

    We discuss the foundations of factor or regression models in the light of the self-consistency condition that the market portfolio (and more generally the risk factors) is (are) constituted of the assets whose returns it is (they are) supposed to explain. As already reported in several articles, self-consistency implies correlations between the return disturbances. As a consequence, the alphas and betas of the factor model are unobservable. Self-consistency leads to renormalized betas with zero effective alphas, which are observable with standard OLS regressions. When the conditions derived from internal consistency are not met, the model is necessarily incomplete, which means that some sources of risk cannot be replicated (or hedged) by a portfolio of stocks traded on the market, even for infinite economies. Analytical derivations and numerical simulations show that, for arbitrary choices of the proxy which are different from the true market portfolio, a modified linear regression holds with a non-zero value αi at the origin between an asset i's return and the proxy's return. Self-consistency also introduces “orthogonality” and “normality” conditions linking the betas, alphas (as well as the residuals) and the weights of the proxy portfolio. Two diagnostics based on these orthogonality and normality conditions are implemented on a basket of 323 assets which have been components of the S&P500 in the period from January 1990 to February 2005. These two diagnostics show interesting departures from dynamical self-consistency starting about 2 years before the end of the Internet bubble. Assuming that the CAPM holds with the self-consistency condition, the OLS method automatically obeys the resulting orthogonality and normality conditions and therefore provides a simple way to self-consistently assess the parameters of the model by using proxy portfolios made only of the assets which are used in the CAPM regressions. Finally, the factor decomposition with the

  2. Towards thermodynamical consistency of quasiparticle picture

    International Nuclear Information System (INIS)

    Biro, T.S.; Shanenko, A.A.; Toneev, V.D.; Research Inst. for Particle and Nuclear Physics, Hungarian Academy of Sciences, Budapest

    2003-01-01

    The purpose of the present article is to call attention to some realistic quasi-particle-based description of the quark/gluon matter and its consistent implementation in thermodynamics. A simple and transparent representation of the thermodynamical consistency conditions is given. This representation allows one to review critically and systemize available phenomenological approaches to the deconfinement problem with respect to their thermodynamical consistency. A particular attention is paid to the development of a method for treating the string screening in the dense matter of unbound color charges. The proposed method yields an integrable effective pair potential, which can be incorporated into the mean-field picture. The results of its application are in reasonable agreement with lattice data on the QCD thermodynamics [ru

  3. Toward thermodynamic consistency of quasiparticle picture

    International Nuclear Information System (INIS)

    Biro, T.S.; Toneev, V.D.; Shanenko, A.A.

    2003-01-01

    The purpose of the present article is to call attention to some realistic quasiparticle-based description of quark/gluon matter and its consistent implementation in thermodynamics. A simple and transparent representation of the thermodynamic consistency conditions is given. This representation allows one to review critically and systemize available phenomenological approaches to the deconfinement problem with respect to their thermodynamic consistency. Particular attention is paid to the development of a method for treating the string screening in the dense matter of unbound color charges. The proposed method yields an integrable effective pair potential that can be incorporated into the mean-field picture. The results of its application are in reasonable agreement with lattice data on the QCD thermodynamics

  4. Toward a consistent RHA-RPA

    International Nuclear Information System (INIS)

    Shepard, J.R.

    1991-01-01

    The authors examine the RPA based on a relativistic Hartree approximation description for nuclear ground states. This model includes contributions from the negative energy sea at the 1-loop level. They emphasize consistency between the treatment of the ground state and the RPA. This consistency is important in the description of low-lying collective levels but less important for the longitudinal (e, e') quasi-elastic response. They also study the effect of imposing a 3-momentum cutoff on negative energy sea contributions. A cutoff of twice the nucleon mass improves agreement with observed spin orbit splittings in nuclei compared to the standard infinite cutoff results, an effect traceable to the fact that imposing the cutoff reduces m*/m. The cutoff is much less important than consistency in the description of low-lying collective levels. The cutoff model provides excellent agreement with quasi-elastic (e, e') data

  5. Personalized recommendation based on unbiased consistence

    Science.gov (United States)

    Zhu, Xuzhen; Tian, Hui; Zhang, Ping; Hu, Zheng; Zhou, Tao

    2015-08-01

    Recently, in physical dynamics, mass-diffusion-based recommendation algorithms on bipartite network provide an efficient solution by automatically pushing possible relevant items to users according to their past preferences. However, traditional mass-diffusion-based algorithms just focus on unidirectional mass diffusion from objects having been collected to those which should be recommended, resulting in a biased causal similarity estimation and not-so-good performance. In this letter, we argue that in many cases, a user's interests are stable, and thus bidirectional mass diffusion abilities, no matter originated from objects having been collected or from those which should be recommended, should be consistently powerful, showing unbiased consistence. We further propose a consistence-based mass diffusion algorithm via bidirectional diffusion against biased causality, outperforming the state-of-the-art recommendation algorithms in disparate real data sets, including Netflix, MovieLens, Amazon and Rate Your Music.

  6. Financial model calibration using consistency hints.

    Science.gov (United States)

    Abu-Mostafa, Y S

    2001-01-01

    We introduce a technique for forcing the calibration of a financial model to produce valid parameters. The technique is based on learning from hints. It converts simple curve fitting into genuine calibration, where broad conclusions can be inferred from parameter values. The technique augments the error function of curve fitting with consistency hint error functions based on the Kullback-Leibler distance. We introduce an efficient EM-type optimization algorithm tailored to this technique. We also introduce other consistency hints, and balance their weights using canonical errors. We calibrate the correlated multifactor Vasicek model of interest rates, and apply it successfully to Japanese Yen swaps market and US dollar yield market.

  7. LOCA testing of high burnup PWR fuel in the HBWR. Additional PIE on the cladding of the segment 650-5

    Energy Technology Data Exchange (ETDEWEB)

    Oberlaender, B.C.; Espeland, M.; Jenssen, H.K.

    2008-07-01

    IFA-650.5, a test with pre-irradiated fuel in the Halden Project LOCA test series, was conducted on October 23rd, 2006. The fuel rod had been used in a commercial PWR and had a high burnup, 83 MWd/kgU. Experimental arrangements of the fifth test were similar to the preceding LOCA tests. The peak cladding temperature (PCT) level was higher than in the third and fourth tests, 1050 C. A peak temperature close to the target was achieved and cladding burst occurred at approx. 750 C. Within the joint programme framework of the Halden Project PIE was done, consisting of gamma scanning, visual inspection, neutron-radiography, hydrogen analysis and metallography / ceramography. An additional extensive PIE including metallography, hydrogen analysis, and hardness measurements of cross-sections at seven axial elevations was done. It was completed to study the high burnup and LOCA induced effects on the Zr-4 cladding, namely the migration of oxygen into the cladding from the inside surface, the cladding distension, and the burst (author)(tk)

  8. Proteolysis and consistency of Meshanger cheese

    NARCIS (Netherlands)

    Jong, de L.

    1978-01-01

    Proteolysis in Meshanger cheese, estimated by quantitative polyacrylamide gel electrophoresis is discussed. The conversion of α s1 -casein was proportional to rennet concentration in the cheese. Changes in consistency, after a maximum, were correlated to breakdown of

  9. Developing consistent pronunciation models for phonemic variants

    CSIR Research Space (South Africa)

    Davel, M

    2006-09-01

    Full Text Available Pronunciation lexicons often contain pronunciation variants. This can create two problems: It can be difficult to define these variants in an internally consistent way and it can also be difficult to extract generalised grapheme-to-phoneme rule sets...

  10. Consistent Valuation across Curves Using Pricing Kernels

    Directory of Open Access Journals (Sweden)

    Andrea Macrina

    2018-03-01

    Full Text Available The general problem of asset pricing when the discount rate differs from the rate at which an asset’s cash flows accrue is considered. A pricing kernel framework is used to model an economy that is segmented into distinct markets, each identified by a yield curve having its own market, credit and liquidity risk characteristics. The proposed framework precludes arbitrage within each market, while the definition of a curve-conversion factor process links all markets in a consistent arbitrage-free manner. A pricing formula is then derived, referred to as the across-curve pricing formula, which enables consistent valuation and hedging of financial instruments across curves (and markets. As a natural application, a consistent multi-curve framework is formulated for emerging and developed inter-bank swap markets, which highlights an important dual feature of the curve-conversion factor process. Given this multi-curve framework, existing multi-curve approaches based on HJM and rational pricing kernel models are recovered, reviewed and generalised and single-curve models extended. In another application, inflation-linked, currency-based and fixed-income hybrid securities are shown to be consistently valued using the across-curve valuation method.

  11. Guided color consistency optimization for image mosaicking

    Science.gov (United States)

    Xie, Renping; Xia, Menghan; Yao, Jian; Li, Li

    2018-01-01

    This paper studies the problem of color consistency correction for sequential images with diverse color characteristics. Existing algorithms try to adjust all images to minimize color differences among images under a unified energy framework, however, the results are prone to presenting a consistent but unnatural appearance when the color difference between images is large and diverse. In our approach, this problem is addressed effectively by providing a guided initial solution for the global consistency optimization, which avoids converging to a meaningless integrated solution. First of all, to obtain the reliable intensity correspondences in overlapping regions between image pairs, we creatively propose the histogram extreme point matching algorithm which is robust to image geometrical misalignment to some extents. In the absence of the extra reference information, the guided initial solution is learned from the major tone of the original images by searching some image subset as the reference, whose color characteristics will be transferred to the others via the paths of graph analysis. Thus, the final results via global adjustment will take on a consistent color similar to the appearance of the reference image subset. Several groups of convincing experiments on both the synthetic dataset and the challenging real ones sufficiently demonstrate that the proposed approach can achieve as good or even better results compared with the state-of-the-art approaches.

  12. Consistent application of codes and standards

    International Nuclear Information System (INIS)

    Scott, M.A.

    1989-01-01

    The guidelines presented in the US Department of Energy, General Design Criteria (DOE 6430.1A), and the Design and Evaluation Guidelines for Department of Energy Facilities Subject to Natural Phenomena Hazards (UCRL-15910) provide a consistent and well defined approach to determine the natural phenomena hazards loads for US Department of Energy site facilities. The guidelines for the application of loads combinations and allowables criteria are not as well defined and are more flexible in interpretation. This flexibility in the interpretation of load combinations can lead to conflict between the designer and overseer. The establishment of an efficient set of acceptable design criteria, based on US Department of Energy guidelines, provides a consistent baseline for analysis, design, and review. Additionally, the proposed method should not limit the design and analytical innovation necessary to analyze or qualify the unique structure. This paper investigates the consistent application of load combinations, analytical methods, and load allowables and suggests a reference path consistent with the US Department of Energy guidelines

  13. Consistency in multi-viewpoint architectural design

    NARCIS (Netherlands)

    Dijkman, R.M.; Dijkman, Remco Matthijs

    2006-01-01

    This thesis presents a framework that aids in preserving consistency in multi-viewpoint designs. In a multi-viewpoint design each stakeholder constructs his own design part. We call each stakeholder’s design part the view of that stakeholder. To construct his view, a stakeholder has a viewpoint.

  14. Consistent Visual Analyses of Intrasubject Data

    Science.gov (United States)

    Kahng, SungWoo; Chung, Kyong-Mee; Gutshall, Katharine; Pitts, Steven C.; Kao, Joyce; Girolami, Kelli

    2010-01-01

    Visual inspection of single-case data is the primary method of interpretation of the effects of an independent variable on a dependent variable in applied behavior analysis. The purpose of the current study was to replicate and extend the results of DeProspero and Cohen (1979) by reexamining the consistency of visual analysis across raters. We…

  15. Consistent Stochastic Modelling of Meteocean Design Parameters

    DEFF Research Database (Denmark)

    Sørensen, John Dalsgaard; Sterndorff, M. J.

    2000-01-01

    Consistent stochastic models of metocean design parameters and their directional dependencies are essential for reliability assessment of offshore structures. In this paper a stochastic model for the annual maximum values of the significant wave height, and the associated wind velocity, current...

  16. On the existence of consistent price systems

    DEFF Research Database (Denmark)

    Bayraktar, Erhan; Pakkanen, Mikko S.; Sayit, Hasanjan

    2014-01-01

    We formulate a sufficient condition for the existence of a consistent price system (CPS), which is weaker than the conditional full support condition (CFS). We use the new condition to show the existence of CPSs for certain processes that fail to have the CFS property. In particular this condition...

  17. Dynamic phonon exchange requires consistent dressing

    International Nuclear Information System (INIS)

    Hahne, F.J.W.; Engelbrecht, C.A.; Heiss, W.D.

    1976-01-01

    It is shown that states with undersirable properties (such as ghosts, states with complex eigenenergies and states with unrestricted normalization) emerge from two-body calculations using dynamic effective interactions if one is not careful in introducing single-particle self-energy insertions in a consistent manner

  18. Consistent feeding positions of great tit parents

    NARCIS (Netherlands)

    Lessells, C.M.; Poelman, E.H.; Mateman, A.C.; Cassey, Ph.

    2006-01-01

    When parent birds arrive at the nest to provision their young, their position on the nest rim may influence which chick or chicks are fed. As a result, the consistency of feeding positions of the individual parents, and the difference in position between the parents, may affect how equitably food is

  19. Consistency of the postulates of special relativity

    International Nuclear Information System (INIS)

    Gron, O.; Nicola, M.

    1976-01-01

    In a recent article in this journal, Kingsley has tried to show that the postulates of special relativity contradict each other. It is shown that the arguments of Kingsley are invalid because of an erroneous appeal to symmetry in a nonsymmetric situation. The consistency of the postulates of special relativity and the relativistic kinematics deduced from them is restated

  20. Consistency of Network Traffic Repositories: An Overview

    NARCIS (Netherlands)

    Lastdrager, E.; Lastdrager, E.E.H.; Pras, Aiko

    2009-01-01

    Traffc repositories with TCP/IP header information are very important for network analysis. Researchers often assume that such repositories reliably represent all traffc that has been flowing over the network; little thoughts are made regarding the consistency of these repositories. Still, for

  1. Consistency analysis of network traffic repositories

    NARCIS (Netherlands)

    Lastdrager, Elmer; Lastdrager, E.E.H.; Pras, Aiko

    Traffic repositories with TCP/IP header information are very important for network analysis. Researchers often assume that such repositories reliably represent all traffic that has been flowing over the network; little thoughts are made regarding the consistency of these repositories. Still, for

  2. Personality and Situation Predictors of Consistent Eating Patterns.

    Science.gov (United States)

    Vainik, Uku; Dubé, Laurette; Lu, Ji; Fellows, Lesley K

    2015-01-01

    A consistent eating style might be beneficial to avoid overeating in a food-rich environment. Eating consistency entails maintaining a similar dietary pattern across different eating situations. This construct is relatively under-studied, but the available evidence suggests that eating consistency supports successful weight maintenance and decreases risk for metabolic syndrome and cardiovascular disease. Yet, personality and situation predictors of consistency have not been studied. A community-based sample of 164 women completed various personality tests, and 139 of them also reported their eating behaviour 6 times/day over 10 observational days. We focused on observations with meals (breakfast, lunch, or dinner). The participants indicated if their momentary eating patterns were consistent with their own baseline eating patterns in terms of healthiness or size of the meal. Further, participants described various characteristics of each eating situation. Eating consistency was positively predicted by trait self-control. Eating consistency was undermined by eating in the evening, eating with others, eating away from home, having consumed alcohol and having undertaken physical exercise. Interactions emerged between personality traits and situations, including punishment sensitivity, restraint, physical activity and alcohol consumption. Trait self-control and several eating situation variables were related to eating consistency. These findings provide a starting point for targeting interventions to improve consistency, suggesting that a focus on self-control skills, together with addressing contextual factors such as social situations and time of day, may be most promising. This work is a first step to provide people with the tools they need to maintain a consistently healthy lifestyle in a food-rich environment.

  3. Personality and Situation Predictors of Consistent Eating Patterns.

    Directory of Open Access Journals (Sweden)

    Uku Vainik

    Full Text Available A consistent eating style might be beneficial to avoid overeating in a food-rich environment. Eating consistency entails maintaining a similar dietary pattern across different eating situations. This construct is relatively under-studied, but the available evidence suggests that eating consistency supports successful weight maintenance and decreases risk for metabolic syndrome and cardiovascular disease. Yet, personality and situation predictors of consistency have not been studied.A community-based sample of 164 women completed various personality tests, and 139 of them also reported their eating behaviour 6 times/day over 10 observational days. We focused on observations with meals (breakfast, lunch, or dinner. The participants indicated if their momentary eating patterns were consistent with their own baseline eating patterns in terms of healthiness or size of the meal. Further, participants described various characteristics of each eating situation.Eating consistency was positively predicted by trait self-control. Eating consistency was undermined by eating in the evening, eating with others, eating away from home, having consumed alcohol and having undertaken physical exercise. Interactions emerged between personality traits and situations, including punishment sensitivity, restraint, physical activity and alcohol consumption.Trait self-control and several eating situation variables were related to eating consistency. These findings provide a starting point for targeting interventions to improve consistency, suggesting that a focus on self-control skills, together with addressing contextual factors such as social situations and time of day, may be most promising. This work is a first step to provide people with the tools they need to maintain a consistently healthy lifestyle in a food-rich environment.

  4. A consistent interpretation of quantum mechanics

    International Nuclear Information System (INIS)

    Omnes, Roland

    1990-01-01

    Some mostly recent theoretical and mathematical advances can be linked together to yield a new consistent interpretation of quantum mechanics. It relies upon a unique and universal interpretative rule of a logical character which is based upon Griffiths consistent history. Some new results in semi-classical physics allow classical physics to be derived from this rule, including its logical aspects, and to prove accordingly the existence of determinism within the quantum framework. Together with decoherence, this can be used to retrieve the existence of facts, despite the probabilistic character of the theory. Measurement theory can then be made entirely deductive. It is accordingly found that wave packet reduction is a logical property, whereas one can always choose to avoid using it. The practical consequences of this interpretation are most often in agreement with the Copenhagen formulation but they can be proved never to give rise to any logical inconsistency or paradox. (author)

  5. Self-consistency in Capital Markets

    Science.gov (United States)

    Benbrahim, Hamid

    2013-03-01

    Capital Markets are considered, at least in theory, information engines whereby traders contribute to price formation with their diverse perspectives. Regardless whether one believes in efficient market theory on not, actions by individual traders influence prices of securities, which in turn influence actions by other traders. This influence is exerted through a number of mechanisms including portfolio balancing, margin maintenance, trend following, and sentiment. As a result market behaviors emerge from a number of mechanisms ranging from self-consistency due to wisdom of the crowds and self-fulfilling prophecies, to more chaotic behavior resulting from dynamics similar to the three body system, namely the interplay between equities, options, and futures. This talk will address questions and findings regarding the search for self-consistency in capital markets.

  6. Consistent Estimation of Partition Markov Models

    Directory of Open Access Journals (Sweden)

    Jesús E. García

    2017-04-01

    Full Text Available The Partition Markov Model characterizes the process by a partition L of the state space, where the elements in each part of L share the same transition probability to an arbitrary element in the alphabet. This model aims to answer the following questions: what is the minimal number of parameters needed to specify a Markov chain and how to estimate these parameters. In order to answer these questions, we build a consistent strategy for model selection which consist of: giving a size n realization of the process, finding a model within the Partition Markov class, with a minimal number of parts to represent the process law. From the strategy, we derive a measure that establishes a metric in the state space. In addition, we show that if the law of the process is Markovian, then, eventually, when n goes to infinity, L will be retrieved. We show an application to model internet navigation patterns.

  7. Internal Branding and Employee Brand Consistent Behaviours

    DEFF Research Database (Denmark)

    Mazzei, Alessandra; Ravazzani, Silvia

    2017-01-01

    constitutive processes. In particular, the paper places emphasis on the role and kinds of communication practices as a central part of the nonnormative and constitutive internal branding process. The paper also discusses an empirical study based on interviews with 32 Italian and American communication managers...... and 2 focus groups with Italian communication managers. Findings show that, in order to enhance employee brand consistent behaviours, the most effective communication practices are those characterised as enablement-oriented. Such a communication creates the organizational conditions adequate to sustain......Employee behaviours conveying brand values, named brand consistent behaviours, affect the overall brand evaluation. Internal branding literature highlights a knowledge gap in terms of communication practices intended to sustain such behaviours. This study contributes to the development of a non...

  8. Self-consistent velocity dependent effective interactions

    International Nuclear Information System (INIS)

    Kubo, Takayuki; Sakamoto, Hideo; Kammuri, Tetsuo; Kishimoto, Teruo.

    1993-09-01

    The field coupling method is extended to a system with a velocity dependent mean potential. By means of this method, we can derive the effective interactions which are consistent with the mean potential. The self-consistent velocity dependent effective interactions are applied to the microscopic analysis of the structures of giant dipole resonances (GDR) of 148,154 Sm, of the first excited 2 + states of Sn isotopes and of the first excited 3 - states of Mo isotopes. It is clarified that the interactions play crucial roles in describing the splitting of the resonant structure of GDR peaks, in restoring the energy weighted sum rule values, and in reducing B (Eλ) values. (author)

  9. Evaluating Temporal Consistency in Marine Biodiversity Hotspots

    OpenAIRE

    Piacenza, Susan E.; Thurman, Lindsey L.; Barner, Allison K.; Benkwitt, Cassandra E.; Boersma, Kate S.; Cerny-Chipman, Elizabeth B.; Ingeman, Kurt E.; Kindinger, Tye L.; Lindsley, Amy J.; Nelson, Jake; Reimer, Jessica N.; Rowe, Jennifer C.; Shen, Chenchen; Thompson, Kevin A.; Heppell, Selina S.

    2015-01-01

    With the ongoing crisis of biodiversity loss and limited resources for conservation, the concept of biodiversity hotspots has been useful in determining conservation priority areas. However, there has been limited research into how temporal variability in biodiversity may influence conservation area prioritization. To address this information gap, we present an approach to evaluate the temporal consistency of biodiversity hotspots in large marine ecosystems. Using a large scale, public monito...

  10. Cloud Standardization: Consistent Business Processes and Information

    Directory of Open Access Journals (Sweden)

    Razvan Daniel ZOTA

    2013-01-01

    Full Text Available Cloud computing represents one of the latest emerging trends in distributed computing that enables the existence of hardware infrastructure and software applications as services. The present paper offers a general approach to the cloud computing standardization as a mean of improving the speed of adoption for the cloud technologies. Moreover, this study tries to show out how organizations may achieve more consistent business processes while operating with cloud computing technologies.

  11. Consistency Analysis of Nearest Subspace Classifier

    OpenAIRE

    Wang, Yi

    2015-01-01

    The Nearest subspace classifier (NSS) finds an estimation of the underlying subspace within each class and assigns data points to the class that corresponds to its nearest subspace. This paper mainly studies how well NSS can be generalized to new samples. It is proved that NSS is strongly consistent under certain assumptions. For completeness, NSS is evaluated through experiments on various simulated and real data sets, in comparison with some other linear model based classifiers. It is also ...

  12. Consistent probabilities in loop quantum cosmology

    International Nuclear Information System (INIS)

    Craig, David A; Singh, Parampreet

    2013-01-01

    A fundamental issue for any quantum cosmological theory is to specify how probabilities can be assigned to various quantum events or sequences of events such as the occurrence of singularities or bounces. In previous work, we have demonstrated how this issue can be successfully addressed within the consistent histories approach to quantum theory for Wheeler–DeWitt-quantized cosmological models. In this work, we generalize that analysis to the exactly solvable loop quantization of a spatially flat, homogeneous and isotropic cosmology sourced with a massless, minimally coupled scalar field known as sLQC. We provide an explicit, rigorous and complete decoherent-histories formulation for this model and compute the probabilities for the occurrence of a quantum bounce versus a singularity. Using the scalar field as an emergent internal time, we show for generic states that the probability for a singularity to occur in this model is zero, and that of a bounce is unity, complementing earlier studies of the expectation values of the volume and matter density in this theory. We also show from the consistent histories point of view that all states in this model, whether quantum or classical, achieve arbitrarily large volume in the limit of infinite ‘past’ or ‘future’ scalar ‘time’, in the sense that the wave function evaluated at any arbitrary fixed value of the volume vanishes in that limit. Finally, we briefly discuss certain misconceptions concerning the utility of the consistent histories approach in these models. (paper)

  13. Orthology and paralogy constraints: satisfiability and consistency.

    Science.gov (United States)

    Lafond, Manuel; El-Mabrouk, Nadia

    2014-01-01

    A variety of methods based on sequence similarity, reconciliation, synteny or functional characteristics, can be used to infer orthology and paralogy relations between genes of a given gene family  G. But is a given set  C of orthology/paralogy constraints possible, i.e., can they simultaneously co-exist in an evolutionary history for  G? While previous studies have focused on full sets of constraints, here we consider the general case where  C does not necessarily involve a constraint for each pair of genes. The problem is subdivided in two parts: (1) Is  C satisfiable, i.e. can we find an event-labeled gene tree G inducing  C? (2) Is there such a G which is consistent, i.e., such that all displayed triplet phylogenies are included in a species tree? Previous results on the Graph sandwich problem can be used to answer to (1), and we provide polynomial-time algorithms for satisfiability and consistency with a given species tree. We also describe a new polynomial-time algorithm for the case of consistency with an unknown species tree and full knowledge of pairwise orthology/paralogy relationships, as well as a branch-and-bound algorithm in the case when unknown relations are present. We show that our algorithms can be used in combination with ProteinOrtho, a sequence similarity-based orthology detection tool, to extract a set of robust orthology/paralogy relationships.

  14. Testing Testing Testing.

    Science.gov (United States)

    Deville, Craig; O'Neill, Thomas; Wright, Benjamin D.; Woodcock, Richard W.; Munoz-Sandoval, Ana; Gershon, Richard C.; Bergstrom, Betty

    1998-01-01

    Articles in this special section consider (1) flow in test taking (Craig Deville); (2) testwiseness (Thomas O'Neill); (3) test length (Benjamin Wright); (4) cross-language test equating (Richard W. Woodcock and Ana Munoz-Sandoval); (5) computer-assisted testing and testwiseness (Richard Gershon and Betty Bergstrom); and (6) Web-enhanced testing…

  15. Do Health Systems Have Consistent Performance Across Locations and Is Consistency Associated With Higher Performance?

    Science.gov (United States)

    Crespin, Daniel J; Christianson, Jon B; McCullough, Jeffrey S; Finch, Michael D

    This study addresses whether health systems have consistent diabetes care performance across their ambulatory clinics and whether increasing consistency is associated with improvements in clinic performance. Study data included 2007 to 2013 diabetes care intermediate outcome measures for 661 ambulatory clinics in Minnesota and bordering states. Health systems provided more consistent performance, as measured by the standard deviation of performance for clinics in a system, relative to propensity score-matched proxy systems created for comparison purposes. No evidence was found that improvements in consistency were associated with higher clinic performance. The combination of high performance and consistent care is likely to enhance a health system's brand reputation, allowing it to better mitigate the financial risks of consumers seeking care outside the organization. These results suggest that larger health systems are most likely to deliver the combination of consistent and high-performance care. Future research should explore the mechanisms that drive consistent care within health systems.

  16. High-performance speech recognition using consistency modeling

    Science.gov (United States)

    Digalakis, Vassilios; Murveit, Hy; Monaco, Peter; Neumeyer, Leo; Sankar, Ananth

    1994-12-01

    The goal of SRI's consistency modeling project is to improve the raw acoustic modeling component of SRI's DECIPHER speech recognition system and develop consistency modeling technology. Consistency modeling aims to reduce the number of improper independence assumptions used in traditional speech recognition algorithms so that the resulting speech recognition hypotheses are more self-consistent and, therefore, more accurate. At the initial stages of this effort, SRI focused on developing the appropriate base technologies for consistency modeling. We first developed the Progressive Search technology that allowed us to perform large-vocabulary continuous speech recognition (LVCSR) experiments. Since its conception and development at SRI, this technique has been adopted by most laboratories, including other ARPA contracting sites, doing research on LVSR. Another goal of the consistency modeling project is to attack difficult modeling problems, when there is a mismatch between the training and testing phases. Such mismatches may include outlier speakers, different microphones and additive noise. We were able to either develop new, or transfer and evaluate existing, technologies that adapted our baseline genonic HMM recognizer to such difficult conditions.

  17. Self-consistent gravitational self-force

    International Nuclear Information System (INIS)

    Pound, Adam

    2010-01-01

    I review the problem of motion for small bodies in general relativity, with an emphasis on developing a self-consistent treatment of the gravitational self-force. An analysis of the various derivations extant in the literature leads me to formulate an asymptotic expansion in which the metric is expanded while a representative worldline is held fixed. I discuss the utility of this expansion for both exact point particles and asymptotically small bodies, contrasting it with a regular expansion in which both the metric and the worldline are expanded. Based on these preliminary analyses, I present a general method of deriving self-consistent equations of motion for arbitrarily structured (sufficiently compact) small bodies. My method utilizes two expansions: an inner expansion that keeps the size of the body fixed, and an outer expansion that lets the body shrink while holding its worldline fixed. By imposing the Lorenz gauge, I express the global solution to the Einstein equation in the outer expansion in terms of an integral over a worldtube of small radius surrounding the body. Appropriate boundary data on the tube are determined from a local-in-space expansion in a buffer region where both the inner and outer expansions are valid. This buffer-region expansion also results in an expression for the self-force in terms of irreducible pieces of the metric perturbation on the worldline. Based on the global solution, these pieces of the perturbation can be written in terms of a tail integral over the body's past history. This approach can be applied at any order to obtain a self-consistent approximation that is valid on long time scales, both near and far from the small body. I conclude by discussing possible extensions of my method and comparing it to alternative approaches.

  18. Consistency Checking of Web Service Contracts

    DEFF Research Database (Denmark)

    Cambronero, M. Emilia; Okika, Joseph C.; Ravn, Anders Peter

    2008-01-01

    Behavioural properties are analyzed for web service contracts formulated in Business Process Execution Language (BPEL) and Choreography Description Language (CDL). The key result reported is an automated technique to check consistency between protocol aspects of the contracts. The contracts...... are abstracted to (timed) automata and from there a simulation is set up, which is checked using automated tools for analyzing networks of finite state processes. Here we use the Concurrency Work Bench. The proposed techniques are illustrated with a case study that include otherwise difficult to analyze fault...

  19. A method for consistent precision radiation therapy

    International Nuclear Information System (INIS)

    Leong, J.

    1985-01-01

    Using a meticulous setup procedure in which repeated portal films were taken before each treatment until satisfactory portal verifications were obtained, a high degree of precision in patient positioning was achieved. A fluctuation from treatment to treatment, over 11 treatments, of less than +-0.10 cm (S.D.) for anatomical points inside the treatment field was obtained. This, however, only applies to specific anatomical points selected for this positioning procedure and does not apply to all points within the portal. We have generalized this procedure and have suggested a means by which any target volume can be consistently positioned which may approach this degree of precision. (orig.)

  20. Gentzen's centenary the quest for consistency

    CERN Document Server

    Rathjen, Michael

    2015-01-01

    Gerhard Gentzen has been described as logic’s lost genius, whom Gödel called a better logician than himself. This work comprises articles by leading proof theorists, attesting to Gentzen’s enduring legacy to mathematical logic and beyond. The contributions range from philosophical reflections and re-evaluations of Gentzen’s original consistency proofs to the most recent developments in proof theory. Gentzen founded modern proof theory. His sequent calculus and natural deduction system beautifully explain the deep symmetries of logic. They underlie modern developments in computer science such as automated theorem proving and type theory.

  1. Two consistent calculations of the Weinberg angle

    International Nuclear Information System (INIS)

    Fairlie, D.B.

    1979-01-01

    The Weinberg-Salam theory is reformulated as a pure Yang-Mills theory in a six-dimensional space, the Higgs field being interpreted as gauge potentials in the additional dimensions. Viewed in this way, the condition that the Higgs field transforms as a U(1) representation of charge one is equivalent to requiring a value of 30 0 C for the Weinberg angle. A second consistent determination comes from the idea borrowed from monopole theory that the electromagnetic field is in the direction of the Higgs field. (Author)

  2. Consistent resolution of some relativistic quantum paradoxes

    International Nuclear Information System (INIS)

    Griffiths, Robert B.

    2002-01-01

    A relativistic version of the (consistent or decoherent) histories approach to quantum theory is developed on the basis of earlier work by Hartle, and used to discuss relativistic forms of the paradoxes of spherical wave packet collapse, Bohm's formulation of the Einstein-Podolsky-Rosen paradox, and Hardy's paradox. It is argued that wave function collapse is not needed for introducing probabilities into relativistic quantum mechanics, and in any case should never be thought of as a physical process. Alternative approaches to stochastic time dependence can be used to construct a physical picture of the measurement process that is less misleading than collapse models. In particular, one can employ a coarse-grained but fully quantum-mechanical description in which particles move along trajectories, with behavior under Lorentz transformations the same as in classical relativistic physics, and detectors are triggered by particles reaching them along such trajectories. States entangled between spacelike separate regions are also legitimate quantum descriptions, and can be consistently handled by the formalism presented here. The paradoxes in question arise because of using modes of reasoning which, while correct for classical physics, are inconsistent with the mathematical structure of quantum theory, and are resolved (or tamed) by using a proper quantum analysis. In particular, there is no need to invoke, nor any evidence for, mysterious long-range superluminal influences, and thus no incompatibility, at least from this source, between relativity theory and quantum mechanics

  3. Self-consistent model of confinement

    International Nuclear Information System (INIS)

    Swift, A.R.

    1988-01-01

    A model of the large-spatial-distance, zero--three-momentum, limit of QCD is developed from the hypothesis that there is an infrared singularity. Single quarks and gluons do not propagate because they have infinite energy after renormalization. The Hamiltonian formulation of the path integral is used to quantize QCD with physical, nonpropagating fields. Perturbation theory in the infrared limit is simplified by the absence of self-energy insertions and by the suppression of large classes of diagrams due to vanishing propagators. Remaining terms in the perturbation series are resummed to produce a set of nonlinear, renormalizable integral equations which fix both the confining interaction and the physical propagators. Solutions demonstrate the self-consistency of the concepts of an infrared singularity and nonpropagating fields. The Wilson loop is calculated to provide a general proof of confinement. Bethe-Salpeter equations for quark-antiquark pairs and for two gluons have finite-energy solutions in the color-singlet channel. The choice of gauge is addressed in detail. Large classes of corrections to the model are discussed and shown to support self-consistency

  4. Subgame consistent cooperation a comprehensive treatise

    CERN Document Server

    Yeung, David W K

    2016-01-01

    Strategic behavior in the human and social world has been increasingly recognized in theory and practice. It is well known that non-cooperative behavior could lead to suboptimal or even highly undesirable outcomes. Cooperation suggests the possibility of obtaining socially optimal solutions and the calls for cooperation are prevalent in real-life problems. Dynamic cooperation cannot be sustainable if there is no guarantee that the agreed upon optimality principle at the beginning is maintained throughout the cooperation duration. It is due to the lack of this kind of guarantees that cooperative schemes fail to last till its end or even fail to get started. The property of subgame consistency in cooperative dynamic games and the corresponding solution mechanism resolve this “classic” problem in game theory. This book is a comprehensive treatise on subgame consistent dynamic cooperation covering the up-to-date state of the art analyses in this important topic. It sets out to provide the theory, solution tec...

  5. Consistent mutational paths predict eukaryotic thermostability

    Directory of Open Access Journals (Sweden)

    van Noort Vera

    2013-01-01

    Full Text Available Abstract Background Proteomes of thermophilic prokaryotes have been instrumental in structural biology and successfully exploited in biotechnology, however many proteins required for eukaryotic cell function are absent from bacteria or archaea. With Chaetomium thermophilum, Thielavia terrestris and Thielavia heterothallica three genome sequences of thermophilic eukaryotes have been published. Results Studying the genomes and proteomes of these thermophilic fungi, we found common strategies of thermal adaptation across the different kingdoms of Life, including amino acid biases and a reduced genome size. A phylogenetics-guided comparison of thermophilic proteomes with those of other, mesophilic Sordariomycetes revealed consistent amino acid substitutions associated to thermophily that were also present in an independent lineage of thermophilic fungi. The most consistent pattern is the substitution of lysine by arginine, which we could find in almost all lineages but has not been extensively used in protein stability engineering. By exploiting mutational paths towards the thermophiles, we could predict particular amino acid residues in individual proteins that contribute to thermostability and validated some of them experimentally. By determining the three-dimensional structure of an exemplar protein from C. thermophilum (Arx1, we could also characterise the molecular consequences of some of these mutations. Conclusions The comparative analysis of these three genomes not only enhances our understanding of the evolution of thermophily, but also provides new ways to engineer protein stability.

  6. Consistent biokinetic models for the actinide elements

    International Nuclear Information System (INIS)

    Leggett, R.W.

    2001-01-01

    The biokinetic models for Th, Np, Pu, Am and Cm currently recommended by the International Commission on Radiological Protection (ICRP) were developed within a generic framework that depicts gradual burial of skeletal activity in bone volume, depicts recycling of activity released to blood and links excretion to retention and translocation of activity. For other actinide elements such as Ac, Pa, Bk, Cf and Es, the ICRP still uses simplistic retention models that assign all skeletal activity to bone surface and depicts one-directional flow of activity from blood to long-term depositories to excreta. This mixture of updated and older models in ICRP documents has led to inconsistencies in dose estimates and interpretation of bioassay for radionuclides with reasonably similar biokinetics. This paper proposes new biokinetic models for Ac, Pa, Bk, Cf and Es that are consistent with the updated models for Th, Np, Pu, Am and Cm. The proposed models are developed within the ICRP's generic model framework for bone-surface-seeking radionuclides, and an effort has been made to develop parameter values that are consistent with results of comparative biokinetic data on the different actinide elements. (author)

  7. Linear augmented plane wave method for self-consistent calculations

    International Nuclear Information System (INIS)

    Takeda, T.; Kuebler, J.

    1979-01-01

    O.K. Andersen has recently introduced a linear augmented plane wave method (LAPW) for the calculation of electronic structure that was shown to be computationally fast. A more general formulation of an LAPW method is presented here. It makes use of a freely disposable number of eigenfunctions of the radial Schroedinger equation. These eigenfunctions can be selected in a self-consistent way. The present formulation also results in a computationally fast method. It is shown that Andersen's LAPW is obtained in a special limit from the present formulation. Self-consistent test calculations for copper show the present method to be remarkably accurate. As an application, scalar-relativistic self-consistent calculations are presented for the band structure of FCC lanthanum. (author)

  8. An assessment of uncertainty on a LOFT L2-5 LBLOCA PCT based on the ACE-RSM approach: complementary work for the OECD BEMUSE PHASE-III program

    International Nuclear Information System (INIS)

    Ahn, Kwang Il; Chung, Bub Dong; Lee, John C.

    2010-01-01

    As pointed out in the OECD BEMUSE Program, when a high computation time is taken to obtain the relevant output values of a complex physical model (or code), the number of statistical samples that must be evaluated through it is a critical factor for the sampling-based uncertainty analysis. Two alternative methods have been utilized to avoid the problem associated with the size of these statistical samples: one is based on Wilks' formula, which is based on simple random sampling, and the other is based on the conventional nonlinear regression approach. While both approaches provide a useful means for drawing conclusions on the resultant uncertainty with a limited number of code runs, there are also some unique corresponding limitations. For example, a conclusion based on the Wilks' formula can be highly affected by the sampled values themselves, while the conventional regression approach requires an a priori estimate on the functional forms of a regression model. The main objective of this paper is to assess the feasibility of the ACE-RSM approach as a complementary method to the Wilks' formula and the conventional regression-based uncertainty analysis. This feasibility was assessed through a practical application of the ACE-RSM approach to the LOFT L2-5 LBLOCA PCT uncertainty analysis, which was implemented as a part of the OECD BEMUSE Phase III program

  9. Consistency of canonical formulation of Horava gravity

    International Nuclear Information System (INIS)

    Soo, Chopin

    2011-01-01

    Both the non-projectable and projectable version of Horava gravity face serious challenges. In the non-projectable version, the constraint algebra is seemingly inconsistent. The projectable version lacks a local Hamiltonian constraint, thus allowing for an extra graviton mode which can be problematic. A new formulation (based on arXiv:1007.1563) of Horava gravity which is naturally realized as a representation of the master constraint algebra (instead of the Dirac algebra) studied by loop quantum gravity researchers is presented. This formulation yields a consistent canonical theory with first class constraints; and captures the essence of Horava gravity in retaining only spatial diffeomorphisms as the physically relevant non-trivial gauge symmetry. At the same time the local Hamiltonian constraint is equivalently enforced by the master constraint.

  10. Consistency of canonical formulation of Horava gravity

    Energy Technology Data Exchange (ETDEWEB)

    Soo, Chopin, E-mail: cpsoo@mail.ncku.edu.tw [Department of Physics, National Cheng Kung University, Tainan, Taiwan (China)

    2011-09-22

    Both the non-projectable and projectable version of Horava gravity face serious challenges. In the non-projectable version, the constraint algebra is seemingly inconsistent. The projectable version lacks a local Hamiltonian constraint, thus allowing for an extra graviton mode which can be problematic. A new formulation (based on arXiv:1007.1563) of Horava gravity which is naturally realized as a representation of the master constraint algebra (instead of the Dirac algebra) studied by loop quantum gravity researchers is presented. This formulation yields a consistent canonical theory with first class constraints; and captures the essence of Horava gravity in retaining only spatial diffeomorphisms as the physically relevant non-trivial gauge symmetry. At the same time the local Hamiltonian constraint is equivalently enforced by the master constraint.

  11. Evaluating the hydrological consistency of evaporation products

    KAUST Repository

    Lopez Valencia, Oliver Miguel; Houborg, Rasmus; McCabe, Matthew

    2017-01-01

    Advances in space-based observations have provided the capacity to develop regional- to global-scale estimates of evaporation, offering insights into this key component of the hydrological cycle. However, the evaluation of large-scale evaporation retrievals is not a straightforward task. While a number of studies have intercompared a range of these evaporation products by examining the variance amongst them, or by comparison of pixel-scale retrievals against ground-based observations, there is a need to explore more appropriate techniques to comprehensively evaluate remote-sensing-based estimates. One possible approach is to establish the level of product agreement between related hydrological components: for instance, how well do evaporation patterns and response match with precipitation or water storage changes? To assess the suitability of this "consistency"-based approach for evaluating evaporation products, we focused our investigation on four globally distributed basins in arid and semi-arid environments, comprising the Colorado River basin, Niger River basin, Aral Sea basin, and Lake Eyre basin. In an effort to assess retrieval quality, three satellite-based global evaporation products based on different methodologies and input data, including CSIRO-PML, the MODIS Global Evapotranspiration product (MOD16), and Global Land Evaporation: the Amsterdam Methodology (GLEAM), were evaluated against rainfall data from the Global Precipitation Climatology Project (GPCP) along with Gravity Recovery and Climate Experiment (GRACE) water storage anomalies. To ensure a fair comparison, we evaluated consistency using a degree correlation approach after transforming both evaporation and precipitation data into spherical harmonics. Overall we found no persistent hydrological consistency in these dryland environments. Indeed, the degree correlation showed oscillating values between periods of low and high water storage changes, with a phase difference of about 2–3 months

  12. Self-consistent modelling of ICRH

    International Nuclear Information System (INIS)

    Hellsten, T.; Hedin, J.; Johnson, T.; Laxaaback, M.; Tennfors, E.

    2001-01-01

    The performance of ICRH is often sensitive to the shape of the high energy part of the distribution functions of the resonating species. This requires self-consistent calculations of the distribution functions and the wave-field. In addition to the wave-particle interactions and Coulomb collisions the effects of the finite orbit width and the RF-induced spatial transport are found to be important. The inward drift dominates in general even for a symmetric toroidal wave spectrum in the centre of the plasma. An inward drift does not necessarily produce a more peaked heating profile. On the contrary, for low concentrations of hydrogen minority in deuterium plasmas it can even give rise to broader profiles. (author)

  13. Non linear self consistency of microtearing modes

    International Nuclear Information System (INIS)

    Garbet, X.; Mourgues, F.; Samain, A.

    1987-01-01

    The self consistency of a microtearing turbulence is studied in non linear regimes where the ergodicity of the flux lines determines the electron response. The current which sustains the magnetic perturbation via the Ampere law results from the combines action of the radial electric field in the frame where the island chains are static and of the thermal electron diamagnetism. Numerical calculations show that at usual values of β pol in Tokamaks the turbulence can create a diffusion coefficient of order ν th p 2 i where p i is the ion larmor radius and ν th the electron ion collision frequency. On the other hand, collisionless regimes involving special profiles of each mode near the resonant surface seem possible

  14. Consistent evolution in a pedestrian flow

    Science.gov (United States)

    Guan, Junbiao; Wang, Kaihua

    2016-03-01

    In this paper, pedestrian evacuation considering different human behaviors is studied by using a cellular automaton (CA) model combined with the snowdrift game theory. The evacuees are divided into two types, i.e. cooperators and defectors, and two different human behaviors, herding behavior and independent behavior, are investigated. It is found from a large amount of numerical simulations that the ratios of the corresponding evacuee clusters are evolved to consistent states despite 11 typically different initial conditions, which may largely owe to self-organization effect. Moreover, an appropriate proportion of initial defectors who are of herding behavior, coupled with an appropriate proportion of initial defectors who are of rationally independent thinking, are two necessary factors for short evacuation time.

  15. Evaluating the hydrological consistency of evaporation products

    KAUST Repository

    Lopez Valencia, Oliver Miguel

    2017-01-18

    Advances in space-based observations have provided the capacity to develop regional- to global-scale estimates of evaporation, offering insights into this key component of the hydrological cycle. However, the evaluation of large-scale evaporation retrievals is not a straightforward task. While a number of studies have intercompared a range of these evaporation products by examining the variance amongst them, or by comparison of pixel-scale retrievals against ground-based observations, there is a need to explore more appropriate techniques to comprehensively evaluate remote-sensing-based estimates. One possible approach is to establish the level of product agreement between related hydrological components: for instance, how well do evaporation patterns and response match with precipitation or water storage changes? To assess the suitability of this "consistency"-based approach for evaluating evaporation products, we focused our investigation on four globally distributed basins in arid and semi-arid environments, comprising the Colorado River basin, Niger River basin, Aral Sea basin, and Lake Eyre basin. In an effort to assess retrieval quality, three satellite-based global evaporation products based on different methodologies and input data, including CSIRO-PML, the MODIS Global Evapotranspiration product (MOD16), and Global Land Evaporation: the Amsterdam Methodology (GLEAM), were evaluated against rainfall data from the Global Precipitation Climatology Project (GPCP) along with Gravity Recovery and Climate Experiment (GRACE) water storage anomalies. To ensure a fair comparison, we evaluated consistency using a degree correlation approach after transforming both evaporation and precipitation data into spherical harmonics. Overall we found no persistent hydrological consistency in these dryland environments. Indeed, the degree correlation showed oscillating values between periods of low and high water storage changes, with a phase difference of about 2–3 months

  16. Thermodynamically consistent model calibration in chemical kinetics

    Directory of Open Access Journals (Sweden)

    Goutsias John

    2011-05-01

    Full Text Available Abstract Background The dynamics of biochemical reaction systems are constrained by the fundamental laws of thermodynamics, which impose well-defined relationships among the reaction rate constants characterizing these systems. Constructing biochemical reaction systems from experimental observations often leads to parameter values that do not satisfy the necessary thermodynamic constraints. This can result in models that are not physically realizable and may lead to inaccurate, or even erroneous, descriptions of cellular function. Results We introduce a thermodynamically consistent model calibration (TCMC method that can be effectively used to provide thermodynamically feasible values for the parameters of an open biochemical reaction system. The proposed method formulates the model calibration problem as a constrained optimization problem that takes thermodynamic constraints (and, if desired, additional non-thermodynamic constraints into account. By calculating thermodynamically feasible values for the kinetic parameters of a well-known model of the EGF/ERK signaling cascade, we demonstrate the qualitative and quantitative significance of imposing thermodynamic constraints on these parameters and the effectiveness of our method for accomplishing this important task. MATLAB software, using the Systems Biology Toolbox 2.1, can be accessed from http://www.cis.jhu.edu/~goutsias/CSS lab/software.html. An SBML file containing the thermodynamically feasible EGF/ERK signaling cascade model can be found in the BioModels database. Conclusions TCMC is a simple and flexible method for obtaining physically plausible values for the kinetic parameters of open biochemical reaction systems. It can be effectively used to recalculate a thermodynamically consistent set of parameter values for existing thermodynamically infeasible biochemical reaction models of cellular function as well as to estimate thermodynamically feasible values for the parameters of new

  17. Consistent creep and rupture properties for creep-fatigue evaluation

    International Nuclear Information System (INIS)

    Schultz, C.C.

    1978-01-01

    The currently accepted practice of using inconsistent representations of creep and rupture behaviors in the prediction of creep-fatigue life is shown to introduce a factor of safety beyond that specified in current ASME Code design rules for 304 stainless steel Class 1 nuclear components. Accurate predictions of creep-fatigue life for uniaxial tests on a given heat of material are obtained by using creep and rupture properties for that same heat of material. The use of a consistent representation of creep and rupture properties for a mininum strength heat is also shown to provide adequate predictions. The viability of using consistent properties (either actual or those of a minimum heat) to predict creep-fatigue life thus identifies significant design uses for the results of characterization tests and improved creep and rupture correlations

  18. Consistent creep and rupture properties for creep-fatigue evaluation

    International Nuclear Information System (INIS)

    Schultz, C.C.

    1979-01-01

    The currently accepted practice of using inconsistent representations of creep and rupture behaviors in the prediction of creep-fatigue life is shown to introduce a factor of safety beyond that specified in current ASME Code design rules for 304 stainless steel Class 1 nuclear components. Accurate predictions of creep-fatigue life for uniaxial tests on a given heat of material are obtained by using creep and rupture properties for that same heat of material. The use of a consistent representation of creep and rupture properties for a minimum strength heat is also shown to provide reasonable predictions. The viability of using consistent properties (either actual or those of a minimum strength heat) to predict creep-fatigue life thus identifies significant design uses for the results of characterization tests and improved creep and rupture correlations. 12 refs

  19. Cognitive consistency and math-gender stereotypes in Singaporean children.

    Science.gov (United States)

    Cvencek, Dario; Meltzoff, Andrew N; Kapur, Manu

    2014-01-01

    In social psychology, cognitive consistency is a powerful principle for organizing psychological concepts. There have been few tests of cognitive consistency in children and no research about cognitive consistency in children from Asian cultures, who pose an interesting developmental case. A sample of 172 Singaporean elementary school children completed implicit and explicit measures of math-gender stereotype (male=math), gender identity (me=male), and math self-concept (me=math). Results showed strong evidence for cognitive consistency; the strength of children's math-gender stereotypes, together with their gender identity, significantly predicted their math self-concepts. Cognitive consistency may be culturally universal and a key mechanism for developmental change in social cognition. We also discovered that Singaporean children's math-gender stereotypes increased as a function of age and that boys identified with math more strongly than did girls despite Singaporean girls' excelling in math. The results reveal both cultural universals and cultural variation in developing social cognition. Copyright © 2013 Elsevier Inc. All rights reserved.

  20. String consistency for unified model building

    International Nuclear Information System (INIS)

    Chaudhuri, S.; Chung, S.W.; Hockney, G.; Lykken, J.

    1995-01-01

    We explore the use of real fermionization as a test case for understanding how specific features of phenomenological interest in the low-energy effective superpotential are realized in exact solutions to heterotic superstring theory. We present pedagogic examples of models which realize SO(10) as a level two current algebra on the world-sheet, and discuss in general how higher level current algebras can be realized in the tensor product of simple constituent conformal field theories. We describe formal developments necessary to compute couplings in models built using real fermionization. This allows us to isolate cases of spin structures where the standard prescription for real fermionization may break down. (orig.)

  1. Validation of consistency of Mendelian sampling variance.

    Science.gov (United States)

    Tyrisevä, A-M; Fikse, W F; Mäntysaari, E A; Jakobsen, J; Aamand, G P; Dürr, J; Lidauer, M H

    2018-03-01

    Experiences from international sire evaluation indicate that the multiple-trait across-country evaluation method is sensitive to changes in genetic variance over time. Top bulls from birth year classes with inflated genetic variance will benefit, hampering reliable ranking of bulls. However, none of the methods available today enable countries to validate their national evaluation models for heterogeneity of genetic variance. We describe a new validation method to fill this gap comprising the following steps: estimating within-year genetic variances using Mendelian sampling and its prediction error variance, fitting a weighted linear regression between the estimates and the years under study, identifying possible outliers, and defining a 95% empirical confidence interval for a possible trend in the estimates. We tested the specificity and sensitivity of the proposed validation method with simulated data using a real data structure. Moderate (M) and small (S) size populations were simulated under 3 scenarios: a control with homogeneous variance and 2 scenarios with yearly increases in phenotypic variance of 2 and 10%, respectively. Results showed that the new method was able to estimate genetic variance accurately enough to detect bias in genetic variance. Under the control scenario, the trend in genetic variance was practically zero in setting M. Testing cows with an average birth year class size of more than 43,000 in setting M showed that tolerance values are needed for both the trend and the outlier tests to detect only cases with a practical effect in larger data sets. Regardless of the magnitude (yearly increases in phenotypic variance of 2 or 10%) of the generated trend, it deviated statistically significantly from zero in all data replicates for both cows and bulls in setting M. In setting S with a mean of 27 bulls in a year class, the sampling error and thus the probability of a false-positive result clearly increased. Still, overall estimated genetic

  2. Exploring the Consistent behavior of Information Services

    Directory of Open Access Journals (Sweden)

    Kapidakis Sarantos

    2016-01-01

    Full Text Available Computer services are normally assumed to work well all the time. This usually happens for crucial services like bank electronic services, but not necessarily so for others, that there is no commercial interest in their operation. In this work we examined the operation and the errors of information services and tried to find clues that will help predicting the consistency of the behavior and the quality of the harvesting, which is harder because of the transient conditions and the many services and the huge amount of harvested information. We found many unexpected situations. The services that always successfully satisfy a request may in fact return part of it. A significant part of the OAI services have ceased working while many other serves occasionally fail to respond. Some services fail in the same way each time, and we pronounce them dead, as we do not see a way to overcome that. Others also always, or sometimes fail, but not in the same way, and we hope that their behavior is affected by temporary factors, that may improve later on. We categorized the services into classes, to study their behavior in more detail.

  3. A Consistent Phylogenetic Backbone for the Fungi

    Science.gov (United States)

    Ebersberger, Ingo; de Matos Simoes, Ricardo; Kupczok, Anne; Gube, Matthias; Kothe, Erika; Voigt, Kerstin; von Haeseler, Arndt

    2012-01-01

    The kingdom of fungi provides model organisms for biotechnology, cell biology, genetics, and life sciences in general. Only when their phylogenetic relationships are stably resolved, can individual results from fungal research be integrated into a holistic picture of biology. However, and despite recent progress, many deep relationships within the fungi remain unclear. Here, we present the first phylogenomic study of an entire eukaryotic kingdom that uses a consistency criterion to strengthen phylogenetic conclusions. We reason that branches (splits) recovered with independent data and different tree reconstruction methods are likely to reflect true evolutionary relationships. Two complementary phylogenomic data sets based on 99 fungal genomes and 109 fungal expressed sequence tag (EST) sets analyzed with four different tree reconstruction methods shed light from different angles on the fungal tree of life. Eleven additional data sets address specifically the phylogenetic position of Blastocladiomycota, Ustilaginomycotina, and Dothideomycetes, respectively. The combined evidence from the resulting trees supports the deep-level stability of the fungal groups toward a comprehensive natural system of the fungi. In addition, our analysis reveals methodologically interesting aspects. Enrichment for EST encoded data—a common practice in phylogenomic analyses—introduces a strong bias toward slowly evolving and functionally correlated genes. Consequently, the generalization of phylogenomic data sets as collections of randomly selected genes cannot be taken for granted. A thorough characterization of the data to assess possible influences on the tree reconstruction should therefore become a standard in phylogenomic analyses. PMID:22114356

  4. Self consistent field theory of virus assembly

    Science.gov (United States)

    Li, Siyu; Orland, Henri; Zandi, Roya

    2018-04-01

    The ground state dominance approximation (GSDA) has been extensively used to study the assembly of viral shells. In this work we employ the self-consistent field theory (SCFT) to investigate the adsorption of RNA onto positively charged spherical viral shells and examine the conditions when GSDA does not apply and SCFT has to be used to obtain a reliable solution. We find that there are two regimes in which GSDA does work. First, when the genomic RNA length is long enough compared to the capsid radius, and second, when the interaction between the genome and capsid is so strong that the genome is basically localized next to the wall. We find that for the case in which RNA is more or less distributed uniformly in the shell, regardless of the length of RNA, GSDA is not a good approximation. We observe that as the polymer-shell interaction becomes stronger, the energy gap between the ground state and first excited state increases and thus GSDA becomes a better approximation. We also present our results corresponding to the genome persistence length obtained through the tangent-tangent correlation length and show that it is zero in case of GSDA but is equal to the inverse of the energy gap when using SCFT.

  5. Consistency based correlations for tailings consolidation

    Energy Technology Data Exchange (ETDEWEB)

    Azam, S.; Paul, A.C. [Regina Univ., Regina, SK (Canada). Environmental Systems Engineering

    2010-07-01

    The extraction of oil, uranium, metals and mineral resources from the earth generates significant amounts of tailings slurry. The tailings are contained in a disposal area with perimeter dykes constructed from the coarser fraction of the slurry. There are many unique challenges pertaining to the management of the containment facilities for several decades beyond mine closure that are a result of the slow settling rates of the fines and the high standing toxic waters. Many tailings dam failures in different parts of the world have been reported to result in significant contaminant releases causing public concern over the conventional practice of tailings disposal. Therefore, in order to reduce and minimize the environmental footprint, the fluid tailings need to undergo efficient consolidation. This paper presented an investigation into the consolidation behaviour of tailings in conjunction with soil consistency that captured physicochemical interactions. The paper discussed the large strain consolidation behaviour (volume compressibility and hydraulic conductivity) of six fine-grained soil slurries based on published data. The paper provided background information on the study and presented the research methodology. The geotechnical index properties of the selected materials were also presented. The large strain consolidation, volume compressibility correlations, and hydraulic conductivity correlations were provided. It was concluded that the normalized void ratio best described volume compressibility whereas liquidity index best explained the hydraulic conductivity. 17 refs., 3 tabs., 4 figs.

  6. Consistency between GRUAN sondes, LBLRTM and IASI

    Directory of Open Access Journals (Sweden)

    X. Calbet

    2017-06-01

    Full Text Available Radiosonde soundings from the GCOS Reference Upper-Air Network (GRUAN data record are shown to be consistent with Infrared Atmospheric Sounding Instrument (IASI-measured radiances via LBLRTM (Line-By-Line Radiative Transfer Model in the part of the spectrum that is mostly affected by water vapour absorption in the upper troposphere (from 700 hPa up. This result is key for climate data records, since GRUAN, IASI and LBLRTM constitute reference measurements or a reference radiative transfer model in each of their fields. This is specially the case for night-time radiosonde measurements. Although the sample size is small (16 cases, daytime GRUAN radiosonde measurements seem to have a small dry bias of 2.5 % in absolute terms of relative humidity, located mainly in the upper troposphere, with respect to LBLRTM and IASI. Full metrological closure is not yet possible and will not be until collocation uncertainties are better characterized and a full uncertainty covariance matrix is clarified for GRUAN.

  7. Self-consistent nuclear energy systems

    International Nuclear Information System (INIS)

    Shimizu, A.; Fujiie, Y.

    1995-01-01

    A concept of self-consistent energy systems (SCNES) has been proposed as an ultimate goal of the nuclear energy system in the coming centuries. SCNES should realize a stable and unlimited energy supply without endangering the human race and the global environment. It is defined as a system that realizes at least the following four objectives simultaneously: (a) energy generation -attain high efficiency in the utilization of fission energy; (b) fuel production - secure inexhaustible energy source: breeding of fissile material with the breeding ratio greater than one and complete burning of transuranium through recycling; (c) burning of radionuclides - zero release of radionuclides from the system: complete burning of transuranium and elimination of radioactive fission products by neutron capture reactions through recycling; (d) system safety - achieve system safety both for the public and experts: eliminate criticality-related safety issues by using natural laws and simple logic. This paper describes the concept of SCNES and discusses the feasibility of the system. Both ''neutron balance'' and ''energbalance'' of the system are introduced as the necessary conditions to be satisfied at least by SCNES. Evaluations made so far indicate that both the neutron balance and the energy balance can be realized by fast reactors but not by thermal reactors. Concerning the system safety, two safety concepts: ''self controllability'' and ''self-terminability'' are introduced to eliminate the criticality-related safety issues in fast reactors. (author)

  8. Toward a consistent model for glass dissolution

    International Nuclear Information System (INIS)

    Strachan, D.M.; McGrail, B.P.; Bourcier, W.L.

    1994-01-01

    Understanding the process of glass dissolution in aqueous media has advanced significantly over the last 10 years through the efforts of many scientists around the world. Mathematical models describing the glass dissolution process have also advanced from simple empirical functions to structured models based on fundamental principles of physics, chemistry, and thermodynamics. Although borosilicate glass has been selected as the waste form for disposal of high-level wastes in at least 5 countries, there is no international consensus on the fundamental methodology for modeling glass dissolution that could be used in assessing the long term performance of waste glasses in a geologic repository setting. Each repository program is developing their own model and supporting experimental data. In this paper, we critically evaluate a selected set of these structured models and show that a consistent methodology for modeling glass dissolution processes is available. We also propose a strategy for a future coordinated effort to obtain the model input parameters that are needed for long-term performance assessments of glass in a geologic repository. (author) 4 figs., tabs., 75 refs

  9. On the internal consistency of the term structure of forecasts of housing starts

    DEFF Research Database (Denmark)

    Pierdzioch, C.; Rulke, J. C.; Stadtmann, G.

    2013-01-01

    We use the term structure of forecasts of housing starts to test for rationality of forecasts. Our test is based on the idea that short-term and long-term forecasts should be internally consistent. We test the internal consistency of forecasts using data for Australia, Canada, Japan and the United...

  10. View from Europe: stability, consistency or pragmatism

    International Nuclear Information System (INIS)

    Dunster, H.J.

    1988-01-01

    The last few years of this decade look like a period of reappraisal of radiation protection standards. The revised risk estimates from Japan will be available, and the United Nations Scientific Committee on the Effects of Atomic Radiation will be publishing new reports on biological topics. The International Commission on Radiological Protection (ICRP) has started a review of its basic recommendations, and the new specification for dose equivalent in radiation fields of the International Commission on Radiation Units and Measurements (ICRU) will be coming into use. All this is occurring at a time when some countries are still trying to catch up with committed dose equivalent and the recently recommended change in the value of the quality factor for neutrons. In Europe, the problems of adapting to new ICRP recommendations are considerable. The European Community, including 12 states and nine languages, takes ICRP recommendations as a basis and develops council directives that are binding on member states, which have then to arrange for their own regulatory changes. Any substantial adjustments could take 5 y or more to work through the system. Clearly, the regulatory preference is for stability. Equally clearly, trade unions and public interest groups favor a rapid response to scientific developments (provided that the change is downward). Organizations such as the ICRP have to balance their desire for internal consistency and intellectual purity against the practical problems of their clients in adjusting to change. This paper indicates some of the changes that might be necessary over the next few years and how, given a pragmatic approach, they might be accommodated in Europe without too much regulatory confusion

  11. Self-consistent meson mass spectrum

    International Nuclear Information System (INIS)

    Balazs, L.A.P.

    1982-01-01

    A dual-topological-unitarization (or dual-fragmentation) approach to the calculation of hadron masses is presented, in which the effect of planar ''sea''-quark loops is taken into account from the beginning. Using techniques based on analyticity and generalized ladder-graph dynamics, we first derive the approximate ''generic'' Regge-trajectory formula α(t) = max (S 1 +S 2 , S 3 +S 4 )-(1/2) +2alpha-circumflex'[s/sub a/ +(1/2)(t-summationm/sub i/ 2 )] for any given hadronic process 1+2→3+4, where S/sub i/ and m/sub i/ are the spins and masses of i = 1,2,3,4, and √s/sub a/ is the effective mass of the lowest nonvanishing contribution (a) exchanged in the crossed channel. By requiring a minimization of secondary (background, etc.) contributions to a, and demanding simultaneous consistency for entire sets of such processes, we are then able to calculate the masses of all the lowest pseudoscalar and vector qq-bar states with q = u,d,s and the Regge trajectories on which they lie. By making certain additional assumptions we are also able to do this with q = u,d,c and q = u,d,b. Our only arbitrary parameters are m/sub rho/, m/sub K/*, m/sub psi/, and m/sub Upsilon/, one of which merely serves to fix the energy scale. In contrast to many other approaches, a small m/sub π/ 2 /m/sub rho/ 2 ratio arises quite naturally in the present scheme

  12. A New Heteroskedastic Consistent Covariance Matrix Estimator using Deviance Measure

    Directory of Open Access Journals (Sweden)

    Nuzhat Aftab

    2016-06-01

    Full Text Available In this article we propose a new heteroskedastic consistent covariance matrix estimator, HC6, based on deviance measure. We have studied and compared the finite sample behavior of the new test and compared it with other this kind of estimators, HC1, HC3 and HC4m, which are used in case of leverage observations. Simulation study is conducted to study the effect of various levels of heteroskedasticity on the size and power of quasi-t test with HC estimators. Results show that the test statistic based on our new suggested estimator has better asymptotic approximation and less size distortion as compared to other estimators for small sample sizes when high level ofheteroskedasticity is present in data.

  13. Bayesian detection of causal rare variants under posterior consistency.

    KAUST Repository

    Liang, Faming

    2013-07-26

    Identification of causal rare variants that are associated with complex traits poses a central challenge on genome-wide association studies. However, most current research focuses only on testing the global association whether the rare variants in a given genomic region are collectively associated with the trait. Although some recent work, e.g., the Bayesian risk index method, have tried to address this problem, it is unclear whether the causal rare variants can be consistently identified by them in the small-n-large-P situation. We develop a new Bayesian method, the so-called Bayesian Rare Variant Detector (BRVD), to tackle this problem. The new method simultaneously addresses two issues: (i) (Global association test) Are there any of the variants associated with the disease, and (ii) (Causal variant detection) Which variants, if any, are driving the association. The BRVD ensures the causal rare variants to be consistently identified in the small-n-large-P situation by imposing some appropriate prior distributions on the model and model specific parameters. The numerical results indicate that the BRVD is more powerful for testing the global association than the existing methods, such as the combined multivariate and collapsing test, weighted sum statistic test, RARECOVER, sequence kernel association test, and Bayesian risk index, and also more powerful for identification of causal rare variants than the Bayesian risk index method. The BRVD has also been successfully applied to the Early-Onset Myocardial Infarction (EOMI) Exome Sequence Data. It identified a few causal rare variants that have been verified in the literature.

  14. Bayesian detection of causal rare variants under posterior consistency.

    Directory of Open Access Journals (Sweden)

    Faming Liang

    Full Text Available Identification of causal rare variants that are associated with complex traits poses a central challenge on genome-wide association studies. However, most current research focuses only on testing the global association whether the rare variants in a given genomic region are collectively associated with the trait. Although some recent work, e.g., the Bayesian risk index method, have tried to address this problem, it is unclear whether the causal rare variants can be consistently identified by them in the small-n-large-P situation. We develop a new Bayesian method, the so-called Bayesian Rare Variant Detector (BRVD, to tackle this problem. The new method simultaneously addresses two issues: (i (Global association test Are there any of the variants associated with the disease, and (ii (Causal variant detection Which variants, if any, are driving the association. The BRVD ensures the causal rare variants to be consistently identified in the small-n-large-P situation by imposing some appropriate prior distributions on the model and model specific parameters. The numerical results indicate that the BRVD is more powerful for testing the global association than the existing methods, such as the combined multivariate and collapsing test, weighted sum statistic test, RARECOVER, sequence kernel association test, and Bayesian risk index, and also more powerful for identification of causal rare variants than the Bayesian risk index method. The BRVD has also been successfully applied to the Early-Onset Myocardial Infarction (EOMI Exome Sequence Data. It identified a few causal rare variants that have been verified in the literature.

  15. Bayesian detection of causal rare variants under posterior consistency.

    KAUST Repository

    Liang, Faming; Xiong, Momiao

    2013-01-01

    Identification of causal rare variants that are associated with complex traits poses a central challenge on genome-wide association studies. However, most current research focuses only on testing the global association whether the rare variants in a given genomic region are collectively associated with the trait. Although some recent work, e.g., the Bayesian risk index method, have tried to address this problem, it is unclear whether the causal rare variants can be consistently identified by them in the small-n-large-P situation. We develop a new Bayesian method, the so-called Bayesian Rare Variant Detector (BRVD), to tackle this problem. The new method simultaneously addresses two issues: (i) (Global association test) Are there any of the variants associated with the disease, and (ii) (Causal variant detection) Which variants, if any, are driving the association. The BRVD ensures the causal rare variants to be consistently identified in the small-n-large-P situation by imposing some appropriate prior distributions on the model and model specific parameters. The numerical results indicate that the BRVD is more powerful for testing the global association than the existing methods, such as the combined multivariate and collapsing test, weighted sum statistic test, RARECOVER, sequence kernel association test, and Bayesian risk index, and also more powerful for identification of causal rare variants than the Bayesian risk index method. The BRVD has also been successfully applied to the Early-Onset Myocardial Infarction (EOMI) Exome Sequence Data. It identified a few causal rare variants that have been verified in the literature.

  16. Thermodynamically consistent data-driven computational mechanics

    Science.gov (United States)

    González, David; Chinesta, Francisco; Cueto, Elías

    2018-05-01

    In the paradigm of data-intensive science, automated, unsupervised discovering of governing equations for a given physical phenomenon has attracted a lot of attention in several branches of applied sciences. In this work, we propose a method able to avoid the identification of the constitutive equations of complex systems and rather work in a purely numerical manner by employing experimental data. In sharp contrast to most existing techniques, this method does not rely on the assumption on any particular form for the model (other than some fundamental restrictions placed by classical physics such as the second law of thermodynamics, for instance) nor forces the algorithm to find among a predefined set of operators those whose predictions fit best to the available data. Instead, the method is able to identify both the Hamiltonian (conservative) and dissipative parts of the dynamics while satisfying fundamental laws such as energy conservation or positive production of entropy, for instance. The proposed method is tested against some examples of discrete as well as continuum mechanics, whose accurate results demonstrate the validity of the proposed approach.

  17. Consistency checks in beam emission modeling for neutral beam injectors

    International Nuclear Information System (INIS)

    Punyapu, Bharathi; Vattipalle, Prahlad; Sharma, Sanjeev Kumar; Baruah, Ujjwal Kumar; Crowley, Brendan

    2015-01-01

    In positive neutral beam systems, the beam parameters such as ion species fractions, power fractions and beam divergence are routinely measured using Doppler shifted beam emission spectrum. The accuracy with which these parameters are estimated depend on the accuracy of the atomic modeling involved in these estimations. In this work, an effective procedure to check the consistency of the beam emission modeling in neutral beam injectors is proposed. As a first consistency check, at a constant beam voltage and current, the intensity of the beam emission spectrum is measured by varying the pressure in the neutralizer. Then, the scaling of measured intensity of un-shifted (target) and Doppler shifted intensities (projectile) of the beam emission spectrum at these pressure values are studied. If the un-shifted component scales with pressure, then the intensity of this component will be used as a second consistency check on the beam emission modeling. As a further check, the modeled beam fractions and emission cross sections of projectile and target are used to predict the intensity of the un-shifted component and then compared with the value of measured target intensity. An agreement between the predicted and measured target intensities provide the degree of discrepancy in the beam emission modeling. In order to test this methodology, a systematic analysis of Doppler shift spectroscopy data obtained on the JET neutral beam test stand data was carried out

  18. Consistency in color parameters of a commonly used shade guide.

    Science.gov (United States)

    Tashkandi, Esam

    2010-01-01

    The use of shade guides to assess the color of natural teeth subjectively remains one of the most common means for dental shade assessment. Any variation in the color parameters of the different shade guides may lead to significant clinical implications. Particularly, since the communication between the clinic and the dental laboratory is based on using the shade guide designation. The purpose of this study was to investigate the consistency of the L∗a∗b∗ color parameters of a sample of a commonly used shade guide. The color parameters of a total of 100 VITAPAN Classical Vacuum shade guide (VITA Zahnfabrik, Bad Säckingen, Germany(were measured using a X-Rite ColorEye 7000A Spectrophotometer (Grand Rapids, Michigan, USA). Each shade guide consists of 16 tabs with different designations. Each shade tab was measured five times and the average values were calculated. The ΔE between the average L∗a∗b∗ value for each shade tab and the average of the 100 shade tabs of the same designation was calculated. Using the Student t-test analysis, no significant differences were found among the measured sample. There is a high consistency level in terms of color parameters of the measured VITAPAN Classical Vacuum shade guide sample tested.

  19. Corrosion Behavior of Alloy 625 in PbSO4-Pb3O4-PbCl2-ZnO-10 Wt Pct CdO Molten Salt Medium

    Science.gov (United States)

    Mohammadi Zahrani, E.; Alfantazi, A. M.

    2012-08-01

    Corrosion behavior and degradation mechanisms of alloy 625 under a 47.288 PbSO4-12.776 Pb3O4-6.844PbCl2-23.108ZnO-10CdO (wt pct) molten salt mixture under air atmosphere were studied at 873 K, 973 K, and 1073 K (600 °C, 700 °C, and 800 °C). Electrochemical impedance spectroscopy (EIS), open circuit potential (OCP) measurements, and potentiodynamic polarization techniques were used to evaluate the degradation mechanisms and characterize the corrosion behavior of the alloy. Morphology, chemical composition, and phase structure of the corrosion products and surface layers of the corroded specimens were studied by scanning electron microscopy/energy-dispersive X-ray (SEM/EDX) and X-ray map analyses. Results confirmed that during the exposure of alloy 625 to the molten salt, chromium was mainly dissolved through an active oxidation process as CrO3, Cr2O3, and CrNbO4, while nickel dissolved only as NiO in the system. Formation of a porous and nonprotective oxide layer with low resistance is responsible for the weak protective properties of the barrier layer at high temperatures of 973 K and 1073 K (700 °C and 800 °C). There were two kinds of attack for INCONEL 625, including general surface corrosion and pitting. Pitting corrosion occurred due to the breakdown of the initial oxide layer by molten salt dissolution of the oxide or oxide cracking.

  20. Quantitative verification of ab initio self-consistent laser theory.

    Science.gov (United States)

    Ge, Li; Tandy, Robert J; Stone, A D; Türeci, Hakan E

    2008-10-13

    We generalize and test the recent "ab initio" self-consistent (AISC) time-independent semiclassical laser theory. This self-consistent formalism generates all the stationary lasing properties in the multimode regime (frequencies, thresholds, internal and external fields, output power and emission pattern) from simple inputs: the dielectric function of the passive cavity, the atomic transition frequency, and the transverse relaxation time of the lasing transition.We find that the theory gives excellent quantitative agreement with full time-dependent simulations of the Maxwell-Bloch equations after it has been generalized to drop the slowly-varying envelope approximation. The theory is infinite order in the non-linear hole-burning interaction; the widely used third order approximation is shown to fail badly.

  1. ROSA-III 50 % break integral test RUN 916

    International Nuclear Information System (INIS)

    Yonomoto, Taisuke; Tasaka, Kanji; Koizumi, Yasuo; Anoda, Yoshinari; Kumamaru, Hiroshige; Nakamura, Hideo; Suzuki, Mitsuhiro; Murata, Hideo

    1985-08-01

    This report presents the experimental data of RUN 916 conducted at the ROSA-III test facility. The facility is a volumetrically scaled (1/424) simulator for a BWR/6 with the electrically heated core, the break simulator and the scaled ECCS(emergency core cooling system). RUN 916 was a 50 % split break test at the recirculation pump suction line with an assumption of HPCS diegel generator failure and conducted as one of the break area parameter tests. A peak cladding temperature (PCT) of 917 K was reached at 190 s after the break during the reflooding phase. Whole core was completely quenched by ECCS, and the effectiveness of ECCS was confermed. The primary test results of RUN 916 are compared in this report with those of RUN 926, which was a 200 % double-ended break test. The initiation of core dryout in RUN 916 was later than that in RUN 926 because of the smaller discharge flow rate. Duration of core dryourt was, however, longer in RUN 916 because of later actuation of ECCSs. PCT in RUN 916 was 133 K higher than that in RUN 926. (author)

  2. ROSA-III 100 % break integral test Run 914

    International Nuclear Information System (INIS)

    Yonomoto, Taisuke; Tasaka, Kanji; Koizumi, Yasuo; Anoda, Yoshinari; Kumamaru, Hiroshige; Nakamura, Hideo; Suzuki, Mitsuhiro; Murata, Hideo

    1987-05-01

    This report presents the experimental data of RUN 914 conducted at the ROSA-III test facility. The facility is a volumetrically scaled (1/424) simulator for a BWR/6 with the electrically heated core, the break simulator and the scaled ECCS (emergency core cooling system). RUN 914 was a 100% split break test at the recirculation pump suction line with an assumption of HPCS diesel generator failure and conducted as one of the break area parameter tests. A peak cladding temperature (PCT) of 851 K was reached at 130 s after the break during the reflooding phase. Whole core was completely quenched by ECCS, and the effectiveness of ECCS was confirmed. The primary test results of RUN 914 are compared in this report with those of RUN 926, which was a 200 % double-ended break test. The initiation of core dryout in RUN 914 was almost the same as that in RUN 926. Duration of core dryourt was, however, longer in RUN 914 because of later actuation of ECCSs. PCT in RUN 914 was 67 K higher than that in RUN 926. (author)

  3. Performance and consistency of indicator groups in two biodiversity hotspots.

    Directory of Open Access Journals (Sweden)

    Joaquim Trindade-Filho

    Full Text Available In a world limited by data availability and limited funds for conservation, scientists and practitioners must use indicator groups to define spatial conservation priorities. Several studies have evaluated the effectiveness of indicator groups, but still little is known about the consistency in performance of these groups in different regions, which would allow their a priori selection.We systematically examined the effectiveness and the consistency of nine indicator groups in representing mammal species in two top-ranked Biodiversity Hotspots (BH: the Brazilian Cerrado and the Atlantic Forest. To test for group effectiveness we first found the best sets of sites able to maximize the representation of each indicator group in the BH and then calculated the average representation of different target species by the indicator groups in the BH. We considered consistent indicator groups whose representation of target species was not statistically different between BH. We called effective those groups that outperformed the target-species representation achieved by random sets of species. Effective indicator groups required the selection of less than 2% of the BH area for representing target species. Restricted-range species were the most effective indicators for the representation of all mammal diversity as well as target species. It was also the only group with high consistency.We show that several indicator groups could be applied as shortcuts for representing mammal species in the Cerrado and the Atlantic Forest to develop conservation plans, however, only restricted-range species consistently held as the most effective indicator group for such a task. This group is of particular importance in conservation planning as it captures high diversity of endemic and endangered species.

  4. Performance and consistency of indicator groups in two biodiversity hotspots.

    Science.gov (United States)

    Trindade-Filho, Joaquim; Loyola, Rafael Dias

    2011-01-01

    In a world limited by data availability and limited funds for conservation, scientists and practitioners must use indicator groups to define spatial conservation priorities. Several studies have evaluated the effectiveness of indicator groups, but still little is known about the consistency in performance of these groups in different regions, which would allow their a priori selection. We systematically examined the effectiveness and the consistency of nine indicator groups in representing mammal species in two top-ranked Biodiversity Hotspots (BH): the Brazilian Cerrado and the Atlantic Forest. To test for group effectiveness we first found the best sets of sites able to maximize the representation of each indicator group in the BH and then calculated the average representation of different target species by the indicator groups in the BH. We considered consistent indicator groups whose representation of target species was not statistically different between BH. We called effective those groups that outperformed the target-species representation achieved by random sets of species. Effective indicator groups required the selection of less than 2% of the BH area for representing target species. Restricted-range species were the most effective indicators for the representation of all mammal diversity as well as target species. It was also the only group with high consistency. We show that several indicator groups could be applied as shortcuts for representing mammal species in the Cerrado and the Atlantic Forest to develop conservation plans, however, only restricted-range species consistently held as the most effective indicator group for such a task. This group is of particular importance in conservation planning as it captures high diversity of endemic and endangered species.

  5. Time-Consistent and Market-Consistent Evaluations (Revised version of 2012-086)

    NARCIS (Netherlands)

    Stadje, M.A.; Pelsser, A.

    2014-01-01

    Abstract: We consider evaluation methods for payoffs with an inherent financial risk as encountered for instance for portfolios held by pension funds and insurance companies. Pricing such payoffs in a way consistent to market prices typically involves combining actuarial techniques with methods from

  6. SIMPLE ESTIMATOR AND CONSISTENT STRONGLY OF STABLE DISTRIBUTIONS

    Directory of Open Access Journals (Sweden)

    Cira E. Guevara Otiniano

    2016-06-01

    Full Text Available Stable distributions are extensively used to analyze earnings of financial assets, such as exchange rates and stock prices assets. In this paper we propose a simple and strongly consistent estimator for the scale parameter of a symmetric stable L´evy distribution. The advantage of this estimator is that your computational time is minimum thus it can be used to initialize intensive computational procedure such as maximum likelihood. With random samples of sized n we tested the efficacy of these estimators by Monte Carlo method. We also included applications for three data sets.

  7. Stereotypes help people connect with others in the community: A situated functional analysis of the stereotype consistency bias in communication

    NARCIS (Netherlands)

    Clark, A.E.; Kashima, Y.

    2007-01-01

    Communicators tend to share more stereotype-consistent than stereotype-inconsistent information. The authors propose and test a situated functional model of this stereotype consistency bias: Stereotype-consistent and inconsistent information differentially serve 2 central functions of

  8. Consistently Showing Your Best Side? Intra-individual Consistency in #Selfie Pose Orientation

    Science.gov (United States)

    Lindell, Annukka K.

    2017-01-01

    Painted and photographic portraits of others show an asymmetric bias: people favor their left cheek. Both experimental and database studies confirm that the left cheek bias extends to selfies. To date all such selfie studies have been cross-sectional; whether individual selfie-takers tend to consistently favor the same pose orientation, or switch between multiple poses, remains to be determined. The present study thus examined intra-individual consistency in selfie pose orientations. Two hundred selfie-taking participants (100 male and 100 female) were identified by searching #selfie on Instagram. The most recent 10 single-subject selfies for the each of the participants were selected and coded for type of selfie (normal; mirror) and pose orientation (left, midline, right), resulting in a sample of 2000 selfies. Results indicated that selfie-takers do tend to consistently adopt a preferred pose orientation (α = 0.72), with more participants showing an overall left cheek bias (41%) than would be expected by chance (overall right cheek bias = 31.5%; overall midline bias = 19.5%; no overall bias = 8%). Logistic regression modellng, controlling for the repeated measure of participant identity, indicated that sex did not affect pose orientation. However, selfie type proved a significant predictor when comparing left and right cheek poses, with a stronger left cheek bias for mirror than normal selfies. Overall, these novel findings indicate that selfie-takers show intra-individual consistency in pose orientation, and in addition, replicate the previously reported left cheek bias for selfies and other types of portrait, confirming that the left cheek bias also presents within individuals’ selfie corpora. PMID:28270790

  9. Consistently Trained Artificial Neural Network for Automatic Ship Berthing Control

    Directory of Open Access Journals (Sweden)

    Y.A. Ahmed

    2015-09-01

    Full Text Available In this paper, consistently trained Artificial Neural Network controller for automatic ship berthing is discussed. Minimum time course changing manoeuvre is utilised to ensure such consistency and a new concept named ‘virtual window’ is introduced. Such consistent teaching data are then used to train two separate multi-layered feed forward neural networks for command rudder and propeller revolution output. After proper training, several known and unknown conditions are tested to judge the effectiveness of the proposed controller using Monte Carlo simulations. After getting acceptable percentages of success, the trained networks are implemented for the free running experiment system to judge the network’s real time response for Esso Osaka 3-m model ship. The network’s behaviour during such experiments is also investigated for possible effect of initial conditions as well as wind disturbances. Moreover, since the final goal point of the proposed controller is set at some distance from the actual pier to ensure safety, therefore a study on automatic tug assistance is also discussed for the final alignment of the ship with actual pier.

  10. EVALUATION OF CONSISTENCY AND SETTING TIME OF IRANIAN DENTAL STONES

    Directory of Open Access Journals (Sweden)

    F GOL BIDI

    2000-09-01

    Full Text Available Introduction. Dental stones are widely used in dentistry and the success or failure of many dental treatments depend on the accuracy of these gypsums. The purpose of this study was the evaluation of Iranian dental stones and comparison between Iranian and foreign ones. In this investigation, consistency and setting time were compared between Pars Dendn, Almas and Hinrizit stones. The latter is accepted by ADA (American Dental Association. Consistency and setting time are 2 of 5 properties that are necessitated by both ADA specification No. 25 and Iranian Standard Organization specification No. 2569 for evaluation of dental stones. Methods. In this study, the number and preparation of specimens and test conditions were done according to the ADA specification No. 25 and all the measurements were done with vicat apparatus. Results. The results of this study showed that the standard consistency of Almas stone was obtained by 42ml water and 100gr powder and the setting time of this stone was 11±0.03 min. Which was with in the limits of ADA specification (12±4 min. The standard consistency of Pars Dandan stone was obrianed by 31ml water and 100 gr powder, but the setting time of this stone was 5± 0.16 min which was nt within the limits of ADA specification. Discussion: Comparison of Iranian and Hinrizit stones properties showed that two probable problems of Iranian stones are:1- Unhemogrnousity of Iranian stoned powder was caused by uncontrolled temperature, pressure and humidity in the production process of stone. 2- Impurities such as sodium chloride was responsible fo shortening of Pars Dendens setting time.

  11. Consistency in performance evaluation reports and medical records.

    Science.gov (United States)

    Lu, Mingshan; Ma, Ching-to Albert

    2002-12-01

    In the health care market managed care has become the latest innovation for the delivery of services. For efficient implementation, the managed care organization relies on accurate information. So clinicians are often asked to report on patients before referrals are approved, treatments authorized, or insurance claims processed. What are clinicians responses to solicitation for information by managed care organizations? The existing health literature has already pointed out the importance of provider gaming, sincere reporting, nudging, and dodging the rules. We assess the consistency of clinicians reports on clients across administrative data and clinical records. For about 1,000 alcohol abuse treatment episodes, we compare clinicians reports across two data sets. The first one, the Maine Addiction Treatment System (MATS), was an administrative data set; the state government used it for program performance monitoring and evaluation. The second was a set of medical record abstracts, taken directly from the clinical records of treatment episodes. A clinician s reporting practice exhibits an inconsistency if the information reported in MATS differs from the information reported in the medical record in a statistically significant way. We look for evidence of inconsistencies in five categories: admission alcohol use frequency, discharge alcohol use frequency, termination status, admission employment status, and discharge employment status. Chi-square tests, Kappa statistics, and sensitivity and specificity tests are used for hypothesis testing. Multiple imputation methods are employed to address the problem of missing values in the record abstract data set. For admission and discharge alcohol use frequency measures, we find, respectively, strong and supporting evidence for inconsistencies. We find equally strong evidence for consistency in reports of admission and discharge employment status, and mixed evidence on report consistency on termination status. Patterns of

  12. Efficient self-consistency for magnetic tight binding

    Science.gov (United States)

    Soin, Preetma; Horsfield, A. P.; Nguyen-Manh, D.

    2011-06-01

    Tight binding can be extended to magnetic systems by including an exchange interaction on an atomic site that favours net spin polarisation. We have used a published model, extended to include long-ranged Coulomb interactions, to study defects in iron. We have found that achieving self-consistency using conventional techniques was either unstable or very slow. By formulating the problem of achieving charge and spin self-consistency as a search for stationary points of a Harris-Foulkes functional, extended to include spin, we have derived a much more efficient scheme based on a Newton-Raphson procedure. We demonstrate the capabilities of our method by looking at vacancies and self-interstitials in iron. Self-consistency can indeed be achieved in a more efficient and stable manner, but care needs to be taken to manage this. The algorithm is implemented in the code PLATO. Program summaryProgram title:PLATO Catalogue identifier: AEFC_v2_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEFC_v2_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 228 747 No. of bytes in distributed program, including test data, etc.: 1 880 369 Distribution format: tar.gz Programming language: C and PERL Computer: Apple Macintosh, PC, Unix machines Operating system: Unix, Linux, Mac OS X, Windows XP Has the code been vectorised or parallelised?: Yes. Up to 256 processors tested RAM: Up to 2 Gbytes per processor Classification: 7.3 External routines: LAPACK, BLAS and optionally ScaLAPACK, BLACS, PBLAS, FFTW Catalogue identifier of previous version: AEFC_v1_0 Journal reference of previous version: Comput. Phys. Comm. 180 (2009) 2616 Does the new version supersede the previous version?: Yes Nature of problem: Achieving charge and spin self-consistency in magnetic tight binding can be very

  13. Measuring consistency of web page design and its effects on performance and satisfaction.

    Science.gov (United States)

    Ozok, A A; Salvendy, G

    2000-04-01

    This study examines the methods for measuring the consistency levels of web pages and the effect of consistency on the performance and satisfaction of the world-wide web (WWW) user. For clarification, a home page is referred to as a single page that is the default page of a web site on the WWW. A web page refers to a single screen that indicates a specific address on the WWW. This study has tested a series of web pages that were mostly hyperlinked. Therefore, the term 'web page' has been adopted for the nomenclature while referring to the objects of which the features were tested. It was hypothesized that participants would perform better and be more satisfied using web pages that have consistent rather than inconsistent interface design; that the overall consistency level of an interface design would significantly correlate with the three elements of consistency, physical, communicational and conceptual consistency; and that physical and communicational consistencies would interact with each other. The hypotheses were tested in a four-group, between-subject design, with 10 participants in each group. The results partially support the hypothesis regarding error rate, but not regarding satisfaction and performance time. The results also support the hypothesis that each of the three elements of consistency significantly contribute to the overall consistency of a web page, and that physical and communicational consistencies interact with each other, while conceptual consistency does not interact with them.

  14. Planck 2013 results. XXXI. Consistency of the Planck data

    DEFF Research Database (Denmark)

    Ade, P. A. R.; Arnaud, M.; Ashdown, M.

    2014-01-01

    The Planck design and scanning strategy provide many levels of redundancy that can be exploited to provide tests of internal consistency. One of the most important is the comparison of the 70 GHz (amplifier) and 100 GHz (bolometer) channels. Based on dierent instrument technologies, with feeds...... in the HFI channels would result in shifts in the posterior distributions of parameters of less than 0.3σ except for As, the amplitude of the primordial curvature perturbations at 0.05 Mpc-1, which changes by about 1.We extend these comparisons to include the sky maps from the complete nine-year mission...... located dierently in the focal plane, analysed independently by dierent teams using dierent software, and near∫ the minimum of diuse foreground emission, these channels are in eect two dierent experiments. The 143 GHz channel has the lowest noise level on Planck, and is near the minimum of unresolved...

  15. Parton Distributions based on a Maximally Consistent Dataset

    Science.gov (United States)

    Rojo, Juan

    2016-04-01

    The choice of data that enters a global QCD analysis can have a substantial impact on the resulting parton distributions and their predictions for collider observables. One of the main reasons for this has to do with the possible presence of inconsistencies, either internal within an experiment or external between different experiments. In order to assess the robustness of the global fit, different definitions of a conservative PDF set, that is, a PDF set based on a maximally consistent dataset, have been introduced. However, these approaches are typically affected by theory biases in the selection of the dataset. In this contribution, after a brief overview of recent NNPDF developments, we propose a new, fully objective, definition of a conservative PDF set, based on the Bayesian reweighting approach. Using the new NNPDF3.0 framework, we produce various conservative sets, which turn out to be mutually in agreement within the respective PDF uncertainties, as well as with the global fit. We explore some of their implications for LHC phenomenology, finding also good consistency with the global fit result. These results provide a non-trivial validation test of the new NNPDF3.0 fitting methodology, and indicate that possible inconsistencies in the fitted dataset do not affect substantially the global fit PDFs.

  16. Consistent three-equation model for thin films

    Science.gov (United States)

    Richard, Gael; Gisclon, Marguerite; Ruyer-Quil, Christian; Vila, Jean-Paul

    2017-11-01

    Numerical simulations of thin films of newtonian fluids down an inclined plane use reduced models for computational cost reasons. These models are usually derived by averaging over the fluid depth the physical equations of fluid mechanics with an asymptotic method in the long-wave limit. Two-equation models are based on the mass conservation equation and either on the momentum balance equation or on the work-energy theorem. We show that there is no two-equation model that is both consistent and theoretically coherent and that a third variable and a three-equation model are required to solve all theoretical contradictions. The linear and nonlinear properties of two and three-equation models are tested on various practical problems. We present a new consistent three-equation model with a simple mathematical structure which allows an easy and reliable numerical resolution. The numerical calculations agree fairly well with experimental measurements or with direct numerical resolutions for neutral stability curves, speed of kinematic waves and of solitary waves and depth profiles of wavy films. The model can also predict the flow reversal at the first capillary trough ahead of the main wave hump.

  17. Planck 2013 results. XXXI. Consistency of the Planck data

    CERN Document Server

    Ade, P A R; Ashdown, M; Aumont, J; Baccigalupi, C; Banday, A.J; Barreiro, R.B; Battaner, E; Benabed, K; Benoit-Levy, A; Bernard, J.P; Bersanelli, M; Bielewicz, P; Bond, J.R; Borrill, J; Bouchet, F.R; Burigana, C; Cardoso, J.F; Catalano, A; Challinor, A; Chamballu, A; Chiang, H.C; Christensen, P.R; Clements, D.L; Colombi, S; Colombo, L.P.L; Couchot, F; Coulais, A; Crill, B.P; Curto, A; Cuttaia, F; Danese, L; Davies, R.D; Davis, R.J; de Bernardis, P; de Rosa, A; de Zotti, G; Delabrouille, J; Desert, F.X; Dickinson, C; Diego, J.M; Dole, H; Donzelli, S; Dore, O; Douspis, M; Dupac, X; Ensslin, T.A; Eriksen, H.K; Finelli, F; Forni, O; Frailis, M; Fraisse, A A; Franceschi, E; Galeotta, S; Ganga, K; Giard, M; Gonzalez-Nuevo, J; Gorski, K.M.; Gratton, S.; Gregorio, A; Gruppuso, A; Gudmundsson, J E; Hansen, F.K; Hanson, D; Harrison, D; Henrot-Versille, S; Herranz, D; Hildebrandt, S.R; Hivon, E; Hobson, M; Holmes, W.A.; Hornstrup, A; Hovest, W.; Huffenberger, K.M; Jaffe, T.R; Jaffe, A.H; Jones, W.C; Keihanen, E; Keskitalo, R; Knoche, J; Kunz, M; Kurki-Suonio, H; Lagache, G; Lahteenmaki, A; Lamarre, J.M; Lasenby, A; Lawrence, C.R; Leonardi, R; Leon-Tavares, J; Lesgourgues, J; Liguori, M; Lilje, P.B; Linden-Vornle, M; Lopez-Caniego, M; Lubin, P.M; Macias-Perez, J.F; Maino, D; Mandolesi, N; Maris, M; Martin, P.G; Martinez-Gonzalez, E; Masi, S; Matarrese, S; Mazzotta, P; Meinhold, P.R; Melchiorri, A; Mendes, L; Mennella, A; Migliaccio, M; Mitra, S; Miville-Deschenes, M.A; Moneti, A; Montier, L; Morgante, G; Mortlock, D; Moss, A; Munshi, D; Murphy, J A; Naselsky, P; Nati, F; Natoli, P; Norgaard-Nielsen, H.U; Noviello, F; Novikov, D; Novikov, I; Oxborrow, C.A; Pagano, L; Pajot, F; Paoletti, D; Partridge, B; Pasian, F; Patanchon, G; Pearson, D; Pearson, T.J; Perdereau, O; Perrotta, F; Piacentini, F; Piat, M; Pierpaoli, E; Pietrobon, D; Plaszczynski, S; Pointecouteau, E; Polenta, G; Ponthieu, N; Popa, L; Pratt, G.W; Prunet, S; Puget, J.L; Rachen, J.P; Reinecke, M; Remazeilles, M; Renault, C; Ricciardi, S.; Ristorcelli, I; Rocha, G.; Roudier, G; Rubino-Martin, J.A; Rusholme, B; Sandri, M; Scott, D; Stolyarov, V; Sudiwala, R; Sutton, D; Suur-Uski, A.S; Sygnet, J.F; Tauber, J.A; Terenzi, L; Toffolatti, L; Tomasi, M; Tristram, M; Tucci, M; Valenziano, L; Valiviita, J; Van Tent, B; Vielva, P; Villa, F; Wade, L.A; Wandelt, B.D; Wehus, I K; White, S D M; Yvon, D; Zacchei, A; Zonca, A

    2014-01-01

    The Planck design and scanning strategy provide many levels of redundancy that can be exploited to provide tests of internal consistency. One of the most important is the comparison of the 70 GHz (amplifier) and 100 GHz (bolometer) channels. Based on different instrument technologies, with feeds located differently in the focal plane, analysed independently by different teams using different software, and near the minimum of diffuse foreground emission, these channels are in effect two different experiments. The 143 GHz channel has the lowest noise level on Planck, and is near the minimum of unresolved foreground emission. In this paper, we analyse the level of consistency achieved in the 2013 Planck data. We concentrate on comparisons between the 70, 100, and 143 GHz channel maps and power spectra, particularly over the angular scales of the first and second acoustic peaks, on maps masked for diffuse Galactic emission and for strong unresolved sources. Difference maps covering angular scales from 8°...

  18. Sex-differences and temporal consistency in stickleback fish boldness.

    Directory of Open Access Journals (Sweden)

    Andrew J King

    Full Text Available Behavioural traits that co-vary across contexts or situations often reflect fundamental trade-offs which individuals experience in different contexts (e.g. fitness trade-offs between exploration and predation risk. Since males tend to experience greater variance in reproductive success than females, there may be considerable fitness benefits associated with "bolder" behavioural types, but only recently have researchers begun to consider sex-specific and life-history strategies associated with these. Here we test the hypothesis that male three-spined sticklebacks (Gasterosteus aculeatus show high risk but potentially high return behaviours compared to females. According to this hypothesis we predicted that male fish would show greater exploration of their environment in a foraging context, and be caught sooner by an experimenter than females. We found that the time fish spent out of cover exploring their environment was correlated over two days, and males spent significantly more time out of cover than females. Also, the order in which fish were net-caught from their holding aquarium by an experimenter prior to experiments was negatively correlated with the time spent out of cover during tests, and males tended to be caught sooner than females. Moreover, we found a positive correlation between the catch number prior to our experiments and nine months after, pointing towards consistent, long-term individual differences in behaviour.

  19. Multiplicative Consistency for Interval Valued Reciprocal Preference Relations

    OpenAIRE

    Wu, Jian; Chiclana, Francisco

    2014-01-01

    The multiplicative consistency (MC) property of interval additive reciprocal preference relations (IARPRs) is explored, and then the consistency index is quantified by the multiplicative consistency estimated IARPR. The MC property is used to measure the level of consistency of the information provided by the experts and also to propose the consistency index induced ordered weighted averaging (CI-IOWA) operator. The novelty of this operator is that it aggregates individual IARPRs in such ...

  20. Assessing distances and consistency of kinematics in Gaia/TGAS

    Science.gov (United States)

    Schönrich, Ralph; Aumer, Michael

    2017-12-01

    We apply the statistical methods by Schönrich, Binney & Asplund to assess the quality of distances and kinematics in the Radial Velocity Experiment (RAVE)-Tycho-Gaia Astrometric Solution (TGAS) and Large Sky Area Multiobject Fiber Spectroscopic Telescope (LAMOST)-TGAS samples of Solar neighbourhood stars. These methods yield a nominal distance accuracy of 1-2 per cent. Other than common tests on parallax accuracy, they directly test distance estimations including the effects of distance priors. We show how to construct these priors including the survey selection functions (SSFs) directly from the data. We demonstrate that neglecting the SSFs causes severe distance biases. Due to the decline of the SSFs in distance, the simple 1/parallax estimate only mildly underestimates distances. We test the accuracy of measured line-of-sight velocities (vlos) by binning the samples in the nominal vlos uncertainties. We find: (i) the LAMOST vlos have a ∼-5 km s-1 offset; (ii) the average LAMOST measurement error for vlos is ∼7 km s-1, significantly smaller than, and nearly uncorrelated with the nominal LAMOST estimates. The RAVE sample shows either a moderate distance underestimate, or an unaccounted source of vlos dispersion (e∥) from measurement errors and binary stars. For a subsample of suspected binary stars in RAVE, our methods indicate significant distance underestimates. Separating a sample in metallicity or kinematics to select thick-disc/halo stars, discriminates between distance bias and e∥. For LAMOST, this separation yields consistency with pure vlos measurement errors. We find an anomaly near longitude l ∼ (300 ± 60)° and distance s ∼ (0.32 ± 0.03) kpc on both sides of the galactic plane, which could be explained by either a localized distance error or a breathing mode.

  1. Skin extensibility and consistency in patients with Ehlers-Danlos syndrome and benign joint hypermobility syndrome

    DEFF Research Database (Denmark)

    Remvig, L; Duhn, P H; Ullman, S

    2009-01-01

    OBJECTIVES: The reproducibility of clinical tests for skin extensibility and consistency, essential for differentiating between types of Ehlers-Danlos syndrome (EDS) and benign joint hypermobility syndrome (BJHS), is unknown. Paraclinical methods may provide objective differential diagnostic...

  2. Can ancestry be consistently determined from the skeleton?

    Directory of Open Access Journals (Sweden)

    Sierp Ingrid

    2015-03-01

    Full Text Available Although the concept of race has been thoroughly criticised in biological anthropology, forensic anthropology still uses a number of methods to determine the ‘race’ of a skeleton. The methods must be evaluated to see how effective they are given large individual variation. This study used 20 cases of skeletons of varied provenance to test whether the nine published methods of ‘race’ determination, using a range of various approaches, were able to consistently identify the ethnic origin. No one individual was identified as belonging to just one ‘major racial class’, e.g. European, meaning that complete consistency across all nine methods was not observed. In 14 cases (70%, various methods identified the same individual as belonging to all three racial classes. This suggests that the existing methods for the determination of ‘race’ are compromised. The very concept of ‘race’ is inapplicable to variation that occurs between populations only in small ways and the methods are limited by the geographic population from which their discriminant functions or observations of morphological traits were derived. Methods of multivariate linear discriminant analysis, e.g. CRANID, are supposed to allocate an individual skull to a specific population rather than a ‘major race’. In our analysis CRANID did not produce convincing allocations of individual skeletons to specific populations. The findings of this study show that great caution must be taken when attempting to ascertain the ‘race’ of a skeleton, as the outcome is not only dependent on which skeletal sites are available for assessment, but also the degree to which the unknown skeleton’s population of origin has been investigated.

  3. Creation of Consistent Burn Wounds: A Rat Model

    Directory of Open Access Journals (Sweden)

    Elijah Zhengyang Cai

    2014-07-01

    Full Text Available Background Burn infliction techniques are poorly described in rat models. An accurate study can only be achieved with wounds that are uniform in size and depth. We describe a simple reproducible method for creating consistent burn wounds in rats. Methods Ten male Sprague-Dawley rats were anesthetized and dorsum shaved. A 100 g cylindrical stainless-steel rod (1 cm diameter was heated to 100℃ in boiling water. Temperature was monitored using a thermocouple. We performed two consecutive toe-pinch tests on different limbs to assess the depth of sedation. Burn infliction was limited to the loin. The skin was pulled upwards, away from the underlying viscera, creating a flat surface. The rod rested on its own weight for 5, 10, and 20 seconds at three different sites on each rat. Wounds were evaluated for size, morphology and depth. Results Average wound size was 0.9957 cm2 (standard deviation [SD] 0.1845 (n=30. Wounds created with duration of 5 seconds were pale, with an indistinct margin of erythema. Wounds of 10 and 20 seconds were well-defined, uniformly brown with a rim of erythema. Average depths of tissue damage were 1.30 mm (SD 0.424, 2.35 mm (SD 0.071, and 2.60 mm (SD 0.283 for duration of 5, 10, 20 seconds respectively. Burn duration of 5 seconds resulted in full-thickness damage. Burn duration of 10 seconds and 20 seconds resulted in full-thickness damage, involving subjacent skeletal muscle. Conclusions This is a simple reproducible method for creating burn wounds consistent in size and depth in a rat burn model.

  4. Trait Implications as a Moderator of Recall of Stereotype-Consistent and Stereotype-Inconsistent Behaviors.

    Science.gov (United States)

    Dijksterhuis, Ap; Knippenberg, Ad van

    1996-01-01

    Tests the assumption that organization in memory of behavior information and recall depends on the descriptive relatedness of consistent information with inconsistent information. Subjects read stereotype-consistent and stereotype-inconsistent behavioral descriptions implying the same trait dimension or different trait dimensions. Predictions were…

  5. HOME COST Pct Units by Rent as Pct of HH Income in 1999 NMSD 2000

    Data.gov (United States)

    Earth Data Analysis Center, University of New Mexico — The 2006 Second Edition TIGER/Line files are an extract of selected geographic and cartographic information from the Census TIGER database. The geographic coverage...

  6. HOME COST Pct All OwnerOcc Units Costs as Pct 1999 HH Income NMHD 2000

    Data.gov (United States)

    Earth Data Analysis Center, University of New Mexico — The 2006 Second Edition TIGER/Line files are an extract of selected geographic and cartographic information from the Census TIGER database. The geographic coverage...

  7. HOME COST Pct Mortgaged Units by Costs as Pct 1999 HH Income NMSD 2000

    Data.gov (United States)

    Earth Data Analysis Center, University of New Mexico — The 2006 Second Edition TIGER/Line files are an extract of selected geographic and cartographic information from the Census TIGER database. The geographic coverage...

  8. Recovery and Precipitate Analysis of 9 Pct Cr-1 Pct MoVNb Steel during Creep

    NARCIS (Netherlands)

    Kabadwal, A.; Tamura, M.; Shinozuka, K.; Esaka, H.

    The effect of tempering temperature and creep exposure on the microstructure of a modified 9Cr steel was investigated. Creep-interrupted specimens, including the grip portion, were investigated precisely using mainly X-ray and inductively coupled plasma (ICP) spectroscopy. After saturation of

  9. Privacy, Time Consistent Optimal Labour Income Taxation and Education Policy

    OpenAIRE

    Konrad, Kai A.

    1999-01-01

    Incomplete information is a commitment device for time consistency problems. In the context of time consistent labour income taxation privacy reduces welfare losses and increases the effectiveness of public education as a second best policy.

  10. Generalized contexts and consistent histories in quantum mechanics

    International Nuclear Information System (INIS)

    Losada, Marcelo; Laura, Roberto

    2014-01-01

    We analyze a restriction of the theory of consistent histories by imposing that a valid description of a physical system must include quantum histories which satisfy the consistency conditions for all states. We prove that these conditions are equivalent to imposing the compatibility conditions of our formalism of generalized contexts. Moreover, we show that the theory of consistent histories with the consistency conditions for all states and the formalism of generalized context are equally useful representing expressions which involve properties at different times

  11. Personality and Situation Predictors of Consistent Eating Patterns

    OpenAIRE

    Vainik, Uku; Dub?, Laurette; Lu, Ji; Fellows, Lesley K.

    2015-01-01

    Introduction A consistent eating style might be beneficial to avoid overeating in a food-rich environment. Eating consistency entails maintaining a similar dietary pattern across different eating situations. This construct is relatively under-studied, but the available evidence suggests that eating consistency supports successful weight maintenance and decreases risk for metabolic syndrome and cardiovascular disease. Yet, personality and situation predictors of consistency have not been studi...

  12. Longitudinal tDCS: Consistency across Working Memory Training Studies

    Directory of Open Access Journals (Sweden)

    Marian E. Berryhill

    2017-04-01

    Full Text Available There is great interest in enhancing and maintaining cognitive function. In recent years, advances in noninvasive brain stimulation devices, such as transcranial direct current stimulation (tDCS, have targeted working memory in particular. Despite controversy surrounding outcomes of single-session studies, a growing field of working memory training studies incorporate multiple sessions of tDCS. It is useful to take stock of these findings because there is a diversity of paradigms employed and the outcomes observed between research groups. This will be important in assessing cognitive training programs paired with stimulation techniques and identifying the more useful and less effective approaches. Here, we treat the tDCS+ working memory training field as a case example, but also survey training benefits in other neuromodulatory techniques (e.g., tRNS, tACS. There are challenges associated with the broad parameter space including: individual differences, stimulation intensity, duration, montage, session number, session spacing, training task selection, timing of follow up testing, near and far transfer tasks. In summary, although the field of assisted cognitive training is young, some design choices are more favorable than others. By way of heuristic, the current evidence supports including more training/tDCS sessions (5+, applying anodal tDCS targeting prefrontal regions, including follow up testing on trained and transfer tasks after a period of no contact. What remains unclear, but important for future translational value is continuing work to pinpoint optimal values for the tDCS parameters on a per cognitive task basis. Importantly the emerging literature shows notable consistency in the application of tDCS for WM across various participant populations compared to single session experimental designs.

  13. Two Impossibility Results on the Converse Consistency Principle in Bargaining

    OpenAIRE

    Youngsub Chun

    1999-01-01

    We present two impossibility results on the converse consistency principle in the context of bargaining. First, we show that there is no solution satis-fying Pareto optimality, contraction independence, and converse consistency. Next, we show that there is no solution satisfying Pareto optimality, strong individual rationality, individual monotonicity, and converse consistency.

  14. Checking Consistency of Pedigree Information is NP-complete

    DEFF Research Database (Denmark)

    Aceto, Luca; Hansen, Jens A.; Ingolfsdottir, Anna

    Consistency checking is a fundamental computational problem in genetics. Given a pedigree and information on the genotypes of some of the individuals in it, the aim of consistency checking is to determine whether these data are consistent with the classic Mendelian laws of inheritance. This probl...

  15. 26 CFR 1.338-8 - Asset and stock consistency.

    Science.gov (United States)

    2010-04-01

    ... that are controlled foreign corporations. (6) Stock consistency. This section limits the application of... 26 Internal Revenue 4 2010-04-01 2010-04-01 false Asset and stock consistency. 1.338-8 Section 1... (CONTINUED) INCOME TAXES Effects on Corporation § 1.338-8 Asset and stock consistency. (a) Introduction—(1...

  16. Internally consistent gamma ray burst time history phenomenology

    International Nuclear Information System (INIS)

    Cline, T.L.

    1985-01-01

    A phenomenology for gamma ray burst time histories is outlined. Order of their generally chaotic appearance is attempted, based on the speculation that any one burst event can be represented above 150 keV as a superposition of similarly shaped increases of varying intensity. The increases can generally overlap, however, confusing the picture, but a given event must at least exhibit its own limiting characteristic rise and decay times if the measurements are made with instruments having adequate temporal resolution. Most catalogued observations may be of doubtful or marginal utility to test this hypothesis, but some time histories from Helios-2, Pioneer Venus Orbiter and other instruments having one-to several-millisecond capabilities appear to provide consistency. Also, recent studies of temporally resolved Solar Maximum Mission burst energy spectra are entirely compatible with this picture. The phenomenology suggested here, if correct, may assist as an analytic tool for modelling of burst processes and possibly in the definition of burst source populations

  17. Consistency check of photon beam physical data after recommissioning process

    International Nuclear Information System (INIS)

    Kadman, B; Chawapun, N; Ua-apisitwong, S; Asakit, T; Chumpu, N; Rueansri, J

    2016-01-01

    In radiotherapy, medical linear accelerator (Linac) is the key system used for radiation treatments delivery. Although, recommissioning was recommended after major modification of the machine by AAPM TG53, but it might not be practical in radiotherapy center with heavy workloads. The main purpose of this study was to compare photon beam physical data between initial commissioning and recommissioning of 6 MV Elekta Precise linac. The parameters for comparing were the percentage depth dose (PDD) and beam profiles. The clinical commissioning test cases followed IAEA-TECDOC-1583 were planned on REF 91230 IMRT Dose Verification Phantom by Philips’ Pinnacle treatment planning system. The Delta 4PT was used for dose distribution verification with 90% passing criteria of the gamma index (3%/3mm). Our results revealed that the PDDs and beam profiles agreed within a tolerance limit recommended by TRS430. Most of the point doses and dose distribution verification passed the acceptance criteria. This study showed the consistency of photon beam physical data after recommissioning process. There was a good agreement between initial commissioning and recommissioning within a tolerance limit, demonstrated that the full recommissioning process might not be required. However, in the complex treatment planning geometry, the initial data should be applied with great caution. (paper)

  18. Globfit: Consistently fitting primitives by discovering global relations

    KAUST Repository

    Li, Yangyan; Wu, Xiaokun; Chrysathou, Yiorgos; Sharf, Andrei Sharf; Cohen-Or, Daniel; Mitra, Niloy J.

    2011-01-01

    Given a noisy and incomplete point set, we introduce a method that simultaneously recovers a set of locally fitted primitives along with their global mutual relations. We operate under the assumption that the data corresponds to a man-made engineering object consisting of basic primitives, possibly repeated and globally aligned under common relations. We introduce an algorithm to directly couple the local and global aspects of the problem. The local fit of the model is determined by how well the inferred model agrees to the observed data, while the global relations are iteratively learned and enforced through a constrained optimization. Starting with a set of initial RANSAC based locally fitted primitives, relations across the primitives such as orientation, placement, and equality are progressively learned and conformed to. In each stage, a set of feasible relations are extracted among the candidate relations, and then aligned to, while best fitting to the input data. The global coupling corrects the primitives obtained in the local RANSAC stage, and brings them to precise global alignment. We test the robustness of our algorithm on a range of synthesized and scanned data, with varying amounts of noise, outliers, and non-uniform sampling, and validate the results against ground truth, where available. © 2011 ACM.

  19. Nonlinear cosmological consistency relations and effective matter stresses

    International Nuclear Information System (INIS)

    Ballesteros, Guillermo; Hollenstein, Lukas; Jain, Rajeev Kumar; Kunz, Martin

    2012-01-01

    We propose a fully nonlinear framework to construct consistency relations for testing generic cosmological scenarios using the evolution of large scale structure. It is based on the covariant approach in combination with a frame that is purely given by the metric, the normal frame. As an example, we apply this framework to the ΛCDM model, by extending the usual first order conditions on the metric potentials to second order, where the two potentials start to differ from each other. We argue that working in the normal frame is not only a practical choice but also helps with the physical interpretation of nonlinear dynamics. In this frame, effective pressures and anisotropic stresses appear at second order in perturbation theory, even for ''pressureless'' dust. We quantify their effect and compare them, for illustration, to the pressure of a generic clustering dark energy fluid and the anisotropic stress in the DGP model. Besides, we also discuss the effect of a mismatch of the potentials on the determination of galaxy bias

  20. Globfit: Consistently fitting primitives by discovering global relations

    KAUST Repository

    Li, Yangyan

    2011-07-01

    Given a noisy and incomplete point set, we introduce a method that simultaneously recovers a set of locally fitted primitives along with their global mutual relations. We operate under the assumption that the data corresponds to a man-made engineering object consisting of basic primitives, possibly repeated and globally aligned under common relations. We introduce an algorithm to directly couple the local and global aspects of the problem. The local fit of the model is determined by how well the inferred model agrees to the observed data, while the global relations are iteratively learned and enforced through a constrained optimization. Starting with a set of initial RANSAC based locally fitted primitives, relations across the primitives such as orientation, placement, and equality are progressively learned and conformed to. In each stage, a set of feasible relations are extracted among the candidate relations, and then aligned to, while best fitting to the input data. The global coupling corrects the primitives obtained in the local RANSAC stage, and brings them to precise global alignment. We test the robustness of our algorithm on a range of synthesized and scanned data, with varying amounts of noise, outliers, and non-uniform sampling, and validate the results against ground truth, where available. © 2011 ACM.

  1. Classical and Quantum Consistency of the DGP Model

    CERN Document Server

    Nicolis, A; Nicolis, Alberto; Rattazzi, Riccardo

    2004-01-01

    We study the Dvali-Gabadadze-Porrati model by the method of the boundary effective action. The truncation of this action to the bending mode \\pi consistently describes physics in a wide range of regimes both at the classical and at the quantum level. The Vainshtein effect, which restores agreement with precise tests of general relativity, follows straightforwardly. We give a simple and general proof of stability, i.e. absence of ghosts in the fluctuations, valid for most of the relevant cases, like for instance the spherical source in asymptotically flat space. However we confirm that around certain interesting self-accelerating cosmological solutions there is a ghost. We consider the issue of quantum corrections. Around flat space \\pi becomes strongly coupled below a macroscopic length of 1000 km, thus impairing the predictivity of the model. Indeed the tower of higher dimensional operators which is expected by a generic UV completion of the model limits predictivity at even larger length scales. We outline ...

  2. Are coping styles consistent in the teleost fish Sparus aurata through sexual maturation and sex reversal?

    Science.gov (United States)

    Castanheira, Maria Filipa; Martínez Páramo, Sonia; Figueiredo, F; Cerqueira, Marco; Millot, Sandie; Oliveira, Catarina C V; Martins, Catarina I M; Conceição, Luís E C

    2016-10-01

    Individual differences in behaviour and physiological responses to stress are associated with evolutionary adaptive variation and thus raw material for evolution. In farmed animals, the interest in consistent trait associations, i.e. coping styles, has increased dramatically over the last years. However, one of limitations of the available knowledge, regarding the temporal consistency, is that it refers always to short-term consistency (usually few weeks). The present study used an escape response during a net restraining test, previously shown to be an indicative of coping styles in seabream, to investigate long-term consistency of coping styles both over time and during different life history stages. Results showed both short-term (14 days) consistency and long-term (8 months) consistency of escape response. However, we did not found consistency in the same behaviour after sexual maturation when the restraining test was repeated 16, 22 and 23 months after the first test was performed. In conclusion, this study showed consistent behaviour traits in seabream when juveniles, and a loss of this behavioural traits when adults. Therefore, these results underline that adding a life story approach to data interpretation as an essential step forward towards coping styles foreground. Furthermore, a fine-tuning of aquaculture rearing strategies to adapt to different coping strategies may need to be adjusted differently at early stages of development and adults to improve the welfare of farmed fish.

  3. Vitrification and Product Testing of AW-101 and AN-107 Pretreated Waste

    Energy Technology Data Exchange (ETDEWEB)

    Smith, Gary L.; Greenwood, Lawrence R.; Piepel, Gregory F.; Schweiger, Michael J.; Smith, Harry D.; Urie, Michael W.; Wagner, Jerome J.

    2000-10-31

    The primary objective for vitrifying the LAW samples is to generate glass products for subsequent product testing. The work presented in this report is divided into 6 work elements: 1) Glass Fabrication, 2) Chemical Composition, 3) Radiochemical Composition, 4) Crystalline and Non-crystalline Phase Determination, and 5) Release Rate (Modified PCT). These work elements will help demonstrate the RPP-WTP projects ability to satisfy the product requirements concerning, chemical and radionuclide reporting, waste loading, identification and quantification of crystalline and non-crystalline phases, and waste form leachability. VOA, SVOA, dioxins, furans, PCBs, and total cyanide analyses will be reported in as separate document (WTP-RPT-005).

  4. A consistency test of white dwarf and main sequence ages: NGC 6791

    Directory of Open Access Journals (Sweden)

    Córsico A.H.

    2013-03-01

    Full Text Available NGC 6791 is an open cluster that it is so close to us that can be imaged down to very faint luminosities. The main sequence turn-off age (∼8 Gyr and the age derived from the cut-off of the white dwarf luminosity function (∼6 Gyr were found to be significantly different. Here we demonstrate that the origin of this age discrepancy lies in an incorrect evaluation of the white dwarf cooling ages, and we show that when the relevant physical separation processes are included in the calculation of white dwarf sequences both ages are coincident.

  5. Visualizing Changes in Pretest and Post-Test Student Responses with Consistency Plots

    Science.gov (United States)

    Wittmann, Michael C.; Black, Katrina E.

    2014-01-01

    Tabular presentations of student data often hide information about the switches in responses by individual students over the course of a semester. We extend unpublished work by Kanim on "escalator diagrams," which show changes in student responses from correct to incorrect (and vice versa) while representing pre- and postinstruction…

  6. Testing the consistency of three-point halo clustering in Fourier and configuration space

    Science.gov (United States)

    Hoffmann, K.; Gaztañaga, E.; Scoccimarro, R.; Crocce, M.

    2018-05-01

    We compare reduced three-point correlations Q of matter, haloes (as proxies for galaxies) and their cross-correlations, measured in a total simulated volume of ˜100 (h-1 Gpc)3, to predictions from leading order perturbation theory on a large range of scales in configuration space. Predictions for haloes are based on the non-local bias model, employing linear (b1) and non-linear (c2, g2) bias parameters, which have been constrained previously from the bispectrum in Fourier space. We also study predictions from two other bias models, one local (g2 = 0) and one in which c2 and g2 are determined by b1 via approximately universal relations. Overall, measurements and predictions agree when Q is derived for triangles with (r1r2r3)1/3 ≳60 h-1 Mpc, where r1 - 3 are the sizes of the triangle legs. Predictions for Qmatter, based on the linear power spectrum, show significant deviations from the measurements at the BAO scale (given our small measurement errors), which strongly decrease when adding a damping term or using the non-linear power spectrum, as expected. Predictions for Qhalo agree best with measurements at large scales when considering non-local contributions. The universal bias model works well for haloes and might therefore be also useful for tightening constraints on b1 from Q in galaxy surveys. Such constraints are independent of the amplitude of matter density fluctuation (σ8) and hence break the degeneracy between b1 and σ8, present in galaxy two-point correlations.

  7. Input Consistency in the Acquisition of Questions in Bulgarian and English: A Hypothesis Testing Model

    Science.gov (United States)

    Tornyova, Lidiya

    2011-01-01

    The goal of this dissertation is to address several major empirical and theoretical issues related to English-speaking children's difficulties with auxiliary use and inversion in questions. The empirical data on English question acquisition are inconsistent due to differences in methods and techniques used. A range of proposals about the source of…

  8. Test research of consistency for amplitude calibration coefficients of pulsed electric field sensor

    International Nuclear Information System (INIS)

    Meng Cui; Guo Xiaoqiang; Chen Xiangyue; Nie Xin; Mao Congguang; Xiang Hui; Cheng Jianping

    2007-01-01

    The amplitude calibration of an electric field sensor is important in the measurement of electromagnetic pulse. In this paper, an arbitrary waveform generator (AWG) is used to generate multi-waveform electric field in the TEM cell and the dipole antenna pulsed electric field sensor is calibrated. In the frequency band of the sensor, the calibrated amplitude coefficients with different waveforms are identical. The coefficient derived from the TEM cell calibration system suits to the measurement of unknown electric field pulse within the frequency band. (authors)

  9. NEW BV(RI)C PHOTOMETRY FOR PRAESEPE: FURTHER TESTS OF BROADBAND PHOTOMETRIC CONSISTENCY

    International Nuclear Information System (INIS)

    Joner, Michael D.; Taylor, Benjamin J.; Laney, C. David; Van Wyk, Francois

    2011-01-01

    New BV(RI) C measurements of Praesepe made at the South African Astronomical Observatory (SAAO) are presented. When those measurements are combined with those reported in previous papers in this series, it is found that they support previously determined V zero points for Praesepe, M67, and the Hyades. Support is also found for joint (V – R) C and (R – I) C zero points established previously for Praesepe and NGC 752. For the SAAO system of standard stars, a B – V correction to the Johnson system of about –9 mmag appears to be reasonably well established. The preferred (though not definitive) V correction is about +7 mmag. For the Landolt V system, zero-point identity with the Johnson system at a 2σ level of 4.8 mmag is found, and no color term as large as 4 mmag (mag) –1 is detected. Updated CDS data files for Praesepe are briefly described.

  10. Testing spatial theories of plant coexistence: no consistent differences in intra- and interspecific interaction distances.

    Science.gov (United States)

    Vogt, Deborah R; Murrell, David J; Stoll, Peter

    2010-01-01

    Plants stand still and interact with their immediate neighbors. Theory has shown that the distances over which these interactions occur may have important consequences for population and community dynamics. In particular, if intraspecific competition occurs over longer distances than interspecific competition (heteromyopia), coexistence can be promoted. We examined how intraspecific and interspecific competition scales with neighbor distance in a target-neighbor greenhouse competition experiment. Individuals from co-occurring forbs from calcareous grasslands were grown in isolation and with single conspecific or heterospecific neighbors at distances of 5, 10, or 15 cm (Plantago lanceolata vs. Plantago media and Hieracium pilosella vs. Prunella grandiflora). Neighbor effects were strong and declined with distance. Interaction distances varied greatly within and between species, but we found no evidence for heteromyopia. Instead, neighbor identity effects were mostly explained by relative size differences between target and neighbor. We found a complex interaction between final neighbor size and identity such that neighbor identity may become important only as the neighbor becomes very large compared with the target individual. Our results suggest that species-specific size differences between neighboring individuals determine both the strength of competitive interactions and the distance over which these interactions occur.

  11. Evaluating the hydrological consistency of satellite based water cycle components

    KAUST Repository

    Lopez Valencia, Oliver Miguel

    2016-06-15

    Advances in multi-satellite based observations of the earth system have provided the capacity to retrieve information across a wide-range of land surface hydrological components and provided an opportunity to characterize terrestrial processes from a completely new perspective. Given the spatial advantage that space-based observations offer, several regional-to-global scale products have been developed, offering insights into the multi-scale behaviour and variability of hydrological states and fluxes. However, one of the key challenges in the use of satellite-based products is characterizing the degree to which they provide realistic and representative estimates of the underlying retrieval: that is, how accurate are the hydrological components derived from satellite observations? The challenge is intrinsically linked to issues of scale, since the availability of high-quality in-situ data is limited, and even where it does exist, is generally not commensurate to the resolution of the satellite observation. Basin-scale studies have shown considerable variability in achieving water budget closure with any degree of accuracy using satellite estimates of the water cycle. In order to assess the suitability of this type of approach for evaluating hydrological observations, it makes sense to first test it over environments with restricted hydrological inputs, before applying it to more hydrological complex basins. Here we explore the concept of hydrological consistency, i.e. the physical considerations that the water budget impose on the hydrologic fluxes and states to be temporally and spatially linked, to evaluate the reproduction of a set of large-scale evaporation (E) products by using a combination of satellite rainfall (P) and Gravity Recovery and Climate Experiment (GRACE) observations of storage change, focusing on arid and semi-arid environments, where the hydrological flows can be more realistically described. Our results indicate no persistent hydrological

  12. Consistent Regulation of Infrastructure Businesses: Some Economic Issues

    OpenAIRE

    Flavio M. Menezes

    2008-01-01

    This paper examines some important economic aspects associated with the notion that consistency in the regulation of infrastructure businesses is a desirable feature. It makes two important points. First, it is not easy to measure consistency. In particular, one cannot simply point to different regulatory parameters as evidence of inconsistent regulatory policy. Second, even if one does observe consistency emerging from decisions made by different regulators, it does not necessarily mean that...

  13. Consistent Code Qualification Process and Application to WWER-1000 NPP

    International Nuclear Information System (INIS)

    Berthon, A.; Petruzzi, A.; Giannotti, W.; D'Auria, F.; Reventos, F.

    2006-01-01

    Calculation analysis by application of the system codes are performed to evaluate the NPP or the facility behavior during a postulated transient or to evaluate the code capability. The calculation analysis constitutes a process that involves the code itself, the data of the reference plant, the data about the transient, the nodalization, and the user. All these elements affect one each other and affect the results. A major issue in the use of mathematical model is constituted by the model capability to reproduce the plant or facility behavior under steady state and transient conditions. These aspects constitute two main checks that must be satisfied during the qualification process. The first of them is related to the realization of a scheme of the reference plant; the second one is related to the capability to reproduce the transient behavior. The aim of this paper is to describe the UMAE (Uncertainty Method based on Accuracy Extrapolation) methodology developed at University of Pisa for qualifying a nodalization and analysing the calculated results and to perform the uncertainty evaluation of the system code by the CIAU code (Code with the capability of Internal Assessment of Uncertainty). The activity consists with the re-analysis of the Experiment BL-44 (SBLOCA) performed in the LOBI facility and the analysis of a Kv-scaling calculation of the WWER-1000 NPP nodalization taking as reference the test BL-44. Relap5/Mod3.3 has been used as thermal-hydraulic system code and the standard procedure adopted at University of Pisa has been applied to show the capability of the code to predict the significant aspects of the transient and to obtain a qualified nodalization of the WWER-1000 through a systematic qualitative and quantitative accuracy evaluation. The qualitative accuracy evaluation is based on the selection of Relevant Thermal-hydraulic Aspects (RTAs) and is a prerequisite to the application of the Fast Fourier Transform Based Method (FFTBM) which quantifies

  14. Consistency of ocular coherence tomography fast macular thickness mapping in diabetic diffuse macular edema

    International Nuclear Information System (INIS)

    Saraiva, Fabio Petersen; Costa, Patricia Grativol; Inomata, Daniela Lumi; Melo, Carlos Sergio Nascimento; Helal Junior, John; Nakashima, Yoshitaka

    2007-01-01

    Objectives: To investigate optical coherence tomography consistency on foveal thickness, foveal volume, and macular volume measurements in patients with and without diffuse diabetic macular edema. Introduction: Optical coherence tomography represents an objective technique that provides cross-sectional tomographs of retinal structure in vivo. However, it is expected that poor fixation ability, as seen in diabetic macular edema, could alter its results. Several authors have discussed the reproducibility of optical coherence tomography, but only a few have addressed the topic with respect to diabetic maculopathy. Methods: The study recruited diabetic patients without clinically evident retinopathy (control group) and with diffuse macular edema (case group). Only one eye of each patient was evaluated. Five consecutive fast macular scans were taken using Ocular Coherence Tomography 3; the 6 mm macular map was chosen. The consistency in measurements of foveal thickness, foveal volume, and total macular volume for both groups was evaluated using the Pearson's coefficient of variation. The T-test for independent samples was used in order to compare measurements of both groups. Results: Each group consisted of 20 patients. All measurements had a coefficient of variation less than 10%. The most consistent parameter for both groups was the total macular volume. Discussion: Consistency in measurement is a mainstay of any test. A test is unreliable if its measurements can not be correctly repeated. We found a good index of consistency, even considering patients with an unstable gaze. Conclusions: Optical coherence tomography is a consistent method for diabetic subjects with diffuse macular edema. (author)

  15. Consistency of ocular coherence tomography fast macular thickness mapping in diabetic diffuse macular edema

    Energy Technology Data Exchange (ETDEWEB)

    Saraiva, Fabio Petersen; Costa, Patricia Grativol; Inomata, Daniela Lumi; Melo, Carlos Sergio Nascimento; Helal Junior, John; Nakashima, Yoshitaka [Universidade de Sao Paulo (USP), SP (Brazil). Hospital das Clinicas. Dept. de Oftalmologia]. E-mail: fabiopetersen@yahoo.com.br

    2007-07-01

    Objectives: To investigate optical coherence tomography consistency on foveal thickness, foveal volume, and macular volume measurements in patients with and without diffuse diabetic macular edema. Introduction: Optical coherence tomography represents an objective technique that provides cross-sectional tomographs of retinal structure in vivo. However, it is expected that poor fixation ability, as seen in diabetic macular edema, could alter its results. Several authors have discussed the reproducibility of optical coherence tomography, but only a few have addressed the topic with respect to diabetic maculopathy. Methods: The study recruited diabetic patients without clinically evident retinopathy (control group) and with diffuse macular edema (case group). Only one eye of each patient was evaluated. Five consecutive fast macular scans were taken using Ocular Coherence Tomography 3; the 6 mm macular map was chosen. The consistency in measurements of foveal thickness, foveal volume, and total macular volume for both groups was evaluated using the Pearson's coefficient of variation. The T-test for independent samples was used in order to compare measurements of both groups. Results: Each group consisted of 20 patients. All measurements had a coefficient of variation less than 10%. The most consistent parameter for both groups was the total macular volume. Discussion: Consistency in measurement is a mainstay of any test. A test is unreliable if its measurements can not be correctly repeated. We found a good index of consistency, even considering patients with an unstable gaze. Conclusions: Optical coherence tomography is a consistent method for diabetic subjects with diffuse macular edema. (author)

  16. Student Consistency and Implications for Feedback in Online Assessment Systems

    Science.gov (United States)

    Madhyastha, Tara M.; Tanimoto, Steven

    2009-01-01

    Most of the emphasis on mining online assessment logs has been to identify content-specific errors. However, the pattern of general "consistency" is domain independent, strongly related to performance, and can itself be a target of educational data mining. We demonstrate that simple consistency indicators are related to student outcomes,…

  17. 26 CFR 301.6224(c)-3 - Consistent settlements.

    Science.gov (United States)

    2010-04-01

    ... 26 Internal Revenue 18 2010-04-01 2010-04-01 false Consistent settlements. 301.6224(c)-3 Section... settlements. (a) In general. If the Internal Revenue Service enters into a settlement agreement with any..., settlement terms consistent with those contained in the settlement agreement entered into. (b) Requirements...

  18. Self-consistent calculation of atomic structure for mixture

    International Nuclear Information System (INIS)

    Meng Xujun; Bai Yun; Sun Yongsheng; Zhang Jinglin; Zong Xiaoping

    2000-01-01

    Based on relativistic Hartree-Fock-Slater self-consistent average atomic model, atomic structure for mixture is studied by summing up component volumes in mixture. Algorithmic procedure for solving both the group of Thomas-Fermi equations and the self-consistent atomic structure is presented in detail, and, some numerical results are discussed

  19. A Preliminary Study toward Consistent Soil Moisture from AMSR2

    NARCIS (Netherlands)

    Parinussa, R.M.; Holmes, T.R.H.; Wanders, N.; Dorigo, W.A.; de Jeu, R.A.M.

    2015-01-01

    A preliminary study toward consistent soil moisture products from the Advanced Microwave Scanning Radiometer 2 (AMSR2) is presented. Its predecessor, the Advanced Microwave Scanning Radiometer for Earth Observing System (AMSR-E), has providedEarth scientists with a consistent and continuous global

  20. Consistency and Inconsistency in PhD Thesis Examination

    Science.gov (United States)

    Holbrook, Allyson; Bourke, Sid; Lovat, Terry; Fairbairn, Hedy

    2008-01-01

    This is a mixed methods investigation of consistency in PhD examination. At its core is the quantification of the content and conceptual analysis of examiner reports for 804 Australian theses. First, the level of consistency between what examiners say in their reports and the recommendation they provide for a thesis is explored, followed by an…

  1. Delimiting Coefficient a from Internal Consistency and Unidimensionality

    Science.gov (United States)

    Sijtsma, Klaas

    2015-01-01

    I discuss the contribution by Davenport, Davison, Liou, & Love (2015) in which they relate reliability represented by coefficient a to formal definitions of internal consistency and unidimensionality, both proposed by Cronbach (1951). I argue that coefficient a is a lower bound to reliability and that concepts of internal consistency and…

  2. Risk aversion vs. the Omega ratio : Consistency results

    NARCIS (Netherlands)

    Balder, Sven; Schweizer, Nikolaus

    This paper clarifies when the Omega ratio and related performance measures are consistent with second order stochastic dominance and when they are not. To avoid consistency problems, the threshold parameter in the ratio should be chosen as the expected return of some benchmark – as is commonly done

  3. Carl Rogers during Initial Interviews: A Moderate and Consistent Therapist.

    Science.gov (United States)

    Edwards, H. P.; And Others

    1982-01-01

    Analyzed two initial interviews by Carl Rogers in their entirety using the Carkhuff scales, Hill's category system, and a brief grammatical analysis to establish the level and consistency with which Rogers provides facilitative conditions. Results indicated his behavior as counselor was stable and consistent within and across interviews. (Author)

  4. Policy consistency and the achievement of Nigeria's foreign policy ...

    African Journals Online (AJOL)

    This study is an attempt to investigate the policy consistency of Nigeria‟s foreign policy and to understand the basis for this consistency; and also to see whether peacekeeping/peace-enforcement is key instrument in the achievement of Nigeria‟s foreign policy goals. The objective of the study was to examine whether the ...

  5. Decentralized Consistency Checking in Cross-organizational Workflows

    NARCIS (Netherlands)

    Wombacher, Andreas

    Service Oriented Architectures facilitate loosely coupled composed services, which are established in a decentralized way. One challenge for such composed services is to guarantee consistency, i.e., deadlock-freeness. This paper presents a decentralized approach to consistency checking, which

  6. Consistency of a system of equations: What does that mean?

    NARCIS (Netherlands)

    Still, Georg J.; Kern, Walter; Koelewijn, Jaap; Bomhoff, M.J.

    2010-01-01

    The concept of (structural) consistency also called structural solvability is an important basic tool for analyzing the structure of systems of equations. Our aim is to provide a sound and practically relevant meaning to this concept. The implications of consistency are expressed in terms of

  7. Quasi-Particle Self-Consistent GW for Molecules.

    Science.gov (United States)

    Kaplan, F; Harding, M E; Seiler, C; Weigend, F; Evers, F; van Setten, M J

    2016-06-14

    We present the formalism and implementation of quasi-particle self-consistent GW (qsGW) and eigenvalue only quasi-particle self-consistent GW (evGW) adapted to standard quantum chemistry packages. Our implementation is benchmarked against high-level quantum chemistry computations (coupled-cluster theory) and experimental results using a representative set of molecules. Furthermore, we compare the qsGW approach for five molecules relevant for organic photovoltaics to self-consistent GW results (scGW) and analyze the effects of the self-consistency on the ground state density by comparing calculated dipole moments to their experimental values. We show that qsGW makes a significant improvement over conventional G0W0 and that partially self-consistent flavors (in particular evGW) can be excellent alternatives.

  8. Putting humans in ecology: consistency in science and management.

    Science.gov (United States)

    Hobbs, Larry; Fowler, Charles W

    2008-03-01

    Normal and abnormal levels of human participation in ecosystems can be revealed through the use of macro-ecological patterns. Such patterns also provide consistent and objective guidance that will lead to achieving and maintaining ecosystem health and sustainability. This paper focuses on the consistency of this type of guidance and management. Such management, in sharp contrast to current management practices, ensures that our actions as individuals, institutions, political groups, societies, and as a species are applied consistently across all temporal, spatial, and organizational scales. This approach supplants management of today, where inconsistency results from debate, politics, and legal and religious polarity. Consistency is achieved when human endeavors are guided by natural patterns. Pattern-based management meets long-standing demands for enlightened management that requires humans to participate in complex systems in consistent and sustainable ways.

  9. Multiplicative renormalizability and self-consistent treatments of the Schwinger-Dyson equations

    International Nuclear Information System (INIS)

    Brown, N.; Dorey, N.

    1989-11-01

    Many approximations to the Schwinger-Dyson equations place constraints on the renormalization constants of a theory. The requirement that the solutions to the equations be multiplicatively renormalizable also places constraints on these constants. Demanding that these two sets of constraints be compatible is an important test of the self-consistency of the approximations made. We illustrate this idea by considering the equation for the fermion propagator in massless quenched quantum electrodynamics, (QED), checking the consistency of various approximations. In particular, we show that the much used 'ladder' approximation is self-consistent, provided that the coupling constant is renormalized in a particular way. We also propose another approximation which satisfies this self-consistency test, but requires that the coupling be unrenormalized, as should be the case in the full quenched approximation. This new approximation admits an exact solution, which also satisfies the renormalization group equation for the quenched approximation. (author)

  10. A fresh look at the thermodynamic consistency of vapour-liquid equilibria data

    International Nuclear Information System (INIS)

    Wisniak, Jaime; Ortega, Juan; Fernández, Luis

    2017-01-01

    Highlights: • The thermodynamic consistency tests commonly used to evaluate VLE data are presented and discussed. • Advantages/disadvantages for each one of test are listed using actual examples. • All manuscripts should include information about VLE variables and test results. • In any case, the Herington test should not be used for VLE data evaluation. • Simultaneous application of several tests is recommended. - Abstract: Design of a separation unit requires real information about the phase equilibrium of the system being handled. Accurate equilibrium data allows the best design from a thermodynamic viewpoint and contributes to a better knowledge about the behaviour of fluids and their mixtures. The principles behind the concept of thermodynamic consistency are presented and discussed. The present state of the art shows that no definite test is available for insuring the quality of the measured values. The main available procedures for testing the consistency of vapour-liquid equilibrium (VLE) data at constant temperature or pressure are reviewed and analysed and recommendations provided for their proper use, for the presentation of VLE results, and also some possible means for determining their quality. Suitable examples are provided about the adequate use of the available tests and about their misuse.

  11. Facial Mimicry and Emotion Consistency: Influences of Memory and Context.

    Science.gov (United States)

    Kirkham, Alexander J; Hayes, Amy E; Pawling, Ralph; Tipper, Steven P

    2015-01-01

    This study investigates whether mimicry of facial emotions is a stable response or can instead be modulated and influenced by memory of the context in which the emotion was initially observed, and therefore the meaning of the expression. The study manipulated emotion consistency implicitly, where a face expressing smiles or frowns was irrelevant and to be ignored while participants categorised target scenes. Some face identities always expressed emotions consistent with the scene (e.g., smiling with a positive scene), whilst others were always inconsistent (e.g., frowning with a positive scene). During this implicit learning of face identity and emotion consistency there was evidence for encoding of face-scene emotion consistency, with slower RTs, a reduction in trust, and inhibited facial EMG for faces expressing incompatible emotions. However, in a later task where the faces were subsequently viewed expressing emotions with no additional context, there was no evidence for retrieval of prior emotion consistency, as mimicry of emotion was similar for consistent and inconsistent individuals. We conclude that facial mimicry can be influenced by current emotion context, but there is little evidence of learning, as subsequent mimicry of emotionally consistent and inconsistent faces is similar.

  12. Facial Mimicry and Emotion Consistency: Influences of Memory and Context.

    Directory of Open Access Journals (Sweden)

    Alexander J Kirkham

    Full Text Available This study investigates whether mimicry of facial emotions is a stable response or can instead be modulated and influenced by memory of the context in which the emotion was initially observed, and therefore the meaning of the expression. The study manipulated emotion consistency implicitly, where a face expressing smiles or frowns was irrelevant and to be ignored while participants categorised target scenes. Some face identities always expressed emotions consistent with the scene (e.g., smiling with a positive scene, whilst others were always inconsistent (e.g., frowning with a positive scene. During this implicit learning of face identity and emotion consistency there was evidence for encoding of face-scene emotion consistency, with slower RTs, a reduction in trust, and inhibited facial EMG for faces expressing incompatible emotions. However, in a later task where the faces were subsequently viewed expressing emotions with no additional context, there was no evidence for retrieval of prior emotion consistency, as mimicry of emotion was similar for consistent and inconsistent individuals. We conclude that facial mimicry can be influenced by current emotion context, but there is little evidence of learning, as subsequent mimicry of emotionally consistent and inconsistent faces is similar.

  13. Quasiparticle self-consistent GW method: a short summary

    International Nuclear Information System (INIS)

    Kotani, Takao; Schilfgaarde, Mark van; Faleev, Sergey V; Chantis, Athanasios

    2007-01-01

    We have developed a quasiparticle self-consistent GW method (QSGW), which is a new self-consistent method to calculate the electronic structure within the GW approximation. The method is formulated based on the idea of a self-consistent perturbation; the non-interacting Green function G 0 , which is the starting point for GWA to obtain G, is determined self-consistently so as to minimize the perturbative correction generated by GWA. After self-consistency is attained, we have G 0 , W (the screened Coulomb interaction) and G self-consistently. This G 0 can be interpreted as the optimum non-interacting propagator for the quasiparticles. We will summarize some theoretical discussions to justify QSGW. Then we will survey results which have been obtained up to now: e.g., band gaps for normal semiconductors are predicted to a precision of 0.1-0.3 eV; the self-consistency including the off-diagonal part is required for NiO and MnO; and so on. There are still some remaining disagreements with experiments; however, they are very systematic, and can be explained from the neglect of excitonic effects

  14. SCALCE: boosting sequence compression algorithms using locally consistent encoding.

    Science.gov (United States)

    Hach, Faraz; Numanagic, Ibrahim; Alkan, Can; Sahinalp, S Cenk

    2012-12-01

    The high throughput sequencing (HTS) platforms generate unprecedented amounts of data that introduce challenges for the computational infrastructure. Data management, storage and analysis have become major logistical obstacles for those adopting the new platforms. The requirement for large investment for this purpose almost signalled the end of the Sequence Read Archive hosted at the National Center for Biotechnology Information (NCBI), which holds most of the sequence data generated world wide. Currently, most HTS data are compressed through general purpose algorithms such as gzip. These algorithms are not designed for compressing data generated by the HTS platforms; for example, they do not take advantage of the specific nature of genomic sequence data, that is, limited alphabet size and high similarity among reads. Fast and efficient compression algorithms designed specifically for HTS data should be able to address some of the issues in data management, storage and communication. Such algorithms would also help with analysis provided they offer additional capabilities such as random access to any read and indexing for efficient sequence similarity search. Here we present SCALCE, a 'boosting' scheme based on Locally Consistent Parsing technique, which reorganizes the reads in a way that results in a higher compression speed and compression rate, independent of the compression algorithm in use and without using a reference genome. Our tests indicate that SCALCE can improve the compression rate achieved through gzip by a factor of 4.19-when the goal is to compress the reads alone. In fact, on SCALCE reordered reads, gzip running time can improve by a factor of 15.06 on a standard PC with a single core and 6 GB memory. Interestingly even the running time of SCALCE + gzip improves that of gzip alone by a factor of 2.09. When compared with the recently published BEETL, which aims to sort the (inverted) reads in lexicographic order for improving bzip2, SCALCE + gzip

  15. Repeatability and consistency of individual behaviour in juvenile and adult Eurasian harvest mice

    Science.gov (United States)

    Schuster, Andrea C.; Carl, Teresa; Foerster, Katharina

    2017-04-01

    Knowledge on animal personality has provided new insights into evolutionary biology and animal ecology, as behavioural types have been shown to affect fitness. Animal personality is characterized by repeatable and consistent between-individual behavioural differences throughout time and across different situations. Behavioural repeatability within life history stages and consistency between life history stages should be checked for the independence of sex and age, as recent data have shown that males and females in some species may differ in the repeatability of behavioural traits, as well as in their consistency. We measured the repeatability and consistency of three behavioural and one cognitive traits in juvenile and adult Eurasian harvest mice ( Micromys minutus). We found that exploration, activity and boldness were repeatable in juveniles and adults. Spatial recognition measured in a Y Maze was only repeatable in adult mice. Exploration, activity and boldness were consistent before and after maturation, as well as before and after first sexual contact. Data on spatial recognition provided little evidence for consistency. Further, we found some evidence for a litter effect on behaviours by comparing different linear mixed models. We concluded that harvest mice express animal personality traits as behaviours were repeatable across sexes and consistent across life history stages. The tested cognitive trait showed low repeatability and was less consistent across life history stages. Given the rising interest in individual variation in cognitive performance, and in its relationship to animal personality, we suggest that it is important to gather more data on the repeatability and consistency of cognitive traits.

  16. Protective Factors, Risk Indicators, and Contraceptive Consistency Among College Women.

    Science.gov (United States)

    Morrison, Leslie F; Sieving, Renee E; Pettingell, Sandra L; Hellerstedt, Wendy L; McMorris, Barbara J; Bearinger, Linda H

    2016-01-01

    To explore risk and protective factors associated with consistent contraceptive use among emerging adult female college students and whether effects of risk indicators were moderated by protective factors. Secondary analysis of National Longitudinal Study of Adolescent to Adult Health Wave III data. Data collected through in-home interviews in 2001 and 2002. National sample of 18- to 25-year-old women (N = 842) attending 4-year colleges. We examined relationships between protective factors, risk indicators, and consistent contraceptive use. Consistent contraceptive use was defined as use all of the time during intercourse in the past 12 months. Protective factors included external supports of parental closeness and relationship with caring nonparental adult and internal assets of self-esteem, confidence, independence, and life satisfaction. Risk indicators included heavy episodic drinking, marijuana use, and depression symptoms. Multivariable logistic regression models were used to evaluate relationships between protective factors and consistent contraceptive use and between risk indicators and contraceptive use. Self-esteem, confidence, independence, and life satisfaction were significantly associated with more consistent contraceptive use. In a final model including all internal assets, life satisfaction was significantly related to consistent contraceptive use. Marijuana use and depression symptoms were significantly associated with less consistent use. With one exception, protective factors did not moderate relationships between risk indicators and consistent use. Based on our findings, we suggest that risk and protective factors may have largely independent influences on consistent contraceptive use among college women. A focus on risk and protective factors may improve contraceptive use rates and thereby reduce unintended pregnancy among college students. Copyright © 2016 AWHONN, the Association of Women's Health, Obstetric and Neonatal Nurses. Published

  17. Credibility test; Vertrauens-Test

    Energy Technology Data Exchange (ETDEWEB)

    Fuhs, Michael

    2009-07-01

    Solar wafer producers must prove their quality standards. Q-Cells has started a marketing office and opened their test center to journalists. They are aware that adherence to standards is going only half the way of quality assurance. The other half consists in gaining the customers' trust. (orig.)

  18. Consistency in boldness, activity and exploration at different stages of life

    Science.gov (United States)

    2013-01-01

    Background Animals show consistent individual behavioural patterns over time and over situations. This phenomenon has been referred to as animal personality or behavioural syndromes. Little is known about consistency of animal personalities over entire life times. We investigated the repeatability of behaviour in common voles (Microtus arvalis) at different life stages, with different time intervals, and in different situations. Animals were tested using four behavioural tests in three experimental groups: 1. before and after maturation over three months, 2. twice as adults during one week, and 3. twice as adult animals over three months, which resembles a substantial part of their entire adult life span of several months. Results Different behaviours were correlated within and between tests and a cluster analysis showed three possible behavioural syndrome-axes, which we name boldness, exploration and activity. Activity and exploration behaviour in all tests was highly repeatable in adult animals tested over one week. In animals tested over maturation, exploration behaviour was consistent whereas activity was not. Voles that were tested as adults with a three-month interval showed the opposite pattern with stable activity but unstable exploration behaviour. Conclusions The consistency in behaviour over time suggests that common voles do express stable personality over short time. Over longer periods however, behaviour is more flexible and depending on life stage (i.e. tested before/after maturation or as adults) of the tested individual. Level of boldness or activity does not differ between tested groups and maintenance of variation in behavioural traits can therefore not be explained by expected future assets as reported in other studies. PMID:24314274

  19. The Consistent Preferences Approach to Deductive Reasoning in Games

    CERN Document Server

    Asheim, Geir B

    2006-01-01

    "The Consistent Preferences Approach to Deductive Reasoning in Games" presents, applies, and synthesizes what my co-authors and I have called the 'consistent preferences' approach to deductive reasoning in games. Briefly described, this means that the object of the analysis is the ranking by each player of his own strategies, rather than his choice. The ranking can be required to be consistent (in different senses) with his beliefs about the opponent's ranking of her strategies. This can be contrasted to the usual 'rational choice' approach where a player's strategy choice is (in dif

  20. Multiphase flows of N immiscible incompressible fluids: A reduction-consistent and thermodynamically-consistent formulation and associated algorithm

    Science.gov (United States)

    Dong, S.

    2018-05-01

    We present a reduction-consistent and thermodynamically consistent formulation and an associated numerical algorithm for simulating the dynamics of an isothermal mixture consisting of N (N ⩾ 2) immiscible incompressible fluids with different physical properties (densities, viscosities, and pair-wise surface tensions). By reduction consistency we refer to the property that if only a set of M (1 ⩽ M ⩽ N - 1) fluids are present in the system then the N-phase governing equations and boundary conditions will exactly reduce to those for the corresponding M-phase system. By thermodynamic consistency we refer to the property that the formulation honors the thermodynamic principles. Our N-phase formulation is developed based on a more general method that allows for the systematic construction of reduction-consistent formulations, and the method suggests the existence of many possible forms of reduction-consistent and thermodynamically consistent N-phase formulations. Extensive numerical experiments have been presented for flow problems involving multiple fluid components and large density ratios and large viscosity ratios, and the simulation results are compared with the physical theories or the available physical solutions. The comparisons demonstrate that our method produces physically accurate results for this class of problems.

  1. On the consistent histories approach to quantum mechanics

    International Nuclear Information System (INIS)

    Dowker, F.; Kent, A.

    1996-01-01

    We review the consistent histories formulations of quantum mechanics developed by Griffiths, Omnes, Gell-Man, and Hartle, and we describe the classifications of consistent sets. We illustrate some general features of consistent sets by a few lemmas and examples. We also consider various interpretations of the formalism, and we examine the new problems which arise in reconstructing the past and predicting the future. It is shown that Omnes characterization of true statements---statements that can be deduced unconditionally in his interpretation---is incorrect. We examine critically Gell-Mann and Hartle's interpretation of the formalism, and in particular, their discussions of communication, prediction, and retrodiction, and we conclude that their explanation of the apparent persistence of quasiclassicality relies on assumptions about an as-yet-unknown theory of experience. Our overall conclusion is that the consistent histories approach illustrates the need to supplement quantum mechanics by some selection principle in order to produce a fundamental theory capable of unconditional predictions

  2. Consistency of Trend Break Point Estimator with Underspecified Break Number

    Directory of Open Access Journals (Sweden)

    Jingjing Yang

    2017-01-01

    Full Text Available This paper discusses the consistency of trend break point estimators when the number of breaks is underspecified. The consistency of break point estimators in a simple location model with level shifts has been well documented by researchers under various settings, including extensions such as allowing a time trend in the model. Despite the consistency of break point estimators of level shifts, there are few papers on the consistency of trend shift break point estimators in the presence of an underspecified break number. The simulation study and asymptotic analysis in this paper show that the trend shift break point estimator does not converge to the true break points when the break number is underspecified. In the case of two trend shifts, the inconsistency problem worsens if the magnitudes of the breaks are similar and the breaks are either both positive or both negative. The limiting distribution for the trend break point estimator is developed and closely approximates the finite sample performance.

  3. Liking for Evaluators: Consistency and Self-Esteem Theories

    Science.gov (United States)

    Regan, Judith Weiner

    1976-01-01

    Consistency and self-esteem theories make contrasting predictions about the relationship between a person's self-evaluation and his liking for an evaluator. Laboratory experiments confirmed predictions about these theories. (Editor/RK)

  4. Sparse PDF Volumes for Consistent Multi-Resolution Volume Rendering

    KAUST Repository

    Sicat, Ronell Barrera; Kruger, Jens; Moller, Torsten; Hadwiger, Markus

    2014-01-01

    This paper presents a new multi-resolution volume representation called sparse pdf volumes, which enables consistent multi-resolution volume rendering based on probability density functions (pdfs) of voxel neighborhoods. These pdfs are defined

  5. Structures, profile consistency, and transport scaling in electrostatic convection

    DEFF Research Database (Denmark)

    Bian, N.H.; Garcia, O.E.

    2005-01-01

    Two mechanisms at the origin of profile consistency in models of electrostatic turbulence in magnetized plasmas are considered. One involves turbulent diffusion in collisionless plasmas and the subsequent turbulent equipartition of Lagrangian invariants. By the very nature of its definition...

  6. 15 CFR 930.36 - Consistency determinations for proposed activities.

    Science.gov (United States)

    2010-01-01

    ... necessity of issuing separate consistency determinations for each incremental action controlled by the major... plans), and that affect any coastal use or resource of more than one State. Many States share common...

  7. Decentralized Consistent Network Updates in SDN with ez-Segway

    KAUST Repository

    Nguyen, Thanh Dang; Chiesa, Marco; Canini, Marco

    2017-01-01

    We present ez-Segway, a decentralized mechanism to consistently and quickly update the network state while preventing forwarding anomalies (loops and black-holes) and avoiding link congestion. In our design, the centralized SDN controller only pre-computes

  8. The utility of theory of planned behavior in predicting consistent ...

    African Journals Online (AJOL)

    admin

    disease. Objective: To examine the utility of theory of planned behavior in predicting consistent condom use intention of HIV .... (24-25), making subjective norms as better predictors of intention ..... Organizational Behavior and Human Decision.

  9. A methodology for the data energy regional consumption consistency analysis

    International Nuclear Information System (INIS)

    Canavarros, Otacilio Borges; Silva, Ennio Peres da

    1999-01-01

    The article introduces a methodology for data energy regional consumption consistency analysis. The work was going based on recent studies accomplished by several cited authors and boarded Brazilian matrices and Brazilian energetics regional balances. The results are compared and analyzed

  10. Island of Stability for Consistent Deformations of Einstein's Gravity

    DEFF Research Database (Denmark)

    Dietrich, Dennis D.; Berkhahn, Felix; Hofmann, Stefan

    2012-01-01

    We construct deformations of general relativity that are consistent and phenomenologically viable, since they respect, in particular, cosmological backgrounds. These deformations have unique symmetries in accordance with their Minkowski cousins (Fierz-Pauli theory for massive gravitons) and incor...

  11. The effect of correct and consistent condom use on chlamydial and gonococcal infection among urban adolescents.

    Science.gov (United States)

    Paz-Bailey, Gabriela; Koumans, Emilia H; Sternberg, Maya; Pierce, Antonya; Papp, John; Unger, Elizabeth R; Sawyer, Mary; Black, Carolyn M; Markowitz, Lauri E

    2005-06-01

    To evaluate the relationship between self-reported correct and consistent condom use and chlamydial and gonococcal infection. Cross-sectional study. An urban adolescent health care clinic. Patients A total of 509 adolescent girls tested for Chlamydia trachomatis and Neisseria gonorrhoeae infection by urine nucleic acid amplification tests. Main Outcome Measure Effect of condom use on infection rates of chlamydia and gonorrhea. Consistent condom use was defined as using condoms for every act of vaginal sex and correct use as consistent use without any of the following: beginning sex without a condom, taking it off before finishing sex, flipping it over, condom breakage, or condom slippage. A total of 95% of the participants were African American, with a mean age of 16.6 years. Chlamydia prevalence was 21% (105/509) and gonorrhea prevalence was 7% (36/509). Condom errors were reported by 316 (71%) of 442 participants who had reported using a condom at least once in the previous 3 months. Consistent use was reported by 176 patients (35%); however, both correct and consistent use was reported by only 80 patients (16%). After adjusting for confounders, correct and consistent use was protective for chlamydia (odds ratio, 0.4; 95% confidence interval, 0.2-1.0) and highly protective for gonorrhea (odds ratio, 0.1; 95% confidence interval, 0-0.7). Our findings indicate that assessing both correctness and consistency of use is important for evaluation of condom effectiveness.

  12. Self-consistent normal ordering of gauge field theories

    International Nuclear Information System (INIS)

    Ruehl, W.

    1987-01-01

    Mean-field theories with a real action of unconstrained fields can be self-consistently normal ordered. This leads to a considerable improvement over standard mean-field theory. This concept is applied to lattice gauge theories. First an appropriate real action mean-field theory is constructed. The equations determining the Gaussian kernel necessary for self-consistent normal ordering of this mean-field theory are derived. (author). 4 refs

  13. Consistency of the least weighted squares under heteroscedasticity

    Czech Academy of Sciences Publication Activity Database

    Víšek, Jan Ámos

    2011-01-01

    Roč. 2011, č. 47 (2011), s. 179-206 ISSN 0023-5954 Grant - others:GA UK(CZ) GA402/09/055 Institutional research plan: CEZ:AV0Z10750506 Keywords : Regression * Consistency * The least weighted squares * Heteroscedasticity Subject RIV: BB - Applied Statistics, Operational Research Impact factor: 0.454, year: 2011 http://library.utia.cas.cz/separaty/2011/SI/visek-consistency of the least weighted squares under heteroscedasticity.pdf

  14. Parquet equations for numerical self-consistent-field theory

    International Nuclear Information System (INIS)

    Bickers, N.E.

    1991-01-01

    In recent years increases in computational power have provided new motivation for the study of self-consistent-field theories for interacting electrons. In this set of notes, the so-called parquet equations for electron systems are derived pedagogically. The principal advantages of the parquet approach are outlined, and its relationship to simpler self-consistent-field methods, including the Baym-Kadanoff technique, is discussed in detail. (author). 14 refs, 9 figs

  15. Consistent Estimation of Pricing Kernels from Noisy Price Data

    OpenAIRE

    Vladislav Kargin

    2003-01-01

    If pricing kernels are assumed non-negative then the inverse problem of finding the pricing kernel is well-posed. The constrained least squares method provides a consistent estimate of the pricing kernel. When the data are limited, a new method is suggested: relaxed maximization of the relative entropy. This estimator is also consistent. Keywords: $\\epsilon$-entropy, non-parametric estimation, pricing kernel, inverse problems.

  16. Consistency of Flashbulb Memories of September 11 over Long Delays: Implications for Consolidation and Wrong Time Slice Hypotheses

    Science.gov (United States)

    Kvavilashvili, Lia; Mirani, Jennifer; Schlagman, Simone; Foley, Kerry; Kornbrot, Diana E.

    2009-01-01

    The consistency of flashbulb memories over long delays provides a test of theories of memory for highly emotional events. This study used September 11, 2001 as the target event, with test-retest delays of 2 and 3 years. The nature and consistency of flashbulb memories were examined as a function of delay between the target event and an initial…

  17. Measuring consistency of autobiographical memory recall in depression.

    LENUS (Irish Health Repository)

    Semkovska, Maria

    2012-05-15

    Autobiographical amnesia assessments in depression need to account for normal changes in consistency over time, contribution of mood and type of memories measured. We report herein validation studies of the Columbia Autobiographical Memory Interview - Short Form (CAMI-SF), exclusively used in depressed patients receiving electroconvulsive therapy (ECT) but without previous published report of normative data. The CAMI-SF was administered twice with a 6-month interval to 44 healthy volunteers to obtain normative data for retrieval consistency of its Semantic, Episodic-Extended and Episodic-Specific components and assess their reliability and validity. Healthy volunteers showed significant large decreases in retrieval consistency on all components. The Semantic and Episodic-Specific components demonstrated substantial construct validity. We then assessed CAMI-SF retrieval consistencies over a 2-month interval in 30 severely depressed patients never treated with ECT compared with healthy controls (n=19). On initial assessment, depressed patients produced less episodic-specific memories than controls. Both groups showed equivalent amounts of consistency loss over a 2-month interval on all components. At reassessment, only patients with persisting depressive symptoms were distinguishable from controls on episodic-specific memories retrieved. Research quantifying retrograde amnesia following ECT for depression needs to control for normal loss in consistency over time and contribution of persisting depressive symptoms.

  18. Measuring consistency of autobiographical memory recall in depression.

    Science.gov (United States)

    Semkovska, Maria; Noone, Martha; Carton, Mary; McLoughlin, Declan M

    2012-05-15

    Autobiographical amnesia assessments in depression need to account for normal changes in consistency over time, contribution of mood and type of memories measured. We report herein validation studies of the Columbia Autobiographical Memory Interview - Short Form (CAMI-SF), exclusively used in depressed patients receiving electroconvulsive therapy (ECT) but without previous published report of normative data. The CAMI-SF was administered twice with a 6-month interval to 44 healthy volunteers to obtain normative data for retrieval consistency of its Semantic, Episodic-Extended and Episodic-Specific components and assess their reliability and validity. Healthy volunteers showed significant large decreases in retrieval consistency on all components. The Semantic and Episodic-Specific components demonstrated substantial construct validity. We then assessed CAMI-SF retrieval consistencies over a 2-month interval in 30 severely depressed patients never treated with ECT compared with healthy controls (n=19). On initial assessment, depressed patients produced less episodic-specific memories than controls. Both groups showed equivalent amounts of consistency loss over a 2-month interval on all components. At reassessment, only patients with persisting depressive symptoms were distinguishable from controls on episodic-specific memories retrieved. Research quantifying retrograde amnesia following ECT for depression needs to control for normal loss in consistency over time and contribution of persisting depressive symptoms. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.

  19. Autonomous Navigation with Constrained Consistency for C-Ranger

    Directory of Open Access Journals (Sweden)

    Shujing Zhang

    2014-06-01

    Full Text Available Autonomous underwater vehicles (AUVs have become the most widely used tools for undertaking complex exploration tasks in marine environments. Their synthetic ability to carry out localization autonomously and build an environmental map concurrently, in other words, simultaneous localization and mapping (SLAM, are considered to be pivotal requirements for AUVs to have truly autonomous navigation. However, the consistency problem of the SLAM system has been greatly ignored during the past decades. In this paper, a consistency constrained extended Kalman filter (EKF SLAM algorithm, applying the idea of local consistency, is proposed and applied to the autonomous navigation of the C-Ranger AUV, which is developed as our experimental platform. The concept of local consistency (LC is introduced after an explicit theoretical derivation of the EKF-SLAM system. Then, we present a locally consistency-constrained EKF-SLAM design, LC-EKF, in which the landmark estimates used for linearization are fixed at the beginning of each local time period, rather than evaluated at the latest landmark estimates. Finally, our proposed LC-EKF algorithm is experimentally verified, both in simulations and sea trials. The experimental results show that the LC-EKF performs well with regard to consistency, accuracy and computational efficiency.

  20. Are prescription drug insurance choices consistent with expected utility theory?

    Science.gov (United States)

    Bundorf, M Kate; Mata, Rui; Schoenbaum, Michael; Bhattacharya, Jay

    2013-09-01

    To determine the extent to which people make choices inconsistent with expected utility theory when choosing among prescription drug insurance plans and whether tabular or graphical presentation format influences the consistency of their choices. Members of an Internet-enabled panel chose between two Medicare prescription drug plans. The "low variance" plan required higher out-of-pocket payments for the drugs respondents usually took but lower out-of-pocket payments for the drugs they might need if they developed a new health condition than the "high variance" plan. The probability of a change in health varied within subjects and the presentation format (text vs. graphical) and the affective salience of the clinical condition (abstract vs. risk related to specific clinical condition) varied between subjects. Respondents were classified based on whether they consistently chose either the low or high variance plan. Logistic regression models were estimated to examine the relationship between decision outcomes and task characteristics. The majority of respondents consistently chose either the low or high variance plan, consistent with expected utility theory. Half of respondents consistently chose the low variance plan. Respondents were less likely to make discrepant choices when information was presented in graphical format. Many people, although not all, make choices consistent with expected utility theory when they have information on differences among plans in the variance of out-of-pocket spending. Medicare beneficiaries would benefit from information on the extent to which prescription drug plans provide risk protection. PsycINFO Database Record (c) 2013 APA, all rights reserved.

  1. Analysis of the SBLOCAs in HANARO pool for the 3-pin fuel test loop

    International Nuclear Information System (INIS)

    Park, S. K.; Chi, D. Y.; Sim, B. S.; Park, K. N.; Ahn, S. H.; Lee, J. M.; Lee, C. Y.; Kim, Y. J.

    2004-09-01

    Fuel Test Loop(FTL) has been developed to meet the increasing demand on fuel irradiation and burn up test required the development of new fuels in Korea. It is designed to provide the test conditions of high pressure and temperature like the commercial PWR and CANDU power plants. And also the FTL have the cooling capability to sufficiently remove the thermal power of the in-pile test section for normal operation, Anticipated Operational Occurrences(AOOs), and Design Basis Accidents(DBAs). This report deals with the Small Break Loss Of Coolant Accidents (SBLOCAs) in HANARO pool for the 3-pin fuel test loop. The MARS code has been used for the prediction of the emergency core cooling capability of the FTL and the peak cladding temperature of the test fuels for the SBLOCAs. The location of the pipe break is assumed at the hill taps connecting the cold and hot legs in HANARO pool to the inlet and outlet nozzles of the In-Pile test Section (IPS). The break size is also assumed less than 20% of the cross section area of the pipe. The test fuels are heated up when the cold leg break occur. However, they are not heated up when the hot leg break occur. The maximum Peak Cladding Temperatures (PCT) are predicted to be about 906.9 .deg. C for the cold leg break accident in PWR fuel test mode and 971.9 .deg. C in CANDU fuel test mode respectively. The critical break size is about the 6% of the cross section area of the pipe for PWR fuel test mode and the 8% for CANDU fuel test mode. The PCTs meet the design criterion of commercial PWR fuel that the maximum PCT is lower than 1204 .deg. C

  2. QUENCH-LOCA program at KIT and results of the QUENCH-L0 bundle test

    International Nuclear Information System (INIS)

    Stuckert, J.; Grosse, M.; Roessger, C.; Steinbrueck, M.; Walter, M.

    2012-01-01

    The current LOCA criteria and their safety goals are applied worldwide with minor modifications since the USNRC release in 1973. The criteria are given as limits on peak cladding temperature (T PCT ≤ 1200 C) and on oxidation level ECR (equivalent cladding reacted) calculated as a percentage of cladding oxidized (ECR ≤ 17% calculated using Baker-Just oxidation correlation). These two rules constitute the criterion of cladding embrittlement due to oxygen uptake. The results elaborated worldwide in the 1980s and 1990s on Zircaloy-4 (Zry-4) cladding tubes behavior (oxidation, deformation and bundle coolability) under LOCA conditions constitute a detailed data base and an important input for the safety assessment of LWRs. In-pile test data (with burn-up up to 35 MWd/kgU) were consistent with the out-of-pile data and did not indicate an influence of the nuclear environment on cladding deformation. At high burn-up, fuel rods fabricated from conventional Zry-4 often exhibit significant oxidation, hydriding, and oxide spallation. Thus, many fuel vendors have proposed the use of recently developed cladding alloys, such as M5 registered , ZIRLO trademark and other. Therefore, it is important to verify the safety margins for high burn-up fuel and fuel claddings with new alloys. Due to long cladding hydriding period for the high fuel burn-up, post-quench ductility is strongly influenced not only by oxidation but also hydrogen uptake. The 17% ECR limit is inadequate to ensure post-quench ductility at hydrogen concentrations higher than ∼500 wppm. Due to so called secondary hydriding (during oxidation of inner cladding surface after burst), which was firstly observed in JAEA, the hydrogen content can reach 4000 wppm in Zircaloy cladding regions around burst. To investigate the influence of these phenomena on the applicability of the embrittlement criteria for the German nuclear reactors it was decided to perform the QUENCH-LOCA bundle test series at the Karlsruhe Institute

  3. QUENCH-LOCA program at KIT and results of the QUENCH-L0 bundle test

    Energy Technology Data Exchange (ETDEWEB)

    Stuckert, J.; Grosse, M.; Roessger, C.; Steinbrueck, M.; Walter, M. [Karlsruhe Institute of Technology (KIT), Eggenstein-Leopoldshafen (Germany)

    2012-11-01

    The current LOCA criteria and their safety goals are applied worldwide with minor modifications since the USNRC release in 1973. The criteria are given as limits on peak cladding temperature (T{sub PCT} {<=} 1200 C) and on oxidation level ECR (equivalent cladding reacted) calculated as a percentage of cladding oxidized (ECR {<=} 17% calculated using Baker-Just oxidation correlation). These two rules constitute the criterion of cladding embrittlement due to oxygen uptake. The results elaborated worldwide in the 1980s and 1990s on Zircaloy-4 (Zry-4) cladding tubes behavior (oxidation, deformation and bundle coolability) under LOCA conditions constitute a detailed data base and an important input for the safety assessment of LWRs. In-pile test data (with burn-up up to 35 MWd/kgU) were consistent with the out-of-pile data and did not indicate an influence of the nuclear environment on cladding deformation. At high burn-up, fuel rods fabricated from conventional Zry-4 often exhibit significant oxidation, hydriding, and oxide spallation. Thus, many fuel vendors have proposed the use of recently developed cladding alloys, such as M5 {sup registered}, ZIRLO trademark and other. Therefore, it is important to verify the safety margins for high burn-up fuel and fuel claddings with new alloys. Due to long cladding hydriding period for the high fuel burn-up, post-quench ductility is strongly influenced not only by oxidation but also hydrogen uptake. The 17% ECR limit is inadequate to ensure post-quench ductility at hydrogen concentrations higher than {approx}500 wppm. Due to so called secondary hydriding (during oxidation of inner cladding surface after burst), which was firstly observed in JAEA, the hydrogen content can reach 4000 wppm in Zircaloy cladding regions around burst. To investigate the influence of these phenomena on the applicability of the embrittlement criteria for the German nuclear reactors it was decided to perform the QUENCH-LOCA bundle test series at the

  4. Martial arts striking hand peak acceleration, accuracy and consistency.

    Science.gov (United States)

    Neto, Osmar Pinto; Marzullo, Ana Carolina De Miranda; Bolander, Richard P; Bir, Cynthia A

    2013-01-01

    The goal of this paper was to investigate the possible trade-off between peak hand acceleration and accuracy and consistency of hand strikes performed by martial artists of different training experiences. Ten male martial artists with training experience ranging from one to nine years volunteered to participate in the experiment. Each participant performed 12 maximum effort goal-directed strikes. Hand acceleration during the strikes was obtained using a tri-axial accelerometer block. A pressure sensor matrix was used to determine the accuracy and consistency of the strikes. Accuracy was estimated by the radial distance between the centroid of each subject's 12 strikes and the target, whereas consistency was estimated by the square root of the 12 strikes mean squared distance from their centroid. We found that training experience was significantly correlated to hand peak acceleration prior to impact (r(2)=0.456, p =0.032) and accuracy (r(2)=0. 621, p=0.012). These correlations suggest that more experienced participants exhibited higher hand peak accelerations and at the same time were more accurate. Training experience, however, was not correlated to consistency (r(2)=0.085, p=0.413). Overall, our results suggest that martial arts training may lead practitioners to achieve higher striking hand accelerations with better accuracy and no change in striking consistency.

  5. Self-consistent electrodynamic scattering in the symmetric Bragg case

    International Nuclear Information System (INIS)

    Campos, H.S.

    1988-01-01

    We have analyzed the symmetric Bragg case, introducing a model of self consistent scattering for two elliptically polarized beams. The crystal is taken as a set of mathematical planes, each of them defined by a surface density of dipoles. We have considered the mesofield and the epifield differently from that of the Ewald's theory and, we assumed a plane of dipoles and the associated fields as a self consistent scattering unit. The exact analytical treatment when applied to any two neighbouring planes, results in a general and self consistent Bragg's equation, in terms of the amplitude and phase variations. The generalized solution for the set of N planes was obtained after introducing an absorption factor in the incident radiation, in two ways: (i) the analytical one, through a rule of field similarity, which says that the incidence occurs in both faces of the all crystal planes and also, through a matricial development with the Chebyshev polynomials; (ii) using the numerical solution we calculated, iteratively, the reflectivity, the reflection phase, the transmissivity, the transmission phase and the energy. The results are showed through reflection and transmission curves, which are characteristics as from kinematical as dynamical theories. The conservation of the energy results from the Ewald's self consistency principle is used. In the absorption case, the results show that it is not the only cause for the asymmetric form in the reflection curves. The model contains basic elements for a unified, microscope, self consistent, vectorial and exact formulation for interpretating the X ray diffraction in perfect crystals. (author)

  6. A consistent response spectrum analysis including the resonance range

    International Nuclear Information System (INIS)

    Schmitz, D.; Simmchen, A.

    1983-01-01

    The report provides a complete consistent Response Spectrum Analysis for any component. The effect of supports with different excitation is taken into consideration, at is the description of the resonance ranges. It includes information explaining how the contributions of the eigenforms with higher eigenfrequencies are to be considered. Stocking of floor response spectra is also possible using the method described here. However, modified floor response spectra must now be calculated for each building mode. Once these have been prepared, the calculation of the dynamic component values is practically no more complicated than with the conventional, non-consistent methods. The consistent Response Spectrum Analysis can supply smaller and larger values than the conventional theory, a fact which can be demonstrated using simple examples. The report contains a consistent Response Spectrum Analysis (RSA), which, as far as we know, has been formulated in this way for the first time. A consistent RSA is so important because today this method is preferentially applied as an important tool for the earthquake proof of components in nuclear power plants. (orig./HP)

  7. GRAVITATIONALLY CONSISTENT HALO CATALOGS AND MERGER TREES FOR PRECISION COSMOLOGY

    International Nuclear Information System (INIS)

    Behroozi, Peter S.; Wechsler, Risa H.; Wu, Hao-Yi; Busha, Michael T.; Klypin, Anatoly A.; Primack, Joel R.

    2013-01-01

    We present a new algorithm for generating merger trees and halo catalogs which explicitly ensures consistency of halo properties (mass, position, and velocity) across time steps. Our algorithm has demonstrated the ability to improve both the completeness (through detecting and inserting otherwise missing halos) and purity (through detecting and removing spurious objects) of both merger trees and halo catalogs. In addition, our method is able to robustly measure the self-consistency of halo finders; it is the first to directly measure the uncertainties in halo positions, halo velocities, and the halo mass function for a given halo finder based on consistency between snapshots in cosmological simulations. We use this algorithm to generate merger trees for two large simulations (Bolshoi and Consuelo) and evaluate two halo finders (ROCKSTAR and BDM). We find that both the ROCKSTAR and BDM halo finders track halos extremely well; in both, the number of halos which do not have physically consistent progenitors is at the 1%-2% level across all halo masses. Our code is publicly available at http://code.google.com/p/consistent-trees. Our trees and catalogs are publicly available at http://hipacc.ucsc.edu/Bolshoi/.

  8. Analysis of th SBLOCAs in the room 1 for the 3-pin fuel test loop

    International Nuclear Information System (INIS)

    Park, S. K.; Chi, D. Y.; Sim, B. S.; Park, K. N.; Ahn, S. H.; Lee, J. M.; Lee, C. Y.; Kim, Y. J.

    2004-10-01

    Fuel Test Loop(FTL) has been developed to meet the increasing demand on fuel irradiation and burn up test required the development of new fuels in Korea. It is designed to provide the test conditions of high pressure and temperature like the commercial PWR and CANDU power plants. And also the FTL have the cooling capability to sufficiently remove the thermal power of the in-pile test section for normal operation, Anticipated Operational Occurrences(AOOs), and Design Basis Accidents(DBAs). This report deals with the Small Break Loss of Coolant Accidents (SBLOCAs) in the Room 1 for the 3-pin fuel test loop. The MARS code has been used for the prediction of the emergency core cooling capability of the FTL and the peak cladding temperature of the test fuels for the SBLOCAs. The location of the pipe break is assumed at the downstream of the main cooling water pump and the upstream of the main cooler in the room 1. The break size is also assumed less than 20% of the cross section area of the pipe. The test fuels are heated up when the cold leg break occur. However, they are not heated up when the hot leg break occur. The maximum Peak Cladding Temperature (PCT) is predicted to be about 931.4 .deg. C for the cold leg break accident in PWR fuel test mode and 931.6 .deg. C in CANDU fuel test mode respectively. The critical break size is about the 8% of the cross section area of the pipe for PWR fuel test mode and the 10% for CANDU fuel test mode. The PCTs meet the design criterion of commercial PWR fuel that the maximum PCT is lower than 1204 .deg. C

  9. Consistent Partial Least Squares Path Modeling via Regularization.

    Science.gov (United States)

    Jung, Sunho; Park, JaeHong

    2018-01-01

    Partial least squares (PLS) path modeling is a component-based structural equation modeling that has been adopted in social and psychological research due to its data-analytic capability and flexibility. A recent methodological advance is consistent PLS (PLSc), designed to produce consistent estimates of path coefficients in structural models involving common factors. In practice, however, PLSc may frequently encounter multicollinearity in part because it takes a strategy of estimating path coefficients based on consistent correlations among independent latent variables. PLSc has yet no remedy for this multicollinearity problem, which can cause loss of statistical power and accuracy in parameter estimation. Thus, a ridge type of regularization is incorporated into PLSc, creating a new technique called regularized PLSc. A comprehensive simulation study is conducted to evaluate the performance of regularized PLSc as compared to its non-regularized counterpart in terms of power and accuracy. The results show that our regularized PLSc is recommended for use when serious multicollinearity is present.

  10. Context-dependent individual behavioral consistency in Daphnia

    DEFF Research Database (Denmark)

    Heuschele, Jan; Ekvall, Mikael T.; Bianco, Giuseppe

    2017-01-01

    The understanding of consistent individual differences in behavior, often termed "personality," for adapting and coping with threats and novel environmental conditions has advanced considerably during the last decade. However, advancements are almost exclusively associated with higher-order animals......, whereas studies focusing on smaller aquatic organisms are still rare. Here, we show individual differences in the swimming behavior of Daphnia magna, a clonal freshwater invertebrate, before, during, and after being exposed to a lethal threat, ultraviolet radiation (UVR). We show consistency in swimming...... that of adults. Overall, we show that aquatic invertebrates are far from being identical robots, but instead they show considerable individual differences in behavior that can be attributed to both ontogenetic development and individual consistency. Our study also demonstrates, for the first time...

  11. Consistent forcing scheme in the cascaded lattice Boltzmann method

    Science.gov (United States)

    Fei, Linlin; Luo, Kai Hong

    2017-11-01

    In this paper, we give an alternative derivation for the cascaded lattice Boltzmann method (CLBM) within a general multiple-relaxation-time (MRT) framework by introducing a shift matrix. When the shift matrix is a unit matrix, the CLBM degrades into an MRT LBM. Based on this, a consistent forcing scheme is developed for the CLBM. The consistency of the nonslip rule, the second-order convergence rate in space, and the property of isotropy for the consistent forcing scheme is demonstrated through numerical simulations of several canonical problems. Several existing forcing schemes previously used in the CLBM are also examined. The study clarifies the relation between MRT LBM and CLBM under a general framework.

  12. Application of consistent fluid added mass matrix to core seismic

    International Nuclear Information System (INIS)

    Koo, K. H.; Lee, J. H.

    2003-01-01

    In this paper, the application algorithm of a consistent fluid added mass matrix including the coupling terms to the core seismic analysis is developed and installed at SAC-CORE3.0 code. As an example, we assumed the 7-hexagon system of the LMR core and carried out the vibration modal analysis and the nonlinear time history seismic response analysis using SAC-CORE3.0. Used consistent fluid added mass matrix is obtained by using the finite element program of the FAMD(Fluid Added Mass and Damping) code. From the results of the vibration modal analysis, the core duct assemblies reveal strongly coupled vibration modes, which are so different from the case of in-air condition. From the results of the time history seismic analysis, it was verified that the effects of the coupled terms of the consistent fluid added mass matrix are significant in impact responses and the dynamic responses

  13. Self-consistent approximations beyond the CPA: Part II

    International Nuclear Information System (INIS)

    Kaplan, T.; Gray, L.J.

    1982-01-01

    This paper concentrates on a self-consistent approximation for random alloys developed by Kaplan, Leath, Gray, and Diehl. The construction of the augmented space formalism for a binary alloy is sketched, and the notation to be used derived. Using the operator methods of the augmented space, the self-consistent approximation is derived for the average Green's function, and for evaluating the self-energy, taking into account the scattering by clusters of excitations. The particular cluster approximation desired is derived by treating the scattering by the excitations with S /SUB T/ exactly. Fourier transforms on the disorder-space clustersite labels solve the self-consistent set of equations. Expansion to short range order in the alloy is also discussed. A method to reduce the problem to a computationally tractable form is described

  14. Self-consistency and coherent effects in nonlinear resonances

    International Nuclear Information System (INIS)

    Hofmann, I.; Franchetti, G.; Qiang, J.; Ryne, R. D.

    2003-01-01

    The influence of space charge on emittance growth is studied in simulations of a coasting beam exposed to a strong octupolar perturbation in an otherwise linear lattice, and under stationary parameters. We explore the importance of self-consistency by comparing results with a non-self-consistent model, where the space charge electric field is kept 'frozen-in' to its initial values. For Gaussian distribution functions we find that the 'frozen-in' model results in a good approximation of the self-consistent model, hence coherent response is practically absent and the emittance growth is self-limiting due to space charge de-tuning. For KV or waterbag distributions, instead, strong coherent response is found, which we explain in terms of absence of Landau damping

  15. A consistent time frame for Chaucer's Canterbury Pilgrimage

    Science.gov (United States)

    Kummerer, K. R.

    2001-08-01

    A consistent time frame for the pilgrimage that Geoffrey Chaucer describes in The Canterbury Tales can be established if the seven celestial assertions related to the journey mentioned in the text can be reconciled with each other and the date of April 18 that is also mentioned. Past attempts to establish such a consistency for all seven celestial assertions have not been successful. The analysis herein, however, indicates that in The Canterbury Tales Chaucer accurately describes the celestial conditions he observed in the April sky above the London(Canterbury region of England in the latter half of the fourteenth century. All seven celestial assertions are in agreement with each other and consistent with the April 18 date. The actual words of Chaucer indicate that the Canterbury journey began during the 'seson' he defines in the General Prologue and ends under the light of the full Moon on the night of April 18, 1391.

  16. An approach to a self-consistent nuclear energy system

    International Nuclear Information System (INIS)

    Fujii-e, Yoichi; Arie, Kazuo; Endo, Hiroshi

    1992-01-01

    A nuclear energy system should provide a stable supply of energy without endangering the environment or humans. If there is fear about exhausting world energy resources, accumulating radionuclides, and nuclear reactor safety, tension is created in human society. Nuclear energy systems of the future should be able to eliminate fear from people's minds. In other words, the whole system, including the nuclear fuel cycle, should be self-consistent. This is the ultimate goal of nuclear energy. If it can be realized, public acceptance of nuclear energy will increase significantly. In a self-consistent nuclear energy system, misunderstandings between experts on nuclear energy and the public should be minimized. The way to achieve this goal is to explain using simple logic. This paper proposes specific targets for self-consistent nuclear energy systems and shows that the fast breeder reactor (FBR) lies on the route to attaining the final goal

  17. Consistent forcing scheme in the cascaded lattice Boltzmann method.

    Science.gov (United States)

    Fei, Linlin; Luo, Kai Hong

    2017-11-01

    In this paper, we give an alternative derivation for the cascaded lattice Boltzmann method (CLBM) within a general multiple-relaxation-time (MRT) framework by introducing a shift matrix. When the shift matrix is a unit matrix, the CLBM degrades into an MRT LBM. Based on this, a consistent forcing scheme is developed for the CLBM. The consistency of the nonslip rule, the second-order convergence rate in space, and the property of isotropy for the consistent forcing scheme is demonstrated through numerical simulations of several canonical problems. Several existing forcing schemes previously used in the CLBM are also examined. The study clarifies the relation between MRT LBM and CLBM under a general framework.

  18. Consistency and Reconciliation Model In Regional Development Planning

    Directory of Open Access Journals (Sweden)

    Dina Suryawati

    2016-10-01

    Full Text Available The aim of this study was to identify the problems and determine the conceptual model of regional development planning. Regional development planning is a systemic, complex and unstructured process. Therefore, this study used soft systems methodology to outline unstructured issues with a structured approach. The conceptual models that were successfully constructed in this study are a model of consistency and a model of reconciliation. Regional development planning is a process that is well-integrated with central planning and inter-regional planning documents. Integration and consistency of regional planning documents are very important in order to achieve the development goals that have been set. On the other hand, the process of development planning in the region involves technocratic system, that is, both top-down and bottom-up system of participation. Both must be balanced, do not overlap and do not dominate each other. regional, development, planning, consistency, reconciliation

  19. Bootstrap-Based Inference for Cube Root Consistent Estimators

    DEFF Research Database (Denmark)

    Cattaneo, Matias D.; Jansson, Michael; Nagasawa, Kenichi

    This note proposes a consistent bootstrap-based distributional approximation for cube root consistent estimators such as the maximum score estimator of Manski (1975) and the isotonic density estimator of Grenander (1956). In both cases, the standard nonparametric bootstrap is known...... to be inconsistent. Our method restores consistency of the nonparametric bootstrap by altering the shape of the criterion function defining the estimator whose distribution we seek to approximate. This modification leads to a generic and easy-to-implement resampling method for inference that is conceptually distinct...... from other available distributional approximations based on some form of modified bootstrap. We offer simulation evidence showcasing the performance of our inference method in finite samples. An extension of our methodology to general M-estimation problems is also discussed....

  20. Self-consistent modelling of resonant tunnelling structures

    DEFF Research Database (Denmark)

    Fiig, T.; Jauho, A.P.

    1992-01-01

    We report a comprehensive study of the effects of self-consistency on the I-V-characteristics of resonant tunnelling structures. The calculational method is based on a simultaneous solution of the effective-mass Schrödinger equation and the Poisson equation, and the current is evaluated...... applied voltages and carrier densities at the emitter-barrier interface. We include the two-dimensional accumulation layer charge and the quantum well charge in our self-consistent scheme. We discuss the evaluation of the current contribution originating from the two-dimensional accumulation layer charges......, and our qualitative estimates seem consistent with recent experimental studies. The intrinsic bistability of resonant tunnelling diodes is analyzed within several different approximation schemes....

  1. Consistent Parameter and Transfer Function Estimation using Context Free Grammars

    Science.gov (United States)

    Klotz, Daniel; Herrnegger, Mathew; Schulz, Karsten

    2017-04-01

    This contribution presents a method for the inference of transfer functions for rainfall-runoff models. Here, transfer functions are defined as parametrized (functional) relationships between a set of spatial predictors (e.g. elevation, slope or soil texture) and model parameters. They are ultimately used for estimation of consistent, spatially distributed model parameters from a limited amount of lumped global parameters. Additionally, they provide a straightforward method for parameter extrapolation from one set of basins to another and can even be used to derive parameterizations for multi-scale models [see: Samaniego et al., 2010]. Yet, currently an actual knowledge of the transfer functions is often implicitly assumed. As a matter of fact, for most cases these hypothesized transfer functions can rarely be measured and often remain unknown. Therefore, this contribution presents a general method for the concurrent estimation of the structure of transfer functions and their respective (global) parameters. Note, that by consequence an estimation of the distributed parameters of the rainfall-runoff model is also undertaken. The method combines two steps to achieve this. The first generates different possible transfer functions. The second then estimates the respective global transfer function parameters. The structural estimation of the transfer functions is based on the context free grammar concept. Chomsky first introduced context free grammars in linguistics [Chomsky, 1956]. Since then, they have been widely applied in computer science. But, to the knowledge of the authors, they have so far not been used in hydrology. Therefore, the contribution gives an introduction to context free grammars and shows how they can be constructed and used for the structural inference of transfer functions. This is enabled by new methods from evolutionary computation, such as grammatical evolution [O'Neill, 2001], which make it possible to exploit the constructed grammar as a

  2. An Explicit Consistent Geometric Stiffness Matrix for the DKT Element

    Directory of Open Access Journals (Sweden)

    Eliseu Lucena Neto

    Full Text Available Abstract A large number of references dealing with the geometric stiffness matrix of the DKT finite element exist in the literature, where nearly all of them adopt an inconsistent form. While such a matrix may be part of the element to treat nonlinear problems in general, it is of crucial importance for linearized buckling analysis. The present work seems to be the first to obtain an explicit expression for this matrix in a consistent way. Numerical results on linear buckling of plates assess the element performance either with the proposed explicit consistent matrix, or with the most commonly used inconsistent matrix.

  3. The cluster bootstrap consistency in generalized estimating equations

    KAUST Repository

    Cheng, Guang

    2013-03-01

    The cluster bootstrap resamples clusters or subjects instead of individual observations in order to preserve the dependence within each cluster or subject. In this paper, we provide a theoretical justification of using the cluster bootstrap for the inferences of the generalized estimating equations (GEE) for clustered/longitudinal data. Under the general exchangeable bootstrap weights, we show that the cluster bootstrap yields a consistent approximation of the distribution of the regression estimate, and a consistent approximation of the confidence sets. We also show that a computationally more efficient one-step version of the cluster bootstrap provides asymptotically equivalent inference. © 2012.

  4. Consistency in the description of diffusion in compacted bentonite

    International Nuclear Information System (INIS)

    Lehikoinen, J.; Muurinen, A.

    2009-01-01

    A macro-level diffusion model, which aims to provide a unifying framework for explaining the experimentally observed co-ion exclusion and greatly controversial counter-ion surface diffusion in a consistent fashion, is presented. It is explained in detail why a term accounting for the non-zero mobility of the counter-ion surface excess is required in the mathematical form of the macroscopic diffusion flux. The prerequisites for the consistency of the model and the problems associated with the interpretation of diffusion in such complex pore geometries as in compacted smectite clays are discussed. (author)

  5. An energetically consistent vertical mixing parameterization in CCSM4

    DEFF Research Database (Denmark)

    Nielsen, Søren Borg; Jochum, Markus; Eden, Carsten

    2018-01-01

    An energetically consistent stratification-dependent vertical mixing parameterization is implemented in the Community Climate System Model 4 and forced with energy conversion from the barotropic tides to internal waves. The structures of the resulting dissipation and diffusivity fields are compared......, however, depends greatly on the details of the vertical mixing parameterizations, where the new energetically consistent parameterization results in low thermocline diffusivities and a sharper and shallower thermocline. It is also investigated if the ocean state is more sensitive to a change in forcing...

  6. The consistency service of the ATLAS Distributed Data Management system

    CERN Document Server

    Serfon, C; The ATLAS collaboration

    2011-01-01

    With the continuously increasing volume of data produced by ATLAS and stored on the WLCG sites, the probability of data corruption or data losses, due to software and hardware failures is increasing. In order to ensure the consistency of all data produced by ATLAS a Consistency Service has been developed as part of the DQ2 Distributed Data Management system. This service is fed by the different ATLAS tools, i.e. the analysis tools, production tools, DQ2 site services or by site administrators that report corrupted or lost files. It automatically corrects the errors reported and informs the users in case of irrecoverable file loss.

  7. The Consistency Service of the ATLAS Distributed Data Management system

    CERN Document Server

    Serfon, C; The ATLAS collaboration

    2010-01-01

    With the continuously increasing volume of data produced by ATLAS and stored on the WLCG sites, the probability of data corruption or data losses, due to software and hardware failure is increasing. In order to ensure the consistency of all data produced by ATLAS a Consistency Service has been developed as part of the DQ2 Distributed Data Management system. This service is fed by the different ATLAS tools, i.e. the analysis tools, production tools, DQ2 site services or by site administrators that report corrupted or lost files. It automatically correct the errors reported and informs the users in case of irrecoverable file loss.

  8. Consistency among integral measurements of aggregate decay heat power

    Energy Technology Data Exchange (ETDEWEB)

    Takeuchi, H.; Sagisaka, M.; Oyamatsu, K.; Kukita, Y. [Nagoya Univ. (Japan)

    1998-03-01

    Persisting discrepancies between summation calculations and integral measurements force us to assume large uncertainties in the recommended decay heat power. In this paper, we develop a hybrid method to calculate the decay heat power of a fissioning system from those of different fissioning systems. Then, this method is applied to examine consistency among measured decay heat powers of {sup 232}Th, {sup 233}U, {sup 235}U, {sup 238}U and {sup 239}Pu at YAYOI. The consistency among the measured values are found to be satisfied for the {beta} component and fairly well for the {gamma} component, except for cooling times longer than 4000 s. (author)

  9. Standard Model Vacuum Stability and Weyl Consistency Conditions

    DEFF Research Database (Denmark)

    Antipin, Oleg; Gillioz, Marc; Krog, Jens

    2013-01-01

    At high energy the standard model possesses conformal symmetry at the classical level. This is reflected at the quantum level by relations between the different beta functions of the model. These relations are known as the Weyl consistency conditions. We show that it is possible to satisfy them...... order by order in perturbation theory, provided that a suitable coupling constant counting scheme is used. As a direct phenomenological application, we study the stability of the standard model vacuum at high energies and compare with previous computations violating the Weyl consistency conditions....

  10. STP: A mathematically and physically consistent library of steam properties

    International Nuclear Information System (INIS)

    Aguilar, F.; Hutter, A.C.; Tuttle, P.G.

    1982-01-01

    A new FORTRAN library of subroutines has been developed from the fundamental equation of Keenan et al. to evaluate a large set of water properties including derivatives such as sound speed and isothermal compressibility. The STP library uses the true saturation envelope of the Keenan et al. fundamental equation. The evaluation of the true envelope by a continuation method is explained. This envelope, along with other design features, imparts an exceptionally high degree of thermodynamic and mathematical consistency to the STP library, even at the critical point. Accuracy and smoothness, library self-consistency, and designed user convenience make the STP library a reliable and versatile water property package

  11. Weyl consistency conditions in non-relativistic quantum field theory

    Energy Technology Data Exchange (ETDEWEB)

    Pal, Sridip; Grinstein, Benjamín [Department of Physics, University of California,San Diego, 9500 Gilman Drive, La Jolla, CA 92093 (United States)

    2016-12-05

    Weyl consistency conditions have been used in unitary relativistic quantum field theory to impose constraints on the renormalization group flow of certain quantities. We classify the Weyl anomalies and their renormalization scheme ambiguities for generic non-relativistic theories in 2+1 dimensions with anisotropic scaling exponent z=2; the extension to other values of z are discussed as well. We give the consistency conditions among these anomalies. As an application we find several candidates for a C-theorem. We comment on possible candidates for a C-theorem in higher dimensions.

  12. A Van Atta reflector consisting of half-wave dipoles

    DEFF Research Database (Denmark)

    Appel-Hansen, Jørgen

    1966-01-01

    The reradiation pattern of a passive Van Atta reflector consisting of half-wave dipoles is investigated. The character of the reradiation pattern first is deduced by qualitative and physical considerations. Various types of array elements are considered and several geometrical configurations...... of these elements are outlined. Following this, an analysis is made of the reradiation pattern of a linear Van Atta array consisting of four equispaced half-wave dipoles. The general form of the reradiation pattern is studied analytically. The influence of scattering and coupling is determined and the dependence...

  13. A self-consistent theory of the magnetic polaron

    International Nuclear Information System (INIS)

    Marvakov, D.I.; Kuzemsky, A.L.; Vlahov, J.P.

    1984-10-01

    A finite temperature self-consistent theory of magnetic polaron in the s-f model of ferromagnetic semiconductors is developed. The calculations are based on the novel approach of the thermodynamic two-time Green function methods. This approach consists in the introduction of the ''irreducible'' Green functions (IGF) and derivation of the exact Dyson equation and exact self-energy operator. It is shown that IGF method gives a unified and natural approach for a calculation of the magnetic polaron states by taking explicitly into account the damping effects and finite lifetime. (author)

  14. Evidence for Consistency of the Glycation Gap in Diabetes

    OpenAIRE

    Nayak, Ananth U.; Holland, Martin R.; Macdonald, David R.; Nevill, Alan; Singh, Baldev M.

    2011-01-01

    OBJECTIVE Discordance between HbA1c and fructosamine estimations in the assessment of glycemia is often encountered. A number of mechanisms might explain such discordance, but whether it is consistent is uncertain. This study aims to coanalyze paired glycosylated hemoglobin (HbA1c)-fructosamine estimations by using fructosamine to determine a predicted HbA1c, to calculate a glycation gap (G-gap) and to determine whether the G-gap is consistent over time. RESEARCH DESIGN AND METHODS We include...

  15. Diagnostic language consistency among multicultural English-speaking nurses.

    Science.gov (United States)

    Wieck, K L

    1996-01-01

    Cultural differences among nurses may influence the choice of terminology applicable to use of a nursing diagnostic statement. This study explored whether defining characteristics are consistently applied by culturally varied nurses in an English language setting. Two diagnoses, pain, and high risk for altered skin integrity, were studied within six cultures: African, Asian, Filipino, East Indian, African-American, and Anglo-American nurses. Overall, there was consistency between the cultural groups. Analysis of variance for the pain scale demonstrated differences among cultures on two characteristics of pain, restlessness and grimace. The only difference on the high risk for altered skin integrity scale was found on the distructor, supple skin.

  16. Numerical investigation of degas performance on impeller of medium-consistency pump

    Directory of Open Access Journals (Sweden)

    Hong Li

    2015-12-01

    Full Text Available Medium-consistency technology is known as the process with high efficiency and low pollution. The gas distribution was simulated in the medium-consistency pump with different degas hole positions. Rheological behaviors of pulp suspension were obtained by experimental test. A modified Herschel–Bulkley model and the Eulerian gas–liquid two-phase flow model were utilized to approximately represent the behaviors of the medium-consistency pulp suspension. The results show that when the relative position is 0.53, the gas volume ratio is less than 0.1% at the pump outlet and 9.8% at the vacuum inlet, and the pump head is at the maximum. Because of the different numbers of the impeller blades and turbulence blades and the asymmetric volute structure, the gas is distributed unevenly in the impeller. In addition, the pump performance was tested in experiment and the results are used to validate computational fluid dynamics outcomes.

  17. Grasping the Second Law of Thermodynamics at University: The Consistency of Macroscopic and Microscopic Explanations

    Science.gov (United States)

    Leinonen, Risto; Asikainen, Mervi A.; Hirvonen, Pekka E.

    2015-01-01

    This study concentrates on evaluating the consistency of upper-division students' use of the second law of thermodynamics at macroscopic and microscopic levels. Data were collected by means of a paper and pencil test (N = 48) focusing on the macroscopic and microscopic features of the second law concerned with heat transfer processes. The data…

  18. The internal consistency and validity of the Self-assessment Parkinson's Disease Disability Scale.

    NARCIS (Netherlands)

    Biemans, M.A.J.E.; Dekker, J.; Woude, L.H.V. van der

    2001-01-01

    OBJECTIVE: To test the consistency and validity of the Self-assessment Parkinson's Disease Disability Scale in patients with Parkinson's disease living at home. DESIGN: Patients with Parkinson's disease responded to a set of questionnaires. In addition, an observation of the performance of daily

  19. Internal consistency and validity of the self-assessment Parkinson's Disease disability scale. Abstract.

    NARCIS (Netherlands)

    Dekker, J.; Biemans, M.A.J.E.; Woude, L.H.V. van der

    2000-01-01

    OBJECTIVE: To test the consistency and validity of the Self-assessment Parkinson's Disease Disability Scale in patients with Parkinson's disease living at home. DESIGN: Patients with Parkinson's disease responded to a set of questionnaires. In addition, an observation of the performance of daily

  20. Time perspective and attitude-behaviour consistency in future-oriented behaviours

    NARCIS (Netherlands)

    Rabinovich, A.; Morton, T.; Postmes, T.

    The authors propose that the salience of a distant-future time perspective, compared to a near-future time perspective, should increase attitude-behaviour and attitude-intention consistency for future-oriented behaviours. To test this prediction, time perspective was experimentally manipulated in