WorldWideScience

Sample records for data validation

  1. Quality data validation: Comprehensive approach to environmental data validation

    International Nuclear Information System (INIS)

    Matejka, L.A. Jr.

    1993-01-01

    Environmental data validation consists of an assessment of three major areas: analytical method validation; field procedures and documentation review; evaluation of the level of achievement of data quality objectives based in part on PARCC parameters analysis and expected applications of data. A program utilizing matrix association of required levels of validation effort and analytical levels versus applications of this environmental data was developed in conjunction with DOE-ID guidance documents to implement actions under the Federal Facilities Agreement and Consent Order in effect at the Idaho National Engineering Laboratory. This was an effort to bring consistent quality to the INEL-wide Environmental Restoration Program and database in an efficient and cost-effective manner. This program, documenting all phases of the review process, is described here

  2. CosmoQuest:Using Data Validation for More Than Just Data Validation

    Science.gov (United States)

    Lehan, C.; Gay, P.

    2016-12-01

    It is often taken for granted that different scientists completing the same task (e.g. mapping geologic features) will get the same results, and data validation is often skipped or under-utilized due to time and funding constraints. Robbins et. al (2014), however, demonstrated that this is a needed step, as large variation can exist even among collaborating team members completing straight-forward tasks like marking craters. Data Validation should be much more than a simple post-project verification of results. The CosmoQuest virtual research facility employs regular data-validation for a variety of benefits, including real-time user feedback, real-time tracking to observe user activity while it's happening, and using pre-solved data to analyze users' progress and to help them retain skills. Some creativity in this area can drastically improve project results. We discuss methods of validating data in citizen science projects and outline the variety of uses for validation, which, when used properly, improves the scientific output of the project and the user experience for the citizens doing the work. More than just a tool for scientists, validation can assist users in both learning and retaining important information and skills, improving the quality and quantity of data gathered. Real-time analysis of user data can give key information in the effectiveness of the project that a broad glance would miss, and properly presenting that analysis is vital. Training users to validate their own data, or the data of others, can significantly improve the accuracy of misinformed or novice users.

  3. PIV Data Validation Software Package

    Science.gov (United States)

    Blackshire, James L.

    1997-01-01

    A PIV data validation and post-processing software package was developed to provide semi-automated data validation and data reduction capabilities for Particle Image Velocimetry data sets. The software provides three primary capabilities including (1) removal of spurious vector data, (2) filtering, smoothing, and interpolating of PIV data, and (3) calculations of out-of-plane vorticity, ensemble statistics, and turbulence statistics information. The software runs on an IBM PC/AT host computer working either under Microsoft Windows 3.1 or Windows 95 operating systems.

  4. ICP-MS Data Validation

    Science.gov (United States)

    Document designed to offer data reviewers guidance in determining the validity ofanalytical data generated through the USEPA Contract Laboratory Program Statement ofWork (SOW) ISM01.X Inorganic Superfund Methods (Multi-Media, Multi-Concentration)

  5. CIPS Validation Data Plan

    International Nuclear Information System (INIS)

    Dinh, Nam

    2012-01-01

    This report documents analysis, findings and recommendations resulted from a task 'CIPS Validation Data Plan (VDP)' formulated as an POR4 activity in the CASL VUQ Focus Area (FA), to develop a Validation Data Plan (VDP) for Crud-Induced Power Shift (CIPS) challenge problem, and provide guidance for the CIPS VDP implementation. The main reason and motivation for this task to be carried at this time in the VUQ FA is to bring together (i) knowledge of modern view and capability in VUQ, (ii) knowledge of physical processes that govern the CIPS, and (iii) knowledge of codes, models, and data available, used, potentially accessible, and/or being developed in CASL for CIPS prediction, to devise a practical VDP that effectively supports the CASL's mission in CIPS applications.

  6. CIPS Validation Data Plan

    Energy Technology Data Exchange (ETDEWEB)

    Nam Dinh

    2012-03-01

    This report documents analysis, findings and recommendations resulted from a task 'CIPS Validation Data Plan (VDP)' formulated as an POR4 activity in the CASL VUQ Focus Area (FA), to develop a Validation Data Plan (VDP) for Crud-Induced Power Shift (CIPS) challenge problem, and provide guidance for the CIPS VDP implementation. The main reason and motivation for this task to be carried at this time in the VUQ FA is to bring together (i) knowledge of modern view and capability in VUQ, (ii) knowledge of physical processes that govern the CIPS, and (iii) knowledge of codes, models, and data available, used, potentially accessible, and/or being developed in CASL for CIPS prediction, to devise a practical VDP that effectively supports the CASL's mission in CIPS applications.

  7. Site characterization and validation - validation drift fracture data, stage 4

    International Nuclear Information System (INIS)

    Bursey, G.; Gale, J.; MacLeod, R.; Straahle, A.; Tiren, S.

    1991-08-01

    This report describes the mapping procedures and the data collected during fracture mapping in the validation drift. Fracture characteristics examined include orientation, trace length, termination mode, and fracture minerals. These data have been compared and analysed together with fracture data from the D-boreholes to determine the adequacy of the borehole mapping procedures and to assess the nature and degree of orientation bias in the borehole data. The analysis of the validation drift data also includes a series of corrections to account for orientation, truncation, and censoring biases. This analysis has identified at least 4 geologically significant fracture sets in the rock mass defined by the validation drift. An analysis of the fracture orientations in both the good rock and the H-zone has defined groups of 7 clusters and 4 clusters, respectively. Subsequent analysis of the fracture patterns in five consecutive sections along the validation drift further identified heterogeneity through the rock mass, with respect to fracture orientations. These results are in stark contrast to the results form the D-borehole analysis, where a strong orientation bias resulted in a consistent pattern of measured fracture orientations through the rock. In the validation drift, fractures in the good rock also display a greater mean variance in length than those in the H-zone. These results provide strong support for a distinction being made between fractures in the good rock and the H-zone, and possibly between different areas of the good rock itself, for discrete modelling purposes. (au) (20 refs.)

  8. Validation of multisource electronic health record data: an application to blood transfusion data.

    Science.gov (United States)

    Hoeven, Loan R van; Bruijne, Martine C de; Kemper, Peter F; Koopman, Maria M W; Rondeel, Jan M M; Leyte, Anja; Koffijberg, Hendrik; Janssen, Mart P; Roes, Kit C B

    2017-07-14

    Although data from electronic health records (EHR) are often used for research purposes, systematic validation of these data prior to their use is not standard practice. Existing validation frameworks discuss validity concepts without translating these into practical implementation steps or addressing the potential influence of linking multiple sources. Therefore we developed a practical approach for validating routinely collected data from multiple sources and to apply it to a blood transfusion data warehouse to evaluate the usability in practice. The approach consists of identifying existing validation frameworks for EHR data or linked data, selecting validity concepts from these frameworks and establishing quantifiable validity outcomes for each concept. The approach distinguishes external validation concepts (e.g. concordance with external reports, previous literature and expert feedback) and internal consistency concepts which use expected associations within the dataset itself (e.g. completeness, uniformity and plausibility). In an example case, the selected concepts were applied to a transfusion dataset and specified in more detail. Application of the approach to a transfusion dataset resulted in a structured overview of data validity aspects. This allowed improvement of these aspects through further processing of the data and in some cases adjustment of the data extraction. For example, the proportion of transfused products that could not be linked to the corresponding issued products initially was 2.2% but could be improved by adjusting data extraction criteria to 0.17%. This stepwise approach for validating linked multisource data provides a basis for evaluating data quality and enhancing interpretation. When the process of data validation is adopted more broadly, this contributes to increased transparency and greater reliability of research based on routinely collected electronic health records.

  9. Mercury and Cyanide Data Validation

    Science.gov (United States)

    Document designed to offer data reviewers guidance in determining the validity ofanalytical data generated through the USEPA Contract Laboratory Program (CLP) Statement ofWork (SOW) ISM01.X Inorganic Superfund Methods (Multi-Media, Multi-Concentration)

  10. Earth Science Enterprise Scientific Data Purchase Project: Verification and Validation

    Science.gov (United States)

    Jenner, Jeff; Policelli, Fritz; Fletcher, Rosea; Holecamp, Kara; Owen, Carolyn; Nicholson, Lamar; Dartez, Deanna

    2000-01-01

    This paper presents viewgraphs on the Earth Science Enterprise Scientific Data Purchase Project's verification,and validation process. The topics include: 1) What is Verification and Validation? 2) Why Verification and Validation? 3) Background; 4) ESE Data Purchas Validation Process; 5) Data Validation System and Ingest Queue; 6) Shipment Verification; 7) Tracking and Metrics; 8) Validation of Contract Specifications; 9) Earth Watch Data Validation; 10) Validation of Vertical Accuracy; and 11) Results of Vertical Accuracy Assessment.

  11. Validation of EAF-2005 data

    International Nuclear Information System (INIS)

    Kopecky, J.

    2005-01-01

    Full text: Validation procedures applied on EAF-2003 starter file, which lead to the production of EAF-2005 library, are described. The results in terms of reactions with assigned quality scores in EAF-20005 are given. Further the extensive validation against the recent integral data is discussed together with the status of the final report 'Validation of EASY-2005 using integral measurements'. Finally, the novel 'cross section trend analysis' is presented with some examples of its use. This action will lead to the release of improved library EAF-2005.1 at the end of 2005, which shall be used as the starter file for EAF-2007. (author)

  12. Improving your real-time data infrastructure using advanced data validation and reconciliation

    Energy Technology Data Exchange (ETDEWEB)

    Wising, Ulrika; Campan, Julien; Vrielynck, Bruno; Anjos, Cristiano dos; Kalitventzeff, Pierre-Boris [Belsim S.A., Awans (Belgium)

    2008-07-01

    'Smart fields', 'e-fields', 'field of the future', 'digital oil fields' and 'field monitoring' are all names of real-time data infrastructures aimed at providing information for decision making. This paper discusses these new real-time data infrastructures that are being developed and deployed in oil and gas production and in particular the challenge of supplying these new systems with high quality data. In order for these infrastructures to be successful and provide efficient and successful performance management and optimization, they need to have access to high quality production data. Advanced Data Validation and Reconciliation is a technology that could meet this data quality challenge. It has been successfully deployed in many different industry sectors and more recently in oil and gas production. Advanced Data Validation and Reconciliation provides a coherent, accurate set of production data and basing these new infrastructures on validated and reconciled data brings a solution to the data quality challenge. There are numerous other benefits by applying advanced data validation and reconciliation in oil and gas production, such as uninterrupted well production, optimized valves opening and water or gas injection, backup values for traditional multiphase flow meters, and the avoidance of production upsets. (author)

  13. Automatic, semi-automatic and manual validation of urban drainage data.

    Science.gov (United States)

    Branisavljević, N; Prodanović, D; Pavlović, D

    2010-01-01

    Advances in sensor technology and the possibility of automated long distance data transmission have made continuous measurements the preferable way of monitoring urban drainage processes. Usually, the collected data have to be processed by an expert in order to detect and mark the wrong data, remove them and replace them with interpolated data. In general, the first step in detecting the wrong, anomaly data is called the data quality assessment or data validation. Data validation consists of three parts: data preparation, validation scores generation and scores interpretation. This paper will present the overall framework for the data quality improvement system, suitable for automatic, semi-automatic or manual operation. The first two steps of the validation process are explained in more detail, using several validation methods on the same set of real-case data from the Belgrade sewer system. The final part of the validation process, which is the scores interpretation, needs to be further investigated on the developed system.

  14. Radiochemical verification and validation in the environmental data collection process

    International Nuclear Information System (INIS)

    Rosano-Reece, D.; Bottrell, D.; Bath, R.J.

    1994-01-01

    A credible and cost effective environmental data collection process should produce analytical data which meets regulatory and program specific requirements. Analytical data, which support the sampling and analysis activities at hazardous waste sites, undergo verification and independent validation before the data are submitted to regulators. Understanding the difference between verification and validation and their respective roles in the sampling and analysis process is critical to the effectiveness of a program. Verification is deciding whether the measurement data obtained are what was requested. The verification process determines whether all the requirements were met. Validation is more complicated than verification. It attempts to assess the impacts on data use, especially when requirements are not met. Validation becomes part of the decision-making process. Radiochemical data consists of a sample result with an associated error. Therefore, radiochemical validation is different and more quantitative than is currently possible for the validation of hazardous chemical data. Radiochemical data include both results and uncertainty that can be statistically compared to identify significance of differences in a more technically defensible manner. Radiochemical validation makes decisions about analyte identification, detection, and uncertainty for a batch of data. The process focuses on the variability of the data in the context of the decision to be made. The objectives of this paper are to present radiochemical verification and validation for environmental data and to distinguish the differences between the two operations

  15. Worldwide Protein Data Bank validation information: usage and trends.

    Science.gov (United States)

    Smart, Oliver S; Horský, Vladimír; Gore, Swanand; Svobodová Vařeková, Radka; Bendová, Veronika; Kleywegt, Gerard J; Velankar, Sameer

    2018-03-01

    Realising the importance of assessing the quality of the biomolecular structures deposited in the Protein Data Bank (PDB), the Worldwide Protein Data Bank (wwPDB) partners established Validation Task Forces to obtain advice on the methods and standards to be used to validate structures determined by X-ray crystallography, nuclear magnetic resonance spectroscopy and three-dimensional electron cryo-microscopy. The resulting wwPDB validation pipeline is an integral part of the wwPDB OneDep deposition, biocuration and validation system. The wwPDB Validation Service webserver (https://validate.wwpdb.org) can be used to perform checks prior to deposition. Here, it is shown how validation metrics can be combined to produce an overall score that allows the ranking of macromolecular structures and domains in search results. The ValTrends DB database provides users with a convenient way to access and analyse validation information and other properties of X-ray crystal structures in the PDB, including investigating trends in and correlations between different structure properties and validation metrics.

  16. Validity of diagnoses, procedures, and laboratory data in Japanese administrative data.

    Science.gov (United States)

    Yamana, Hayato; Moriwaki, Mutsuko; Horiguchi, Hiromasa; Kodan, Mariko; Fushimi, Kiyohide; Yasunaga, Hideo

    2017-10-01

    Validation of recorded data is a prerequisite for studies that utilize administrative databases. The present study evaluated the validity of diagnoses and procedure records in the Japanese Diagnosis Procedure Combination (DPC) data, along with laboratory test results in the newly-introduced Standardized Structured Medical Record Information Exchange (SS-MIX) data. Between November 2015 and February 2016, we conducted chart reviews of 315 patients hospitalized between April 2014 and March 2015 in four middle-sized acute-care hospitals in Shizuoka, Kochi, Fukuoka, and Saga Prefectures and used them as reference standards. The sensitivity and specificity of DPC data in identifying 16 diseases and 10 common procedures were identified. The accuracy of SS-MIX data for 13 laboratory test results was also examined. The specificity of diagnoses in the DPC data exceeded 96%, while the sensitivity was below 50% for seven diseases and variable across diseases. When limited to primary diagnoses, the sensitivity and specificity were 78.9% and 93.2%, respectively. The sensitivity of procedure records exceeded 90% for six procedures, and the specificity exceeded 90% for nine procedures. Agreement between the SS-MIX data and the chart reviews was above 95% for all 13 items. The validity of diagnoses and procedure records in the DPC data and laboratory results in the SS-MIX data was high in general, supporting their use in future studies. Copyright © 2017 The Authors. Production and hosting by Elsevier B.V. All rights reserved.

  17. Hanford Environmental Restoration data validation process for chemical and radiochemical analyses

    International Nuclear Information System (INIS)

    Adams, M.R.; Bechtold, R.A.; Clark, D.E.; Angelos, K.M.; Winter, S.M.

    1993-10-01

    Detailed procedures for validation of chemical and radiochemical data are used to assure consistent application of validation principles and support a uniform database of quality environmental data. During application of these procedures, it was determined that laboratory data packages were frequently missing certain types of documentation causing subsequent delays in meeting critical milestones in the completion of validation activities. A quality improvement team was assembled to address the problems caused by missing documentation and streamline the entire process. The result was the development of a separate data package verification procedure and revisions to the data validation procedures. This has resulted in a system whereby deficient data packages are immediately identified and corrected prior to validation and revised validation procedures which more closely match the common analytical reporting practices of laboratory service vendors

  18. Validation of Structures in the Protein Data Bank.

    Science.gov (United States)

    Gore, Swanand; Sanz García, Eduardo; Hendrickx, Pieter M S; Gutmanas, Aleksandras; Westbrook, John D; Yang, Huanwang; Feng, Zukang; Baskaran, Kumaran; Berrisford, John M; Hudson, Brian P; Ikegawa, Yasuyo; Kobayashi, Naohiro; Lawson, Catherine L; Mading, Steve; Mak, Lora; Mukhopadhyay, Abhik; Oldfield, Thomas J; Patwardhan, Ardan; Peisach, Ezra; Sahni, Gaurav; Sekharan, Monica R; Sen, Sanchayita; Shao, Chenghua; Smart, Oliver S; Ulrich, Eldon L; Yamashita, Reiko; Quesada, Martha; Young, Jasmine Y; Nakamura, Haruki; Markley, John L; Berman, Helen M; Burley, Stephen K; Velankar, Sameer; Kleywegt, Gerard J

    2017-12-05

    The Worldwide PDB recently launched a deposition, biocuration, and validation tool: OneDep. At various stages of OneDep data processing, validation reports for three-dimensional structures of biological macromolecules are produced. These reports are based on recommendations of expert task forces representing crystallography, nuclear magnetic resonance, and cryoelectron microscopy communities. The reports provide useful metrics with which depositors can evaluate the quality of the experimental data, the structural model, and the fit between them. The validation module is also available as a stand-alone web server and as a programmatically accessible web service. A growing number of journals require the official wwPDB validation reports (produced at biocuration) to accompany manuscripts describing macromolecular structures. Upon public release of the structure, the validation report becomes part of the public PDB archive. Geometric quality scores for proteins in the PDB archive have improved over the past decade. Copyright © 2017 The Authors. Published by Elsevier Ltd.. All rights reserved.

  19. [Data validation methods and discussion on Chinese materia medica resource survey].

    Science.gov (United States)

    Zhang, Yue; Ma, Wei-Feng; Zhang, Xiao-Bo; Zhu, Shou-Dong; Guo, Lan-Ping; Wang, Xing-Xing

    2013-07-01

    From the beginning of the fourth national survey of the Chinese materia medica resources, there were 22 provinces have conducted pilots. The survey teams have reported immense data, it put forward the very high request to the database system construction. In order to ensure the quality, it is necessary to check and validate the data in database system. Data validation is important methods to ensure the validity, integrity and accuracy of census data. This paper comprehensively introduce the data validation system of the fourth national survey of the Chinese materia medica resources database system, and further improve the design idea and programs of data validation. The purpose of this study is to promote the survey work smoothly.

  20. CASL Validation Data: An Initial Review

    Energy Technology Data Exchange (ETDEWEB)

    Nam Dinh

    2011-01-01

    The study aims to establish a comprehensive view of “data” needed for supporting implementation of the Consortium of Advanced Simulation of LWRs (CASL). Insights from this review (and its continual refinement), together with other elements developed in CASL, should provide the foundation for developing the CASL Validation Data Plan (VDP). VDP is instrumental to the development and assessment of CASL simulation tools as predictive capability. Most importantly, to be useful for CASL, the VDP must be devised (and agreed upon by all participating stakeholders) with appropriate account for nature of nuclear engineering applications, the availability, types and quality of CASL-related data, and novelty of CASL goals and its approach to the selected challenge problems. The initial review (summarized on the January 2011 report version) discusses a broad range of methodological issues in data review and Validation Data Plan. Such a top-down emphasis in data review is both needed to see a big picture on CASL data and appropriate when the actual data are not available for detailed scrutiny. As the data become available later in 2011, a revision of data review (and regular update) should be performed. It is expected that the basic framework for review laid out in this report will help streamline the CASL data review in a way that most pertinent to CASL VDP.

  1. Processing and validation of intermediate energy evaluated data files

    International Nuclear Information System (INIS)

    2000-01-01

    Current accelerator-driven and other intermediate energy technologies require accurate nuclear data to model the performance of the target/blanket assembly, neutron production, activation, heating and damage. In a previous WPEC subgroup, SG13 on intermediate energy nuclear data, various aspects of intermediate energy data, such as nuclear data needs, experiments, model calculations and file formatting issues were investigated and categorized to come to a joint evaluation effort. The successor of SG13, SG14 on the processing and validation of intermediate energy evaluated data files, goes one step further. The nuclear data files that have been created with the aforementioned information need to be processed and validated in order to be applicable in realistic intermediate energy simulations. We emphasize that the work of SG14 excludes the 0-20 MeV data part of the neutron evaluations, which is supposed to be covered elsewhere. This final report contains the following sections: section 2: a survey of the data files above 20 MeV that have been considered for validation in SG14; section 3: a summary of the review of the 150 MeV intermediate energy data files for ENDF/B-VI and, more briefly, the other libraries; section 4: validation of the data library against an integral experiment with MCNPX; section 5: conclusions. (author)

  2. Opportunities and Pitfalls in the Definition of Data Validity.

    Science.gov (United States)

    Stausberg, Jürgen; Kuklik, Nils; Jöckel, Karl-Heinz

    2018-01-01

    Several dimensions of data quality are described in the literature. One overriding aspect is considered to be the extent to which data represent the truth which is captured by data validity. Unfortunately, a common terminology, well defined concepts, and approved measures are missing in regard to data validity. In particular, there is a need to discuss the gold standard as reference for the data at hand and respective measures. Ultimate gold standard would be the state of the patient which itself is subjected to human and personal interpretations. Usually, an often diverse form of source data is used as gold standard. Based on the concept of the measure, it might be inappropriate differentiating between present and absent while calculating precision and recall. Due to the complexity and uncertainty of many health care related issues, a more sophisticated comparison might be necessary in order to establish relevant and general figures of data quality. Unfortunately, a harmonization in this field is not visible. Further research is needed to establish validated standards to measure data quality.

  3. Treatment and Combination of Data Quality Monitoring Histograms to Perform Data vs. Monte Carlo Validation

    CERN Document Server

    Colin, Nolan

    2013-01-01

    In CMS's automated data quality validation infrastructure, it is not currently possible to assess how well Monte Carlo simulations describe data from collisions, if at all. In order to guarantee high quality data, a novel work flow was devised to perform `data vs. Monte Carlo' validation. Support for this comparison was added by allowing distributions from several Monte Carlo samples to be combined, matched to the data and then displayed in a histogram stack, overlaid with the experimental data.

  4. Valid, legally defensible data from your analytical laboratories

    International Nuclear Information System (INIS)

    Gay, D.D.; Allen, V.C.

    1989-01-01

    This paper discusses the definition of valid, legally defensible data. The authors describe the expectations of project managers and what should be gleaned from the laboratory in regard to analytical data

  5. Gaia Data Release 1. Catalogue validation

    NARCIS (Netherlands)

    Arenou, F.; Luri, X.; Babusiaux, C.; Fabricius, C.; Helmi, A.; Robin, A. C.; Vallenari, A.; Blanco-Cuaresma, S.; Cantat-Gaudin, T.; Findeisen, K.; Reylé, C.; Ruiz-Dern, L.; Sordo, R.; Turon, C.; Walton, N. A.; Shih, I.-C.; Antiche, E.; Barache, C.; Barros, M.; Breddels, M.; Carrasco, J. M.; Costigan, G.; Diakité, S.; Eyer, L.; Figueras, F.; Galluccio, L.; Heu, J.; Jordi, C.; Krone-Martins, A.; Lallement, R.; Lambert, S.; Leclerc, N.; Marrese, P. M.; Moitinho, A.; Mor, R.; Romero-Gómez, M.; Sartoretti, P.; Soria, S.; Soubiran, C.; Souchay, J.; Veljanoski, J.; Ziaeepour, H.; Giuffrida, G.; Pancino, E.; Bragaglia, A.

    Context. Before the publication of the Gaia Catalogue, the contents of the first data release have undergone multiple dedicated validation tests. Aims: These tests aim to provide in-depth analysis of the Catalogue content in order to detect anomalies and individual problems in specific objects or in

  6. Validation of argo data in the Indian Ocean

    Digital Repository Service at National Institute of Oceanography (India)

    Pankajakshan, T.; Muraleedharan, P.M.; Gopalakrishna, V.V.; Reddy, G.V.; Ratnakaran, L.; Revichandran, C.; Murty, V.S.N.

    Gayana (Concepción) - VALIDATION OF ARGO DATA IN THE INDIAN OCEA... 8/11/2006http://www.scielo.cl/scielo.php?script=sci_arttext&pid=S0717-65382004000300025&lng=... susceptible to fouling and associated drift because of the possible change... initial profiles (profiles 1 and 2). Figure-1b represents the same as that in figure-1a, but for 29 match-ups involving profile numbers 5 and above. Page 2 of 5Gayana (Concepción) - VALIDATION OF ARGO DATA IN THE INDIAN OCEA... 8/11/2006http://www...

  7. An introduction to use of the USACE HTRW program's data validation guidelines engineering manual

    International Nuclear Information System (INIS)

    Becker, L.D.; Coats, K.H.

    1994-01-01

    Data validation has been defined by regulatory agencies as a systematic process (consisting of data editing, screening, checking, auditing, verification, certification, and review) for comparing data to established criteria in order to provide assurance that data are adequate for their intended use. A problem for the USACE HTRW Program was that clearly defined data validation guidelines were available only for analytical data quality level IV. These functional data validation guidelines were designed for validation of data produced using protocols from the US E.P.A.'s Contract Laboratory Program (CLP). Unfortunately, USACE experience demonstrates that these level IV functional data validation guidelines were being used to validate data not produced under the CLP. The resulting data validation product was less than satisfactory for USACE HTRW needs. Therefore, the HTRW-MCX initiated an Engineering Manual (EM) for validation of analytical data quality levels other than IV. This EM is entitle ''USACE HTRW Data Validation Guidelines.'' Use of the EM is required for validation of analytical data relating to projects under the jurisdiction of the Department of the Army, Corps of Engineers, Hazardous, Toxic, and Radioactive Waste Program. These data validation guidelines include procedures and checklists for technical review of analytical data at quality levels I, II, III, and V

  8. A Comprehensive Validation Methodology for Sparse Experimental Data

    Science.gov (United States)

    Norman, Ryan B.; Blattnig, Steve R.

    2010-01-01

    A comprehensive program of verification and validation has been undertaken to assess the applicability of models to space radiation shielding applications and to track progress as models are developed over time. The models are placed under configuration control, and automated validation tests are used so that comparisons can readily be made as models are improved. Though direct comparisons between theoretical results and experimental data are desired for validation purposes, such comparisons are not always possible due to lack of data. In this work, two uncertainty metrics are introduced that are suitable for validating theoretical models against sparse experimental databases. The nuclear physics models, NUCFRG2 and QMSFRG, are compared to an experimental database consisting of over 3600 experimental cross sections to demonstrate the applicability of the metrics. A cumulative uncertainty metric is applied to the question of overall model accuracy, while a metric based on the median uncertainty is used to analyze the models from the perspective of model development by analyzing subsets of the model parameter space.

  9. A project manager's primer on data validation

    International Nuclear Information System (INIS)

    Ramos, S.J.

    1991-01-01

    While carrying out their central responsibility of conducting an environmental investigation in a high- quality, timely, and cost-effective manner, project managers also face a significant challenge due to the many inherent uncertainties associated with characterizing and remediating sites. From all aspects and considerations (health and financial risks; and technical, professional, and legal defensibility/credibility), the project manager must minimize the uncertainty associated with making decisions based on environmental data. A key objective for every project manager is to produce sufficient data of known and acceptable quality. In simple terms, the level of confidence in the gathered data directly relates to: (1) the upfront determination of the types and uses of the data needed (which drives the required quality of the data); (2) the ongoing verification that the prescribed methods by which the data are to be obtained and analyzed are being followed; and (3) the validation of the verified data to determine whether the preestablished data quality objectives have been met, therefore making the data adequate for their intended use(s). This paper focuses on the third clement of the equation for data quality, therefore implying that the first two elements (planning and verification) have been accomplished. The open-quotes Who,close quotes open-quotes What,close quotes open-quotes Why,close quotes open-quotes Whenclose quotes and open-quotes Howclose quotes of data validation are discussed in general terms

  10. Likelihood ratio data to report the validation of a forensic fingerprint evaluation method

    Directory of Open Access Journals (Sweden)

    Daniel Ramos

    2017-02-01

    Full Text Available Data to which the authors refer to throughout this article are likelihood ratios (LR computed from the comparison of 5–12 minutiae fingermarks with fingerprints. These LRs data are used for the validation of a likelihood ratio (LR method in forensic evidence evaluation. These data present a necessary asset for conducting validation experiments when validating LR methods used in forensic evidence evaluation and set up validation reports. These data can be also used as a baseline for comparing the fingermark evidence in the same minutiae configuration as presented in (D. Meuwly, D. Ramos, R. Haraksim, [1], although the reader should keep in mind that different feature extraction algorithms and different AFIS systems used may produce different LRs values. Moreover, these data may serve as a reproducibility exercise, in order to train the generation of validation reports of forensic methods, according to [1]. Alongside the data, a justification and motivation for the use of methods is given. These methods calculate LRs from the fingerprint/mark data and are subject to a validation procedure. The choice of using real forensic fingerprint in the validation and simulated data in the development is described and justified. Validation criteria are set for the purpose of validation of the LR methods, which are used to calculate the LR values from the data and the validation report. For privacy and data protection reasons, the original fingerprint/mark images cannot be shared. But these images do not constitute the core data for the validation, contrarily to the LRs that are shared.

  11. Implementing an X-ray validation pipeline for the Protein Data Bank

    International Nuclear Information System (INIS)

    Gore, Swanand; Velankar, Sameer; Kleywegt, Gerard J.

    2012-01-01

    The implementation of a validation pipeline, based on community recommendations, for future depositions of X-ray crystal structures in the Protein Data Bank is described. There is an increasing realisation that the quality of the biomacromolecular structures deposited in the Protein Data Bank (PDB) archive needs to be assessed critically using established and powerful validation methods. The Worldwide Protein Data Bank (wwPDB) organization has convened several Validation Task Forces (VTFs) to advise on the methods and standards that should be used to validate all of the entries already in the PDB as well as all structures that will be deposited in the future. The recommendations of the X-ray VTF are currently being implemented in a software pipeline. Here, ongoing work on this pipeline is briefly described as well as ways in which validation-related information could be presented to users of structural data

  12. Implementing an X-ray validation pipeline for the Protein Data Bank

    Energy Technology Data Exchange (ETDEWEB)

    Gore, Swanand; Velankar, Sameer; Kleywegt, Gerard J., E-mail: gerard@ebi.ac.uk [EMBL–EBI, Wellcome Trust Genome Campus, Hinxton, Cambridge CB10 1SD (United Kingdom)

    2012-04-01

    The implementation of a validation pipeline, based on community recommendations, for future depositions of X-ray crystal structures in the Protein Data Bank is described. There is an increasing realisation that the quality of the biomacromolecular structures deposited in the Protein Data Bank (PDB) archive needs to be assessed critically using established and powerful validation methods. The Worldwide Protein Data Bank (wwPDB) organization has convened several Validation Task Forces (VTFs) to advise on the methods and standards that should be used to validate all of the entries already in the PDB as well as all structures that will be deposited in the future. The recommendations of the X-ray VTF are currently being implemented in a software pipeline. Here, ongoing work on this pipeline is briefly described as well as ways in which validation-related information could be presented to users of structural data.

  13. Collecting the data but missing the point: validity of hand hygiene audit data.

    Science.gov (United States)

    Jeanes, A; Coen, P G; Wilson, A P; Drey, N S; Gould, D J

    2015-06-01

    Monitoring of hand hygiene compliance (HHC) by observation has been used in healthcare for more than a decade to provide assurance of infection control practice. The validity of this information is rarely tested. To examine the process and validity of collecting and reporting HHC data based on direct observation of compliance. Five years of HHC data routinely collected in one large National Health Service hospital trust were examined. The data collection process was reviewed by survey and interview of the auditors. HHC data collected for other research purposes undertaken during this period were compared with the organizational data set. After an initial increase, the reported HHC remained unchanged close to its intended target throughout this period. Examination of the data collection process revealed changes, including local interpretations of the data collection system, which invalidated the results. A minority of auditors had received formal training in observation and feedback of results. Whereas observation of HHC is the current gold standard, unless data collection definitions and methods are unambiguous, published, carefully supervised, and regularly monitored, variations may occur which affect the validity of the data. If the purpose of HHC monitoring is to improve practice and minimize transmission of infection, then a focus on progressively improving performance rather than on achieving a target may offer greater opportunities to achieve this. Copyright © 2015 The Healthcare Infection Society. Published by Elsevier Ltd. All rights reserved.

  14. Integrated Syntactic/Semantic XML Data Validation with a Reusable Software Component

    Science.gov (United States)

    Golikov, Steven

    2013-01-01

    Data integration is a critical component of enterprise system integration, and XML data validation is the foundation for sound data integration of XML-based information systems. Since B2B e-commerce relies on data validation as one of the critical components for enterprise integration, it is imperative for financial industries and e-commerce…

  15. Methods for Geometric Data Validation of 3d City Models

    Science.gov (United States)

    Wagner, D.; Alam, N.; Wewetzer, M.; Pries, M.; Coors, V.

    2015-12-01

    Geometric quality of 3D city models is crucial for data analysis and simulation tasks, which are part of modern applications of the data (e.g. potential heating energy consumption of city quarters, solar potential, etc.). Geometric quality in these contexts is however a different concept as it is for 2D maps. In the latter case, aspects such as positional or temporal accuracy and correctness represent typical quality metrics of the data. They are defined in ISO 19157 and should be mentioned as part of the metadata. 3D data has a far wider range of aspects which influence their quality, plus the idea of quality itself is application dependent. Thus, concepts for definition of quality are needed, including methods to validate these definitions. Quality on this sense means internal validation and detection of inconsistent or wrong geometry according to a predefined set of rules. A useful starting point would be to have correct geometry in accordance with ISO 19107. A valid solid should consist of planar faces which touch their neighbours exclusively in defined corner points and edges. No gaps between them are allowed, and the whole feature must be 2-manifold. In this paper, we present methods to validate common geometric requirements for building geometry. Different checks based on several algorithms have been implemented to validate a set of rules derived from the solid definition mentioned above (e.g. water tightness of the solid or planarity of its polygons), as they were developed for the software tool CityDoctor. The method of each check is specified, with a special focus on the discussion of tolerance values where they are necessary. The checks include polygon level checks to validate the correctness of each polygon, i.e. closeness of the bounding linear ring and planarity. On the solid level, which is only validated if the polygons have passed validation, correct polygon orientation is checked, after self-intersections outside of defined corner points and edges

  16. Mining Twitter Data to Augment NASA GPM Validation

    Science.gov (United States)

    Teng, Bill; Albayrak, Arif; Huffman, George; Vollmer, Bruce; Loeser, Carlee; Acker, Jim

    2017-01-01

    The Twitter data stream is an important new source of real-time and historical global information for potentially augmenting the validation program of NASA's Global Precipitation Measurement (GPM) mission. There have been other similar uses of Twitter, though mostly related to natural hazards monitoring and management. The validation of satellite precipitation estimates is challenging, because many regions lack data or access to data, especially outside of the U.S. and in remote and developing areas. The time-varying set of "precipitation" tweets can be thought of as an organic network of rain gauges, potentially providing a widespread view of precipitation occurrence. Twitter provides a large source of crowd for crowdsourcing. During a 24-hour period in the middle of the snow storm this past March in the U.S. Northeast, we collected more than 13,000 relevant precipitation tweets with exact geolocation. The overall objective of our project is to determine the extent to which processed tweets can provide additional information that improves the validation of GPM data. Though our current effort focuses on tweets and precipitation, our approach is general and applicable to other social media and other geophysical measurements. Specifically, we have developed an operational infrastructure for processing tweets, in a format suitable for analysis with GPM data; engaged with potential participants, both passive and active, to "enrich" the Twitter stream; and inter-compared "precipitation" tweet data, ground station data, and GPM retrievals. In this presentation, we detail the technical capabilities of our tweet processing infrastructure, including data abstraction, feature extraction, search engine, context-awareness, real-time processing, and high volume (big) data processing; various means for "enriching" the Twitter stream; and results of inter-comparisons. Our project should bring a new kind of visibility to Twitter and engender a new kind of appreciation of the value

  17. Internal Cluster Validation on Earthquake Data in the Province of Bengkulu

    Science.gov (United States)

    Rini, D. S.; Novianti, P.; Fransiska, H.

    2018-04-01

    K-means method is an algorithm for cluster n object based on attribute to k partition, where k < n. There is a deficiency of algorithms that is before the algorithm is executed, k points are initialized randomly so that the resulting data clustering can be different. If the random value for initialization is not good, the clustering becomes less optimum. Cluster validation is a technique to determine the optimum cluster without knowing prior information from data. There are two types of cluster validation, which are internal cluster validation and external cluster validation. This study aims to examine and apply some internal cluster validation, including the Calinski-Harabasz (CH) Index, Sillhouette (S) Index, Davies-Bouldin (DB) Index, Dunn Index (D), and S-Dbw Index on earthquake data in the Bengkulu Province. The calculation result of optimum cluster based on internal cluster validation is CH index, S index, and S-Dbw index yield k = 2, DB Index with k = 6 and Index D with k = 15. Optimum cluster (k = 6) based on DB Index gives good results for clustering earthquake in the Bengkulu Province.

  18. A rating system for post pulse data validation

    International Nuclear Information System (INIS)

    Buceti, G.; Centioli, Cristina; Iannone, F.; Panella, M.; Rizzo, A.; Vitale, V.

    2003-01-01

    The aim of an automatic data validation system in a fusion experiment is to account--after every shot--for any occurrence of faulty sensors and unreliable measurements, thus preventing the proliferation of poor pulse data. In the past years a prototype has been successfully developed at Frascati Tokamak Upgrade (FTU) on a small set of density measurements. The results have shown that the model can be further extended to plant and diagnostic data, and that the same system can be used to assign to raw data a quality factor, to be stored in the archive and to be used in the post-shot elaboration phase as a selection criterion. In this way, a data validation system can also provide data analysts with an useful tool to be used as a key--together with other significant parameters, like plasma current, or magnetic field--to search the archive for quality data. This paper will describe how, using soft computing techniques, both these functions have been implemented on FTU, providing the users with a simple interface for fault detection developed in an open source environment (PHP-MySQL), to be finalised into the realisation of an overall rating system for FTU data

  19. A rating system for post pulse data validation

    Energy Technology Data Exchange (ETDEWEB)

    Buceti, G.; Centioli, Cristina E-mail: centioli@frascati.enea.it; Iannone, F.; Panella, M.; Rizzo, A.; Vitale, V

    2003-09-01

    The aim of an automatic data validation system in a fusion experiment is to account--after every shot--for any occurrence of faulty sensors and unreliable measurements, thus preventing the proliferation of poor pulse data. In the past years a prototype has been successfully developed at Frascati Tokamak Upgrade (FTU) on a small set of density measurements. The results have shown that the model can be further extended to plant and diagnostic data, and that the same system can be used to assign to raw data a quality factor, to be stored in the archive and to be used in the post-shot elaboration phase as a selection criterion. In this way, a data validation system can also provide data analysts with an useful tool to be used as a key--together with other significant parameters, like plasma current, or magnetic field--to search the archive for quality data. This paper will describe how, using soft computing techniques, both these functions have been implemented on FTU, providing the users with a simple interface for fault detection developed in an open source environment (PHP-MySQL), to be finalised into the realisation of an overall rating system for FTU data.

  20. Implementing an X-ray validation pipeline for the Protein Data Bank.

    Science.gov (United States)

    Gore, Swanand; Velankar, Sameer; Kleywegt, Gerard J

    2012-04-01

    There is an increasing realisation that the quality of the biomacromolecular structures deposited in the Protein Data Bank (PDB) archive needs to be assessed critically using established and powerful validation methods. The Worldwide Protein Data Bank (wwPDB) organization has convened several Validation Task Forces (VTFs) to advise on the methods and standards that should be used to validate all of the entries already in the PDB as well as all structures that will be deposited in the future. The recommendations of the X-ray VTF are currently being implemented in a software pipeline. Here, ongoing work on this pipeline is briefly described as well as ways in which validation-related information could be presented to users of structural data.

  1. Nuclear data to support computer code validation

    International Nuclear Information System (INIS)

    Fisher, S.E.; Broadhead, B.L.; DeHart, M.D.; Primm, R.T. III

    1997-04-01

    The rate of plutonium disposition will be a key parameter in determining the degree of success of the Fissile Materials Disposition Program. Estimates of the disposition rate are dependent on neutronics calculations. To ensure that these calculations are accurate, the codes and data should be validated against applicable experimental measurements. Further, before mixed-oxide (MOX) fuel can be fabricated and loaded into a reactor, the fuel vendors, fabricators, fuel transporters, reactor owners and operators, regulatory authorities, and the Department of Energy (DOE) must accept the validity of design calculations. This report presents sources of neutronics measurements that have potential application for validating reactor physics (predicting the power distribution in the reactor core), predicting the spent fuel isotopic content, predicting the decay heat generation rate, certifying criticality safety of fuel cycle facilities, and ensuring adequate radiation protection at the fuel cycle facilities and the reactor. The U.S. in-reactor experience with MOX fuel is first presented, followed by information related to other aspects of the MOX fuel performance information that is valuable to this program, but the data base remains largely proprietary. Thus, this information is not reported here. It is expected that the selected consortium will make the necessary arrangements to procure or have access to the requisite information

  2. The Copernicus S5P Mission Performance Centre / Validation Data Analysis Facility for TROPOMI operational atmospheric data products

    Science.gov (United States)

    Compernolle, Steven; Lambert, Jean-Christopher; Langerock, Bavo; Granville, José; Hubert, Daan; Keppens, Arno; Rasson, Olivier; De Mazière, Martine; Fjæraa, Ann Mari; Niemeijer, Sander

    2017-04-01

    Sentinel-5 Precursor (S5P), to be launched in 2017 as the first atmospheric composition satellite of the Copernicus programme, carries as payload the TROPOspheric Monitoring Instrument (TROPOMI) developed by The Netherlands in close cooperation with ESA. Designed to measure Earth radiance and solar irradiance in the ultraviolet, visible and near infrared, TROPOMI will provide Copernicus with observational data on atmospheric composition at unprecedented geographical resolution. The S5P Mission Performance Center (MPC) provides an operational service-based solution for various QA/QC tasks, including the validation of S5P Level-2 data products and the support to algorithm evolution. Those two tasks are to be accomplished by the MPC Validation Data Analysis Facility (VDAF), one MPC component developed and operated at BIRA-IASB with support from S[&]T and NILU. The routine validation to be ensured by VDAF is complemented by a list of validation AO projects carried out by ESA's S5P Validation Team (S5PVT), with whom interaction is essential. Here we will introduce the general architecture of VDAF, its relation to the other MPC components, the generic and specific validation strategies applied for each of the official TROPOMI data products, and the expected output of the system. The S5P data products to be validated by VDAF are diverse: O3 (vertical profile, total column, tropospheric column), NO2 (total and tropospheric column), HCHO (tropospheric column), SO2 (column), CO (column), CH4 (column), aerosol layer height and clouds (fractional cover, cloud-top pressure and optical thickness). Starting from a generic validation protocol meeting community-agreed standards, a set of specific validation settings is associated with each data product, as well as the appropriate set of Fiducial Reference Measurements (FRM) to which it will be compared. VDAF collects FRMs from ESA's Validation Data Centre (EVDC) and from other sources (e.g., WMO's GAW, NDACC and TCCON). Data

  3. Integration and validation of a data grid software

    Science.gov (United States)

    Carenton-Madiec, Nicolas; Berger, Katharina; Cofino, Antonio

    2014-05-01

    The Earth System Grid Federation (ESGF) Peer-to-Peer (P2P) is a software infrastructure for the management, dissemination, and analysis of model output and observational data. The ESGF grid is composed with several types of nodes which have different roles. About 40 data nodes host model outputs and datasets using thredds catalogs. About 25 compute nodes offer remote visualization and analysis tools. About 15 index nodes crawl data nodes catalogs and implement faceted and federated search in a web interface. About 15 Identity providers nodes manage accounts, authentication and authorization. Here we will present an actual size test federation spread across different institutes in different countries and a python test suite that were started in December 2013. The first objective of the test suite is to provide a simple tool that helps to test and validate a single data node and its closest index, compute and identity provider peer. The next objective will be to run this test suite on every data node of the federation and therefore test and validate every single node of the whole federation. The suite already implements nosetests, requests, myproxy-logon, subprocess, selenium and fabric python libraries in order to test both web front ends, back ends and security services. The goal of this project is to improve the quality of deliverable in a small developers team context. Developers are widely spread around the world working collaboratively and without hierarchy. This kind of working organization context en-lighted the need of a federated integration test and validation process.

  4. STATISTICS. The reusable holdout: Preserving validity in adaptive data analysis.

    Science.gov (United States)

    Dwork, Cynthia; Feldman, Vitaly; Hardt, Moritz; Pitassi, Toniann; Reingold, Omer; Roth, Aaron

    2015-08-07

    Misapplication of statistical data analysis is a common cause of spurious discoveries in scientific research. Existing approaches to ensuring the validity of inferences drawn from data assume a fixed procedure to be performed, selected before the data are examined. In common practice, however, data analysis is an intrinsically adaptive process, with new analyses generated on the basis of data exploration, as well as the results of previous analyses on the same data. We demonstrate a new approach for addressing the challenges of adaptivity based on insights from privacy-preserving data analysis. As an application, we show how to safely reuse a holdout data set many times to validate the results of adaptively chosen analyses. Copyright © 2015, American Association for the Advancement of Science.

  5. Using linked electronic data to validate algorithms for health outcomes in administrative databases.

    Science.gov (United States)

    Lee, Wan-Ju; Lee, Todd A; Pickard, Alan Simon; Shoaibi, Azadeh; Schumock, Glen T

    2015-08-01

    The validity of algorithms used to identify health outcomes in claims-based and administrative data is critical to the reliability of findings from observational studies. The traditional approach to algorithm validation, using medical charts, is expensive and time-consuming. An alternative method is to link the claims data to an external, electronic data source that contains information allowing confirmation of the event of interest. In this paper, we describe this external linkage validation method and delineate important considerations to assess the feasibility and appropriateness of validating health outcomes using this approach. This framework can help investigators decide whether to pursue an external linkage validation method for identifying health outcomes in administrative/claims data.

  6. Fire Intensity Data for Validation of the Radiative Transfer Equation

    Energy Technology Data Exchange (ETDEWEB)

    Blanchat, Thomas K. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Jernigan, Dann A. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2016-01-01

    A set of experiments and test data are outlined in this report that provides radiation intensity data for the validation of models for the radiative transfer equation. The experiments were performed with lightly-sooting liquid hydrocarbon fuels that yielded fully turbulent fires 2 m diameter). In addition, supplemental measurements of air flow and temperature, fuel temperature and burn rate, and flame surface emissive power, wall heat, and flame height and width provide a complete set of boundary condition data needed for validation of models used in fire simulations.

  7. A new generation of crystallographic validation tools for the protein data bank.

    Science.gov (United States)

    Read, Randy J; Adams, Paul D; Arendall, W Bryan; Brunger, Axel T; Emsley, Paul; Joosten, Robbie P; Kleywegt, Gerard J; Krissinel, Eugene B; Lütteke, Thomas; Otwinowski, Zbyszek; Perrakis, Anastassis; Richardson, Jane S; Sheffler, William H; Smith, Janet L; Tickle, Ian J; Vriend, Gert; Zwart, Peter H

    2011-10-12

    This report presents the conclusions of the X-ray Validation Task Force of the worldwide Protein Data Bank (PDB). The PDB has expanded massively since current criteria for validation of deposited structures were adopted, allowing a much more sophisticated understanding of all the components of macromolecular crystals. The size of the PDB creates new opportunities to validate structures by comparison with the existing database, and the now-mandatory deposition of structure factors creates new opportunities to validate the underlying diffraction data. These developments highlighted the need for a new assessment of validation criteria. The Task Force recommends that a small set of validation data be presented in an easily understood format, relative to both the full PDB and the applicable resolution class, with greater detail available to interested users. Most importantly, we recommend that referees and editors judging the quality of structural experiments have access to a concise summary of well-established quality indicators. Copyright © 2011 Elsevier Ltd. All rights reserved.

  8. NPOESS Preparatory Project Validation Program for Atmsophere Data Products from VIIRS

    Science.gov (United States)

    Starr, D.; Wong, E.

    2009-12-01

    The National Polar-orbiting Operational Environmental Satellite Suite (NPOESS) Program, in partnership with National Aeronautical Space Administration (NASA), will launch the NPOESS Preparatory Project (NPP), a risk reduction and data continuity mission, prior to the first operational NPOESS launch. The NPOESS Program, in partnership with Northrop Grumman Aerospace Systems (NGAS), will execute the NPP Validation program to ensure the data products comply with the requirements of the sponsoring agencies. Data from the NPP Visible/Infrared Imager/Radiometer Suite (VIIRS) will be used to produce Environmental Data Records (EDR's) for aerosol and clouds, specifically Aerosol Optical Thickness (AOT), Aerosol Particle Size Parameter (APSP), and Suspended Matter (SM); and Cloud Optical Thickness (COT), Cloud Effective Particle Size (CEPS), Cloud Top Temperature (CTT), Height (CTH) and Pressure (CTP), and Cloud Base Height (CBH). The Aerosol and Cloud EDR Validation Program is a multifaceted effort to characterize and validate these data products. The program involves systematic comparison to heritage data products, e.g., MODIS, and ground-based correlative data, such as AERONET and ARM data products, and potentially airborne field measurements. To the extent possible, the domain is global. The program leverages various investments that have and are continuing to be made by national funding agencies in such resources, as well as the operational user community and the broad Earth science user community. This presentation will provide an overview of the approaches, data and schedule for the validation of the NPP VIIRS Aerosol and Cloud environmental data products.

  9. Validation Techniques for Sensor Data in Mobile Health Applications

    Directory of Open Access Journals (Sweden)

    Ivan Miguel Pires

    2016-01-01

    Full Text Available Mobile applications have become a must in every user’s smart device, and many of these applications make use of the device sensors’ to achieve its goal. Nevertheless, it remains fairly unknown to the user to which extent the data the applications use can be relied upon and, therefore, to which extent the output of a given application is trustworthy or not. To help developers and researchers and to provide a common ground of data validation algorithms and techniques, this paper presents a review of the most commonly used data validation algorithms, along with its usage scenarios, and proposes a classification for these algorithms. This paper also discusses the process of achieving statistical significance and trust for the desired output.

  10. Validation of CERBRRS against Gentilly-2 station data

    International Nuclear Information System (INIS)

    Varin, E.; Mao, J.; Kaveh, S.; Baudouin, A.

    2008-01-01

    CERBRRS models the CANDU 6 reactor regulating system (RRS) in CERBERUS transient core calculations. The validation of CERBRRS described here is performed against Gentilly-2 reactor data. The present validation focuses on the liquid zone controller (LZC) modelling, specifically the LZC fill rates. Two core transient measurements were compared to CERBRRS results: a shutdown system manual trip and the 1995 adjuster bank test. The standard CERBRRS model is found to produce a slower fill rate than observed at G2 in the SDS1 trip test. A modified LZC model was developed to reproduce the SDS1 trip test results, and compared to the 1995 adjuster bank test data. The CERBRRS results compared to the measurements show good agreement of the average LZC fill level, which is a key parameter in the RRS logic. These comparisons against Gentilly-2 station data also provide additional confidence in the overall performance of the CERBRRS. (author)

  11. Multimission satellite altimetric data validation in the Baltic Sea

    DEFF Research Database (Denmark)

    Cheng, Yongcun; Andersen, Ole Baltazar; Knudsen, Per

    2014-01-01

    The assessment of altimetric data is crucial for investigating the regional sea level variability. Few works has been performed to validate the altimetric data [1, 2] in the Baltic Sea. The exploring of multi-mission altimetric data in the Baltic Sea has yet to be published. The number of available...

  12. Citizen science networks in natural history and the collective validation of biodiversity data.

    Science.gov (United States)

    Turnhout, Esther; Lawrence, Anna; Turnhout, Sander

    2016-06-01

    Biodiversity data are in increasing demand to inform policy and management. A substantial portion of these data is generated in citizen science networks. To ensure the quality of biodiversity data, standards and criteria for validation have been put in place. We used interviews and document analysis from the United Kingdom and The Netherlands to examine how data validation serves as a point of connection between the diverse people and practices in natural history citizen science networks. We found that rather than a unidirectional imposition of standards, validation was performed collectively. Specifically, it was enacted in ongoing circulations of biodiversity records between recorders and validators as they jointly negotiated the biodiversity that was observed and the validity of the records. These collective validation practices contributed to the citizen science character or natural history networks and tied these networks together. However, when biodiversity records were included in biodiversity-information initiatives on different policy levels and scales, the circulation of records diminished. These initiatives took on a more extractive mode of data use. Validation ceased to be collective with important consequences for the natural history networks involved and citizen science more generally. © 2016 The Authors. Conservation Biology published by Wiley Periodicals, Inc. on behalf of Society for Conservation Biology.

  13. Mining Twitter Data Stream to Augment NASA GPM Validation

    Science.gov (United States)

    Teng, W. L.; Albayrak, A.; Huffman, G. J.; Vollmer, B.

    2017-12-01

    The Twitter data stream is an important new source of real-time and historical global information for potentially augmenting the validation program of NASA's Global Precipitation Measurement (GPM) mission. There have been other similar uses of Twitter, though mostly related to natural hazards monitoring and management. The validation of satellite precipitation estimates is challenging, because many regions lack data or access to data, especially outside of the U.S. and in remote and developing areas. The time-varying set of "precipitation" tweets can be thought of as an organic network of rain gauges, potentially providing a widespread view of precipitation occurrence. Twitter provides a large source of crowd for crowdsourcing. During a 24-hour period in the middle of the snow storm this past March in the U.S. Northeast, we collected more than 13,000 relevant precipitation tweets with exact geolocation. The overall objective of our project is to determine the extent to which processed tweets can provide additional information that improves the validation of GPM data. Though our current effort focuses on tweets and precipitation, our approach is general and applicable to other social media and other geophysical measurements. Specifically, we have developed an operational infrastructure for processing tweets, in a format suitable for analysis with GPM data; engaged with potential participants, both passive and active, to "enrich" the Twitter stream; and inter-compared "precipitation" tweet data, ground station data, and GPM retrievals. In this presentation, we detail the technical capabilities of our tweet processing infrastructure, including data abstraction, feature extraction, search engine, context-awareness, real-time processing, and high volume (big) data processing; various means for "enriching" the Twitter stream; and results of inter-comparisons. Our project should bring a new kind of visibility to Twitter and engender a new kind of appreciation of the value

  14. Learning from biomedical linked data to suggest valid pharmacogenes.

    Science.gov (United States)

    Dalleau, Kevin; Marzougui, Yassine; Da Silva, Sébastien; Ringot, Patrice; Ndiaye, Ndeye Coumba; Coulet, Adrien

    2017-04-20

    A standard task in pharmacogenomics research is identifying genes that may be involved in drug response variability, i.e., pharmacogenes. Because genomic experiments tended to generate many false positives, computational approaches based on the use of background knowledge have been proposed. Until now, only molecular networks or the biomedical literature were used, whereas many other resources are available. We propose here to consume a diverse and larger set of resources using linked data related either to genes, drugs or diseases. One of the advantages of linked data is that they are built on a standard framework that facilitates the joint use of various sources, and thus facilitates considering features of various origins. We propose a selection and linkage of data sources relevant to pharmacogenomics, including for example DisGeNET and Clinvar. We use machine learning to identify and prioritize pharmacogenes that are the most probably valid, considering the selected linked data. This identification relies on the classification of gene-drug pairs as either pharmacogenomically associated or not and was experimented with two machine learning methods -random forest and graph kernel-, which results are compared in this article. We assembled a set of linked data relative to pharmacogenomics, of 2,610,793 triples, coming from six distinct resources. Learning from these data, random forest enables identifying valid pharmacogenes with a F-measure of 0.73, on a 10 folds cross-validation, whereas graph kernel achieves a F-measure of 0.81. A list of top candidates proposed by both approaches is provided and their obtention is discussed.

  15. Indirect Validation of Probe Speed Data on Arterial Corridors

    Energy Technology Data Exchange (ETDEWEB)

    Eshragh, Sepideh [Center for Advanced Transportation Technology, University of Maryland, College Park, 5000 College Avenue, Suite 2206, College Park, MD 20742; Young, Stanley E. [National Renewable Energy Laboratory, 15013 Denver West Parkway, Golden, CO 80401; Sharifi, Elham [Center for Advanced Transportation Technology, University of Maryland, College Park, 5000 College Avenue, Suite 2206, College Park, MD 20742; Hamedi, Masoud [Center for Advanced Transportation Technology, University of Maryland, College Park, 5000 College Avenue, Suite 2206, College Park, MD 20742; Sadabadi, Kaveh Farokhi [Center for Advanced Transportation Technology, University of Maryland, College Park, 5000 College Avenue, Suite 2206, College Park, MD 20742

    2017-01-01

    This study aimed to estimate the accuracy of probe speed data on arterial corridors on the basis of roadway geometric attributes and functional classification. It was assumed that functional class (medium and low) along with other road characteristics (such as weighted average of the annual average daily traffic, average signal density, average access point density, and average speed) were available as correlation factors to estimate the accuracy of probe traffic data. This study tested these factors as predictors of the fidelity of probe traffic data by using the results of an extensive validation exercise. This study showed strong correlations between these geometric attributes and the accuracy of probe data when they were assessed by using average absolute speed error. Linear models were regressed to existing data to estimate appropriate models for medium- and low-type arterial corridors. The proposed models for medium- and low-type arterials were validated further on the basis of the results of a slowdown analysis. These models can be used to predict the accuracy of probe data indirectly in medium and low types of arterial corridors.

  16. Using 'big data' to validate claims made in the pharmaceutical approval process.

    Science.gov (United States)

    Wasser, Thomas; Haynes, Kevin; Barron, John; Cziraky, Mark

    2015-01-01

    Big Data in the healthcare setting refers to the storage, assimilation, and analysis of large quantities of information regarding patient care. These data can be collected and stored in a wide variety of ways including electronic medical records collected at the patient bedside, or through medical records that are coded and passed to insurance companies for reimbursement. When these data are processed it is possible to validate claims as a part of the regulatory review process regarding the anticipated performance of medications and devices. In order to analyze properly claims by manufacturers and others, there is a need to express claims in terms that are testable in a timeframe that is useful and meaningful to formulary committees. Claims for the comparative benefits and costs, including budget impact, of products and devices need to be expressed in measurable terms, ideally in the context of submission or validation protocols. Claims should be either consistent with accessible Big Data or able to support observational studies where Big Data identifies target populations. Protocols should identify, in disaggregated terms, key variables that would lead to direct or proxy validation. Once these variables are identified, Big Data can be used to query massive quantities of data in the validation process. Research can be passive or active in nature. Passive, where the data are collected retrospectively; active where the researcher is prospectively looking for indicators of co-morbid conditions, side-effects or adverse events, testing these indicators to determine if claims are within desired ranges set forth by the manufacturer. Additionally, Big Data can be used to assess the effectiveness of therapy through health insurance records. This, for example, could indicate that disease or co-morbid conditions cease to be treated. Understanding the basic strengths and weaknesses of Big Data in the claim validation process provides a glimpse of the value that this research

  17. Validation of elk resource selection models with spatially independent data

    Science.gov (United States)

    Priscilla K. Coe; Bruce K. Johnson; Michael J. Wisdom; John G. Cook; Marty Vavra; Ryan M. Nielson

    2011-01-01

    Knowledge of how landscape features affect wildlife resource use is essential for informed management. Resource selection functions often are used to make and validate predictions about landscape use; however, resource selection functions are rarely validated with data from landscapes independent of those from which the models were built. This problem has severely...

  18. Seismic Data Gathering and Validation

    Energy Technology Data Exchange (ETDEWEB)

    Coleman, Justin [Idaho National Lab. (INL), Idaho Falls, ID (United States)

    2015-02-01

    Three recent earthquakes in the last seven years have exceeded their design basis earthquake values (so it is implied that damage to SSC’s should have occurred). These seismic events were recorded at North Anna (August 2011, detailed information provided in [Virginia Electric and Power Company Memo]), Fukushima Daichii and Daini (March 2011 [TEPCO 1]), and Kaswazaki-Kariwa (2007, [TEPCO 2]). However, seismic walk downs at some of these plants indicate that very little damage occurred to safety class systems and components due to the seismic motion. This report presents seismic data gathered for two of the three events mentioned above and recommends a path for using that data for two purposes. One purpose is to determine what margins exist in current industry standard seismic soil-structure interaction (SSI) tools. The second purpose is the use the data to validated seismic site response tools and SSI tools. The gathered data represents free field soil and in-structure acceleration time histories data. Gathered data also includes elastic and dynamic soil properties and structural drawings. Gathering data and comparing with existing models has potential to identify areas of uncertainty that should be removed from current seismic analysis and SPRA approaches. Removing uncertainty (to the extent possible) from SPRA’s will allow NPP owners to make decisions on where to reduce risk. Once a realistic understanding of seismic response is established for a nuclear power plant (NPP) then decisions on needed protective measures, such as SI, can be made.

  19. Validating data quality during wet weather monitoring of wastewater treatment plant influents

    DEFF Research Database (Denmark)

    Alferes, Janelcy; Lynggaard-Jensen, Anders; Munk-Nielsen, Thomas

    Efficient monitoring of water systems and proper use of the collected data in further applications such as modelling, forecasting influent water quality and real-time control depends on careful data quality control. Given the size of the data sets produced nowadays in online water quality...... monitoring schemes, automated data validation is the only feasible option. In this paper, software tools for automatic data quality assessment with a practical orientation are presented. The developments from three organizations ranging from simple to more complex methods for automated data validation...

  20. GPM Ground Validation Navigation Data ER-2 OLYMPEX V1

    Data.gov (United States)

    National Aeronautics and Space Administration — The GPM Ground Validation NASA ER-2 Navigation Data OLYMPEX dataset supplies navigation data collected by the NASA ER-2 aircraft for flights that occurred during...

  1. In-Flight Validation of Mid and Thermal Infrared Remotely Sensed Data Using the Lake Tahoe and Salton Sea Automated Validation Sites

    Science.gov (United States)

    Hook, Simon J.

    2008-01-01

    The presentation includes an introduction, Lake Tahoe site layout and measurements, Salton Sea site layout and measurements, field instrument calibration and cross-calculations, data reduction methodology and error budgets, and example results for MODIS. Summary and conclusions are: 1) Lake Tahoe CA/NV automated validation site was established in 1999 to assess radiometric accuracy of satellite and airborne mid and thermal infrared data and products. Water surface temperatures range from 4-25C.2) Salton Sea CA automated validation site was established in 2008 to broaden range of available water surface temperatures and atmospheric water vapor test cases. Water surface temperatures range from 15-35C. 3) Sites provide all information necessary for validation every 2 mins (bulk temperature, skin temperature, air temperature, wind speed, wind direction, net radiation, relative humidity). 4) Sites have been used to validate mid and thermal infrared data and products from: ASTER, AATSR, ATSR2, MODIS-Terra, MODIS-Aqua, Landsat 5, Landsat 7, MTI, TES, MASTER, MAS. 5) Approximately 10 years of data available to help validate AVHRR.

  2. Validating firn compaction model with remote sensing data

    DEFF Research Database (Denmark)

    Simonsen, S. B.; Stenseng, Lars; Sørensen, Louise Sandberg

    A comprehensive understanding of firn processes is of outmost importance, when estimating present and future changes of the Greenland Ice Sheet. Especially, when remote sensing altimetry is used to assess the state of ice sheets and their contribution to global sea level rise, firn compaction...... models have been shown to be a key component. Now, remote sensing data can also be used to validate the firn models. Radar penetrating the upper part of the firn column in the interior part of Greenland shows a clear layering. The observed layers from the radar data can be used as an in-situ validation...... correction relative to the changes in the elevation of the surface observed with remote sensing altimetry? What model time resolution is necessary to resolved the observed layering? What model refinements are necessary to give better estimates of the surface mass balance of the Greenland ice sheet from...

  3. Validation of a case definition to define hypertension using administrative data.

    Science.gov (United States)

    Quan, Hude; Khan, Nadia; Hemmelgarn, Brenda R; Tu, Karen; Chen, Guanmin; Campbell, Norm; Hill, Michael D; Ghali, William A; McAlister, Finlay A

    2009-12-01

    We validated the accuracy of case definitions for hypertension derived from administrative data across time periods (year 2001 versus 2004) and geographic regions using physician charts. Physician charts were randomly selected in rural and urban areas from Alberta and British Columbia, Canada, during years 2001 and 2004. Physician charts were linked with administrative data through unique personal health number. We reviewed charts of approximately 50 randomly selected patients >35 years of age from each clinic within 48 urban and 16 rural family physician clinics to identify physician diagnoses of hypertension during the years 2001 and 2004. The validity indices were estimated for diagnosed hypertension using 3 years of administrative data for the 8 case-definition combinations. Of the 3,362 patient charts reviewed, the prevalence of hypertension ranged from 18.8% to 33.3%, depending on the year and region studied. The administrative data hypertension definition of "2 claims within 2 years or 1 hospitalization" had the highest validity relative to the other definitions evaluated (sensitivity 75%, specificity 94%, positive predictive value 81%, negative predictive value 92%, and kappa 0.71). After adjustment for age, sex, and comorbid conditions, the sensitivities between regions, years, and provinces were not significantly different, but the positive predictive value varied slightly across geographic regions. These results provide evidence that administrative data can be used as a relatively valid source of data to define cases of hypertension for surveillance and research purposes.

  4. 78 FR 77718 - Comment Request for Information Collection for Information Collection for the Data Validation...

    Science.gov (United States)

    2013-12-24

    ... Collection for Information Collection for the Data Validation Requirement for Employment and Training... collection of data validation information for the following employment and training programs: Workforce... information on program activities and outcomes is available. Data validation is intended to accomplish the...

  5. Towards Seamless Validation of Land Cover Data

    Science.gov (United States)

    Chuprikova, Ekaterina; Liebel, Lukas; Meng, Liqiu

    2018-05-01

    This article demonstrates the ability of the Bayesian Network analysis for the recognition of uncertainty patterns associated with the fusion of various land cover data sets including GlobeLand30, CORINE (CLC2006, Germany) and land cover data derived from Volunteered Geographic Information (VGI) such as Open Street Map (OSM). The results of recognition are expressed as probability and uncertainty maps which can be regarded as a by-product of the GlobeLand30 data. The uncertainty information may guide the quality improvement of GlobeLand30 by involving the ground truth data, information with superior quality, the know-how of experts and the crowd intelligence. Such an endeavor aims to pave a way towards a seamless validation of global land cover data on the one hand and a targeted knowledge discovery in areas with higher uncertainty values on the other hand.

  6. Validation of the TIARA code to tritium inventory data

    International Nuclear Information System (INIS)

    Billone, M.C.

    1994-03-01

    The TIARA code has been developed to predict tritium inventory in Li 2 O breeder ceramic and to predict purge exit flow rate and composition. Inventory predictions are based on models for bulk diffusion, surface desorption, solubility and precipitation. Parameters for these models are determined from the results of laboratory annealing studies on unirradiated and irradiated Li 2 O. Inventory data from in-reactor purge flow tests are used for model improvement, fine-tuning of model parameters and validation. In this current work, the inventory measurement near the purge inlet from the BEATRIX-II thin-ring sample is used to fine tune the surface desorption model parameters for T > 470 degrees C, and the inventory measurement near the midplane from VOM-15H is used to fine tune the moisture solubility model parameters. predictions are then validated to the remaining inventory data from EXOTIC-2 (1 point), SIBELIUS (3 axial points), VOM-15H (2 axial points), CRITIC-1 (4 axial points), BEATRIX-II thin ring (3 axial points) and BEATRIX-II thick pellet (5 radial points). Thus. of the 20 data points, two we re used for fine tuning model parameters and 18 were used for validation. The inventory data span the range of 0.05--1.44 wppm with an average of 0.48 wppm. The data pertain to samples whose end-of-life temperatures were in the range of 490--1000 degrees C. On the average, the TIARA predictions agree quite well with the data (< 0.02 wppm difference). However, the root-mean-square deviation is 0.44 wppm, mostly due to over-predictions for the SIBELIUS samples and the higher-temperature radial samples from the BEATRIX-11 thick-pellet

  7. A validation system for data preservation in HEP

    International Nuclear Information System (INIS)

    Kemp, Yves; Strutz, Marco; Hessling, Hermann

    2012-01-01

    Preserving data from past experiments and preserving the ability to perform analysis with old data is of growing importance in many domains of science, including High Energy Physics (HEP). A study group on this issue, DPHEP, has been established in this field to provide guidelines and a structure for international collaboration on data preservation projects in HEP. This contribution presents a framework that allows experimentalists to validate their software against a previously defined set of tests in an automated way. The framework has been designed with a special focus for longevity, as it makes use of open protocols, has a modular design and is based on simple communication mechanisms. On the fabrics side, tests are carried out in a virtual environment using a cloud infrastructure. Within the framework, it is easy to run validation tests on different hardware platforms, or different major or minor versions of operating systems. Experts from IT or the experiments can automatically detect failures in the test procedure by the help of reporting tools. Hence, appropriate actions can be taken in a timely manner. The design and important implementation aspects of the framework are shown and first experiences from early-bird users will be presented.

  8. Validation and augmentation of Inrix arterial travel time data using independent sources : [research summary].

    Science.gov (United States)

    2015-02-01

    Although the freeway travel time data has been validated extensively in recent : years, the quality of arterial travel time data is not well known. This project : presents a comprehensive validation scheme for arterial travel time data based : on GPS...

  9. Text data extraction for a prospective, research-focused data mart: implementation and validation.

    Science.gov (United States)

    Hinchcliff, Monique; Just, Eric; Podlusky, Sofia; Varga, John; Chang, Rowland W; Kibbe, Warren A

    2012-09-13

    Translational research typically requires data abstracted from medical records as well as data collected specifically for research. Unfortunately, many data within electronic health records are represented as text that is not amenable to aggregation for analyses. We present a scalable open source SQL Server Integration Services package, called Regextractor, for including regular expression parsers into a classic extract, transform, and load workflow. We have used Regextractor to abstract discrete data from textual reports from a number of 'machine generated' sources. To validate this package, we created a pulmonary function test data mart and analyzed the quality of the data mart versus manual chart review. Eleven variables from pulmonary function tests performed closest to the initial clinical evaluation date were studied for 100 randomly selected subjects with scleroderma. One research assistant manually reviewed, abstracted, and entered relevant data into a database. Correlation with data obtained from the automated pulmonary function test data mart within the Northwestern Medical Enterprise Data Warehouse was determined. There was a near perfect (99.5%) agreement between results generated from the Regextractor package and those obtained via manual chart abstraction. The pulmonary function test data mart has been used subsequently to monitor disease progression of patients in the Northwestern Scleroderma Registry. In addition to the pulmonary function test example presented in this manuscript, the Regextractor package has been used to create cardiac catheterization and echocardiography data marts. The Regextractor package was released as open source software in October 2009 and has been downloaded 552 times as of 6/1/2012. Collaboration between clinical researchers and biomedical informatics experts enabled the development and validation of a tool (Regextractor) to parse, abstract and assemble structured data from text data contained in the electronic health

  10. Process data validation according VDI 2048 in conventional and nuclear power plants

    International Nuclear Information System (INIS)

    Langenstein, M.; Laipple, B.; Schmid, F.

    2004-01-01

    Process data validation according VDI 2048 in conventional and nuclear power plants is required for acceptance testing, process and component monitoring, and status-oriented maintenance. Once a validation system like VALI III has been certified according to VDI 2048, power plant owners can use the data obtained for efficiency increase. Further, all control variables can be adjusted so as to ensure maximum plant efficiency. (orig.)

  11. Do coder characteristics influence validity of ICD-10 hospital discharge data?

    Directory of Open Access Journals (Sweden)

    Beck Cynthia A

    2010-04-01

    Full Text Available Abstract Background Administrative data are widely used to study health systems and make important health policy decisions. Yet little is known about the influence of coder characteristics on administrative data validity in these studies. Our goal was to describe the relationship between several measures of validity in coded hospital discharge data and 1 coders' volume of coding (≥13,000 vs. Methods This descriptive study examined 6 indicators of face validity in ICD-10 coded discharge records from 4 hospitals in Calgary, Canada between April 2002 and March 2007. Specifically, mean number of coded diagnoses, procedures, complications, Z-codes, and codes ending in 8 or 9 were compared by coding volume and employment status, as well as hospital type. The mean number of diagnoses was also compared across coder characteristics for 6 major conditions of varying complexity. Next, kappa statistics were computed to assess agreement between discharge data and linked chart data reabstracted by nursing chart reviewers. Kappas were compared across coder characteristics. Results 422,618 discharge records were coded by 59 coders during the study period. The mean number of diagnoses per record decreased from 5.2 in 2002/2003 to 3.9 in 2006/2007, while the number of records coded annually increased from 69,613 to 102,842. Coders at the tertiary hospital coded the most diagnoses (5.0 compared with 3.9 and 3.8 at other sites. There was no variation by coder or site characteristics for any other face validity indicator. The mean number of diagnoses increased from 1.5 to 7.9 with increasing complexity of the major diagnosis, but did not vary with coder characteristics. Agreement (kappa between coded data and chart review did not show any consistent pattern with respect to coder characteristics. Conclusions This large study suggests that coder characteristics do not influence the validity of hospital discharge data. Other jurisdictions might benefit from

  12. EASY-2005: Validation and new tools for data checking

    International Nuclear Information System (INIS)

    Forrest, R.A.; Kopecky, J.

    2007-01-01

    The European Activation System (EASY) has been developed for activation calculations within the Fusion Technology programme. It comprises the EAF nuclear data libraries and the FISPACT inventory code. EASY-2005 contains about five times more neutron-induced reactions (62,637) than the previous version; due to an increase in the upper energy limit from 20 to 60 MeV. Many measurements of activation in well-defined neutron spectra have been made; these integral results are used to validate EAF. Details of the validation covering 453 reactions are given. Model calculations were used to produce the majority of the data, and the large number of reactions, and the lack of experimental data, make it necessary to develop new checking tools. Within the SAFEPAQ-II code, the new methodology of statistical analysis of cross sections (SACS) has been developed. This enables various quantities such as maximum cross section (σ max ) to be plotted as a function of asymmetry parameter (s) for each reaction type. Such plots show well-defined trends and inconsistent data for a particular reaction can readily be identified. The use of integral data and SACS for improvement are discussed, in relation to the maintenance version, EASY-2005.1 and the planned new version, EASY-2007

  13. Spatial and Semantic Validation of Secondary Food Source Data

    DEFF Research Database (Denmark)

    Lyseen, Anders Knørr; Hansen, Henning Sten

    2014-01-01

    Governmental and commercial lists of food retailers are often used to measure food environments and foodscapes for health and nutritional research. Information about the validity of such secondary food source data is relevant to understanding the potential and limitations of its application....... This study assesses the validity of two government lists of food retailer locations and types by comparing them to direct field observations, including an assessment of whether pre-classification of the directories can reduce the need for field observation. Lists of food retailers were obtained from......-classification was measured through the calculation of PPV, sensitivity and negative prediction value (NPV). The application of either CVR or Smiley as a measure of the food environment would result in a misrepresentation. The pre-classification based on the food retailer names was found to be a valid method for identifying...

  14. Sample size determination for disease prevalence studies with partially validated data.

    Science.gov (United States)

    Qiu, Shi-Fang; Poon, Wai-Yin; Tang, Man-Lai

    2016-02-01

    Disease prevalence is an important topic in medical research, and its study is based on data that are obtained by classifying subjects according to whether a disease has been contracted. Classification can be conducted with high-cost gold standard tests or low-cost screening tests, but the latter are subject to the misclassification of subjects. As a compromise between the two, many research studies use partially validated datasets in which all data points are classified by fallible tests, and some of the data points are validated in the sense that they are also classified by the completely accurate gold-standard test. In this article, we investigate the determination of sample sizes for disease prevalence studies with partially validated data. We use two approaches. The first is to find sample sizes that can achieve a pre-specified power of a statistical test at a chosen significance level, and the second is to find sample sizes that can control the width of a confidence interval with a pre-specified confidence level. Empirical studies have been conducted to demonstrate the performance of various testing procedures with the proposed sample sizes. The applicability of the proposed methods are illustrated by a real-data example. © The Author(s) 2012.

  15. Model Validation and Verification of Data Mining from the ...

    African Journals Online (AJOL)

    Michael Horsfall

    In this paper, we seek to present a hybrid method for Model Validation and Verification of Data Mining from the ... This model generally states the numerical value of knowledge .... procedures found in the field of software engineering should be ...

  16. When Assessment Data Are Words: Validity Evidence for Qualitative Educational Assessments.

    Science.gov (United States)

    Cook, David A; Kuper, Ayelet; Hatala, Rose; Ginsburg, Shiphra

    2016-10-01

    Quantitative scores fail to capture all important features of learner performance. This awareness has led to increased use of qualitative data when assessing health professionals. Yet the use of qualitative assessments is hampered by incomplete understanding of their role in forming judgments, and lack of consensus in how to appraise the rigor of judgments therein derived. The authors articulate the role of qualitative assessment as part of a comprehensive program of assessment, and translate the concept of validity to apply to judgments arising from qualitative assessments. They first identify standards for rigor in qualitative research, and then use two contemporary assessment validity frameworks to reorganize these standards for application to qualitative assessment.Standards for rigor in qualitative research include responsiveness, reflexivity, purposive sampling, thick description, triangulation, transparency, and transferability. These standards can be reframed using Messick's five sources of validity evidence (content, response process, internal structure, relationships with other variables, and consequences) and Kane's four inferences in validation (scoring, generalization, extrapolation, and implications). Evidence can be collected and evaluated for each evidence source or inference. The authors illustrate this approach using published research on learning portfolios.The authors advocate a "methods-neutral" approach to assessment, in which a clearly stated purpose determines the nature of and approach to data collection and analysis. Increased use of qualitative assessments will necessitate more rigorous judgments of the defensibility (validity) of inferences and decisions. Evidence should be strategically sought to inform a coherent validity argument.

  17. Text data extraction for a prospective, research-focused data mart: implementation and validation

    Directory of Open Access Journals (Sweden)

    Hinchcliff Monique

    2012-09-01

    Full Text Available Abstract Background Translational research typically requires data abstracted from medical records as well as data collected specifically for research. Unfortunately, many data within electronic health records are represented as text that is not amenable to aggregation for analyses. We present a scalable open source SQL Server Integration Services package, called Regextractor, for including regular expression parsers into a classic extract, transform, and load workflow. We have used Regextractor to abstract discrete data from textual reports from a number of ‘machine generated’ sources. To validate this package, we created a pulmonary function test data mart and analyzed the quality of the data mart versus manual chart review. Methods Eleven variables from pulmonary function tests performed closest to the initial clinical evaluation date were studied for 100 randomly selected subjects with scleroderma. One research assistant manually reviewed, abstracted, and entered relevant data into a database. Correlation with data obtained from the automated pulmonary function test data mart within the Northwestern Medical Enterprise Data Warehouse was determined. Results There was a near perfect (99.5% agreement between results generated from the Regextractor package and those obtained via manual chart abstraction. The pulmonary function test data mart has been used subsequently to monitor disease progression of patients in the Northwestern Scleroderma Registry. In addition to the pulmonary function test example presented in this manuscript, the Regextractor package has been used to create cardiac catheterization and echocardiography data marts. The Regextractor package was released as open source software in October 2009 and has been downloaded 552 times as of 6/1/2012. Conclusions Collaboration between clinical researchers and biomedical informatics experts enabled the development and validation of a tool (Regextractor to parse, abstract and assemble

  18. Good validity of the international spinal cord injury quality of life basic data set

    DEFF Research Database (Denmark)

    Post, M W M; Adriaansen, J J E; Charlifue, S

    2016-01-01

    STUDY DESIGN: Cross-sectional validation study. OBJECTIVES: To examine the construct and concurrent validity of the International Spinal Cord Injury (SCI) Quality of Life (QoL) Basic Data Set. SETTING: Dutch community. PARTICIPANTS: People 28-65 years of age, who obtained their SCI between 18...... and 35 years of age, were at least 10 years post SCI and were wheelchair users in daily life.Measure(s):The International SCI QoL Basic Data Set consists of three single items on satisfaction with life as a whole, physical health and psychological health (0=complete dissatisfaction; 10=complete...... and psychological health (0.70). CONCLUSIONS: This first validity study of the International SCI QoL Basic Data Set shows that it appears valid for persons with SCI....

  19. 48 CFR 252.227-7037 - Validation of restrictive markings on technical data.

    Science.gov (United States)

    2010-10-01

    .... However, this clause neither creates nor implies privity of contract between the Government and... AND CONTRACT CLAUSES Text of Provisions And Clauses 252.227-7037 Validation of restrictive markings on... following clause: Validation of Restrictive Markings on Technical Data (SEP 1999) (a) Definitions. The terms...

  20. EASY-2005: validation and new tools for data checking

    International Nuclear Information System (INIS)

    Forrest, R.; Kopecky, J.

    2006-01-01

    The European Activation System (EASY) is the tool developed for activation calculations as part of the Fusion Technology programme. It comprises the EAF nuclear data libraries and the FISPACT inventory code. The version EASY-2005, released last year contains many more neutron-induced reactions than the previous version; an increase of about a factor of five to 62,637. The reason for this was the increase in the upper energy limit from 20 to 60 MeV. A major strand of work within Europe has been the measurement of activation in a series of well-defined neutron spectra, these integral results can then be used to validate a subset of the EAF-2005 library, demonstrating that it can be relied on for activation predictions in similar spectra. A validation report has been produced covering 453 reactions; this shows the integral results and the available differential data from EXFOR for these reactions. For 202 reactions there is good agreement, indicating that the data for the reaction are adequate. In cases of disagreement the report indicates where additional measurements are required or where the library data can be improved. The majority of the data above 20 MeV were obtained from model calculations, and this is also the case for many of the reactions at all energies. Because of the large number of reactions and the lack of experimental data, there was a need to develop new tools that could help in the checking of such a large library. Within the SAFEPAQ-II code that is used for evaluation and production of the EAF libraries, the new methodology of Statistical Analysis of Cross Sections (SACS) has been developed. This enables various quantities such as maximum cross section (σ max ), energy at maximum (E max ) and width at half maximum (W 0.5 ) to be plotted as functions of atomic mass (A) and asymmetry parameter (s) for each type of reaction. Such plots show well defined trends and inconsistent data for a particular reaction can easily be identified. Using the

  1. Assessing the validity of commercial and municipal food environment data sets in Vancouver, Canada.

    Science.gov (United States)

    Daepp, Madeleine Ig; Black, Jennifer

    2017-10-01

    The present study assessed systematic bias and the effects of data set error on the validity of food environment measures in two municipal and two commercial secondary data sets. Sensitivity, positive predictive value (PPV) and concordance were calculated by comparing two municipal and two commercial secondary data sets with ground-truthed data collected within 800 m buffers surrounding twenty-six schools. Logistic regression examined associations of sensitivity and PPV with commercial density and neighbourhood socio-economic deprivation. Kendall's τ estimated correlations between density and proximity of food outlets near schools constructed with secondary data sets v. ground-truthed data. Vancouver, Canada. Food retailers located within 800 m of twenty-six schools RESULTS: All data sets scored relatively poorly across validity measures, although, overall, municipal data sets had higher levels of validity than did commercial data sets. Food outlets were more likely to be missing from municipal health inspections lists and commercial data sets in neighbourhoods with higher commercial density. Still, both proximity and density measures constructed from all secondary data sets were highly correlated (Kendall's τ>0·70) with measures constructed from ground-truthed data. Despite relatively low levels of validity in all secondary data sets examined, food environment measures constructed from secondary data sets remained highly correlated with ground-truthed data. Findings suggest that secondary data sets can be used to measure the food environment, although estimates should be treated with caution in areas with high commercial density.

  2. cluML: A markup language for clustering and cluster validity assessment of microarray data.

    Science.gov (United States)

    Bolshakova, Nadia; Cunningham, Pádraig

    2005-01-01

    cluML is a new markup language for microarray data clustering and cluster validity assessment. The XML-based format has been designed to address some of the limitations observed in traditional formats, such as inability to store multiple clustering (including biclustering) and validation results within a dataset. cluML is an effective tool to support biomedical knowledge representation in gene expression data analysis. Although cluML was developed for DNA microarray analysis applications, it can be effectively used for the representation of clustering and for the validation of other biomedical and physical data that has no limitations.

  3. Geophysical validation of MIPAS-ENVISAT operational ozone data

    Directory of Open Access Journals (Sweden)

    U. Cortesi

    2007-09-01

    Full Text Available The Michelson Interferometer for Passive Atmospheric Sounding (MIPAS, on-board the European ENVIronmental SATellite (ENVISAT launched on 1 March 2002, is a middle infrared Fourier Transform spectrometer measuring the atmospheric emission spectrum in limb sounding geometry. The instrument is capable to retrieve the vertical distribution of temperature and trace gases, aiming at the study of climate and atmospheric chemistry and dynamics, and at applications to data assimilation and weather forecasting. MIPAS operated in its standard observation mode for approximately two years, from July 2002 to March 2004, with scans performed at nominal spectral resolution of 0.025 cm−1 and covering the altitude range from the mesosphere to the upper troposphere with relatively high vertical resolution (about 3 km in the stratosphere. Only reduced spectral resolution measurements have been performed subsequently. MIPAS data were re-processed by ESA using updated versions of the Instrument Processing Facility (IPF v4.61 and v4.62 and provided a complete set of level-2 operational products (geo-located vertical profiles of temperature and volume mixing ratio of H2O, O3, HNO3, CH4, N2O and NO2 with quasi continuous and global coverage in the period of MIPAS full spectral resolution mission. In this paper, we report a detailed description of the validation of MIPAS-ENVISAT operational ozone data, that was based on the comparison between MIPAS v4.61 (and, to a lesser extent, v4.62 O3 VMR profiles and a comprehensive set of correlative data, including observations from ozone sondes, ground-based lidar, FTIR and microwave radiometers, remote-sensing and in situ instruments on-board stratospheric aircraft and balloons, concurrent satellite sensors and ozone fields assimilated by the European Center for Medium-range Weather Forecasting.

    A coordinated effort was carried out

  4. Vulnerability in Acquisition, Language Impairments in Dutch: Creating a VALID Data Archive

    NARCIS (Netherlands)

    Klatter, J.; van Hout, R.; van den Heuvel, H.; Fikkert, P.; Baker, A.; de Jong, J.; Wijnen, F.; Sanders, E.; Trilsbeek, P.; Calzolari, N.; Choukri, K.; Declerck, T.; Maegaard, H.B.; Mariani, J.; Moreno, A.; Odijk, J.; Piperidis, S.

    2014-01-01

    The VALID Data Archive is an open multimedia data archive (under construction) with data from speakers suffering from language impairments. We report on a pilot project in the CLARIN-NL framework in which five data resources were curated. For all data sets concerned, written informed consent from

  5. Data validation summary report for the 100-HR-3 Round 8, Phases 1 and 2 groundwater sampling task

    International Nuclear Information System (INIS)

    1996-01-01

    This report presents a summary of data validation results on groundwater samples collected for the 100-HR-3 Round 8 Groundwater Sampling task. The analyses performed for this project consisted of: metals, general chemistry, and radiochemistry. The laboratories conducting the analyses were Quanterra Environmental Services (QES) and Lockheed Analytical Services. As required by the contract and the WHC statement of work (WHC 1994), data validation was conducted using the Westinghouse data validation procedures for chemical and radiochemical analyses (WHC 1993a and 1993b). Sample results were validated to levels A and D as described in the data validation procedures. At the completion of validation and verification of each data package, a data validation summary was prepared and transmitted with the original documentation to Environmental Restoration Contract (ERC) for inclusion in the project QA record

  6. Validation of NEPTUNE-CFD two-phase flow models using experimental data

    International Nuclear Information System (INIS)

    Perez-Manes, Jorge; Sanchez Espinoza, Victor Hugo; Bottcher, Michael; Stieglitz, Robert; Sergio Chiva Vicent

    2014-01-01

    This paper deals with the validation of the two-phase flow models of the CFD code NEPTUNE-CFD using experimental data provided by the OECD BWR BFBT and PSBT Benchmark. Since the two-phase models of CFD codes are extensively being improved, the validation is a key step for the acceptability of such codes. The validation work is performed in the frame of the European NURISP Project and it was focused on the steady state and transient void fraction tests. The influence of different NEPTUNE-CFD model parameters on the void fraction prediction is investigated and discussed in detail. Due to the coupling of heat conduction solver SYRTHES with NEPTUNE-CFD, the description of the coupled fluid dynamics and heat transfer between the fuel rod and the fluid is improved significantly. The averaged void fraction predicted by NEPTUNE-CFD for selected PSBT and BFBT tests is in good agreement with the experimental data. Finally, areas for future improvements of the NEPTUNE-CFD code were identified, too. (authors)

  7. Validity and efficiency of conformal anomaly detection on big distributed data

    Directory of Open Access Journals (Sweden)

    Ilia Nouretdinov

    2017-05-01

    Full Text Available Conformal Prediction is a recently developed framework for reliable confident predictions. In this work we discuss its possible application to big data coming from different, possibly heterogeneous data sources. On example of anomaly detection problem, we study the question of saving validity of Conformal Prediction in this case. We show that the straight forward averaging approach is invalid, while its easy alternative of maximizing is not very efficient because of its conservativeness. We propose the third compromised approach that is valid, but much less conservative. It is supported by both theoretical justification and experimental results in the area of energy engineering.

  8. Discharge data from 50 selected rivers for GCM validation

    International Nuclear Information System (INIS)

    Duemenil, L.; Isele, K.; Liebscher, H.J.; Schroeder, U.; Schumacher, M.; Wilke, K.

    1993-01-01

    This Technical Report refers to a joint project between GRDC Koblenz and MPI Hamburg. The Global Runoff Data Centre operates under the auspieces of WMO at the Federal Institute of Hydrology (Bundesanstalt fuer Gewaesserkunde) in Koblenz. River discharge data of the 50 largest rivers provide an independent data source for the validation of the hydrological cycle in general circulation models. This type of data is particularly valuable, because in some cases the available time series are exceptionally long. The data are presented as time series of annual average discharge (averaged over the period for which data is available, see below for caveats) and as annual cycles of monthly mean discharge averaged over the length of the time series available. (orig./KW)

  9. ValidatorDB: database of up-to-date validation results for ligands and non-standard residues from the Protein Data Bank.

    Science.gov (United States)

    Sehnal, David; Svobodová Vařeková, Radka; Pravda, Lukáš; Ionescu, Crina-Maria; Geidl, Stanislav; Horský, Vladimír; Jaiswal, Deepti; Wimmerová, Michaela; Koča, Jaroslav

    2015-01-01

    Following the discovery of serious errors in the structure of biomacromolecules, structure validation has become a key topic of research, especially for ligands and non-standard residues. ValidatorDB (freely available at http://ncbr.muni.cz/ValidatorDB) offers a new step in this direction, in the form of a database of validation results for all ligands and non-standard residues from the Protein Data Bank (all molecules with seven or more heavy atoms). Model molecules from the wwPDB Chemical Component Dictionary are used as reference during validation. ValidatorDB covers the main aspects of validation of annotation, and additionally introduces several useful validation analyses. The most significant is the classification of chirality errors, allowing the user to distinguish between serious issues and minor inconsistencies. Other such analyses are able to report, for example, completely erroneous ligands, alternate conformations or complete identity with the model molecules. All results are systematically classified into categories, and statistical evaluations are performed. In addition to detailed validation reports for each molecule, ValidatorDB provides summaries of the validation results for the entire PDB, for sets of molecules sharing the same annotation (three-letter code) or the same PDB entry, and for user-defined selections of annotations or PDB entries. © The Author(s) 2014. Published by Oxford University Press on behalf of Nucleic Acids Research.

  10. Calibration And Validation Of CryoSat-2 Low Resolution Mode Data

    Science.gov (United States)

    Naeije, M.; Schrama, E.; Scharroo, R.

    2011-02-01

    Running ahead of the continuously growing need for operational use of sea level products, TUDelft started off the Radar Altimeter Database System RADS many years ago. This system attends to a global international sea- level service. It supports, on one hand, science, like studies on ocean circulation, El Nio, sea level change, and ice topography, and on the other hand (offshore) operations, like delivery of ocean current information, wind and wave statistics, ice detection and ice classification. At present, the database is used by a large scientific community throughout the world, and is daily maintained and developed by Altimetrics LLC, TUDelft and NOAA. It contains all historic altimeter data, and now has to be up- dated with the data from ESAs ice mission CryoSat-2, which was launched successfully in April 2010. These new data are important to augment the data set and by that to improve the estimates of sea level change and its contributors. For this the data have to be validated and calibrated, necessary corrections added and improved (including modelling of corrections that are not directly available from the CryoSat-2 platform), and the orbit ac- curacy verified and if possible the orbits brushed up. Subsequently, value-added ocean and ice products need to be developed in synergy with all the other satellite altimeter data. During the commissioning phase we primarily looked at the sanity of the available level-1b and level-2 Low Resolution Mode (LRM) data. Here, for the 2011 CryoSat Validation Workshop, we present the results of our calibration and validation of LRM L2 data by internal comparison of CryoSat-2 and external comparison with other satellites. We have established a range bias of 3.77 (measurement range too long) and a timing bias of 8.2ms (measurement range too late).

  11. Tools for the Validation of Genomes and Transcriptomes with Proteomics data

    DEFF Research Database (Denmark)

    Pang, Chi Nam Ignatius; Aya, Carlos; Tay, Aidan

    data generated from protein mass spectrometry. We are developing a set of tools which allow users to: •Co-visualise genomics, transcriptomics, and proteomics data using the Integrated Genomics Viewer (IGV).1 •Validate the existence of genes and mRNAs using peptides identified from mass spectrometry...

  12. Use of standard vocabulary services in validation of water resources data

    Science.gov (United States)

    Yu, Jonathan; Cox, Simon; Ratcliffe, David

    2010-05-01

    Ontology repositories are increasingly being exposed through vocabulary and concept services. Primarily this is in support of resource discovery. Thesaurus functionality and even more sophisticated reasoning offers the possibility of overcoming the limitations of simple text-matching and tagging which is the basis of most search. However, controlled vocabularies have other important roles in distributed systems: in particular in constraining content validity. A national water information system established by the Australian Bureau of Meterorology ('the Bureau') has deployed a system for ingestion of data from multiple providers. This uses a http interface onto separately maintained vocabulary services as part of the quality assurance chain. With over 200 data providers potentially transferring data to the Bureau, a standard XML-based Water Data Transfer Format (WDTF) was developed for receipt of data into an integrated national water information system. The WDTF schema was built upon standards from the Open Geospatial Consortium (OGC). The structure and syntax specified by a W3C XML Schema is complemented by additional constraints described using Schematron. These implement important content requirements and business rules including: • Restricted cardinality: where optional elements and attributes inherited from the base standards become mandatory in the application, or repeatable elements or attributes are limited to one or omitted. For example, the sampledFeature element from O&M is optional but is mandatory for a samplingPoint element in WDTF. • Vocabulary checking: WDTF data use seventeen vocabularies or code lists derived from Regulations under the Commonwealth Water Act 2007. Examples of codelists are the Australian Water Regulations list, observed property vocabulary, and units of measures. • Contextual constraints: in many places, the permissible value is dependent on the value of another field. For example, within observations the unit of measure

  13. The Role of Laboratory Experiments in the Validation of Field Data

    DEFF Research Database (Denmark)

    Mouneyrac, Catherine; Lagarde, Fabienne; Chatel, Amelie

    2017-01-01

    The ubiquitous presence and persistency of microplastics (MPs) in aquatic environments are of particular concern, since they constitute a potential threat to marine organisms and ecosystems. However, evaluating this threat and the impacts of MP on aquatic organisms is challenging. MPs form a very...... and to what degree these complexities are addressed in the current literature, to: (1) evaluate how well laboratory studies, investigated so far, represent environmentally relevant processes and scenarios and (2) suggest directions for future research The Role of Laboratory Experiments in the Validation...... of Field Data | Request PDF. Available from: https://www.researchgate.net/publication/310360438_The_Role_of_Laboratory_Experiments_in_the_Validation_of_Field_Data [accessed Jan 15 2018]....

  14. 100 Area excavation treatability study data validation report

    International Nuclear Information System (INIS)

    Frain, J.M.

    1994-01-01

    This report presents the results of sampling and chemical analyses at Hanford Reservation. The samples were analyzed by Thermo-Analytic Laboratories and Roy F. Weston Laboratories using US Environmental Protection Agency CLP protocols. Sample analyses included: volatile organics; semivolatile organics; inorganics; and general chemical parameters. The data from the chemical analyses were reviewed and validated to verify that reported sample results were of sufficient quality to support decisions regarding remedial actions performed at this site

  15. Data validation report for the 100-HR-3 Operable Unit first quarter 1994 groundwater sampling data

    Energy Technology Data Exchange (ETDEWEB)

    Biggerstaff, R.L.

    1994-06-24

    Westinghouse-Hanford has requested that a minimum of 20% of the total number of Sample Delivery Groups be validated for the 100-HR-3 Operable Unit First Quarter 1994 Groundwater Sampling Investigation. Therefore, the data from the chemical analysis of twenty-four samples from this sampling event and their related quality assurance samples were reviewed and validated to verify that reported sample results were of sufficient quality to support decisions regarding remedial actions performed at this site. The samples were analyzed by Thermo-Analytic Laboratories (TMA) and Roy F. Weston Laboratories (WESTON) using US Environmental Protection Agency (EPA) CLP protocols. Sample analyses included: inorganics; and general chemical parameters. Forty-two samples were validated for radiochemical parameters by TMA and Teledyne.

  16. Data validation report for the 100-HR-3 Operable Unit first quarter 1994 groundwater sampling data

    International Nuclear Information System (INIS)

    Biggerstaff, R.L.

    1994-01-01

    Westinghouse-Hanford has requested that a minimum of 20% of the total number of Sample Delivery Groups be validated for the 100-HR-3 Operable Unit First Quarter 1994 Groundwater Sampling Investigation. Therefore, the data from the chemical analysis of twenty-four samples from this sampling event and their related quality assurance samples were reviewed and validated to verify that reported sample results were of sufficient quality to support decisions regarding remedial actions performed at this site. The samples were analyzed by Thermo-Analytic Laboratories (TMA) and Roy F. Weston Laboratories (WESTON) using US Environmental Protection Agency (EPA) CLP protocols. Sample analyses included: inorganics; and general chemical parameters. Forty-two samples were validated for radiochemical parameters by TMA and Teledyne

  17. Irradiated food: validity of extrapolating wholesomeness data

    International Nuclear Information System (INIS)

    Taub, I.A.; Angelini, P.; Merritt, C. Jr.

    1976-01-01

    Criteria are considered for validly extrapolating the conclusions reached on the wholesomeness of an irradiated food receiving high doses to the same food receiving a lower dose. A consideration first is made of the possible chemical mechanisms that could give rise to different functional dependences of radiolytic products on dose. It is shown that such products should increase linearly with dose and the ratio of products should be constant throughout the dose range considered. The assumption, generally accepted in pharmacology, then is made that if any adverse effects related to the food are discerned in the test animals, then the intensity of these effects would increase with the concentration of radiolytic products in the food. Lastly, the need to compare data from animal studies with foods irradiated to several doses against chemical evidence obtained over a comparable dose range is considered. It is concluded that if the products depend linearly on dose and if feeding studies indicate no adverse effects, then an extrapolation to lower doses is clearly valid. This approach is illustrated for irradiated codfish. The formation of selected volatile products in samples receiving between 0.1 and 3 Mrads was examined, and their concentrations were found to increase linearly at least up to 1 Mrad. These data were compared with results from animal feeding studies establishing the wholesomeness of codfish and haddock irradiated to 0.2, 0.6 and 2.8 Mrads. It is stated, therefore, that if ocean fish, currently under consideration for onboard processing, were irradiated to 0.1 Mrad, it would be correspondingly wholesome

  18. Validation of a Hot Water Distribution Model Using Laboratory and Field Data

    Energy Technology Data Exchange (ETDEWEB)

    Backman, C.; Hoeschele, M.

    2013-07-01

    Characterizing the performance of hot water distribution systems is a critical step in developing best practice guidelines for the design and installation of high performance hot water systems. Developing and validating simulation models is critical to this effort, as well as collecting accurate input data to drive the models. In this project, the ARBI team validated the newly developed TRNSYS Type 604 pipe model against both detailed laboratory and field distribution system performance data. Validation efforts indicate that the model performs very well in handling different pipe materials, insulation cases, and varying hot water load conditions. Limitations of the model include the complexity of setting up the input file and long simulation run times. In addition to completing validation activities, this project looked at recent field hot water studies to better understand use patterns and potential behavioral changes as homeowners convert from conventional storage water heaters to gas tankless units. Based on these datasets, we conclude that the current Energy Factor test procedure overestimates typical use and underestimates the number of hot water draws. This has implications for both equipment and distribution system performance. Gas tankless water heaters were found to impact how people use hot water, but the data does not necessarily suggest an increase in usage. Further study in hot water usage and patterns is needed to better define these characteristics in different climates and home vintages.

  19. Validation of data and indicators in the Danish Cholecystectomy Database

    DEFF Research Database (Denmark)

    Harboe, Kirstine Moll; Anthonsen, Kristian; Bardram, Linda

    2009-01-01

    and postoperative complications RESEARCH DESIGN AND SUBJECTS: Data from 1360 medical records of patients undergoing cholecystectomy were compared to the relevant administrative data from the National Patient Registry. The medical records served as the "gold standard". The association between the individual...... indicators and the occurrence of a postoperative complication was assessed. MEASURES: Validation of administrative data against the gold standard was done by the calculation of per cent agreement (including kappa-values) sensitivity/specificity and predictive values. The association between indicators...

  20. Cloud detection algorithm comparison and validation for operational Landsat data products

    Science.gov (United States)

    Foga, Steven Curtis; Scaramuzza, Pat; Guo, Song; Zhu, Zhe; Dilley, Ronald; Beckmann, Tim; Schmidt, Gail L.; Dwyer, John L.; Hughes, MJ; Laue, Brady

    2017-01-01

    Clouds are a pervasive and unavoidable issue in satellite-borne optical imagery. Accurate, well-documented, and automated cloud detection algorithms are necessary to effectively leverage large collections of remotely sensed data. The Landsat project is uniquely suited for comparative validation of cloud assessment algorithms because the modular architecture of the Landsat ground system allows for quick evaluation of new code, and because Landsat has the most comprehensive manual truth masks of any current satellite data archive. Currently, the Landsat Level-1 Product Generation System (LPGS) uses separate algorithms for determining clouds, cirrus clouds, and snow and/or ice probability on a per-pixel basis. With more bands onboard the Landsat 8 Operational Land Imager (OLI)/Thermal Infrared Sensor (TIRS) satellite, and a greater number of cloud masking algorithms, the U.S. Geological Survey (USGS) is replacing the current cloud masking workflow with a more robust algorithm that is capable of working across multiple Landsat sensors with minimal modification. Because of the inherent error from stray light and intermittent data availability of TIRS, these algorithms need to operate both with and without thermal data. In this study, we created a workflow to evaluate cloud and cloud shadow masking algorithms using cloud validation masks manually derived from both Landsat 7 Enhanced Thematic Mapper Plus (ETM +) and Landsat 8 OLI/TIRS data. We created a new validation dataset consisting of 96 Landsat 8 scenes, representing different biomes and proportions of cloud cover. We evaluated algorithm performance by overall accuracy, omission error, and commission error for both cloud and cloud shadow. We found that CFMask, C code based on the Function of Mask (Fmask) algorithm, and its confidence bands have the best overall accuracy among the many algorithms tested using our validation data. The Artificial Thermal-Automated Cloud Cover Algorithm (AT-ACCA) is the most accurate

  1. Validation of IT-based Data Communication Protocol for Nuclear Power Plant

    International Nuclear Information System (INIS)

    Jeong, K. I.; Kim, D. H.; Lee, J. C.

    2009-12-01

    The communication network designed to transmit control and processing signals in digital Instrument and Control (I and C) systems in Nuclear Power Plant (NPP), should provide a high level of safety and reliability. There are different features between the communication networks of NPPs and other commercial communication networks. Safety and reliability are the most important factors in the communication networks of an NPP rather than efficiency which are important factors of a commercial communication network design. To develop Data Communication Protocol for Nuclear Power Plant, We analyze the design criteria and performance requirements of existing commercial communication protocols based on Information Technology(IT). And also, we examine the adaptability to the communication protocol of an NPP. Based on these results, we developed our own protocol(Nuclear power plant Safety Communication Protocol : NSCP) for NPP I and C, which meet the required specifications through design overall protocol architecture and data frame format, definition of functional requirements and specifications. NSCP is the communication protocol designed for a safety-grade control network in the nuclear power plant. In this report, we had specified NSCP protocol by FDT(Formal Description Technique) and established validation procedures based on the validation methodology. It was confirmed specification error, major function's validity and reachability of NSCP by performing simulation and the validation process using Telelogic Tau tool

  2. Data validation summary report for the 100-BC-5 Operable Unit Round 9 Groundwater Sampling. Revision 0

    International Nuclear Information System (INIS)

    Kearney, A.T.

    1996-03-01

    The information provided in this validation summary report includes chemical analyses of samples from 100-BC-5 Operable Unit Round 9 Groundwater sampling data. Data from this sampling event and their related quality assurance (QA) samples were reviewed and validated in accordance with Westinghouse Hanford Company (WHC) guidelines at the requested level. Sample analyses included metals, general chemistry, and radiochemistry. Sixty metals samples were analyzed by Quanterra Environmental Services (QES) and Lockheed Analytical Services (LAS). The metals samples were validated using WHC protocols specified in Data Validation Procedures for Chemical Analyses. All qualifiers assigned to the metals data were based on this guidance. The Table 1.1 lists the metals sample delivery group (SDG) that were validated for this sampling event

  3. Validating safeguards effectiveness given inherently limited test data

    International Nuclear Information System (INIS)

    Sicherman, A.

    1987-01-01

    A key issue in designing and evaluating nuclear safeguards systems is how to validate safeguards effectiveness against a spectrum of potential threats. Safeguards effectiveness is measured by a performance indicator such as the probability of defeating an adversary attempting a malevolent act. Effectiveness validation means a testing program that provides sufficient evidence that the performance indicator is at an acceptable level. Traditional statistical program when numerous independent system trials are possible. However, within the safeguards environment, many situations arise for which traditional statistical approaches may be neither feasible nor appropriate. Such situations can occur, for example, when there are obvious constraints on the number of possible tests due to operational impacts and testing costs. Furthermore, these tests are usually simulations (e.g., staged force-on-force exercises) rather than actual tests, and the system is often modified after each test. Under such circumstances, it is difficult to make and justify inferences about system performance by using traditional statistical techniques. In this paper, the authors discuss several alternative quantitative techniques for validating system effectiveness. The techniques include: (1) minimizing the number of required tests using sequential testing; (2) combining data from models inspections and exercises using Bayesian statistics to improve inferences about system performance; and (3) using reliability growth and scenario modeling to help specify which safeguards elements and scenarios to test

  4. Examples of Use of SINBAD Database for Nuclear Data and Code Validation

    Science.gov (United States)

    Kodeli, Ivan; Žerovnik, Gašper; Milocco, Alberto

    2017-09-01

    The SINBAD database currently contains compilations and evaluations of over 100 shielding benchmark experiments. The SINBAD database is widely used for code and data validation. Materials covered include: Air, N. O, H2O, Al, Be, Cu, graphite, concrete, Fe, stainless steel, Pb, Li, Ni, Nb, SiC, Na, W, V and mixtures thereof. Over 40 organisations from 14 countries and 2 international organisations have contributed data and work in support of SINBAD. Examples of the use of the database in the scope of different international projects, such as the Working Party on Evaluation Cooperation of the OECD and the European Fusion Programme demonstrate the merit and possible usage of the database for the validation of modern nuclear data evaluations and new computer codes.

  5. Driving simulator validation of driver behavior with limited safe vantage points for data collection in work zones.

    Science.gov (United States)

    Bham, Ghulam H; Leu, Ming C; Vallati, Manoj; Mathur, Durga R

    2014-06-01

    This study is aimed at validating a driving simulator (DS) for the study of driver behavior in work zones. A validation study requires field data collection. For studies conducted in highway work zones, the availability of safe vantage points for data collection at critical locations can be a significant challenge. A validation framework is therefore proposed in this paper, demonstrated using a fixed-based DS that addresses the issue by using a global positioning system (GPS). The validation of the DS was conducted using objective and subjective evaluations. The objective validation was divided into qualitative and quantitative evaluations. The DS was validated by comparing the results of simulation with the field data, which were collected using a GPS along the highway and video recordings at specific locations in a work zone. The constructed work zone scenario in the DS was subjectively evaluated with 46 participants. The objective evaluation established the absolute and relative validity of the DS. The mean speeds from the DS data showed excellent agreement with the field data. The subjective evaluation indicated realistic driving experience by the participants. The use of GPS showed that continuous data collected along the highway can overcome the challenges of unavailability of safe vantage points especially at critical locations. Further, a validated DS can be used for examining driver behavior in complex situations by replicating realistic scenarios. Copyright © 2014 Elsevier Ltd. All rights reserved.

  6. Generation, Validation, and Application of Abundance Map Reference Data for Spectral Unmixing

    Science.gov (United States)

    Williams, McKay D.

    Reference data ("ground truth") maps traditionally have been used to assess the accuracy of imaging spectrometer classification algorithms. However, these reference data can be prohibitively expensive to produce, often do not include sub-pixel abundance estimates necessary to assess spectral unmixing algorithms, and lack published validation reports. Our research proposes methodologies to efficiently generate, validate, and apply abundance map reference data (AMRD) to airborne remote sensing scenes. We generated scene-wide AMRD for three different remote sensing scenes using our remotely sensed reference data (RSRD) technique, which spatially aggregates unmixing results from fine scale imagery (e.g., 1-m Ground Sample Distance (GSD)) to co-located coarse scale imagery (e.g., 10-m GSD or larger). We validated the accuracy of this methodology by estimating AMRD in 51 randomly-selected 10 m x 10 m plots, using seven independent methods and observers, including field surveys by two observers, imagery analysis by two observers, and RSRD using three algorithms. Results indicated statistically-significant differences between all versions of AMRD, suggesting that all forms of reference data need to be validated. Given these significant differences between the independent versions of AMRD, we proposed that the mean of all (MOA) versions of reference data for each plot and class were most likely to represent true abundances. We then compared each version of AMRD to MOA. Best case accuracy was achieved by a version of imagery analysis, which had a mean coverage area error of 2.0%, with a standard deviation of 5.6%. One of the RSRD algorithms was nearly as accurate, achieving a mean error of 3.0%, with a standard deviation of 6.3%, showing the potential of RSRD-based AMRD generation. Application of validated AMRD to specific coarse scale imagery involved three main parts: 1) spatial alignment of coarse and fine scale imagery, 2) aggregation of fine scale abundances to produce

  7. Quantifying the foodscape: A systematic review and meta-analysis of the validity of commercially available business data

    Science.gov (United States)

    Lebel, Alexandre; Daepp, Madeleine I. G.; Block, Jason P.; Walker, Renée; Lalonde, Benoît; Kestens, Yan; Subramanian, S. V.

    2017-01-01

    This paper reviews studies of the validity of commercially available business (CAB) data on food establishments (“the foodscape”), offering a meta-analysis of characteristics associated with CAB quality and a case study evaluating the performance of commonly-used validity indicators describing the foodscape. Existing validation studies report a broad range in CAB data quality, although most studies conclude that CAB quality is “moderate” to “substantial”. We conclude that current studies may underestimate the quality of CAB data. We recommend that future validation studies use density-adjusted and exposure measures to offer a more meaningful characterization of the relationship of data error with spatial exposure. PMID:28358819

  8. A multi-source satellite data approach for modelling Lake Turkana water level: calibration and validation using satellite altimetry data

    Directory of Open Access Journals (Sweden)

    N. M. Velpuri

    2012-01-01

    Full Text Available Lake Turkana is one of the largest desert lakes in the world and is characterized by high degrees of inter- and intra-annual fluctuations. The hydrology and water balance of this lake have not been well understood due to its remote location and unavailability of reliable ground truth datasets. Managing surface water resources is a great challenge in areas where in-situ data are either limited or unavailable. In this study, multi-source satellite-driven data such as satellite-based rainfall estimates, modelled runoff, evapotranspiration, and a digital elevation dataset were used to model Lake Turkana water levels from 1998 to 2009. Due to the unavailability of reliable lake level data, an approach is presented to calibrate and validate the water balance model of Lake Turkana using a composite lake level product of TOPEX/Poseidon, Jason-1, and ENVISAT satellite altimetry data. Model validation results showed that the satellite-driven water balance model can satisfactorily capture the patterns and seasonal variations of the Lake Turkana water level fluctuations with a Pearson's correlation coefficient of 0.90 and a Nash-Sutcliffe Coefficient of Efficiency (NSCE of 0.80 during the validation period (2004–2009. Model error estimates were within 10% of the natural variability of the lake. Our analysis indicated that fluctuations in Lake Turkana water levels are mainly driven by lake inflows and over-the-lake evaporation. Over-the-lake rainfall contributes only up to 30% of lake evaporative demand. During the modelling time period, Lake Turkana showed seasonal variations of 1–2 m. The lake level fluctuated in the range up to 4 m between the years 1998–2009. This study demonstrated the usefulness of satellite altimetry data to calibrate and validate the satellite-driven hydrological model for Lake Turkana without using any in-situ data. Furthermore, for Lake Turkana, we identified and outlined opportunities and challenges of using a calibrated

  9. A multi-source satellite data approach for modelling Lake Turkana water level: Calibration and validation using satellite altimetry data

    Science.gov (United States)

    Velpuri, N.M.; Senay, G.B.; Asante, K.O.

    2012-01-01

    Lake Turkana is one of the largest desert lakes in the world and is characterized by high degrees of interand intra-annual fluctuations. The hydrology and water balance of this lake have not been well understood due to its remote location and unavailability of reliable ground truth datasets. Managing surface water resources is a great challenge in areas where in-situ data are either limited or unavailable. In this study, multi-source satellite-driven data such as satellite-based rainfall estimates, modelled runoff, evapotranspiration, and a digital elevation dataset were used to model Lake Turkana water levels from 1998 to 2009. Due to the unavailability of reliable lake level data, an approach is presented to calibrate and validate the water balance model of Lake Turkana using a composite lake level product of TOPEX/Poseidon, Jason-1, and ENVISAT satellite altimetry data. Model validation results showed that the satellitedriven water balance model can satisfactorily capture the patterns and seasonal variations of the Lake Turkana water level fluctuations with a Pearson's correlation coefficient of 0.90 and a Nash-Sutcliffe Coefficient of Efficiency (NSCE) of 0.80 during the validation period (2004-2009). Model error estimates were within 10% of the natural variability of the lake. Our analysis indicated that fluctuations in Lake Turkana water levels are mainly driven by lake inflows and over-the-lake evaporation. Over-the-lake rainfall contributes only up to 30% of lake evaporative demand. During the modelling time period, Lake Turkana showed seasonal variations of 1-2m. The lake level fluctuated in the range up to 4m between the years 1998-2009. This study demonstrated the usefulness of satellite altimetry data to calibrate and validate the satellite-driven hydrological model for Lake Turkana without using any in-situ data. Furthermore, for Lake Turkana, we identified and outlined opportunities and challenges of using a calibrated satellite-driven water balance

  10. Genomic Prediction in Animals and Plants: Simulation of Data, Validation, Reporting, and Benchmarking

    Science.gov (United States)

    Daetwyler, Hans D.; Calus, Mario P. L.; Pong-Wong, Ricardo; de los Campos, Gustavo; Hickey, John M.

    2013-01-01

    The genomic prediction of phenotypes and breeding values in animals and plants has developed rapidly into its own research field. Results of genomic prediction studies are often difficult to compare because data simulation varies, real or simulated data are not fully described, and not all relevant results are reported. In addition, some new methods have been compared only in limited genetic architectures, leading to potentially misleading conclusions. In this article we review simulation procedures, discuss validation and reporting of results, and apply benchmark procedures for a variety of genomic prediction methods in simulated and real example data. Plant and animal breeding programs are being transformed by the use of genomic data, which are becoming widely available and cost-effective to predict genetic merit. A large number of genomic prediction studies have been published using both simulated and real data. The relative novelty of this area of research has made the development of scientific conventions difficult with regard to description of the real data, simulation of genomes, validation and reporting of results, and forward in time methods. In this review article we discuss the generation of simulated genotype and phenotype data, using approaches such as the coalescent and forward in time simulation. We outline ways to validate simulated data and genomic prediction results, including cross-validation. The accuracy and bias of genomic prediction are highlighted as performance indicators that should be reported. We suggest that a measure of relatedness between the reference and validation individuals be reported, as its impact on the accuracy of genomic prediction is substantial. A large number of methods were compared in example simulated and real (pine and wheat) data sets, all of which are publicly available. In our limited simulations, most methods performed similarly in traits with a large number of quantitative trait loci (QTL), whereas in traits

  11. Validating Machine Learning Algorithms for Twitter Data Against Established Measures of Suicidality.

    Science.gov (United States)

    Braithwaite, Scott R; Giraud-Carrier, Christophe; West, Josh; Barnes, Michael D; Hanson, Carl Lee

    2016-05-16

    One of the leading causes of death in the United States (US) is suicide and new methods of assessment are needed to track its risk in real time. Our objective is to validate the use of machine learning algorithms for Twitter data against empirically validated measures of suicidality in the US population. Using a machine learning algorithm, the Twitter feeds of 135 Mechanical Turk (MTurk) participants were compared with validated, self-report measures of suicide risk. Our findings show that people who are at high suicidal risk can be easily differentiated from those who are not by machine learning algorithms, which accurately identify the clinically significant suicidal rate in 92% of cases (sensitivity: 53%, specificity: 97%, positive predictive value: 75%, negative predictive value: 93%). Machine learning algorithms are efficient in differentiating people who are at a suicidal risk from those who are not. Evidence for suicidality can be measured in nonclinical populations using social media data.

  12. Validation Data and Model Development for Fuel Assembly Response to Seismic Loads

    International Nuclear Information System (INIS)

    Bardet, Philippe; Ricciardi, Guillaume

    2016-01-01

    Vibrations are inherently present in nuclear reactors, especially in cores and steam generators of pressurized water reactors (PWR). They can have significant effects on local heat transfer and wear and tear in the reactor and often set safety margins. The simulation of these multiphysics phenomena from first principles requires the coupling of several codes, which is one the most challenging task in modern computer simulation. Here an ambitious multiphysics multidisciplinary validation campaign is conducted. It relied on an integrated team of experimentalists and code developers to acquire benchmark and validation data for fluid-structure interaction codes. Data are focused on PWR fuel bundle behavior during seismic transients.

  13. Validation Data and Model Development for Fuel Assembly Response to Seismic Loads

    Energy Technology Data Exchange (ETDEWEB)

    Bardet, Philippe [George Washington Univ., Washington, DC (United States); Ricciardi, Guillaume [Atomic Energy Commission (CEA) (France)

    2016-01-31

    Vibrations are inherently present in nuclear reactors, especially in cores and steam generators of pressurized water reactors (PWR). They can have significant effects on local heat transfer and wear and tear in the reactor and often set safety margins. The simulation of these multiphysics phenomena from first principles requires the coupling of several codes, which is one the most challenging tasks in modern computer simulation. Here an ambitious multiphysics multidisciplinary validation campaign is conducted. It relied on an integrated team of experimentalists and code developers to acquire benchmark and validation data for fluid-structure interaction codes. Data are focused on PWR fuel bundle behavior during seismic transients.

  14. Validation of virtual instrument for data analysis in metrology of time and frequency

    International Nuclear Information System (INIS)

    Jordao, Bruno; Quaresma, Daniel; Rocha, Pedro; Carvalho, Ricardo; Peixoto, Jose Guilherme

    2016-01-01

    Commercial Software (CS) for collection, analysis and plot time and frequency data plots are being increasingly used in reference laboratories worldwide. With this, it has greatly improved the results of calculations of uncertainty for these values. We propose the creation of a collection of software and data analysis using Virtual Instruments (VI) developed the Primary Laboratory Time and frequency of the National Observatory - ON and validation of this instrument. To validate the instrument developed, it made a comparative analysis between the results obtained (VI) with the results obtained by (CS) widely used in many metrology laboratories. From these results we can conclude that there was equivalence between the analyzed data. (author)

  15. Enhanced data validation strategy of air quality monitoring network.

    Science.gov (United States)

    Harkat, Mohamed-Faouzi; Mansouri, Majdi; Nounou, Mohamed; Nounou, Hazem

    2018-01-01

    Quick validation and detection of faults in measured air quality data is a crucial step towards achieving the objectives of air quality networks. Therefore, the objectives of this paper are threefold: (i) to develop a modeling technique that can be used to predict the normal behavior of air quality variables and help provide accurate reference for monitoring purposes; (ii) to develop fault detection method that can effectively and quickly detect any anomalies in measured air quality data. For this purpose, a new fault detection method that is based on the combination of generalized likelihood ratio test (GLRT) and exponentially weighted moving average (EWMA) will be developed. GLRT is a well-known statistical fault detection method that relies on maximizing the detection probability for a given false alarm rate. In this paper, we propose to develop GLRT-based EWMA fault detection method that will be able to detect the changes in the values of certain air quality variables; (iii) to develop fault isolation and identification method that allows defining the fault source(s) in order to properly apply appropriate corrective actions. In this paper, reconstruction approach that is based on Midpoint-Radii Principal Component Analysis (MRPCA) model will be developed to handle the types of data and models associated with air quality monitoring networks. All air quality modeling, fault detection, fault isolation and reconstruction methods developed in this paper will be validated using real air quality data (such as particulate matter, ozone, nitrogen and carbon oxides measurement). Copyright © 2017 Elsevier Inc. All rights reserved.

  16. Ensemble Kalman filter regularization using leave-one-out data cross-validation

    KAUST Repository

    Rayo Schiappacasse, Lautaro Jerónimo

    2012-09-19

    In this work, the classical leave-one-out cross-validation method for selecting a regularization parameter for the Tikhonov problem is implemented within the EnKF framework. Following the original concept, the regularization parameter is selected such that it minimizes the predictive error. Some ideas about the implementation, suitability and conceptual interest of the method are discussed. Finally, what will be called the data cross-validation regularized EnKF (dCVr-EnKF) is implemented in a 2D 2-phase synthetic oil reservoir experiment and the results analyzed.

  17. Use of operational data for the validation of the SOPHT thermal-hydraulic code

    Energy Technology Data Exchange (ETDEWEB)

    Ho, S F; Martin, G; Shoukas, L; Siddiqui, Z; Phillips, B [Ontario Hydro, Bowmanville, ON (Canada). Darlington Nuclear Generating Station

    1996-12-31

    The primary objective of this paper is to describe the validation process of the SOPHT and MINI-SOPHT codes with the use of reactor operational data. The secondary objective is to illustrative the effectiveness of the code as a performance monitoring tool by discussing the discoveries that were made during the validation process. (author). 2 refs.

  18. Validating Quantitative Measurement Using Qualitative Data: Combining Rasch Scaling and Latent Semantic Analysis in Psychiatry

    Science.gov (United States)

    Lange, Rense

    2015-02-01

    An extension of concurrent validity is proposed that uses qualitative data for the purpose of validating quantitative measures. The approach relies on Latent Semantic Analysis (LSA) which places verbal (written) statements in a high dimensional semantic space. Using data from a medical / psychiatric domain as a case study - Near Death Experiences, or NDE - we established concurrent validity by connecting NDErs qualitative (written) experiential accounts with their locations on a Rasch scalable measure of NDE intensity. Concurrent validity received strong empirical support since the variance in the Rasch measures could be predicted reliably from the coordinates of their accounts in the LSA derived semantic space (R2 = 0.33). These coordinates also predicted NDErs age with considerable precision (R2 = 0.25). Both estimates are probably artificially low due to the small available data samples (n = 588). It appears that Rasch scalability of NDE intensity is a prerequisite for these findings, as each intensity level is associated (at least probabilistically) with a well- defined pattern of item endorsements.

  19. Implicit structural inversion of gravity data using linear programming, a validation study

    NARCIS (Netherlands)

    Zon, A.T. van; Roy Chowdhury, K.

    2010-01-01

    In this study, a regional scale gravity data set has been inverted to infer the structure (topography) of the top of the basement underlying sub-horizontal strata. We apply our method to this real data set for further proof of concept, validation and benchmarking against results from an earlier

  20. Applicability of U.S. Army tracer test data to model validation needs of ERDA

    International Nuclear Information System (INIS)

    Shearer, D.L.; Minott, D.H.

    1976-06-01

    This report covers the first phase of an atmospheric dispersion model validation project sponsored by the Energy Research and Development Administration (ERDA). The project will employ dispersion data generated during an extensive series of field tracer experiments that were part of a meteorological research program which was conducted by the U. S. Army Dugway Proving Ground, Utah, from the late 1950's to the early 1970's. The tests were conducted at several locations in the U. S., South America, Germany, and Norway chosen to typify the effects of certain environmental factors on atmospheric dispersion. The purpose of the Phase I work of this project was to identify applicable portions of the Army data, obtain and review that data, and make recommendations for its uses for atmospheric dispersion model validations. This report presents key information in three formats. The first is a tabular listing of the Army dispersion test reports summarizing the test data contained in each report. This listing is presented in six separate tables with each tabular list representing a different topical area that is based on model validation requirements and the nature of the Army data base. The second format for presenting key information is a series of discussions of the Army test information assigned to each of the six topical areas. These discussions relate the extent and quality of the available data, as well as its prospective use for model validation. The third format is a series of synopses for each Army test report

  1. Validation of MODIS Data for localized spatio-temporal evapotranspiration mapping

    International Nuclear Information System (INIS)

    Nadzri, M I; Hashim, M

    2014-01-01

    Advancement in satellite remote sensing sensors allows evapo-transpiration (ET) from land surfaces to be derived from selected reflectance and emmitance in visible and thermal infrared wavelengths, such as using Moderate Solution Imaging Spectrometer (MODIS). In this paper, we report the validation of recent MODIS-generated higher-order global terrestrial ET product 16A2. The main focus of this paper is to devise the follow-up calibration for the localised region covering the entire Malaysia peninsular. The validation is carried out locally by dividing the study area into 3 distinct climatological regions based on the influence to monsoons, and using multi-temporal MODIS data acquired in 2000-2009. The results, evidently show the local effects still inherit in the MODIS 16A2 products; with varying R2 within the 3 local climatological regions established (Northwest = 0.49 South = 0.47, and Southwest = 0.52; all with P < 0.001). The accuracy of each region validated is within + RMSE 43mm for monthly ET. With P value in acceptable range, the correction is useable for further usage

  2. Good validity of the international spinal cord injury quality of life basic data set

    NARCIS (Netherlands)

    Post, M. W. M.; Adriaansen, J. J. E.; Charlifue, S.; Biering-Sorensen, F.; van Asbeck, F. W. A.

    Study design: Cross-sectional validation study. Objectives: To examine the construct and concurrent validity of the International Spinal Cord Injury (SCI) Quality of Life (QoL) Basic Data Set. Setting: Dutch community. Participants: People 28-65 years of age, who obtained their SCI between 18 and 35

  3. Funding for the 2ND IAEA technical meeting on fusion data processing, validation and analysis

    Energy Technology Data Exchange (ETDEWEB)

    Greenwald, Martin

    2017-06-02

    The International Atomic Energy Agency (IAEA) will organize the second Technical Meeting on Fusion Da Processing, Validation and Analysis from 30 May to 02 June, 2017, in Cambridge, MA USA. The meeting w be hosted by the MIT Plasma Science and Fusion Center (PSFC). The objective of the meeting is to provide a platform where a set of topics relevant to fusion data processing, validation and analysis are discussed with the view of extrapolation needs to next step fusion devices such as ITER. The validation and analysis of experimental data obtained from diagnostics used to characterize fusion plasmas are crucial for a knowledge based understanding of the physical processes governing the dynamics of these plasmas. The meeting will aim at fostering, in particular, discussions of research and development results that set out or underline trends observed in the current major fusion confinement devices. General information on the IAEA, including its mission and organization, can be found at the IAEA websit Uncertainty quantification (UQ) Model selection, validation, and verification (V&V) Probability theory and statistical analysis Inverse problems & equilibrium reconstru ction Integrated data analysis Real time data analysis Machine learning Signal/image proc essing & pattern recognition Experimental design and synthetic diagnostics Data management

  4. Visual Tools for Crowdsourcing Data Validation Within the GLOBELAND30 Geoportal

    Science.gov (United States)

    Chuprikova, E.; Wu, H.; Murphy, C. E.; Meng, L.

    2016-06-01

    This research aims to investigate the role of visualization of the user generated data that can empower the geoportal of GlobeLand30 produced by NGCC (National Geomatics Center of China). The focus is set on the development of a concept of tools that can extend the Geo-tagging functionality and make use of it for different target groups. The anticipated tools should improve the continuous data validation, updating and efficient use of the remotely-sensed data distributed within GlobeLand30.

  5. Validation of the newborn larynx modeling with aerodynamical experimental data.

    Science.gov (United States)

    Nicollas, R; Giordano, J; Garrel, R; Medale, M; Caminat, P; Giovanni, A; Ouaknine, M; Triglia, J M

    2009-06-01

    Many authors have studied adult's larynx modelization, but the mechanisms of newborn's voice production have very rarely been investigated. After validating a numerical model with acoustic data, studies were performed on larynges of human fetuses in order to validate this model with aerodynamical experiments. Anatomical measurements were performed and a simplified numerical model was built using Fluent((R)) with the vocal folds in phonatory position. The results obtained are in good agreement with those obtained by laser Doppler velocimetry (LDV) and high-frame rate particle image velocimetry (HFR-PIV), on an experimental bench with excised human fetus larynges. It appears that computing with first cry physiological parameters leads to a model which is close to those obtained in experiments with real organs.

  6. Likelihood ratio data to report the validation of a forensic fingerprint evaluation method

    NARCIS (Netherlands)

    Ramos, Daniel; Haraksim, Rudolf; Meuwly, Didier

    2017-01-01

    Data to which the authors refer to throughout this article are likelihood ratios (LR) computed from the comparison of 5–12 minutiae fingermarks with fingerprints. These LRs data are used for the validation of a likelihood ratio (LR) method in forensic evidence evaluation. These data present a

  7. K-means clustering versus validation measures: a data-distribution perspective.

    Science.gov (United States)

    Xiong, Hui; Wu, Junjie; Chen, Jian

    2009-04-01

    K-means is a well-known and widely used partitional clustering method. While there are considerable research efforts to characterize the key features of the K-means clustering algorithm, further investigation is needed to understand how data distributions can have impact on the performance of K-means clustering. To that end, in this paper, we provide a formal and organized study of the effect of skewed data distributions on K-means clustering. Along this line, we first formally illustrate that K-means tends to produce clusters of relatively uniform size, even if input data have varied "true" cluster sizes. In addition, we show that some clustering validation measures, such as the entropy measure, may not capture this uniform effect and provide misleading information on the clustering performance. Viewed in this light, we provide the coefficient of variation (CV) as a necessary criterion to validate the clustering results. Our findings reveal that K-means tends to produce clusters in which the variations of cluster sizes, as measured by CV, are in a range of about 0.3-1.0. Specifically, for data sets with large variation in "true" cluster sizes (e.g., CV > 1.0), K-means reduces variation in resultant cluster sizes to less than 1.0. In contrast, for data sets with small variation in "true" cluster sizes (e.g., CV K-means increases variation in resultant cluster sizes to greater than 0.3. In other words, for the earlier two cases, K-means produces the clustering results which are away from the "true" cluster distributions.

  8. Sensor Selection and Data Validation for Reliable Integrated System Health Management

    Science.gov (United States)

    Garg, Sanjay; Melcher, Kevin J.

    2008-01-01

    For new access to space systems with challenging mission requirements, effective implementation of integrated system health management (ISHM) must be available early in the program to support the design of systems that are safe, reliable, highly autonomous. Early ISHM availability is also needed to promote design for affordable operations; increased knowledge of functional health provided by ISHM supports construction of more efficient operations infrastructure. Lack of early ISHM inclusion in the system design process could result in retrofitting health management systems to augment and expand operational and safety requirements; thereby increasing program cost and risk due to increased instrumentation and computational complexity. Having the right sensors generating the required data to perform condition assessment, such as fault detection and isolation, with a high degree of confidence is critical to reliable operation of ISHM. Also, the data being generated by the sensors needs to be qualified to ensure that the assessments made by the ISHM is not based on faulty data. NASA Glenn Research Center has been developing technologies for sensor selection and data validation as part of the FDDR (Fault Detection, Diagnosis, and Response) element of the Upper Stage project of the Ares 1 launch vehicle development. This presentation will provide an overview of the GRC approach to sensor selection and data quality validation and will present recent results from applications that are representative of the complexity of propulsion systems for access to space vehicles. A brief overview of the sensor selection and data quality validation approaches is provided below. The NASA GRC developed Systematic Sensor Selection Strategy (S4) is a model-based procedure for systematically and quantitatively selecting an optimal sensor suite to provide overall health assessment of a host system. S4 can be logically partitioned into three major subdivisions: the knowledge base, the down

  9. VISUAL TOOLS FOR CROWDSOURCING DATA VALIDATION WITHIN THE GLOBELAND30 GEOPORTAL

    Directory of Open Access Journals (Sweden)

    E. Chuprikova

    2016-06-01

    Full Text Available This research aims to investigate the role of visualization of the user generated data that can empower the geoportal of GlobeLand30 produced by NGCC (National Geomatics Center of China. The focus is set on the development of a concept of tools that can extend the Geo-tagging functionality and make use of it for different target groups. The anticipated tools should improve the continuous data validation, updating and efficient use of the remotely-sensed data distributed within GlobeLand30.

  10. A psychophysically validated metric for bidirectional texture data reduction

    Czech Academy of Sciences Publication Activity Database

    Filip, Jiří; Chantler, M.J.; Green, P.R.; Haindl, Michal

    2008-01-01

    Roč. 27, č. 5 (2008), s. 138:1-138:11 ISSN 0730-0301 R&D Projects: GA AV ČR 1ET400750407; GA ČR GA102/08/0593 Institutional research plan: CEZ:AV0Z10750506 Keywords : Bidirectional Texture Functions * texture compression Subject RIV: BD - Theory of Information Impact factor: 3.383, year: 2008 http://library.utia.cas.cz/separaty/2008/RO/haindl-a psychophysically validated metric for bidirectional texture data reduction.pdf

  11. The tissue microarray data exchange specification: A document type definition to validate and enhance XML data

    Science.gov (United States)

    Nohle, David G; Ayers, Leona W

    2005-01-01

    Background The Association for Pathology Informatics (API) Extensible Mark-up Language (XML) TMA Data Exchange Specification (TMA DES) proposed in April 2003 provides a community-based, open source tool for sharing tissue microarray (TMA) data in a common format. Each tissue core within an array has separate data including digital images; therefore an organized, common approach to produce, navigate and publish such data facilitates viewing, sharing and merging TMA data from different laboratories. The AIDS and Cancer Specimen Resource (ACSR) is a HIV/AIDS tissue bank consortium sponsored by the National Cancer Institute (NCI) Division of Cancer Treatment and Diagnosis (DCTD). The ACSR offers HIV-related malignancies and uninfected control tissues in microarrays (TMA) accompanied by de-identified clinical data to approved researchers. Exporting our TMA data into the proposed API specified format offers an opportunity to evaluate the API specification in an applied setting and to explore its usefulness. Results A document type definition (DTD) that governs the allowed common data elements (CDE) in TMA DES export XML files was written, tested and evolved and is in routine use by the ACSR. This DTD defines TMA DES CDEs which are implemented in an external file that can be supplemented by internal DTD extensions for locally defined TMA data elements (LDE). Conclusion ACSR implementation of the TMA DES demonstrated the utility of the specification and allowed application of a DTD to validate the language of the API specified XML elements and to identify possible enhancements within our TMA data management application. Improvements to the specification have additionally been suggested by our experience in importing other institution's exported TMA data. Enhancements to TMA DES to remove ambiguous situations and clarify the data should be considered. Better specified identifiers and hierarchical relationships will make automatic use of the data possible. Our tool can be

  12. A New Statistical Method to Determine the Degree of Validity of Health Economic Model Outcomes against Empirical Data.

    Science.gov (United States)

    Corro Ramos, Isaac; van Voorn, George A K; Vemer, Pepijn; Feenstra, Talitha L; Al, Maiwenn J

    2017-09-01

    The validation of health economic (HE) model outcomes against empirical data is of key importance. Although statistical testing seems applicable, guidelines for the validation of HE models lack guidance on statistical validation, and actual validation efforts often present subjective judgment of graphs and point estimates. To discuss the applicability of existing validation techniques and to present a new method for quantifying the degrees of validity statistically, which is useful for decision makers. A new Bayesian method is proposed to determine how well HE model outcomes compare with empirical data. Validity is based on a pre-established accuracy interval in which the model outcomes should fall. The method uses the outcomes of a probabilistic sensitivity analysis and results in a posterior distribution around the probability that HE model outcomes can be regarded as valid. We use a published diabetes model (Modelling Integrated Care for Diabetes based on Observational data) to validate the outcome "number of patients who are on dialysis or with end-stage renal disease." Results indicate that a high probability of a valid outcome is associated with relatively wide accuracy intervals. In particular, 25% deviation from the observed outcome implied approximately 60% expected validity. Current practice in HE model validation can be improved by using an alternative method based on assessing whether the model outcomes fit to empirical data at a predefined level of accuracy. This method has the advantage of assessing both model bias and parameter uncertainty and resulting in a quantitative measure of the degree of validity that penalizes models predicting the mean of an outcome correctly but with overly wide credible intervals. Copyright © 2017 International Society for Pharmacoeconomics and Outcomes Research (ISPOR). Published by Elsevier Inc. All rights reserved.

  13. Validation of a Hot Water Distribution Model Using Laboratory and Field Data

    Energy Technology Data Exchange (ETDEWEB)

    Backman, C. [Alliance for Residential Building Innovation (ARBI), Davis, CA (United States); Hoeschele, M. [Alliance for Residential Building Innovation (ARBI), Davis, CA (United States)

    2013-07-01

    Characterizing the performance of hot water distribution systems is a critical step in developing best practice guidelines for the design and installation of high performance hot water systems. Developing and validating simulation models is critical to this effort, as well as collecting accurate input data to drive the models. In this project, the Building America research team ARBI validated the newly developed TRNSYS Type 604 pipe model against both detailed laboratory and field distribution system performance data. Validation efforts indicate that the model performs very well in handling different pipe materials, insulation cases, and varying hot water load conditions. Limitations of the model include the complexity of setting up the input file and long simulation run times. This project also looked at recent field hot water studies to better understand use patterns and potential behavioral changes as homeowners convert from conventional storage water heaters to gas tankless units. The team concluded that the current Energy Factor test procedure overestimates typical use and underestimates the number of hot water draws, which has implications for both equipment and distribution system performance. Gas tankless water heaters were found to impact how people use hot water, but the data does not necessarily suggest an increase in usage. Further study in hot water usage and patterns is needed to better define these characteristics in different climates and home vintages.

  14. Quantifying the foodscape: A systematic review and meta-analysis of the validity of commercially available business data

    OpenAIRE

    Lebel, Alexandre; Daepp, Madeleine I. G.; Block, Jason P.; Walker, Renée; Lalonde, Benoît; Kestens, Yan; Subramanian, S. V.

    2017-01-01

    This paper reviews studies of the validity of commercially available business (CAB) data on food establishments (?the foodscape?), offering a meta-analysis of characteristics associated with CAB quality and a case study evaluating the performance of commonly-used validity indicators describing the foodscape. Existing validation studies report a broad range in CAB data quality, although most studies conclude that CAB quality is ?moderate? to ?substantial?. We conclude that current studies may ...

  15. A Comprehensive Plan for the Long-Term Calibration and Validation of Oceanic Biogeochemical Satellite Data

    Science.gov (United States)

    Hooker, Stanford B.; McClain, Charles R.; Mannino, Antonio

    2007-01-01

    The primary objective of this planning document is to establish a long-term capability and validating oceanic biogeochemical satellite data. It is a pragmatic solution to a practical problem based primarily o the lessons learned from prior satellite missions. All of the plan's elements are seen to be interdependent, so a horizontal organizational scheme is anticipated wherein the overall leadership comes from the NASA Ocean Biology and Biogeochemistry (OBB) Program Manager and the entire enterprise is split into two components of equal sature: calibration and validation plus satellite data processing. The detailed elements of the activity are based on the basic tasks of the two main components plus the current objectives of the Carbon Cycle and Ecosystems Roadmap. The former is distinguished by an internal core set of responsibilities and the latter is facilitated through an external connecting-core ring of competed or contracted activities. The core elements for the calibration and validation component include a) publish protocols and performance metrics; b) verify uncertainty budgets; c) manage the development and evaluation of instrumentation; and d) coordinate international partnerships. The core elements for the satellite data processing component are e) process and reprocess multisensor data; f) acquire, distribute, and archive data products; and g) implement new data products. Both components have shared responsibilities for initializing and temporally monitoring satellite calibration. Connecting-core elements include (but are not restricted to) atmospheric correction and characterization, standards and traceability, instrument and analysis round robins, field campaigns and vicarious calibration sites, in situ database, bio-optical algorithm (and product) validation, satellite characterization and vicarious calibration, and image processing software. The plan also includes an accountability process, creating a Calibration and Validation Team (to help manage

  16. Internal validation of risk models in clustered data: a comparison of bootstrap schemes

    NARCIS (Netherlands)

    Bouwmeester, W.; Moons, K.G.M.; Kappen, T.H.; van Klei, W.A.; Twisk, J.W.R.; Eijkemans, M.J.C.; Vergouwe, Y.

    2013-01-01

    Internal validity of a risk model can be studied efficiently with bootstrapping to assess possible optimism in model performance. Assumptions of the regular bootstrap are violated when the development data are clustered. We compared alternative resampling schemes in clustered data for the estimation

  17. Validation of Australian data in the Australasian Vascular Audit.

    Science.gov (United States)

    Beiles, Charles Barry; Bourke, Bernie M

    2014-09-01

    Accuracy of data is important in any clinical audit. It is necessary to determine how complete the dataset is as well as the accuracy of the data that have been entered. The Australasian Vascular Audit has been operational for 4 years and a data validation process has been undertaken. An independent data source is available, which is collected by the Australian Institute of Health and Welfare. This collects all public and private hospital data and is available for interrogation. Similarly, private-only data are available from the Medicare website. This has been compared with the Australasian Vascular Audit dataset to establish completeness of data collection. Quality of data collected has been verified by comparing accuracy of data fields with that present in patient records in a 5% random sample. For the 2 years studied, there was a 63% capture rate in Australia for all patients. In the private sector, only 50% of patients were captured with a significant decrease noted in 2013. The quality of data entered had a 2.6% error rate. There is a need to increase compliance with vascular audit in Australia and data accuracy is acceptable but could be improved. © 2014 Royal Australasian College of Surgeons.

  18. Spent reactor fuel benchmark composition data for code validation

    International Nuclear Information System (INIS)

    Bierman, S.R.

    1991-09-01

    To establish criticality safety margins utilizing burnup credit in the storage and transport of spent reactor fuels requires a knowledge of the uncertainty in the calculated fuel composition used in making the reactivity assessment. To provide data for validating such calculated burnup fuel compositions, radiochemical assays are being obtained as part of the United States Department of Energy From-Reactor Cask Development Program. Destructive assay data are being obtained from representative reactor fuels having experienced irradiation exposures up to about 55 GWD/MTM. Assay results and associated operating histories on the initial three samples analyzed in this effort are presented. The three samples were taken from different axial regions of the same fuel rod and represent radiation exposures of about 27, 37, and 44 GWD/MTM. The data are presented in a benchmark type format to facilitate identification/referencing and computer code input

  19. Soil Moisture Active Passive Mission L4_SM Data Product Assessment (Version 2 Validated Release)

    Science.gov (United States)

    Reichle, Rolf Helmut; De Lannoy, Gabrielle J. M.; Liu, Qing; Ardizzone, Joseph V.; Chen, Fan; Colliander, Andreas; Conaty, Austin; Crow, Wade; Jackson, Thomas; Kimball, John; hide

    2016-01-01

    During the post-launch SMAP calibration and validation (Cal/Val) phase there are two objectives for each science data product team: 1) calibrate, verify, and improve the performance of the science algorithm, and 2) validate the accuracy of the science data product as specified in the science requirements and according to the Cal/Val schedule. This report provides an assessment of the SMAP Level 4 Surface and Root Zone Soil Moisture Passive (L4_SM) product specifically for the product's public Version 2 validated release scheduled for 29 April 2016. The assessment of the Version 2 L4_SM data product includes comparisons of SMAP L4_SM soil moisture estimates with in situ soil moisture observations from core validation sites and sparse networks. The assessment further includes a global evaluation of the internal diagnostics from the ensemble-based data assimilation system that is used to generate the L4_SM product. This evaluation focuses on the statistics of the observation-minus-forecast (O-F) residuals and the analysis increments. Together, the core validation site comparisons and the statistics of the assimilation diagnostics are considered primary validation methodologies for the L4_SM product. Comparisons against in situ measurements from regional-scale sparse networks are considered a secondary validation methodology because such in situ measurements are subject to up-scaling errors from the point-scale to the grid cell scale of the data product. Based on the limited set of core validation sites, the wide geographic range of the sparse network sites, and the global assessment of the assimilation diagnostics, the assessment presented here meets the criteria established by the Committee on Earth Observing Satellites for Stage 2 validation and supports the validated release of the data. An analysis of the time average surface and root zone soil moisture shows that the global pattern of arid and humid regions are captured by the L4_SM estimates. Results from the

  20. Data validation summary report for the 100-BC-5 Operable Unit Round 8 Groundwater Sampling

    International Nuclear Information System (INIS)

    Kearney, A.T.

    1996-03-01

    The information provided in this validation summary report includes data from the chemical analyses of samples from the 100-BC-5 Operable Unit Round 8 Groundwater Sampling Investigation. All of the data from this sampling event and their related quality assurance samples were reviewed and validated to verify that the reported sample results were of sufficient quality to support decisions regarding remedial actions performed at this site. Sample analyses included metals, general chemistry and radiochemistry

  1. Validation of the H-SAF precipitation product H03 over Greece using rain gauge data

    Science.gov (United States)

    Feidas, H.; Porcu, F.; Puca, S.; Rinollo, A.; Lagouvardos, C.; Kotroni, V.

    2018-01-01

    This paper presents an extensive validation of the combined infrared/microwave H-SAF (EUMETSAT Satellite Application Facility on Support to Operational Hydrology and Water Management) precipitation product H03, for a 1-year period, using gauge observations from a relatively dense network of 233 stations over Greece. First, the quality of the interpolated data used to validate the precipitation product is assessed and a quality index is constructed based on parameters such as the density of the station network and the orography. Then, a validation analysis is conducted based on comparisons of satellite (H03) with interpolated rain gauge data to produce continuous and multi-categorical statistics at monthly and annual timescales by taking into account the different geophysical characteristics of the terrain (land, coast, sea, elevation). Finally, the impact of the quality of interpolated data on the validation statistics is examined in terms of different configurations of the interpolation model and the rain gauge network characteristics used in the interpolation. The possibility of using a quality index of the interpolated data as a filter in the validation procedure is also investigated. The continuous validation statistics show yearly root mean squared error (RMSE) and mean absolute error (MAE) corresponding to the 225 and 105 % of the mean rain rate, respectively. Mean error (ME) indicates a slight overall tendency for underestimation of the rain gauge rates, which takes large values for the high rain rates. In general, the H03 algorithm cannot retrieve very well the light (10 mm/h) precipitation. The poor correlation between satellite and gauge data points to algorithm problems in co-locating precipitation patterns. Seasonal comparison shows that retrieval errors are lower for cold months than in the summer months of the year. The multi-categorical statistics indicate that the H03 algorithm is able to discriminate efficiently the rain from the no rain events

  2. Validation of simulation codes for future systems: motivations, approach, and the role of nuclear data

    International Nuclear Information System (INIS)

    Palmiotti, G.; Salvatores, M.; Aliberti, G.

    2007-01-01

    The validation of advanced simulation tools will still play a very significant role in several areas of reactor system analysis. This is the case of reactor physics and neutronics, where nuclear data uncertainties still play a crucial role for many core and fuel cycle parameters. The present paper gives a summary of validation motivations, objectives and approach. A validation effort is in particular necessary in the frame of advanced (e.g. Generation-IV or GNEP) reactors and associated fuel cycles assessment and design. Validation of simulation codes is complementary to the 'verification' process. In fact, 'verification' addresses the question 'are we solving the equations correctly' while validation addresses the question 'are we solving the correct equations with the correct parameters'. Verification implies comparisons with 'reference' equation solutions or with analytical solutions, when they exist. Most of what is called 'numerical validation' falls in this category. Validation strategies differ according to the relative weight of the methods and of the parameters that enter into the simulation tools. Most validation is based on experiments, and the field of neutronics where a 'robust' physics description model exists and which is function of 'input' parameters not fully known, will be the focus of this paper. In fact, in the case of reactor core, shielding and fuel cycle physics the model (theory) is well established (the Boltzmann and Bateman equations) and the parameters are the nuclear cross-sections, decay data etc. Two types of validation approaches can and have been used: (a) Mock-up experiments ('global' validation): need for a very close experimental simulation of a reference configuration. Bias factors cannot be extrapolated beyond reference configuration; (b) Use of 'clean', 'representative' integral experiments ('bias factor and adjustment' method). Allows to define bias factors, uncertainties and can be used for a wide range of applications. It

  3. A New Statistical Method to Determine the Degree of Validity of Health Economic Model Outcomes against Empirical Data.

    NARCIS (Netherlands)

    Corro Ramos, Isaac; van Voorn, George A K; Vemer, Pepijn; Feenstra, Talitha L; Al, Maiwenn J

    2017-01-01

    The validation of health economic (HE) model outcomes against empirical data is of key importance. Although statistical testing seems applicable, guidelines for the validation of HE models lack guidance on statistical validation, and actual validation efforts often present subjective judgment of

  4. Data validation report for the 100-KR-4 operable unit first quarter, 1994

    International Nuclear Information System (INIS)

    Krug, A.D.

    1994-01-01

    Samples were obtained from the 100-KR-4 Operable Unit first Quarter 1994 Groundwater Sampling event. The data from the chemical analysis of fifty-eight samples from this sampling event and their related quality assurance samples were reviewed and validated to verify that reported samples results were of sufficient quality to support decisions regarding remedial actions performed at this site. Information fro the sampling event and the information validation processes are presented in this document

  5. Endogenous protein "barcode" for data validation and normalization in quantitative MS analysis.

    Science.gov (United States)

    Lee, Wooram; Lazar, Iulia M

    2014-07-01

    Quantitative proteomic experiments with mass spectrometry detection are typically conducted by using stable isotope labeling and label-free quantitation approaches. Proteins with housekeeping functions and stable expression level such actin, tubulin, and glyceraldehyde-3-phosphate dehydrogenase are frequently used as endogenous controls. Recent studies have shown that the expression level of such common housekeeping proteins is, in fact, dependent on various factors such as cell type, cell cycle, or disease status and can change in response to a biochemical stimulation. The interference of such phenomena can, therefore, substantially compromise their use for data validation, alter the interpretation of results, and lead to erroneous conclusions. In this work, we advance the concept of a protein "barcode" for data normalization and validation in quantitative proteomic experiments. The barcode comprises a novel set of proteins that was generated from cell cycle experiments performed with MCF7, an estrogen receptor positive breast cancer cell line, and MCF10A, a nontumorigenic immortalized breast cell line. The protein set was selected from a list of ~3700 proteins identified in different cellular subfractions and cell cycle stages of MCF7/MCF10A cells, based on the stability of spectral count data generated with an LTQ ion trap mass spectrometer. A total of 11 proteins qualified as endogenous standards for the nuclear and 62 for the cytoplasmic barcode, respectively. The validation of the protein sets was performed with a complementary SKBR3/Her2+ cell line.

  6. Validating hospital antibiotic purchasing data as a metric of inpatient antibiotic use.

    Science.gov (United States)

    Tan, Charlie; Ritchie, Michael; Alldred, Jason; Daneman, Nick

    2016-02-01

    Antibiotic purchasing data are a widely used, but unsubstantiated, measure of antibiotic consumption. To validate this source, we compared purchasing data from hospitals and external medical databases with patient-level dispensing data. Antibiotic purchasing and dispensing data from internal hospital records and purchasing data from IMS Health were obtained for two hospitals between May 2013 and April 2015. Internal purchasing data were validated against dispensing data, and IMS data were compared with both internal metrics. Scatterplots of individual antimicrobial data points were generated; Pearson's correlation and linear regression coefficients were computed. A secondary analysis re-examined these correlations over shorter calendar periods. Internal purchasing data were strongly correlated with dispensing data, with correlation coefficients of 0.90 (95% CI = 0.83-0.95) and 0.98 (95% CI = 0.95-0.99) at hospitals A and B, respectively. Although dispensing data were consistently lower than purchasing data, this was attributed to a single antibiotic at both hospitals. IMS data were favourably correlated with, but underestimated, internal purchasing and dispensing data. This difference was accounted for by eight antibiotics for which direct sales from some manufacturers were not included in the IMS database. The correlation between purchasing and dispensing data was consistent across periods as short as 3 months, but not at monthly intervals. Both internal and external antibiotic purchasing data are strongly correlated with dispensing data. If outliers are accounted for appropriately, internal purchasing data could be used for cost-effective evaluation of antimicrobial stewardship programmes, and external data sets could be used for surveillance and research across geographical regions. © The Author 2015. Published by Oxford University Press on behalf of the British Society for Antimicrobial Chemotherapy. All rights reserved. For Permissions, please e

  7. Data validation and risk assessment -- some pitfalls when evaluating VOC measurements

    International Nuclear Information System (INIS)

    Korte, N.; Kearl, P.

    1993-01-01

    Data validation, as described in Environmental Protection Agency (EPA) protocols under the Contract Laboratory Program (CLP), yields false confidence in the data and drives up costs while providing little benefit (Korte and Brown 1992). Commonly, these data are then used to perform a risk assessment. Much of the published guidance for risk assessments in and arid soils is inadequate because it does not take into account vapor migration due to density-driven flow (Korte and others 1992). Investigations into both of these problems have been performed by personnel of Oak Ridge National Laboratory (ORNL) and are described in this presentation

  8. Modification and Validation of an Automotive Data Processing Unit, Compessed Video System, and Communications Equipment

    Energy Technology Data Exchange (ETDEWEB)

    Carter, R.J.

    1997-04-01

    The primary purpose of the "modification and validation of an automotive data processing unit (DPU), compressed video system, and communications equipment" cooperative research and development agreement (CRADA) was to modify and validate both hardware and software, developed by Scientific Atlanta, Incorporated (S-A) for defense applications (e.g., rotary-wing airplanes), for the commercial sector surface transportation domain (i.e., automobiles and trucks). S-A also furnished a state-of-the-art compressed video digital storage and retrieval system (CVDSRS), and off-the-shelf data storage and transmission equipment to support the data acquisition system for crash avoidance research (DASCAR) project conducted by Oak Ridge National Laboratory (ORNL). In turn, S-A received access to hardware and technology related to DASCAR. DASCAR was subsequently removed completely and installation was repeated a number of times to gain an accurate idea of complete installation, operation, and removal of DASCAR. Upon satisfactory completion of the DASCAR construction and preliminary shakedown, ORNL provided NHTSA with an operational demonstration of DASCAR at their East Liberty, OH test facility. The demonstration included an on-the-road demonstration of the entire data acquisition system using NHTSA'S test track. In addition, the demonstration also consisted of a briefing, containing the following: ORNL generated a plan for validating the prototype data acquisition system with regard to: removal of DASCAR from an existing vehicle, and installation and calibration in other vehicles; reliability of the sensors and systems; data collection and transmission process (data integrity); impact on the drivability of the vehicle and obtrusiveness of the system to the driver; data analysis procedures; conspicuousness of the vehicle to other drivers; and DASCAR installation and removal training and documentation. In order to identify any operational problems not captured by the systems

  9. Impacts of Sample Design for Validation Data on the Accuracy of Feedforward Neural Network Classification

    Directory of Open Access Journals (Sweden)

    Giles M. Foody

    2017-08-01

    Full Text Available Validation data are often used to evaluate the performance of a trained neural network and used in the selection of a network deemed optimal for the task at-hand. Optimality is commonly assessed with a measure, such as overall classification accuracy. The latter is often calculated directly from a confusion matrix showing the counts of cases in the validation set with particular labelling properties. The sample design used to form the validation set can, however, influence the estimated magnitude of the accuracy. Commonly, the validation set is formed with a stratified sample to give balanced classes, but also via random sampling, which reflects class abundance. It is suggested that if the ultimate aim is to accurately classify a dataset in which the classes do vary in abundance, a validation set formed via random, rather than stratified, sampling is preferred. This is illustrated with the classification of simulated and remotely-sensed datasets. With both datasets, statistically significant differences in the accuracy with which the data could be classified arose from the use of validation sets formed via random and stratified sampling (z = 2.7 and 1.9 for the simulated and real datasets respectively, for both p < 0.05%. The accuracy of the classifications that used a stratified sample in validation were smaller, a result of cases of an abundant class being commissioned into a rarer class. Simple means to address the issue are suggested.

  10. Criteria of the validation of experimental and evaluated covariance data

    International Nuclear Information System (INIS)

    Badikov, S.

    2008-01-01

    The criteria of the validation of experimental and evaluated covariance data are reviewed. In particular: a) the criterion of the positive definiteness for covariance matrices, b) the relationship between the 'integral' experimental and estimated uncertainties, c) the validity of the statistical invariants, d) the restrictions imposed to correlations between experimental errors, are described. Applying these criteria in nuclear data evaluation was considered and 4 particular points have been examined. First preserving positive definiteness of covariance matrices in case of arbitrary transformation of a random vector was considered, properties of the covariance matrices in operations widely used in neutron and reactor physics (splitting and collapsing energy groups, averaging the physical values over energy groups, estimation parameters on the basis of measurements by means of generalized least squares method) were studied. Secondly, an algorithm for comparison of experimental and estimated 'integral' uncertainties was developed, square root of determinant of a covariance matrix is recommended for use in nuclear data evaluation as a measure of 'integral' uncertainty for vectors of experimental and estimated values. Thirdly, a set of statistical invariants-values which are preserved in statistical processing was presented. And fourthly, the inequality that signals a correlation between experimental errors that leads to unphysical values is given. An application is given concerning the cross-section of the (n,t) reaction on Li 6 with a neutron incident energy comprised between 1 and 100 keV

  11. Generation of integral experiment covariance data and their impact on criticality safety validation

    Energy Technology Data Exchange (ETDEWEB)

    Stuke, Maik; Peters, Elisabeth; Sommer, Fabian

    2016-11-15

    The quantification of statistical dependencies in data of critical experiments and how to account for them properly in validation procedures has been discussed in the literature by various groups. However, these subjects are still an active topic in the Expert Group on Uncertainty Analysis for Criticality Safety Assessment (UACSA) of the OECDNEA Nuclear Science Committee. The latter compiles and publishes the freely available experimental data collection, the International Handbook of Evaluated Criticality Safety Benchmark Experiments, ICSBEP. Most of the experiments were performed as series and share parts of experimental setups, consequently leading to correlation effects in the results. The correct consideration of correlated data seems to be inevitable if the experimental data in a validation procedure is limited or one cannot rely on a sufficient number of uncorrelated data sets, e.g. from different laboratories using different setups. The general determination of correlations and the underlying covariance data as well as the consideration of them in a validation procedure is the focus of the following work. We discuss and demonstrate possible effects on calculated k{sub eff}'s, their uncertainties, and the corresponding covariance matrices due to interpretation of evaluated experimental data and its translation into calculation models. The work shows effects of various modeling approaches, varying distribution functions of parameters and compares and discusses results from the applied Monte-Carlo sampling method with available data on correlations. Our findings indicate that for the reliable determination of integral experimental covariance matrices or the correlation coefficients a detailed study of the underlying experimental data, the modeling approach and assumptions made, and the resulting sensitivity analysis seems to be inevitable. Further, a Bayesian method is discussed to include integral experimental covariance data when estimating an

  12. Generation of integral experiment covariance data and their impact on criticality safety validation

    International Nuclear Information System (INIS)

    Stuke, Maik; Peters, Elisabeth; Sommer, Fabian

    2016-11-01

    The quantification of statistical dependencies in data of critical experiments and how to account for them properly in validation procedures has been discussed in the literature by various groups. However, these subjects are still an active topic in the Expert Group on Uncertainty Analysis for Criticality Safety Assessment (UACSA) of the OECDNEA Nuclear Science Committee. The latter compiles and publishes the freely available experimental data collection, the International Handbook of Evaluated Criticality Safety Benchmark Experiments, ICSBEP. Most of the experiments were performed as series and share parts of experimental setups, consequently leading to correlation effects in the results. The correct consideration of correlated data seems to be inevitable if the experimental data in a validation procedure is limited or one cannot rely on a sufficient number of uncorrelated data sets, e.g. from different laboratories using different setups. The general determination of correlations and the underlying covariance data as well as the consideration of them in a validation procedure is the focus of the following work. We discuss and demonstrate possible effects on calculated k eff 's, their uncertainties, and the corresponding covariance matrices due to interpretation of evaluated experimental data and its translation into calculation models. The work shows effects of various modeling approaches, varying distribution functions of parameters and compares and discusses results from the applied Monte-Carlo sampling method with available data on correlations. Our findings indicate that for the reliable determination of integral experimental covariance matrices or the correlation coefficients a detailed study of the underlying experimental data, the modeling approach and assumptions made, and the resulting sensitivity analysis seems to be inevitable. Further, a Bayesian method is discussed to include integral experimental covariance data when estimating an application

  13. WASP (Write a Scientific Paper) using Excel - 1: Data entry and validation.

    Science.gov (United States)

    Grech, Victor

    2018-02-01

    Data collection for the purposes of analysis, after the planning and execution of a research study, commences with data input and validation. The process of data entry and analysis may appear daunting to the uninitiated, but as pointed out in the 1970s in a series of papers by British Medical Journal Deputy Editor TDV Swinscow, modern hardware and software (he was then referring to the availability of hand calculators) permits the performance of statistical testing outside a computer laboratory. In this day and age, modern software, such as the ubiquitous and almost universally familiar Microsoft Excel™ greatly facilitates this process. This first paper comprises the first of a collection of papers which will emulate Swinscow's series, in his own words, "addressed to readers who want to start at the beginning, not to those who are already skilled statisticians." These papers will have less focus on the actual arithmetic, and more emphasis on how to actually implement simple statistics, step by step, using Excel, thereby constituting the equivalent of Swinscow's papers in the personal computer age. Data entry can be facilitated by several underutilised features in Excel. This paper will explain Excel's little-known form function, data validation implementation at input stage, simple coding tips and data cleaning tools. Copyright © 2018 Elsevier B.V. All rights reserved.

  14. Validating emergency department vital signs using a data quality engine for data warehouse.

    Science.gov (United States)

    Genes, N; Chandra, D; Ellis, S; Baumlin, K

    2013-01-01

    Vital signs in our emergency department information system were entered into free-text fields for heart rate, respiratory rate, blood pressure, temperature and oxygen saturation. We sought to convert these text entries into a more useful form, for research and QA purposes, upon entry into a data warehouse. We derived a series of rules and assigned quality scores to the transformed values, conforming to physiologic parameters for vital signs across the age range and spectrum of illness seen in the emergency department. Validating these entries revealed that 98% of free-text data had perfect quality scores, conforming to established vital sign parameters. Average vital signs varied as expected by age. Degradations in quality scores were most commonly attributed logging temperature in Fahrenheit instead of Celsius; vital signs with this error could still be transformed for use. Errors occurred more frequently during periods of high triage, though error rates did not correlate with triage volume. In developing a method for importing free-text vital sign data from our emergency department information system, we now have a data warehouse with a broad array of quality-checked vital signs, permitting analysis and correlation with demographics and outcomes.

  15. A systematic review of validated methods to capture acute bronchospasm using administrative or claims data.

    Science.gov (United States)

    Sharifi, Mona; Krishanswami, Shanthi; McPheeters, Melissa L

    2013-12-30

    To identify and assess billing, procedural, or diagnosis code, or pharmacy claim-based algorithms used to identify acute bronchospasm in administrative and claims databases. We searched the MEDLINE database from 1991 to September 2012 using controlled vocabulary and key terms related to bronchospasm, wheeze and acute asthma. We also searched the reference lists of included studies. Two investigators independently assessed the full text of studies against pre-determined inclusion criteria. Two reviewers independently extracted data regarding participant and algorithm characteristics. Our searches identified 677 citations of which 38 met our inclusion criteria. In these 38 studies, the most commonly used ICD-9 code was 493.x. Only 3 studies reported any validation methods for the identification of bronchospasm, wheeze or acute asthma in administrative and claims databases; all were among pediatric populations and only 2 offered any validation statistics. Some of the outcome definitions utilized were heterogeneous and included other disease based diagnoses, such as bronchiolitis and pneumonia, which are typically of an infectious etiology. One study offered the validation of algorithms utilizing Emergency Department triage chief complaint codes to diagnose acute asthma exacerbations with ICD-9 786.07 (wheezing) revealing the highest sensitivity (56%), specificity (97%), PPV (93.5%) and NPV (76%). There is a paucity of studies reporting rigorous methods to validate algorithms for the identification of bronchospasm in administrative data. The scant validated data available are limited in their generalizability to broad-based populations. Copyright © 2013 Elsevier Ltd. All rights reserved.

  16. 75 FR 59294 - Comment Request for Information Collection for The Data Validation Requirement for Employment and...

    Science.gov (United States)

    2010-09-27

    ... and reliable information on program activities and outcomes is available. Data validation is intended... handbooks provide detailed information on software installation, building and importing a validation file... DEPARTMENT OF LABOR Employment and Training Administration Comment Request for Information...

  17. A Systematic Review of Validated Methods for Identifying Cerebrovascular Accident or Transient Ischemic Attack Using Administrative Data

    Science.gov (United States)

    Andrade, Susan E.; Harrold, Leslie R.; Tjia, Jennifer; Cutrona, Sarah L.; Saczynski, Jane S.; Dodd, Katherine S.; Goldberg, Robert J.; Gurwitz, Jerry H.

    2012-01-01

    Purpose To perform a systematic review of the validity of algorithms for identifying cerebrovascular accidents (CVAs) or transient ischemic attacks (TIAs) using administrative and claims data. Methods PubMed and Iowa Drug Information Service (IDIS) searches of the English language literature were performed to identify studies published between 1990 and 2010 that evaluated the validity of algorithms for identifying CVAs (ischemic and hemorrhagic strokes, intracranial hemorrhage and subarachnoid hemorrhage) and/or TIAs in administrative data. Two study investigators independently reviewed the abstracts and articles to determine relevant studies according to pre-specified criteria. Results A total of 35 articles met the criteria for evaluation. Of these, 26 articles provided data to evaluate the validity of stroke, 7 reported the validity of TIA, 5 reported the validity of intracranial bleeds (intracerebral hemorrhage and subarachnoid hemorrhage), and 10 studies reported the validity of algorithms to identify the composite endpoints of stroke/TIA or cerebrovascular disease. Positive predictive values (PPVs) varied depending on the specific outcomes and algorithms evaluated. Specific algorithms to evaluate the presence of stroke and intracranial bleeds were found to have high PPVs (80% or greater). Algorithms to evaluate TIAs in adult populations were generally found to have PPVs of 70% or greater. Conclusions The algorithms and definitions to identify CVAs and TIAs using administrative and claims data differ greatly in the published literature. The choice of the algorithm employed should be determined by the stroke subtype of interest. PMID:22262598

  18. Critical Evaluation of Validation Rules Automated Extraction from Data

    Directory of Open Access Journals (Sweden)

    David Pejcoch

    2014-10-01

    Full Text Available The goal of this article is to critically evaluate a possibility of automatic extraction of such kind of rules which could be later used within a Data Quality Management process for validation of records newly incoming to Information System. For practical demonstration the 4FT-Miner procedure implemented in LISpMiner System was chosen. A motivation for this task is the potential simplification of projects focused on Data Quality Management. Initially, this article is going to critically evaluate a possibility of fully automated extraction with the aim to identify strengths and weaknesses of this approach in comparison to its alternative, when at least some a priori knowledge is available. As a result of practical implementation, this article provides design of recommended process which would be used as a guideline for future projects. Also the question of how to store and maintain extracted rules and how to integrate them with existing tools supporting Data Quality Management is discussed

  19. Validation of Simulation Codes for Future Systems: Motivations, Approach and the Role of Nuclear Data

    International Nuclear Information System (INIS)

    G. Palmiotti; M. Salvatores; G. Aliberti

    2007-01-01

    The validation of advanced simulation tools will still play a very significant role in several areas of reactor system analysis. This is the case of reactor physics and neutronics, where nuclear data uncertainties still play a crucial role for many core and fuel cycle parameters. The present paper gives a summary of validation motivations, objectives and approach. A validation effort is in particular necessary in the frame of advanced (e.g. Generation-IV or GNEP) reactors and associated fuel cycles assessment and design

  20. Stochastic modeling of oligodendrocyte generation in cell culture: model validation with time-lapse data

    Directory of Open Access Journals (Sweden)

    Noble Mark

    2006-05-01

    Full Text Available Abstract Background The purpose of this paper is two-fold. The first objective is to validate the assumptions behind a stochastic model developed earlier by these authors to describe oligodendrocyte generation in cell culture. The second is to generate time-lapse data that may help biomathematicians to build stochastic models of cell proliferation and differentiation under other experimental scenarios. Results Using time-lapse video recording it is possible to follow the individual evolutions of different cells within each clone. This experimental technique is very laborious and cannot replace model-based quantitative inference from clonal data. However, it is unrivalled in validating the structure of a stochastic model intended to describe cell proliferation and differentiation at the clonal level. In this paper, such data are reported and analyzed for oligodendrocyte precursor cells cultured in vitro. Conclusion The results strongly support the validity of the most basic assumptions underpinning the previously proposed model of oligodendrocyte development in cell culture. However, there are some discrepancies; the most important is that the contribution of progenitor cell death to cell kinetics in this experimental system has been underestimated.

  1. Preliminary Validation of the MATRA-LMR Code Using Existing Sodium-Cooled Experimental Data

    International Nuclear Information System (INIS)

    Choi, Sun Rock; Kim, Sangji

    2014-01-01

    The main objective of the SFR prototype plant is to verify TRU metal fuel performance, reactor operation, and transmutation ability of high-level wastes. The core thermal-hydraulic design is used to ensure the safe fuel performance during the whole plant operation. The fuel design limit is highly dependent on both the maximum cladding temperature and the uncertainties of the design parameters. Therefore, an accurate temperature calculation in each subassembly is highly important to assure a safe and reliable operation of the reactor systems. The current core thermalhydraulic design is mainly performed using the SLTHEN (Steady-State LMR Thermal-Hydraulic Analysis Code Based on ENERGY Model) code, which has been already validated using the existing sodium-cooled experimental data. In addition to the SLTHEN code, a detailed analysis is performed using the MATRA-LMR (Multichannel Analyzer for Transient and steady-state in Rod Array-Liquid Metal Reactor) code. In this work, the MATRA-LMR code is validated for a single subassembly evaluation using the previous experimental data. The MATRA-LMR code has been validated using existing sodium-cooled experimental data. The results demonstrate that the design code appropriately predicts the temperature distributions compared with the experimental values. Major differences are observed in the experiments with the large pin number due to the radial-wise mixing difference

  2. Analysis and classification of data sets for calibration and validation of agro-ecosystem models

    DEFF Research Database (Denmark)

    Kersebaum, K C; Boote, K J; Jorgenson, J S

    2015-01-01

    Experimental field data are used at different levels of complexity to calibrate, validate and improve agro-ecosystem models to enhance their reliability for regional impact assessment. A methodological framework and software are presented to evaluate and classify data sets into four classes regar...

  3. ACE-FTS version 3.0 data set: validation and data processing update

    Directory of Open Access Journals (Sweden)

    Claire Waymark

    2014-01-01

    Full Text Available On 12 August 2003, the Canadian-led Atmospheric Chemistry Experiment (ACE was launched into a 74° inclination orbit at 650 km with the mission objective to measure atmospheric composition using infrared and UV-visible spectroscopy (Bernath et al. 2005. The ACE mission consists of two main instruments, ACE-FTS and MAESTRO (McElroy et al. 2007, which are being used to investigate the chemistry and dynamics of the Earth’s atmosphere.  Here, we focus on the high resolution (0.02 cm-1 infrared Fourier Transform Spectrometer, ACE-FTS, that measures in the 750-4400 cm-1 (2.2 to 13.3 µm spectral region.  This instrument has been making regular solar occultation observations for more than nine years.  The current ACE-FTS data version (version 3.0 provides profiles of temperature and volume mixing ratios (VMRs of more than 30 atmospheric trace gas species, as well as 20 subsidiary isotopologues of the most abundant trace atmospheric constituents over a latitude range of ~85°N to ~85°S.  This letter describes the current data version and recent validation comparisons and provides a description of our planned updates for the ACE-FTS data set. [...

  4. San Onofre PWR Data for Code Validation of MOX Fuel Depletion Analyses - Revision 1

    International Nuclear Information System (INIS)

    Hermann, O.W.

    2000-01-01

    The isotopic composition of mixed-oxide fuel (fabricated with both uranium and plutonium isotopes) discharged from reactors is of interest to the Fissile Material Disposition Program. The validation of depletion codes used to predict isotopic compositions of MOX fuel, similar to studies concerning uranium-only fueled reactors, thus, is very important. The EEI-Westinghouse Plutonium Recycle Demonstration Program was conducted to examine the use of MOX fuel in the San Onofre PWR, Unit I, during cycles 2 and 3. The data, usually required as input to depletion codes, either one-dimensional or lattice codes, were taken from various sources and compiled into this report. Where data were either lacking or determined inadequate, the appropriate data were supplied from other references. The scope of the reactor operations and design data, in addition to the isotopic analyses, was considered to be of sufficient quality for depletion code validation

  5. Savannah River Site offsite hazardous waste shipment data validation report. Revision 1

    International Nuclear Information System (INIS)

    Casey, C.; Kudera, D.E.; Page, L.A.; Rohe, M.J.

    1995-05-01

    The objective of this data validation is to verify that waste shipments reported in response to the US Department of Energy Headquarters data request are properly categorized according to DOE-HQ definitions. This report documents all findings and actions resulting from the independent review of the Savannah River Site data submittal, and provides a summary of the SRS data submittal and data validation strategy. The overall hazardous waste management and offsite release process from 1987--1991 is documented, along with an identification and description of the hazardous waste generation facilities. SRS did not ship any hazardous waste offsite before 1987. Sampling and analysis and surface surveying procedures and techniques used in determining offsite releasability of the shipments are also described in this report. SRS reported 150 manifested waste shipments from 1984 to 1991 that included 4,755 drums or lab packs and 13 tankers. Of these waste items, this report categorizes 4,251 as clean (including 12 tankers), 326 as likely clean, 138 as likely radioactive, and 55 as radioactive (including one tanker). Although outside the original scope of this report, 14 manifests from 1992 and 1993 are included, covering 393 drums or lab packs and seven tankers. From the 1992--1993 shipments, 58 drums or lab packs are categorized as radioactive and 16 drums are categorized as likely radioactive. The remainder are categorized as clean

  6. Savannah River Site offsite hazardous waste shipment data validation report. Revision 1

    Energy Technology Data Exchange (ETDEWEB)

    Casey, C.; Kudera, D.E.; Page, L.A.; Rohe, M.J.

    1995-05-01

    The objective of this data validation is to verify that waste shipments reported in response to the US Department of Energy Headquarters data request are properly categorized according to DOE-HQ definitions. This report documents all findings and actions resulting from the independent review of the Savannah River Site data submittal, and provides a summary of the SRS data submittal and data validation strategy. The overall hazardous waste management and offsite release process from 1987--1991 is documented, along with an identification and description of the hazardous waste generation facilities. SRS did not ship any hazardous waste offsite before 1987. Sampling and analysis and surface surveying procedures and techniques used in determining offsite releasability of the shipments are also described in this report. SRS reported 150 manifested waste shipments from 1984 to 1991 that included 4,755 drums or lab packs and 13 tankers. Of these waste items, this report categorizes 4,251 as clean (including 12 tankers), 326 as likely clean, 138 as likely radioactive, and 55 as radioactive (including one tanker). Although outside the original scope of this report, 14 manifests from 1992 and 1993 are included, covering 393 drums or lab packs and seven tankers. From the 1992--1993 shipments, 58 drums or lab packs are categorized as radioactive and 16 drums are categorized as likely radioactive. The remainder are categorized as clean.

  7. Data validation report for the 100-HR-3 Operable Unit, fifth round groundwater samples

    International Nuclear Information System (INIS)

    Vukelich, S.E.

    1994-01-01

    The data from the chemical analysis of 68 samples from the 100-HR-3 Operable Unit Third Quarter 1993 Groundwater Sampling Investigation and their related quality assurance samples were reviewed and validated to verify that reported sample results were of sufficient quality to support decisions regarding remedial actions performed at the site. Sample analysis included inorganics and general chemical parameters. Fifty three samples were validated for radiochemical parameters

  8. Data validation report for the 100-D Ponds Operable Unit: 100-D ponds sampling

    International Nuclear Information System (INIS)

    Stankovich, M.T.

    1994-01-01

    Westinghouse-Hanford has requested that 100 percent of the Sample Delivery Groups be validated for the 100-D Ponds Operable Unit Sampling Investigation. Therefore the data from the chemical analysis of all 30 samples from this sampling event and their related quality assurance samples were reviewed and validated to verify that reported sample results were of sufficient quality to support decisions regarding remedial actions performed at this site

  9. Data Visualization and Analysis Tools for the Global Precipitation Measurement (GPM) Validation Network

    Science.gov (United States)

    Morris, Kenneth R.; Schwaller, Mathew

    2010-01-01

    The Validation Network (VN) prototype for the Global Precipitation Measurement (GPM) Mission compares data from the Tropical Rainfall Measuring Mission (TRMM) satellite Precipitation Radar (PR) to similar measurements from U.S. and international operational weather radars. This prototype is a major component of the GPM Ground Validation System (GVS). The VN provides a means for the precipitation measurement community to identify and resolve significant discrepancies between the ground radar (GR) observations and similar satellite observations. The VN prototype is based on research results and computer code described by Anagnostou et al. (2001), Bolen and Chandrasekar (2000), and Liao et al. (2001), and has previously been described by Morris, et al. (2007). Morris and Schwaller (2009) describe the PR-GR volume-matching algorithm used to create the VN match-up data set used for the comparisons. This paper describes software tools that have been developed for visualization and statistical analysis of the original and volume matched PR and GR data.

  10. GPM GROUND VALIDATION NASA ER-2 NAVIGATION DATA MC3E V1

    Data.gov (United States)

    National Aeronautics and Space Administration — The GPM Ground Validation NASA ER-2 Navigation Data MC3E dataset contains information recorded by an on board navigation recorder (NavRec). In addition to typical...

  11. CRED Optical Validation Data in the Auau Channel, Hawaii, 2007, to Support Benthic Habitat Mapping

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — Optical validation data were collected using a RCV-150 remotely operated vehicle (ROV) operated by the Hawaii Undersea Research Laboratory (HURL). Data were...

  12. Towards a Generic Information Data Model for Verification, Validation & Accreditation VV&A

    NARCIS (Netherlands)

    Roza, Z.C.; Voogd, J.M.; Giannoulis, C.

    2008-01-01

    The Generic Methodology for Verification, Validation and Acceptance (GM-VV) is intended to provide a common generic framework for making formal and well balanced acceptance decisions on a specific usage of models, simulations and data. GM-VV will offer the international M&S community with a

  13. Prediction, Detection, and Validation of Isotope Clusters in Mass Spectrometry Data

    Directory of Open Access Journals (Sweden)

    Hendrik Treutler

    2016-10-01

    Full Text Available Mass spectrometry is a key analytical platform for metabolomics. The precise quantification and identification of small molecules is a prerequisite for elucidating the metabolism and the detection, validation, and evaluation of isotope clusters in LC-MS data is important for this task. Here, we present an approach for the improved detection of isotope clusters using chemical prior knowledge and the validation of detected isotope clusters depending on the substance mass using database statistics. We find remarkable improvements regarding the number of detected isotope clusters and are able to predict the correct molecular formula in the top three ranks in 92 % of the cases. We make our methodology freely available as part of the Bioconductor packages xcms version 1.50.0 and CAMERA version 1.30.0.

  14. Validation of SCALE-4 criticality sequences using ENDF/B-V data

    International Nuclear Information System (INIS)

    Bowman, S.M.; Wright, R.Q.; DeHart, M.D.; Taniuchi, H.

    1993-01-01

    The SCALE code system developed at Oak Ridge National Laboratory contains criticality safety analysis sequences that include the KENO V.a Monte Carlo code for calculation of the effective multiplication factor. These sequences are widely used for criticality safety analyses performed both in the United States and abroad. The purpose of the current work is to validate the SCALE-4 criticality sequences with an ENDF/B-V cross-section library for future distribution with SCALE-4. The library used for this validation is a broad-group library (44 groups) collapsed from the 238-group SCALE library. Extensive data testing of both the 238-group and the 44-group libraries included 10 fast and 18 thermal CSEWG benchmarks and 5 other fast benchmarks. Both libraries contain approximately 300 nuclides and are, therefore, capable of modeling most systems, including those containing spent fuel or radioactive waste. The validation of the broad-group library used 93 critical experiments as benchmarks. The range of experiments included 60 light-water-reactor fuel rod lattices, 13 mixed-oxide fuel rod lattice, and 15 other low- and high-enriched uranium critical assemblies

  15. The jmzQuantML programming interface and validator for the mzQuantML data standard.

    Science.gov (United States)

    Qi, Da; Krishna, Ritesh; Jones, Andrew R

    2014-03-01

    The mzQuantML standard from the HUPO Proteomics Standards Initiative has recently been released, capturing quantitative data about peptides and proteins, following analysis of MS data. We present a Java application programming interface (API) for mzQuantML called jmzQuantML. The API provides robust bridges between Java classes and elements in mzQuantML files and allows random access to any part of the file. The API provides read and write capabilities, and is designed to be embedded in other software packages, enabling mzQuantML support to be added to proteomics software tools (http://code.google.com/p/jmzquantml/). The mzQuantML standard is designed around a multilevel validation system to ensure that files are structurally and semantically correct for different proteomics quantitative techniques. In this article, we also describe a Java software tool (http://code.google.com/p/mzquantml-validator/) for validating mzQuantML files, which is a formal part of the data standard. © 2014 The Authors. Proteomics published by Wiley-VCH Verlag GmbH & Co. KGaA, Weinheim.

  16. User's Manual for Data for Validating Models for PV Module Performance

    Energy Technology Data Exchange (ETDEWEB)

    Marion, W.; Anderberg, A.; Deline, C.; Glick, S.; Muller, M.; Perrin, G.; Rodriguez, J.; Rummel, S.; Terwilliger, K.; Silverman, T. J.

    2014-04-01

    This user's manual describes performance data measured for flat-plate photovoltaic (PV) modules installed in Cocoa, Florida, Eugene, Oregon, and Golden, Colorado. The data include PV module current-voltage curves and associated meteorological data for approximately one-year periods. These publicly available data are intended to facilitate the validation of existing models for predicting the performance of PV modules, and for the development of new and improved models. For comparing different modeling approaches, using these public data will provide transparency and more meaningful comparisons of the relative benefits.

  17. Validity and reliability of stillbirth data using linked self-reported and administrative datasets.

    Science.gov (United States)

    Hure, Alexis J; Chojenta, Catherine L; Powers, Jennifer R; Byles, Julie E; Loxton, Deborah

    2015-01-01

    A high rate of stillbirth was previously observed in the Australian Longitudinal Study of Women's Health (ALSWH). Our primary objective was to test the validity and reliability of self-reported stillbirth data linked to state-based administrative datasets. Self-reported data, collected as part of the ALSWH cohort born in 1973-1978, were linked to three administrative datasets for women in New South Wales, Australia (n = 4374): the Midwives Data Collection; Admitted Patient Data Collection; and Perinatal Death Review Database. Linkages were obtained from the Centre for Health Record Linkage for the period 1996-2009. True cases of stillbirth were defined by being consistently recorded in two or more independent data sources. Sensitivity, specificity, positive predictive value, negative predictive value, percent agreement, and kappa statistics were calculated for each dataset. Forty-nine women reported 53 stillbirths. No dataset was 100% accurate. The administrative datasets performed better than self-reported data, with high accuracy and agreement. Self-reported data showed high sensitivity (100%) but low specificity (30%), meaning women who had a stillbirth always reported it, but there was also over-reporting of stillbirths. About half of the misreported cases in the ALSWH were able to be removed by identifying inconsistencies in longitudinal data. Data linkage provides great opportunity to assess the validity and reliability of self-reported study data. Conversely, self-reported study data can help to resolve inconsistencies in administrative datasets. Quantifying the strengths and limitations of both self-reported and administrative data can improve epidemiological research, especially by guiding methods and interpretation of findings.

  18. Analytic Validation of Immunohistochemistry Assays: New Benchmark Data From a Survey of 1085 Laboratories.

    Science.gov (United States)

    Stuart, Lauren N; Volmar, Keith E; Nowak, Jan A; Fatheree, Lisa A; Souers, Rhona J; Fitzgibbons, Patrick L; Goldsmith, Jeffrey D; Astles, J Rex; Nakhleh, Raouf E

    2017-09-01

    - A cooperative agreement between the College of American Pathologists (CAP) and the United States Centers for Disease Control and Prevention was undertaken to measure laboratories' awareness and implementation of an evidence-based laboratory practice guideline (LPG) on immunohistochemical (IHC) validation practices published in 2014. - To establish new benchmark data on IHC laboratory practices. - A 2015 survey on IHC assay validation practices was sent to laboratories subscribed to specific CAP proficiency testing programs and to additional nonsubscribing laboratories that perform IHC testing. Specific questions were designed to capture laboratory practices not addressed in a 2010 survey. - The analysis was based on responses from 1085 laboratories that perform IHC staining. Ninety-six percent (809 of 844) always documented validation of IHC assays. Sixty percent (648 of 1078) had separate procedures for predictive and nonpredictive markers, 42.7% (220 of 515) had procedures for laboratory-developed tests, 50% (349 of 697) had procedures for testing cytologic specimens, and 46.2% (363 of 785) had procedures for testing decalcified specimens. Minimum case numbers were specified by 85.9% (720 of 838) of laboratories for nonpredictive markers and 76% (584 of 768) for predictive markers. Median concordance requirements were 95% for both types. For initial validation, 75.4% (538 of 714) of laboratories adopted the 20-case minimum for nonpredictive markers and 45.9% (266 of 579) adopted the 40-case minimum for predictive markers as outlined in the 2014 LPG. The most common method for validation was correlation with morphology and expected results. Laboratories also reported which assay changes necessitated revalidation and their minimum case requirements. - Benchmark data on current IHC validation practices and procedures may help laboratories understand the issues and influence further refinement of LPG recommendations.

  19. ISOTHERMAL AIR INGRESS VALIDATION EXPERIMENTS AT IDAHO NATIONAL LABORATORY: DESCRIPTION AND SUMMARY OF DATA

    International Nuclear Information System (INIS)

    Oh, Chang H.; Kim, Eung S.

    2010-01-01

    Idaho National Laboratory performed air ingress experiments as part of validating computational fluid dynamics code (CFD). An isothermal stratified flow experiment was designed and set to understand stratified flow phenomena in the very high temperature gas cooled reactor (VHTR) and to provide experimental data for validating computer codes. The isothermal experiment focused on three flow characteristics unique in the VHTR air-ingress accident: stratified flow in the horizontal pipe, stratified flow expansion at the pipe and vessel junction, and stratified flow around supporting structures. Brine and sucrose were used as heavy fluids and water was used as light fluids. The density ratios were changed between 0.87 and 0.98. This experiment clearly showed that a stratified flow between heavy and light fluids is generated even for very small density differences. The code was validated by conducting blind CFD simulations and comparing the results to the experimental data. A grid sensitivity study was also performed based on the Richardson extrapolation and the grid convergence index method for modeling confidence. As a result, the calculated current speed showed very good agreement with the experimental data, indicating that the current CFD methods are suitable for predicting density gradient stratified flow phenomena in the air-ingress accident.

  20. Data validation report for the 100-FR-3 Operable Unit, third round groundwater samples

    International Nuclear Information System (INIS)

    Ayres, J.M.

    1994-01-01

    Westinghouse-Hanford has requested that a minimum of 20% of the total number of Sample Delivery Groups be validated for the 100-FR-3 operable Unit Third Round Groundwater sampling investigation. Therefore, the data from the chemical analysis of 51 samples from this sampling event and their related quality assurance samples were reviewed and validated to verify that reported sample results were of sufficient quality to support decisions regarding remedial actions performed at this site. The report is broken down into sections for each chemical analysis and radiochemical analysis type. Each section addresses the data package completeness, holding time adherence, instrument calibration and tuning acceptability, blank results, accuracy, precision, system performance, as well as the compound identification and quantitation. In addition, each section has an overall assessment and summary for the data packages reviewed for the particular chemical/radiochemical analyses. Detailed backup information is provided to the reader by SDG No. and sample number. For each data package, a matrix of chemical analyses per sample number is presented, as well as data qualification summaries

  1. Determination of clouds in MSG data for the validation of clouds in a regional climate model

    OpenAIRE

    Huckle, Roger

    2009-01-01

    Regional climate models (e.g. CLM) can help to asses the influence of the antropogenic climate change on the different regions of the earth. Validation of these models is very important. Satellite data are of great benefit, as data on a global scale and high temporal resolution is available. In this thesis a cloud detection and object based cloud classification for Meteosat Second Generation (MSG) was developed and used to validate CLM clouds. Results show sometimes too many clouds in the CLM.

  2. Model validation of solar PV plant with hybrid data dynamic simulation based on fast-responding generator method

    Directory of Open Access Journals (Sweden)

    Zhao Dawei

    2016-01-01

    Full Text Available In recent years, a significant number of large-scale solar photovoltaic (PV plants have been put into operation or been under planning around the world. The model accuracy of solar PV plant is the key factor to investigate the mutual influences between solar PV plants and a power grid. However, this problem has not been well solved, especially in how to apply the real measurements to validate the models of the solar PV plants. Taking fast-responding generator method as an example, this paper presents a model validation methodology for solar PV plant via the hybrid data dynamic simulation. First, the implementation scheme of hybrid data dynamic simulation suitable for DIgSILENT PowerFactory software is proposed, and then an analysis model of solar PV plant integration based on IEEE 9 system is established. At last, model validation of solar PV plant is achieved by employing hybrid data dynamic simulation. The results illustrate the effectiveness of the proposed method in solar PV plant model validation.

  3. Computer assistance in the collection, validation and manipulation of data for epidemiological studies

    International Nuclear Information System (INIS)

    Salmon, L.; Venn, J.B.

    1987-01-01

    The difficulties encountered in assembling and storing adequate data for a large cohort study of 50,000 radiation and other workers are discussed. A computer database management system was designed to permit the storage of information that could be conflicting and incomplete. The way in which it was used to validate data and to match records from a variety of sources is described. (author)

  4. Landspotting: Social gaming to collect vast amounts of data for satellite validation

    Science.gov (United States)

    Fritz, S.; Purgathofer, P.; Kayali, F.; Fellner, M.; Wimmer, M.; Sturn, T.; Triebnig, G.; Krause, S.; Schindler, F.; Kollegger, M.; Perger, C.; Dürauer, M.; Haberl, W.; See, L.; McCallum, I.

    2012-04-01

    At present there is no single satellite-derived global land cover product that is accurate enough to provide reliable estimates of forest or cropland area to determine, e.g., how much additional land is available to grow biofuels or to tackle problems of food security. The Landspotting Project aims to improve the quality of this land cover information by vastly increasing the amount of in-situ validation data available for calibration and validation of satellite-derived land cover. The Geo-Wiki (Geo-Wiki.org) system currently allows users to compare three satellite derived land cover products and validate them using Google Earth. However, there is presently no incentive for anyone to provide this data so the amount of validation through Geo-Wiki has been limited. However, recent competitions have proven that incentive driven campaigns can rapidly create large amounts of input. The LandSpotting Project is taking a truly innovative approach through the development of the Landspotting game. The game engages users whilst simultaneously collecting a large amount of in-situ land cover information. The development of the game is informed by the current raft of successful social gaming that is available on the internet and as mobile applications, many of which are geo-spatial in nature. Games that are integrated within a social networking site such as Facebook illustrate the power to reach and continually engage a large number of individuals. The number of active Facebook users is estimated to be greater than 400 million, where 100 million are accessing Facebook from mobile devices. The Landspotting Game has similar game mechanics as the famous strategy game "Civilization" (i.e. build, harvest, research, war, diplomacy, etc.). When a player wishes to make a settlement, they must first classify the land cover over the area they wish to settle. As the game is played on the earth surface with Google Maps, we are able to record and store this land cover/land use classification

  5. Validation of iron nuclear data for the neutron calculation of nuclear reactors

    International Nuclear Information System (INIS)

    Vaglio-Gaudard, C.

    2010-01-01

    The GEN-III and GEN-IV reactors will be equipped with heavy reflectors. However, the existing integral validation of the iron nuclear data in the latest JEFF3 European library in the frame of the neutron calculation of the heavy reflector is very partial: some results exist concerning fast reactors but there is no result corresponding to the LWR heavy reflector. No clear trend on the JEFF3 iron cross sections was brought into evidence up to now for fission reactor calculations. Iron nuclear data were completely re-evaluated in the JEFF3 library. Despite the fact that iron is widely used in the nuclear industry, large uncertainties are still associated with its nuclear data, particularly its inelastic cross section which is very important in the neutron slowing down. A validation of 56 Fe nuclear data was performed on the basis of the analysis of integral experiments. Two major critical experiments, the PERLE experiment and the Gas Benchmark, were interpreted with 3D reference Monte-Carlo calculations and the JEFF3.1.1 library. The PERLE experiment was recently performed in the EOLE zero-power facility (CEA Cadarache). This experiment is dedicated to heavy reflector physics in GEN-III light water reactors. It was especially conceived for the validation of iron nuclear data. The Gas Benchmark is representative of a Gas Fast Reactor with a stainless steel reflector (with no fertile blanket) in the MASURCA facility (CEA Cadarache). Radial traverses of reaction rates were measured to characterize flux attenuation at various energies in the reflector. The results of the analysis of both experiments show good agreement between the calculations and the measurements, which is confirmed by the analysis of complementary experiments (ZR-6M, MISTRAL4, CIRANO-ZONA2B). A process of re-estimating the 56 Fe nuclear data was implemented on the basis of feedback from these two experiments and the RDN code. This code relies on a non-linear regression method using an iterative

  6. A multisite validation of whole slide imaging for primary diagnosis using standardized data collection and analysis.

    Science.gov (United States)

    Wack, Katy; Drogowski, Laura; Treloar, Murray; Evans, Andrew; Ho, Jonhan; Parwani, Anil; Montalto, Michael C

    2016-01-01

    Text-based reporting and manual arbitration for whole slide imaging (WSI) validation studies are labor intensive and do not allow for consistent, scalable, and repeatable data collection or analysis. The objective of this study was to establish a method of data capture and analysis using standardized codified checklists and predetermined synoptic discordance tables and to use these methods in a pilot multisite validation study. Fifteen case report form checklists were generated from the College of American Pathology cancer protocols. Prior to data collection, all hypothetical pairwise comparisons were generated, and a level of harm was determined for each possible discordance. Four sites with four pathologists each generated 264 independent reads of 33 cases. Preestablished discordance tables were applied to determine site by site and pooled accuracy, intrareader/intramodality, and interreader intramodality error rates. Over 10,000 hypothetical pairwise comparisons were evaluated and assigned harm in discordance tables. The average difference in error rates between WSI and glass, as compared to ground truth, was 0.75% with a lower bound of 3.23% (95% confidence interval). Major discordances occurred on challenging cases, regardless of modality. The average inter-reader agreement across sites for glass was 76.5% (weighted kappa of 0.68) and for digital it was 79.1% (weighted kappa of 0.72). These results demonstrate the feasibility and utility of employing standardized synoptic checklists and predetermined discordance tables to gather consistent, comprehensive diagnostic data for WSI validation studies. This method of data capture and analysis can be applied in large-scale multisite WSI validations.

  7. Interlaboratory validation data on real-time polymerase chain reaction detection for unauthorized genetically modified papaya line PRSV-YK

    Directory of Open Access Journals (Sweden)

    Kosuke Nakamura

    2016-06-01

    Real-time polymerase chain reaction (PCR detection method for unauthorized genetically modified (GM papaya (Carica papaya L. line PRSV-YK (PRSV-YK detection method was developed using whole genome sequence data (DDBJ Sequenced Read Archive under accession No. PRJDB3976. Interlaboratory validation datasets for PRSV-YK detection method were provided. Data indicating homogeneity of samples prepared for interlaboratory validation were included. Specificity and sensitivity test data for PRSV-YK detection method were also provided.

  8. International integral experiments databases in support of nuclear data and code validation

    International Nuclear Information System (INIS)

    Briggs, J. Blair; Gado, Janos; Hunter, Hamilton; Kodeli, Ivan; Salvatores, Massimo; Sartori, Enrico

    2002-01-01

    The OECD/NEA Nuclear Science Committee (NSC) has identified the need to establish international databases containing all the important experiments that are available for sharing among the specialists. The NSC has set up or sponsored specific activities to achieve this. The aim is to preserve them in an agreed standard format in computer accessible form, to use them for international activities involving validation of current and new calculational schemes including computer codes and nuclear data libraries, for assessing uncertainties, confidence bounds and safety margins, and to record measurement methods and techniques. The databases so far established or in preparation related to nuclear data validation cover the following areas: SINBAD - A Radiation Shielding Experiments database encompassing reactor shielding, fusion blanket neutronics, and accelerator shielding. ICSBEP - International Criticality Safety Benchmark Experiments Project Handbook, with more than 2500 critical configurations with different combination of materials and spectral indices. IRPhEP - International Reactor Physics Experimental Benchmarks Evaluation Project. The different projects are described in the following including results achieved, work in progress and planned. (author)

  9. High spatial resolution satellite observations for validation of MODIS land products: IKONOS observations acquired under the NASA scientific data purchase.

    Science.gov (United States)

    Jeffrey T. Morisette; Jaime E. Nickeson; Paul Davis; Yujie Wang; Yuhong Tian; Curtis E. Woodcock; Nikolay Shabanov; Matthew Hansen; Warren B. Cohen; Doug R. Oetter; Robert E. Kennedy

    2003-01-01

    Phase 1I of the Scientific Data Purchase (SDP) has provided NASA investigators access to data from four different satellite and airborne data sources. The Moderate Resolution Imaging Spectrometer (MODIS) land discipline team (MODLAND) sought to utilize these data in support of land product validation activities with a lbcus on tile EOS Land Validation Core Sites. These...

  10. Validation of a Robust Neural Real-Time Voltage Estimator for Active Distribution Grids on Field Data

    DEFF Research Database (Denmark)

    Pertl, Michael; Douglass, Philip James; Heussen, Kai

    2018-01-01

    network approach for voltage estimation in active distribution grids by means of measured data from two feeders of a real low voltage distribution grid. The approach enables a real-time voltage estimation at locations in the distribution grid, where otherwise only non-real-time measurements are available......The installation of measurements in distribution grids enables the development of data driven methods for the power system. However, these methods have to be validated in order to understand the limitations and capabilities for their use. This paper presents a systematic validation of a neural...

  11. Mathematical model validation of a thermal architecture system connecting east/west radiators by flight data

    International Nuclear Information System (INIS)

    Torres, Alejandro; Mishkinis, Donatas; Kaya, Tarik

    2014-01-01

    A novel satellite thermal architecture connecting the east and west radiators of a geostationary telecommunication satellite via loop heat pipes (LHPs) is flight tested on board the satellite Hispasat 1E. The LHP operating temperature is regulated by using pressure regulating valves (PRVs). The flight data demonstrated the successful operation of the proposed concept. A transient numerical model specifically developed for the design of this system satisfactorily simulated the flight data. The validated mathematical model can be used to design and analyze the thermal behavior of more complex architectures. - Highlights: •A novel spacecraft thermal control architecture is presented. •The east–west radiators of a GEO communications satellite are connected using LHPs. •A transient mathematical model is validated with flight data. •The space flight data proved successful in-orbit operation of the novel architecture. •The model can be used to design/analyze LHP based complex thermal architectures

  12. Validation of US3D for Capsule Aerodynamics using 05-CA Wind Tunnel Test Data

    Science.gov (United States)

    Schwing, Alan

    2012-01-01

    Several comparisons of computational fluid dynamics to wind tunnel test data are shown for the purpose of code validation. The wind tunnel test, 05-CA, uses a 7.66% model of NASA's Multi-Purpose Crew Vehicle in the 11-foot test section of the Ames Unitary Plan Wind tunnel. A variety of freestream conditions over four Mach numbers and three angles of attack are considered. Test data comparisons include time-averaged integrated forces and moments, time-averaged static pressure ports on the surface, and Strouhal Number. The applicability of the US3D code to subsonic and transonic flow over a bluff body is assessed on a comprehensive data set. With close comparison, this work validates US3D for highly separated flows similar to those examined here.

  13. GPM GROUND VALIDATION NOAA S-BAND PROFILER MINUTE DATA MC3E V1

    Data.gov (United States)

    National Aeronautics and Space Administration — The GPM Ground Validation NOAA S-Band Profiler Minute Data MC3E dataset was gathered during the Midlatitude Continental Convective Clouds Experiment (MC3E) in...

  14. The potential of ground gravity measurements to validate GRACE data

    Directory of Open Access Journals (Sweden)

    D. Crossley

    2003-01-01

    Full Text Available New satellite missions are returning high precision, time-varying, satellite measurements of the Earth’s gravity field. The GRACE mission is now in its calibration/- validation phase and first results of the gravity field solutions are imminent. We consider here the possibility of external validation using data from the superconducting gravimeters in the European sub-array of the Global Geodynamics Project (GGP as ‘ground truth’ for comparison with GRACE. This is a pilot study in which we use 14 months of 1-hour data from the beginning of GGP (1 July 1997 to 30 August 1998, when the Potsdam instrument was relocated to South Africa. There are 7 stations clustered in west central Europe, and one station, Metsahovi in Finland. We remove local tides, polar motion, local and global air pressure, and instrument drift and then decimate to 6-hour samples. We see large variations in the time series of 5–10µgal between even some neighboring stations, but there are also common features that correlate well over the 427-day period. The 8 stations are used to interpolate a minimum curvature (gridded surface that extends over the geographical region. This surface shows time and spatial coherency at the level of 2– 4µgal over the first half of the data and 1–2µgal over the latter half. The mean value of the surface clearly shows a rise in European gravity of about 3µgal over the first 150 days and a fairly constant value for the rest of the data. The accuracy of this mean is estimated at 1µgal, which compares favorably with GRACE predictions for wavelengths of 500 km or less. Preliminary studies of hydrology loading over Western Europe shows the difficulty of correlating the local hydrology, which can be highly variable, with large-scale gravity variations.Key words. GRACE, satellite gravity, superconducting gravimeter, GGP, ground truth

  15. Development of an automated testing system for verification and validation of nuclear data

    International Nuclear Information System (INIS)

    Triplett, B. S.; Anghaie, S.; White, M. C.

    2008-01-01

    Verification and validation of nuclear data is critical to the accuracy of both stochastic and deterministic particle transport codes. In order to effectively test a set of nuclear data, the data must be applied to a wide variety of transport problems. Performing this task in a timely, efficient manner is tedious. The nuclear data team at Los Alamos National Laboratory (LANL) in collaboration with the University of Florida is developing a methodology to automate the process of nuclear data verification and validation. The International Criticality Safety Benchmark Experiment Project (ICSBEP) provides a set of criticality problems that may be used to evaluate nuclear data. This process tests a number of data libraries using cases from the ICSBEP benchmark set to demonstrate how automation of these tasks may reduce errors and increase efficiency. The process is driven by an integrated set of Python scripts. Material and geometry data may be read from an existing code input file to generate a standardized template or the template may be generated directly by the user The user specifies the desired precision and other vital problem parameters. The Python scripts generate input decks for multiple transport codes from these templates, run and monitor individual jobs, and parse the relevant output. This output can then be used to generate reports directly or can be stored into a database for later analysis. This methodology eases the burden on the user by reducing the amount of time and effort required for obtaining and compiling calculation results. (authors)

  16. A multisite validation of whole slide imaging for primary diagnosis using standardized data collection and analysis

    Directory of Open Access Journals (Sweden)

    Katy Wack

    2016-01-01

    Full Text Available Context: Text-based reporting and manual arbitration for whole slide imaging (WSI validation studies are labor intensive and do not allow for consistent, scalable, and repeatable data collection or analysis. Objective: The objective of this study was to establish a method of data capture and analysis using standardized codified checklists and predetermined synoptic discordance tables and to use these methods in a pilot multisite validation study. Methods and Study Design: Fifteen case report form checklists were generated from the College of American Pathology cancer protocols. Prior to data collection, all hypothetical pairwise comparisons were generated, and a level of harm was determined for each possible discordance. Four sites with four pathologists each generated 264 independent reads of 33 cases. Preestablished discordance tables were applied to determine site by site and pooled accuracy, intrareader/intramodality, and interreader intramodality error rates. Results: Over 10,000 hypothetical pairwise comparisons were evaluated and assigned harm in discordance tables. The average difference in error rates between WSI and glass, as compared to ground truth, was 0.75% with a lower bound of 3.23% (95% confidence interval. Major discordances occurred on challenging cases, regardless of modality. The average inter-reader agreement across sites for glass was 76.5% (weighted kappa of 0.68 and for digital it was 79.1% (weighted kappa of 0.72. Conclusion: These results demonstrate the feasibility and utility of employing standardized synoptic checklists and predetermined discordance tables to gather consistent, comprehensive diagnostic data for WSI validation studies. This method of data capture and analysis can be applied in large-scale multisite WSI validations.

  17. aMAP is a validated pipeline for registration and segmentation of high-resolution mouse brain data

    Science.gov (United States)

    Niedworok, Christian J.; Brown, Alexander P. Y.; Jorge Cardoso, M.; Osten, Pavel; Ourselin, Sebastien; Modat, Marc; Margrie, Troy W.

    2016-01-01

    The validation of automated image registration and segmentation is crucial for accurate and reliable mapping of brain connectivity and function in three-dimensional (3D) data sets. While validation standards are necessarily high and routinely met in the clinical arena, they have to date been lacking for high-resolution microscopy data sets obtained from the rodent brain. Here we present a tool for optimized automated mouse atlas propagation (aMAP) based on clinical registration software (NiftyReg) for anatomical segmentation of high-resolution 3D fluorescence images of the adult mouse brain. We empirically evaluate aMAP as a method for registration and subsequent segmentation by validating it against the performance of expert human raters. This study therefore establishes a benchmark standard for mapping the molecular function and cellular connectivity of the rodent brain. PMID:27384127

  18. Distribution and Validation of CERES Irradiance Global Data Products Via Web Based Tools

    Science.gov (United States)

    Rutan, David; Mitrescu, Cristian; Doelling, David; Kato, Seiji

    2016-01-01

    The CERES SYN1deg product provides climate quality 3-hourly globally gridded and temporally complete maps of top of atmosphere, in atmosphere, and surface fluxes. This product requires efficient release to the public and validation to maintain quality assurance. The CERES team developed web-tools for the distribution of both the global gridded products and grid boxes that contain long term validation sites that maintain high quality flux observations at the Earth's surface. These are found at: http://ceres.larc.nasa.gov/order_data.php. In this poster we explore the various tools available to users to sub-set, download, and validate using surface observations the SYN1Deg and Surface-EBAF products. We also analyze differences found in long-term records from well-maintained land surface sites such as the ARM central facility and high quality buoy radiometers, which due to their isolated nature cannot be maintained in a similar manner to their land based counterparts.

  19. SPECTRa: the deposition and validation of primary chemistry research data in digital repositories.

    Science.gov (United States)

    Downing, Jim; Murray-Rust, Peter; Tonge, Alan P; Morgan, Peter; Rzepa, Henry S; Cotterill, Fiona; Day, Nick; Harvey, Matt J

    2008-08-01

    The SPECTRa (Submission, Preservation and Exposure of Chemistry Teaching and Research Data) project has investigated the practices of chemists in archiving and disseminating primary chemical data from academic research laboratories. To redress the loss of the large amount of data never archived or disseminated, we have developed software for data publication into departmental and institutional Open Access digital repositories (DSpace). Data adhering to standard formats in selected disciplines (crystallography, NMR, computational chemistry) is transformed to XML (CML, Chemical Markup Language) which provides added validation. Context-specific chemical metadata and persistent Handle identifiers are added to enable long-term data reuse. It was found essential to provide an embargo mechanism, and policies for operating this and other processes are presented.

  20. Using hospital discharge data for determining neonatal morbidity and mortality: a validation study

    Directory of Open Access Journals (Sweden)

    Algert Charles S

    2007-11-01

    Full Text Available Abstract Background Despite widespread use of neonatal hospital discharge data, there are few published reports on the accuracy of population health data with neonatal diagnostic or procedure codes. The aim of this study was to assess the accuracy of using routinely collected hospital discharge data in identifying neonatal morbidity during the birth admission compared with data from a statewide audit of selected neonatal intensive care (NICU admissions. Methods Validation study of population-based linked hospital discharge/birth data against neonatal intensive care audit data from New South Wales, Australia for 2,432 babies admitted to NICUs, 1994–1996. Sensitivity, specificity and positive predictive values (PPV with exact binomial confidence intervals were calculated for 12 diagnoses and 6 procedures. Results Sensitivities ranged from 37.0% for drainage of an air leak to 97.7% for very low birthweight, specificities all exceeded 85% and PPVs ranged from 70.9% to 100%. In-hospital mortality, low birthweight (≤1500 g, retinopathy of prematurity, respiratory distress syndrome, meconium aspiration, pneumonia, pulmonary hypertension, selected major anomalies, any mechanical ventilation (including CPAP, major surgery and surgery for patent ductus arteriosus or necrotizing enterocolitis were accurately identified with PPVs over 92%. Transient tachypnea of the newborn and drainage of an air leak had the lowest PPVs, 70.9% and 83.6% respectively. Conclusion Although under-ascertained, routinely collected hospital discharge data had high PPVs for most validated items and would be suitable for risk factor analyses of neonatal morbidity. Procedures tended to be more accurately recorded than diagnoses.

  1. Using hospital discharge data for determining neonatal morbidity and mortality: a validation study

    Science.gov (United States)

    Ford, Jane B; Roberts, Christine L; Algert, Charles S; Bowen, Jennifer R; Bajuk, Barbara; Henderson-Smart, David J

    2007-01-01

    Background Despite widespread use of neonatal hospital discharge data, there are few published reports on the accuracy of population health data with neonatal diagnostic or procedure codes. The aim of this study was to assess the accuracy of using routinely collected hospital discharge data in identifying neonatal morbidity during the birth admission compared with data from a statewide audit of selected neonatal intensive care (NICU) admissions. Methods Validation study of population-based linked hospital discharge/birth data against neonatal intensive care audit data from New South Wales, Australia for 2,432 babies admitted to NICUs, 1994–1996. Sensitivity, specificity and positive predictive values (PPV) with exact binomial confidence intervals were calculated for 12 diagnoses and 6 procedures. Results Sensitivities ranged from 37.0% for drainage of an air leak to 97.7% for very low birthweight, specificities all exceeded 85% and PPVs ranged from 70.9% to 100%. In-hospital mortality, low birthweight (≤1500 g), retinopathy of prematurity, respiratory distress syndrome, meconium aspiration, pneumonia, pulmonary hypertension, selected major anomalies, any mechanical ventilation (including CPAP), major surgery and surgery for patent ductus arteriosus or necrotizing enterocolitis were accurately identified with PPVs over 92%. Transient tachypnea of the newborn and drainage of an air leak had the lowest PPVs, 70.9% and 83.6% respectively. Conclusion Although under-ascertained, routinely collected hospital discharge data had high PPVs for most validated items and would be suitable for risk factor analyses of neonatal morbidity. Procedures tended to be more accurately recorded than diagnoses. PMID:18021458

  2. Validation of Remote Sensing Retrieval Products using Data from a Wireless Sensor-Based Online Monitoring in Antarctica

    Science.gov (United States)

    Li, Xiuhong; Cheng, Xiao; Yang, Rongjin; Liu, Qiang; Qiu, Yubao; Zhang, Jialin; Cai, Erli; Zhao, Long

    2016-01-01

    Of the modern technologies in polar-region monitoring, the remote sensing technology that can instantaneously form large-scale images has become much more important in helping acquire parameters such as the freezing and melting of ice as well as the surface temperature, which can be used in the research of global climate change, Antarctic ice sheet responses, and cap formation and evolution. However, the acquirement of those parameters is impacted remarkably by the climate and satellite transit time which makes it almost impossible to have timely and continuous observation data. In this research, a wireless sensor-based online monitoring platform (WSOOP) for the extreme polar environment is applied to obtain a long-term series of data which is site-specific and continuous in time. Those data are compared and validated with the data from a weather station at Zhongshan Station Antarctica and the result shows an obvious correlation. Then those data are used to validate the remote sensing products of the freezing and melting of ice and the surface temperature and the result also indicated a similar correlation. The experiment in Antarctica has proven that WSOOP is an effective system to validate remotely sensed data in the polar region. PMID:27869668

  3. Detailed validation in PCDDF analysis. ISO17025 data from Brazil

    Energy Technology Data Exchange (ETDEWEB)

    Kernick Carvalhaes, G.; Azevedo, J.A.; Azevedo, G.; Machado, M.; Brooks, P. [Analytical Solutions, Rio de Janeiro (Brazil)

    2004-09-15

    When we define validation method we can use the ISO standard 8402, in reference to this, 'validation' is the 'confirmation by the examination and supplying of objective evidences that the particular requirements for a specific intended use are fulfilled'. This concept is extremely important to guarantee the quality of results. Validation method is based on the combined use of different validation procedures, but in this selection we have to analyze the cost benefit conditions. We must focus on the critical elements, and these critical factors must be the essential elements for providing good properties and results. If we have a solid validation methodology and a research of the source of uncertainty of our analytical method, we can generate results with confidence and veracity. When analyzing these two considerations, validation method and uncertainty calculations, we found out that there are very few articles and papers about these subjects, and it is even more difficult to find such materials on dioxins and furans. This short paper describes a validation and uncertainty calculation methodology using traditional studies with a few adaptations, yet it shows a new idea of recovery study as a source of uncertainty.

  4. Validity of data in the Danish colorectal cancer screening database

    DEFF Research Database (Denmark)

    Thomsen, Mette Kielsholm; Njor, Sisse Helle; Rasmussen, Morten

    2017-01-01

    Background: In Denmark, a nationwide screening program for colorectal cancer was implemented in March 2014. Along with this, a clinical database for program monitoring and research purposes was established. Objective: The aim of this study was to estimate the agreement and validity of diagnosis...... and procedure codes in the Danish Colorectal Cancer Screening Database (DCCSD). Methods: All individuals with a positive immunochemical fecal occult blood test (iFOBT) result who were invited to screening in the first 3 months since program initiation were identified. From these, a sample of 150 individuals...... was selected using stratified random sampling by age, gender and region of residence. Data from the DCCSD were compared with data from hospital records, which were used as the reference. Agreement, sensitivity, specificity and positive and negative predictive values were estimated for categories of codes...

  5. Computer code ENDSAM for random sampling and validation of the resonance parameters covariance matrices of some major nuclear data libraries

    International Nuclear Information System (INIS)

    Plevnik, Lucijan; Žerovnik, Gašper

    2016-01-01

    Highlights: • Methods for random sampling of correlated parameters. • Link to open-source code for sampling of resonance parameters in ENDF-6 format. • Validation of the code on realistic and artificial data. • Validation of covariances in three major contemporary nuclear data libraries. - Abstract: Methods for random sampling of correlated parameters are presented. The methods are implemented for sampling of resonance parameters in ENDF-6 format and a link to the open-source code ENDSAM is given. The code has been validated on realistic data. Additionally, consistency of covariances of resonance parameters of three major contemporary nuclear data libraries (JEFF-3.2, ENDF/B-VII.1 and JENDL-4.0u2) has been checked.

  6. Soil Moisture Active Passive Mission L4_C Data Product Assessment (Version 2 Validated Release)

    Science.gov (United States)

    Kimball, John S.; Jones, Lucas A.; Glassy, Joseph; Stavros, E. Natasha; Madani, Nima; Reichle, Rolf H.; Jackson, Thomas; Colliander, Andreas

    2016-01-01

    The SMAP satellite was successfully launched January 31st 2015, and began acquiring Earth observation data following in-orbit sensor calibration. Global data products derived from the SMAP L-band microwave measurements include Level 1 calibrated and geolocated radiometric brightness temperatures, Level 23 surface soil moisture and freezethaw geophysical retrievals mapped to a fixed Earth grid, and model enhanced Level 4 data products for surface to root zone soil moisture and terrestrial carbon (CO2) fluxes. The post-launch SMAP mission CalVal Phase had two primary objectives for each science product team: 1) calibrate, verify, and improve the performance of the science algorithms, and 2) validate accuracies of the science data products as specified in the L1 science requirements. This report provides analysis and assessment of the SMAP Level 4 Carbon (L4_C) product pertaining to the validated release. The L4_C validated product release effectively replaces an earlier L4_C beta-product release (Kimball et al. 2015). The validated release described in this report incorporates a longer data record and benefits from algorithm and CalVal refinements acquired during the SMAP post-launch CalVal intensive period. The SMAP L4_C algorithms utilize a terrestrial carbon flux model informed by SMAP soil moisture inputs along with optical remote sensing (e.g. MODIS) vegetation indices and other ancillary biophysical data to estimate global daily net ecosystem CO2 exchange (NEE) and component carbon fluxes for vegetation gross primary production (GPP) and ecosystem respiration (Reco). Other L4_C product elements include surface (10 cm depth) soil organic carbon (SOC) stocks and associated environmental constraints to these processes, including soil moisture and landscape freeze/thaw (FT) controls on GPP and respiration (Kimball et al. 2012). The L4_C product encapsulates SMAP carbon cycle science objectives by: 1) providing a direct link between terrestrial carbon fluxes and

  7. Validating module network learning algorithms using simulated data.

    Science.gov (United States)

    Michoel, Tom; Maere, Steven; Bonnet, Eric; Joshi, Anagha; Saeys, Yvan; Van den Bulcke, Tim; Van Leemput, Koenraad; van Remortel, Piet; Kuiper, Martin; Marchal, Kathleen; Van de Peer, Yves

    2007-05-03

    In recent years, several authors have used probabilistic graphical models to learn expression modules and their regulatory programs from gene expression data. Despite the demonstrated success of such algorithms in uncovering biologically relevant regulatory relations, further developments in the area are hampered by a lack of tools to compare the performance of alternative module network learning strategies. Here, we demonstrate the use of the synthetic data generator SynTReN for the purpose of testing and comparing module network learning algorithms. We introduce a software package for learning module networks, called LeMoNe, which incorporates a novel strategy for learning regulatory programs. Novelties include the use of a bottom-up Bayesian hierarchical clustering to construct the regulatory programs, and the use of a conditional entropy measure to assign regulators to the regulation program nodes. Using SynTReN data, we test the performance of LeMoNe in a completely controlled situation and assess the effect of the methodological changes we made with respect to an existing software package, namely Genomica. Additionally, we assess the effect of various parameters, such as the size of the data set and the amount of noise, on the inference performance. Overall, application of Genomica and LeMoNe to simulated data sets gave comparable results. However, LeMoNe offers some advantages, one of them being that the learning process is considerably faster for larger data sets. Additionally, we show that the location of the regulators in the LeMoNe regulation programs and their conditional entropy may be used to prioritize regulators for functional validation, and that the combination of the bottom-up clustering strategy with the conditional entropy-based assignment of regulators improves the handling of missing or hidden regulators. We show that data simulators such as SynTReN are very well suited for the purpose of developing, testing and improving module network

  8. Collection, validation, and description of data for the Oak Ridge nuclear industry mortality study

    International Nuclear Information System (INIS)

    Watkins, J.; Reagan, J.; Cragle, D.; West, C.; Tankersley, W.; Frome, E.; Watson, J.

    1992-01-01

    In response to the continuing controversy over the long-term health effects of protracted occupational exposure to low levels of ionizing radiation, a mortality study was initiated pooling data for all workers hired before December 31, 1982, and employed for at least 30 days by one of the three DOE facilities in Oak Ridge, Tennessee. Topics of discussion will include issues involving the collection and validation of this large, complex data set and characteristics of demographic and exposure data for the study cohort

  9. Aplikace pro penetrační testování webových zranitelností typu Data Validation flaws

    OpenAIRE

    Němec, Václav

    2011-01-01

    Tato bakalářské práce se zabývá detekcí webových zranitelností typu Data validation flaws. V práci jsou popsány běžné útoky, obrana před těmito útoky a postupy při automatické detekci. Hlavním cílem je návrh a implementace nástroje pro automatickou detekci zranitelností typu Data validation flaws, jeho otestování a srovnání výsledků s podobnými nástroji jako například Paros Proxy nebo Burp Suite. This bachelor's thesis deals with detection of web vulnerabilities such as data validation fla...

  10. Validating a proxy for disease progression in metastatic cancer patients using prescribing and dispensing data.

    Science.gov (United States)

    Joshi, Vikram; Adelstein, Barbara-Ann; Schaffer, Andrea; Srasuebkul, Preeyaporn; Dobbins, Timothy; Pearson, Sallie-Anne

    2017-10-01

    Routine data collections are used increasingly to examine outcomes of real-world cancer drug use. These datasets lack clinical details about important endpoints such as disease progression. To validate a proxy for disease progression in metastatic cancer patients using prescribing and dispensing claims. We used data from a cohort study of patients undergoing chemotherapy who provided informed consent to the collection of cancer-treatment data from medical records and linkage to pharmaceutical claims. We derived proxy decision rules based on changes to drug treatment in prescription histories (n = 36 patients) and validated the proxy in prescribing data (n = 62 patients). We adapted the decision rules and validated the proxy in dispensing data (n = 109). Our gold standard was disease progression ascertained in patient medical records. Individual progression episodes were the unit of analysis for sensitivity and Positive Predictive Value (PPV) calculations and specificity and Negative Predictive Value (NPV) were calculated at the patient level. The sensitivity of our proxy in prescribing data was 74.3% (95% Confidence Interval (CI), 55.6-86.6%) and PPV 61.2% (95% CI, 45.0-75.3%); specificity and NPV were 87.8% (95% CI, 73.8-95.9%) and 100% (95% CI, 90.3-100%), respectively. In dispensing data, the sensitivity of our proxy was 64% (95% CI, 55.0-77.0%) and PPV 56.0% (95% CI, 43.0-69.0%); specificity and NPV were 81% (95% CI, 70.05-89.0%) and 91.0% (95% CI, 82.0-97.0%), respectively. Our proxy overestimated episodes of disease progression. The proxy's performance is likely to improve if the date of prescribing is used instead of date of dispensing in claims data and by incorporating medical service claims (such as imaging prior to drug changes) in the algorithm. Our proxy is not sufficiently robust for use in real world comparative effectiveness research for cancer medicines. © 2016 John Wiley & Sons Australia, Ltd.

  11. Validation of internet-based self-reported anthropometric, demographic data and participant identity in the Food4Me study

    Science.gov (United States)

    BACKGROUND In e-health intervention studies, there are concerns about the reliability of internet-based, self-reported (SR) data and about the potential for identity fraud. This study introduced and tested a novel procedure for assessing the validity of internet-based, SR identity and validated anth...

  12. CFD Code Validation against Stratified Air-Water Flow Experimental Data

    International Nuclear Information System (INIS)

    Terzuoli, F.; Galassi, M.C.; Mazzini, D.; D'Auria, F.

    2008-01-01

    Pressurized thermal shock (PTS) modelling has been identified as one of the most important industrial needs related to nuclear reactor safety. A severe PTS scenario limiting the reactor pressure vessel (RPV) lifetime is the cold water emergency core cooling (ECC) injection into the cold leg during a loss of coolant accident (LOCA). Since it represents a big challenge for numerical simulations, this scenario was selected within the European Platform for Nuclear Reactor Simulations (NURESIM) Integrated Project as a reference two-phase problem for computational fluid dynamics (CFDs) code validation. This paper presents a CFD analysis of a stratified air-water flow experimental investigation performed at the Institut de Mecanique des Fluides de Toulouse in 1985, which shares some common physical features with the ECC injection in PWR cold leg. Numerical simulations have been carried out with two commercial codes (Fluent and Ansys CFX), and a research code (NEPTUNE CFD). The aim of this work, carried out at the University of Pisa within the NURESIM IP, is to validate the free surface flow model implemented in the codes against experimental data, and to perform code-to-code benchmarking. Obtained results suggest the relevance of three-dimensional effects and stress the importance of a suitable interface drag modelling

  13. CFD Code Validation against Stratified Air-Water Flow Experimental Data

    Directory of Open Access Journals (Sweden)

    F. Terzuoli

    2008-01-01

    Full Text Available Pressurized thermal shock (PTS modelling has been identified as one of the most important industrial needs related to nuclear reactor safety. A severe PTS scenario limiting the reactor pressure vessel (RPV lifetime is the cold water emergency core cooling (ECC injection into the cold leg during a loss of coolant accident (LOCA. Since it represents a big challenge for numerical simulations, this scenario was selected within the European Platform for Nuclear Reactor Simulations (NURESIM Integrated Project as a reference two-phase problem for computational fluid dynamics (CFDs code validation. This paper presents a CFD analysis of a stratified air-water flow experimental investigation performed at the Institut de Mécanique des Fluides de Toulouse in 1985, which shares some common physical features with the ECC injection in PWR cold leg. Numerical simulations have been carried out with two commercial codes (Fluent and Ansys CFX, and a research code (NEPTUNE CFD. The aim of this work, carried out at the University of Pisa within the NURESIM IP, is to validate the free surface flow model implemented in the codes against experimental data, and to perform code-to-code benchmarking. Obtained results suggest the relevance of three-dimensional effects and stress the importance of a suitable interface drag modelling.

  14. Valid and complete data on endometrial cancer in the Danish Gynaecological Cancer Database

    DEFF Research Database (Denmark)

    Juhl, Caroline S; Hansen, Estrid S; Høgdall, Claus K

    2014-01-01

    INTRODUCTION: It is a comparative register study designed for data validation of surgery, pathology and recurrence for endometrial cancer in the Danish Gynaecological Cancer Database (DGCD) in the 2005-2009 period. The main outcomes were completeness of the data registered in the DGCD, agreement...... concerning data reported and comparability between the DGCD and a definite reference. MATERIAL AND METHODS: DGCD data on women with endometrial cancer or adenomatous hyperplasia supplemented with patient charts for data on recurrence were retrieved and compared with a definite reference (the pathology report...... and clinical journals). RESULTS: The completeness of data on pathology and surgery reported to the DGCD was 97.3%. The comparability between the DGCG and the definite reference was 94.4%. The agreement for the reported data in the DGCD was 88.3%. For recurrence, the comparability was 94.5% and the agreement...

  15. FastaValidator: an open-source Java library to parse and validate FASTA formatted sequences.

    Science.gov (United States)

    Waldmann, Jost; Gerken, Jan; Hankeln, Wolfgang; Schweer, Timmy; Glöckner, Frank Oliver

    2014-06-14

    Advances in sequencing technologies challenge the efficient importing and validation of FASTA formatted sequence data which is still a prerequisite for most bioinformatic tools and pipelines. Comparative analysis of commonly used Bio*-frameworks (BioPerl, BioJava and Biopython) shows that their scalability and accuracy is hampered. FastaValidator represents a platform-independent, standardized, light-weight software library written in the Java programming language. It targets computer scientists and bioinformaticians writing software which needs to parse quickly and accurately large amounts of sequence data. For end-users FastaValidator includes an interactive out-of-the-box validation of FASTA formatted files, as well as a non-interactive mode designed for high-throughput validation in software pipelines. The accuracy and performance of the FastaValidator library qualifies it for large data sets such as those commonly produced by massive parallel (NGS) technologies. It offers scientists a fast, accurate and standardized method for parsing and validating FASTA formatted sequence data.

  16. Validation of neutron data libraries by backscattered spectra of Pu-Be Neutrons

    CERN Document Server

    El-Agib, I

    1999-01-01

    Elastically backscattered spectra of Pu-Be neutrons have been measured for SiO sub 2 , water, graphite, paraffin oil and Al slabs using a proton recoil spectrometer. The results were compared with the calculated spectra obtained by the three-dimensional Monte-Carlo transport code MCNP-4B and point-wise cross sections from the ENDF/B-V, ENDF/B-VI, JENDL-3.1 and BROND-2 data libraries. The good agreement between the measured and calculated results indicates that this procedure can be used for validation of different data libraries. This simple method renders possible the detection of oxygen, carbon and hydrogen in bulk samples. (author)

  17. Validation of the PESTLA model: Field test using data from a sandy soil in Schaijk (the Netherlands)

    NARCIS (Netherlands)

    Boekhold AE; Swartjes FA; Hoogenboom FGG; van der Linden AMA

    1993-01-01

    Within the framework of the project "Validation of PESTLA" the Schaijk data set was used to analyse PESTLA model performance. The Schaijk data set contains field data on bentazon behaviour in a coarse textured humic gley soil cropped with maize. PESTLA model input parameters were derived

  18. Integration of data validation and user interface concerns in a DSL for web applications

    NARCIS (Netherlands)

    Groenewegen, D.M.; Visser, E.

    2009-01-01

    This paper is a pre-print of: Danny M. Groenewegen, Eelco Visser. Integration of Data Validation and User Interface Concerns in a DSL for Web Applications. In Mark G. J. van den Brand, Jeff Gray, editors, Software Language Engineering, Second International Conference, SLE 2009, Denver, USA, October,

  19. Exploring Deep Computing in CMS for Automated Data Validation in DQM

    CERN Document Server

    Fernandez Madrazo, Celia

    2017-01-01

    This project has explored the possibility of inclusion of a variational autoencoder in Automated Data Validation in DQM. The analysis has been carried out only with muon features. The main goal is to reconstruct the given lumisections and check if they can be separated between good and bad lumisections by means of the latent space representation given by the developed autoencoder. At the end, many features of good lumisections seem to be correctly reconstructed but the latent space representation does not give a proper distintion between both types of samples.

  20. Validating Smoking Data From the Veteran’s Affairs Health Factors Dataset, an Electronic Data Source

    Science.gov (United States)

    Brandt, Cynthia A.; Skanderson, Melissa; Justice, Amy C.; Shahrir, Shahida; Butt, Adeel A.; Brown, Sheldon T.; Freiberg, Matthew S.; Gibert, Cynthia L.; Goetz, Matthew Bidwell; Kim, Joon Woo; Pisani, Margaret A.; Rimland, David; Rodriguez-Barradas, Maria C.; Sico, Jason J.; Tindle, Hilary A.; Crothers, Kristina

    2011-01-01

    Introduction: We assessed smoking data from the Veterans Health Administration (VHA) electronic medical record (EMR) Health Factors dataset. Methods: To assess the validity of the EMR Health Factors smoking data, we first created an algorithm to convert text entries into a 3-category smoking variable (never, former, and current). We compared this EMR smoking variable to 2 different sources of patient self-reported smoking survey data: (a) 6,816 HIV-infected and -uninfected participants in the 8-site Veterans Aging Cohort Study (VACS-8) and (b) a subset of 13,689 participants from the national VACS Virtual Cohort (VACS-VC), who also completed the 1999 Large Health Study (LHS) survey. Sensitivity, specificity, and kappa statistics were used to evaluate agreement of EMR Health Factors smoking data with self-report smoking data. Results: For the EMR Health Factors and VACS-8 comparison of current, former, and never smoking categories, the kappa statistic was .66. For EMR Health Factors and VACS-VC/LHS comparison of smoking, the kappa statistic was .61. Conclusions: Based on kappa statistics, agreement between the EMR Health Factors and survey sources is substantial. Identification of current smokers nationally within the VHA can be used in future studies to track smoking status over time, to evaluate smoking interventions, and to adjust for smoking status in research. Our methodology may provide insights for other organizations seeking to use EMR data for accurate determination of smoking status. PMID:21911825

  1. Groundwater Model Validation

    Energy Technology Data Exchange (ETDEWEB)

    Ahmed E. Hassan

    2006-01-24

    Models have an inherent uncertainty. The difficulty in fully characterizing the subsurface environment makes uncertainty an integral component of groundwater flow and transport models, which dictates the need for continuous monitoring and improvement. Building and sustaining confidence in closure decisions and monitoring networks based on models of subsurface conditions require developing confidence in the models through an iterative process. The definition of model validation is postulated as a confidence building and long-term iterative process (Hassan, 2004a). Model validation should be viewed as a process not an end result. Following Hassan (2004b), an approach is proposed for the validation process of stochastic groundwater models. The approach is briefly summarized herein and detailed analyses of acceptance criteria for stochastic realizations and of using validation data to reduce input parameter uncertainty are presented and applied to two case studies. During the validation process for stochastic models, a question arises as to the sufficiency of the number of acceptable model realizations (in terms of conformity with validation data). Using a hierarchical approach to make this determination is proposed. This approach is based on computing five measures or metrics and following a decision tree to determine if a sufficient number of realizations attain satisfactory scores regarding how they represent the field data used for calibration (old) and used for validation (new). The first two of these measures are applied to hypothetical scenarios using the first case study and assuming field data consistent with the model or significantly different from the model results. In both cases it is shown how the two measures would lead to the appropriate decision about the model performance. Standard statistical tests are used to evaluate these measures with the results indicating they are appropriate measures for evaluating model realizations. The use of validation

  2. Utilizing Chamber Data for Developing and Validating Climate Change Models

    Science.gov (United States)

    Monje, Oscar

    2012-01-01

    Controlled environment chambers (e.g. growth chambers, SPAR chambers, or open-top chambers) are useful for measuring plant ecosystem responses to climatic variables and CO2 that affect plant water relations. However, data from chambers was found to overestimate responses of C fluxes to CO2 enrichment. Chamber data may be confounded by numerous artifacts (e.g. sidelighting, edge effects, increased temperature and VPD, etc) and this limits what can be measured accurately. Chambers can be used to measure canopy level energy balance under controlled conditions and plant transpiration responses to CO2 concentration can be elucidated. However, these measurements cannot be used directly in model development or validation. The response of stomatal conductance to CO2 will be the same as in the field, but the measured response must be recalculated in such a manner to account for differences in aerodynamic conductance, temperature and VPD between the chamber and the field.

  3. Validation of new 240Pu cross section and covariance data via criticality calculation

    International Nuclear Information System (INIS)

    Kim, Do Heon; Gil, Choong-Sup; Kim, Hyeong Il; Lee, Young-Ouk; Leal, Luiz C.; Dunn, Michael E.

    2011-01-01

    Recent collaboration between KAERI and ORNL has completed an evaluation for 240 Pu neutron cross section with covariance data. The new 240 Pu cross section data has been validated through 28 criticality safety benchmark problems taken from the ICSBEP and/or CSEWG specifications with MCNP calculations. The calculation results based on the new evaluation have been compared with those based on recent evaluations such as ENDF/B-VII.0, JEFF-3.1.1, and JENDL-4.0. In addition, the new 240 Pu covariance data has been tested for some criticality benchmarks via the DANTSYS/SUSD3D-based nuclear data sensitivity and uncertainty analysis of k eff . The k eff uncertainty estimates by the new covariance data has been compared with those by JENDL-4.0, JENDL-3.3, and Low-Fidelity covariance data. (author)

  4. Experimental Peptide Identification Repository (EPIR): an integrated peptide-centric platform for validation and mining of tandem mass spectrometry data

    DEFF Research Database (Denmark)

    Kristensen, Dan Bach; Brønd, Jan Christian; Nielsen, Peter Aagaard

    2004-01-01

    LC MS/MS has become an established technology in proteomic studies, and with the maturation of the technology the bottleneck has shifted from data generation to data validation and mining. To address this bottleneck we developed Experimental Peptide Identification Repository (EPIR), which...... is an integrated software platform for storage, validation, and mining of LC MS/MS-derived peptide evidence. EPIR is a cumulative data repository where precursor ions are linked to peptide assignments and protein associations returned by a search engine (e.g. Mascot, Sequest, or PepSea). Any number of datasets can...

  5. Capitalizing on Citizen Science Data for Validating Models and Generating Hypotheses Describing Meteorological Drivers of Mosquito-Borne Disease Risk

    Science.gov (United States)

    Boger, R. A.; Low, R.; Paull, S.; Anyamba, A.; Soebiyanto, R. P.

    2017-12-01

    Temperature and precipitation are important drivers of mosquito population dynamics, and a growing set of models have been proposed to characterize these relationships. Validation of these models, and development of broader theories across mosquito species and regions could nonetheless be improved by comparing observations from a global dataset of mosquito larvae with satellite-based measurements of meteorological variables. Citizen science data can be particularly useful for two such aspects of research into the meteorological drivers of mosquito populations: i) Broad-scale validation of mosquito distribution models and ii) Generation of quantitative hypotheses regarding changes to mosquito abundance and phenology across scales. The recently released GLOBE Observer Mosquito Habitat Mapper (GO-MHM) app engages citizen scientists in identifying vector taxa, mapping breeding sites and decommissioning non-natural habitats, and provides a potentially useful new tool for validating mosquito ubiquity projections based on the analysis of remotely sensed environmental data. Our early work with GO-MHM data focuses on two objectives: validating citizen science reports of Aedes aegypti distribution through comparison with accepted scientific data sources, and exploring the relationship between extreme temperature and precipitation events and subsequent observations of mosquito larvae. Ultimately the goal is to develop testable hypotheses regarding the shape and character of this relationship between mosquito species and regions.

  6. Intelligibility in Context Scale: Normative and Validation Data for English-Speaking Preschoolers.

    Science.gov (United States)

    McLeod, Sharynne; Crowe, Kathryn; Shahaeian, Ameneh

    2015-07-01

    The purpose of this study was to describe normative and validation data on the Intelligibility in Context Scale (ICS; McLeod, Harrison, & McCormack, 2012c) for English-speaking children. The ICS is a 7-item, parent-report measure of children's speech intelligibility with a range of communicative partners. Data were collected from the parents of 803 Australian English-speaking children ranging in age from 4;0 (years;months) to 5;5 (37.0% were multilingual). The mean ICS score was 4.4 (SD = 0.7) out of a possible total score of 5. Children's speech was reported to be most intelligible to their parents, followed by their immediate family, friends, and teachers; children's speech was least intelligible to strangers. The ICS had high internal consistency (α = .94). Significant differences in scores were identified on the basis of sex and age but not on the basis of socioeconomic status or the number of languages spoken. There were significant differences in scores between children whose parents had concerns about their child's speech (M = 3.9) and those who did not (M = 4.6). A sensitivity of .82 and a specificity of .58 were established as the optimal cutoff. Test-retest reliability and criterion validity were established for 184 children with a speech sound disorder. There was a significant low correlation between the ICS mean score and percentage of phonemes correct (r = .30), percentage of consonants correct (r = .24), and percentage of vowels correct (r = .30) on the Diagnostic Evaluation of Articulation and Phonology (Dodd, Hua, Crosbie, Holm, & Ozanne, 2002). Thirty-one parents completed the ICS related to English and another language spoken by their child with a speech sound disorder. The significant correlations between the scores suggest that the ICS may be robust between languages. This article provides normative ICS data for English-speaking children and additional validation of the psychometric properties of the ICS. The robustness of the ICS was suggested

  7. Validation of a realistic powder sample using data from DMC at PSI

    International Nuclear Information System (INIS)

    Willendrup, Peter; Filges, Uwe; Keller, Lukas; Farhi, Emmanuel; Lefmann, Kim

    2006-01-01

    We present results of a virtual experiment, carried out by means of a McStas simulation of the powder diffractometer DMC at PSI, using the new powder sample component PowderN. This powder component takes tabulated crystallographic input to define realistic powder lines. The simulated output data from the virtual experiment on the compound Na 2 Ca 3 Al 2 F 14 are compared to real measurement data from the DMC instrument. The agreement is very good with respect to peak positions, widths, background intensity and relative peak intensities. This work represents an important step towards reliable virtual experiments and also act as a validation of the PowderN sample component in McStas

  8. Validation of a realistic powder sample using data from DMC at PSI

    Energy Technology Data Exchange (ETDEWEB)

    Willendrup, Peter [Riso National Laboratory, Frederiksborgvej 399, DK-4000 Roskilde (Denmark)]. E-mail: peter.willendrup@risoe.dk; Filges, Uwe [Laboratory for Development and Methods ETHZ and PSI CH-5232 Villigen PSI (Switzerland); Keller, Lukas [Laboratory for Neutron Scattering ETHZ and PSI CH-5232 Villigen PSI (Switzerland); Farhi, Emmanuel [Institut Laue-Langevin (ILL) Grenoble, 6 rue J. Horowitz, BP 156, 38042 Grenoble Cedex 9 (France); Lefmann, Kim [Riso National Laboratory, Frederiksborgvej 399, DK-4000 Roskilde (Denmark)

    2006-11-15

    We present results of a virtual experiment, carried out by means of a McStas simulation of the powder diffractometer DMC at PSI, using the new powder sample component PowderN. This powder component takes tabulated crystallographic input to define realistic powder lines. The simulated output data from the virtual experiment on the compound Na{sub 2}Ca{sub 3}Al{sub 2}F{sub 14} are compared to real measurement data from the DMC instrument. The agreement is very good with respect to peak positions, widths, background intensity and relative peak intensities. This work represents an important step towards reliable virtual experiments and also act as a validation of the PowderN sample component in McStas.

  9. Complementary techniques: validation of gene expression data by quantitative real time PCR.

    Science.gov (United States)

    Provenzano, Maurizio; Mocellin, Simone

    2007-01-01

    Microarray technology can be considered the most powerful tool for screening gene expression profiles of biological samples. After data mining, results need to be validated with highly reliable biotechniques allowing for precise quantitation of transcriptional abundance of identified genes. Quantitative real time PCR (qrt-PCR) technology has recently reached a level of sensitivity, accuracy and practical ease that support its use as a routine bioinstrumentation for gene level measurement. Currently, qrt-PCR is considered by most experts the most appropriate method to confirm or confute microarray-generated data. The knowledge of the biochemical principles underlying qrt-PCR as well as some related technical issues must be beard in mind when using this biotechnology.

  10. Validation of a realistic powder sample using data from DMC at PSI

    DEFF Research Database (Denmark)

    Willendrup, Peter Kjær; Filges, U.; Keller, L.

    2006-01-01

    We present results of a virtual experiment, carried out by means of a McStas simulation of the powder diffractometer DMC at PSI, using the new powder sample component PowderN. This powder component takes tabulated crystallographic input to define realistic powder lines. The simulated output data...... from the virtual experiment on the compound Na2Ca3Al2F14 are compared to real measurement data from the DMC instrument. The agreement is very good with respect to peak positions, widths, background intensity and relative peak intensities. This work represents an important step towards reliable virtual...... experiments and also act as a validation of the PowderN sample component in McStas....

  11. Reliability, Validity, Comparability and Practical Utility of Cybercrime-Related Data, Metrics, and Information

    OpenAIRE

    Nir Kshetri

    2013-01-01

    With an increasing pervasiveness, prevalence and severity of cybercrimes, various metrics, measures and statistics have been developed and used to measure various aspects of this phenomenon. Cybercrime-related data, metrics, and information, however, pose important and difficult dilemmas regarding the issues of reliability, validity, comparability and practical utility. While many of the issues of the cybercrime economy are similar to other underground and underworld industries, this economy ...

  12. Validity of Principal Diagnoses in Discharge Summaries and ICD-10 Coding Assessments Based on National Health Data of Thailand.

    Science.gov (United States)

    Sukanya, Chongthawonsatid

    2017-10-01

    This study examined the validity of the principal diagnoses on discharge summaries and coding assessments. Data were collected from the National Health Security Office (NHSO) of Thailand in 2015. In total, 118,971 medical records were audited. The sample was drawn from government hospitals and private hospitals covered by the Universal Coverage Scheme in Thailand. Hospitals and cases were selected using NHSO criteria. The validity of the principal diagnoses listed in the "Summary and Coding Assessment" forms was established by comparing data from the discharge summaries with data obtained from medical record reviews, and additionally, by comparing data from the coding assessments with data in the computerized ICD (the data base used for reimbursement-purposes). The summary assessments had low sensitivities (7.3%-37.9%), high specificities (97.2%-99.8%), low positive predictive values (9.2%-60.7%), and high negative predictive values (95.9%-99.3%). The coding assessments had low sensitivities (31.1%-69.4%), high specificities (99.0%-99.9%), moderate positive predictive values (43.8%-89.0%), and high negative predictive values (97.3%-99.5%). The discharge summaries and codings often contained mistakes, particularly the categories "Endocrine, nutritional, and metabolic diseases", "Symptoms, signs, and abnormal clinical and laboratory findings not elsewhere classified", "Factors influencing health status and contact with health services", and "Injury, poisoning, and certain other consequences of external causes". The validity of the principal diagnoses on the summary and coding assessment forms was found to be low. The training of physicians and coders must be strengthened to improve the validity of discharge summaries and codings.

  13. Validation of the CATHARE2 code against experimental data from Brayton-cycle plants

    International Nuclear Information System (INIS)

    Bentivoglio, Fabrice; Tauveron, Nicolas; Geffraye, Genevieve; Gentner, Herve

    2008-01-01

    In recent years the Commissariat a l'Energie Atomique (CEA) has commissioned a wide range of feasibility studies of future-advanced nuclear reactors, in particular gas-cooled reactors (GCR). The thermohydraulic behaviour of these systems is a key issue for, among other things, the design of the core, the assessment of thermal stresses, and the design of decay heat removal systems. These studies therefore require efficient and reliable simulation tools capable of modelling the whole reactor, including the core, the core vessel, piping, heat exchangers and turbo-machinery. CATHARE2 is a thermal-hydraulic 1D reference safety code developed and extensively validated for the French pressurized water reactors. It has been recently adapted to deal also with gas-cooled reactor applications. In order to validate CATHARE2 for these new applications, CEA has initiated an ambitious long-term experimental program. The foreseen experimental facilities range from small-scale loops for physical correlations, to component technology and system demonstration loops. In the short-term perspective, CATHARE2 is being validated against existing experimental data. And in particular from the German power plants Oberhausen I and II. These facilities have both been operated by the German utility Energie Versorgung Oberhausen (E.V.O.) and their power conversion systems resemble to the high-temperature reactor concepts: Oberhausen I is a 13.75-MWe Brayton-cycle air turbine plant, and Oberhausen II is a 50-MWe Brayton-cycle helium turbine plant. The paper presents these two plants, the adopted CATHARE2 modelling and a comparison between experimental data and code results for both steady state and transient cases

  14. Validating hierarchical verbal autopsy expert algorithms in a large data set with known causes of death.

    Science.gov (United States)

    Kalter, Henry D; Perin, Jamie; Black, Robert E

    2016-06-01

    Physician assessment historically has been the most common method of analyzing verbal autopsy (VA) data. Recently, the World Health Organization endorsed two automated methods, Tariff 2.0 and InterVA-4, which promise greater objectivity and lower cost. A disadvantage of the Tariff method is that it requires a training data set from a prior validation study, while InterVA relies on clinically specified conditional probabilities. We undertook to validate the hierarchical expert algorithm analysis of VA data, an automated, intuitive, deterministic method that does not require a training data set. Using Population Health Metrics Research Consortium study hospital source data, we compared the primary causes of 1629 neonatal and 1456 1-59 month-old child deaths from VA expert algorithms arranged in a hierarchy to their reference standard causes. The expert algorithms were held constant, while five prior and one new "compromise" neonatal hierarchy, and three former child hierarchies were tested. For each comparison, the reference standard data were resampled 1000 times within the range of cause-specific mortality fractions (CSMF) for one of three approximated community scenarios in the 2013 WHO global causes of death, plus one random mortality cause proportions scenario. We utilized CSMF accuracy to assess overall population-level validity, and the absolute difference between VA and reference standard CSMFs to examine particular causes. Chance-corrected concordance (CCC) and Cohen's kappa were used to evaluate individual-level cause assignment. Overall CSMF accuracy for the best-performing expert algorithm hierarchy was 0.80 (range 0.57-0.96) for neonatal deaths and 0.76 (0.50-0.97) for child deaths. Performance for particular causes of death varied, with fairly flat estimated CSMF over a range of reference values for several causes. Performance at the individual diagnosis level was also less favorable than that for overall CSMF (neonatal: best CCC = 0.23, range 0

  15. GPM GROUND VALIDATION NOAA UHF 449 PROFILER RAW DATA SPC FORMAT MC3E V1

    Data.gov (United States)

    National Aeronautics and Space Administration — The GPM Ground Validation NOAA UHF 449 Profiler Raw Data SPC Format MC3E dataset was collected during the NASA supported Midlatitude Continental Convective Clouds...

  16. EXAMINATION OF A PROPOSED VALIDATION DATA SET USING CFD CALCULATIONS

    International Nuclear Information System (INIS)

    Johnson, Richard W.

    2009-01-01

    The United States Department of Energy is promoting the resurgence of nuclear power in the U. S. for both electrical power generation and production of process heat required for industrial processes such as the manufacture of hydrogen for use as a fuel in automobiles. The DOE project is called the next generation nuclear plant (NGNP) and is based on a Generation IV reactor concept called the very high temperature reactor (VHTR), which will use helium as the coolant at temperatures ranging from 450 C to perhaps 1000 C. While computational fluid dynamics (CFD) has not been used for past safety analysis for nuclear reactors in the U. S., it is being considered for such for future reactors. It is fully recognized that CFD simulation codes will have to be validated for flow physics reasonably close to actual fluid dynamic conditions expected in normal and accident operational situations. To this end, experimental data have been obtained in a scaled model of a narrow slice of the lower plenum of a prismatic VHTR. The present article presents new results of CFD examinations of these data to explore potential issues with the geometry, the initial conditions, the flow dynamics and the data needed to fully specify the inlet and boundary conditions; results for several turbulence models are examined. Issues are addressed and recommendations about the data are made

  17. Production and validation of nuclear data for reactor and fuel cycle applications

    International Nuclear Information System (INIS)

    Trakas, C.; Verwaerde, D.; Toubon, H.

    2002-01-01

    The aim of this technical meeting is the improvement of the existing nuclear data and the production of new data of interest for the upstream and downstream of the fuel cycle (enrichment, fabrication, management, storage, transport, reprocessing), for the industrial reactors, the research reactors and the new reactor concepts (criticality, dimensioning, exploitation), for the instrumentation systems (external and internal sensors), the radioprotection, the residual power, the structures (neutron bombardment effect on vessels, rods etc..), and for the activation of steel structures (Fr, Ni, Co). The expected result is the collection of more reliable and accurate data in a wider spectrum of energies and temperatures thanks to more precise computer codes and measurement techniques. This document brings together the communications presented at this meeting and dealing with: the process of production and validation of nuclear data; the measurement facilities and the big international programs; the users needs and the industrial priorities; the basic nuclear data (BND) needs at Cogema; the expression and evaluation of BND; the evaluation work: the efficient cross-sections; the processing of data and the creation of activation libraries; from the integral measurement to the qualification and the feedback on nuclear data. (J.S.)

  18. Verification and Validation of FAARR Model and Data Envelopment Analysis Models for United States Army Recruiting

    National Research Council Canada - National Science Library

    Piskator, Gene

    1998-01-01

    ...) model and to develop a Data Envelopment Analysis (DEA) modeling strategy. First, the FAARR model was verified using a simulation of a known production function and validated using sensitivity analysis and ex-post forecasts...

  19. New approach for validating the segmentation of 3D data applied to individual fibre extraction

    DEFF Research Database (Denmark)

    Emerson, Monica Jane; Dahl, Anders Bjorholm; Dahl, Vedrana Andersen

    2017-01-01

    We present two approaches for validating the segmentation of 3D data. The first approach consists on comparing the amount of estimated material to a value provided by the manufacturer. The second approach consists on comparing the segmented results to those obtained from imaging modalities...

  20. CERESVis: A QC Tool for CERES that Leverages Browser Technology for Data Validation

    Science.gov (United States)

    Chu, C.; Sun-Mack, S.; Heckert, E.; Chen, Y.; Doelling, D.

    2015-12-01

    In this poster, we are going to present three user interfaces that CERES team uses to validate pixel-level data. Besides our home grown tools, we will aslo present the browser technology that we use to provide interactive interfaces, such as jquery, HighCharts and Google Earth. We pass data to the users' browsers and use the browsers to do some simple computations. The three user interfaces are: Thumbnails -- it displays hundrends images to allow users to browse 24-hour data files in few seconds. Multiple-synchronized cursors -- it allows users to compare multiple images side by side. Bounding Boxes and Histograms -- it allows users to draw multiple bounding boxes on an image and the browser computes/display the histograms.

  1. Validating administrative data for the detection of adverse events in older hospitalized patients

    Directory of Open Access Journals (Sweden)

    Ackroyd-Stolarz S

    2014-08-01

    Full Text Available Stacy Ackroyd-Stolarz,1,2 Susan K Bowles,3–5 Lorri Giffin6 1Performance Excellence Portfolio, Capital District Health Authority, Halifax, Nova Scotia, Canada; 2Department of Emergency Medicine, Dalhousie University, Halifax, Nova Scotia, Canada; 3Geriatric Medicine, Capital District Health Authority, Halifax, Nova Scotia, Canada; 4College of Pharmacy and Division of Geriatric Medicine, Dalhousie University, Halifax, Nova Scotia, Canada; 5Department of Pharmacy at Capital District Health Authority, Halifax, Nova Scotia, Canada; 6South Shore Family Health, Bridgewater, Nova Scotia, Canada Abstract: Older hospitalized patients are at risk of experiencing adverse events including, but not limited to, hospital-acquired pressure ulcers, fall-related injuries, and adverse drug events. A significant challenge in monitoring and managing adverse events is lack of readily accessible information on their occurrence. Purpose: The objective of this retrospective cross-sectional study was to validate diagnostic codes for pressure ulcers, fall-related injuries, and adverse drug events found in routinely collected administrative hospitalization data. Methods: All patients 65 years of age or older discharged between April 1, 2009 and March 31, 2011 from a provincial academic health sciences center in Canada were eligible for inclusion in the validation study. For each of the three types of adverse events, a random sample of 50 patients whose records were positive and 50 patients whose records were not positive for an adverse event was sought for review in the validation study (n=300 records in total. A structured health record review was performed independently by two health care providers with experience in geriatrics, both of whom were unaware of the patient's status with respect to adverse event coding. A physician reviewed 40 records (20 reviewed by each health care provider to establish interrater agreement. Results: A total of 39 pressure ulcers, 56 fall

  2. Validation of GPS atmospheric water vapor with WVR data in satellite tracking mode

    Science.gov (United States)

    Shangguan, M.; Heise, S.; Bender, M.; Dick, G.; Ramatschi, M.; Wickert, J.

    2015-01-01

    Slant-integrated water vapor (SIWV) data derived from GPS STDs (slant total delays), which provide the spatial information on tropospheric water vapor, have a high potential for assimilation to weather models or for nowcasting or reconstruction of the 3-D humidity field with tomographic techniques. Therefore, the accuracy of GPS STD is important, and independent observations are needed to estimate the quality of GPS STD. In 2012 the GFZ (German Research Centre for Geosciences) started to operate a microwave radiometer in the vicinity of the Potsdam GPS station. The water vapor content along the line of sight between a ground station and a GPS satellite can be derived from GPS data and directly measured by a water vapor radiometer (WVR) at the same time. In this study we present the validation results of SIWV observed by a ground-based GPS receiver and a WVR. The validation covers 184 days of data with dry and wet humidity conditions. SIWV data from GPS and WVR generally show good agreement with a mean bias of -0.4 kg m-2 and an rms (root mean square) of 3.15 kg m-2. The differences in SIWV show an elevation dependent on an rms of 7.13 kg m-2 below 15° but of 1.76 kg m-2 above 15°. Nevertheless, this elevation dependence is not observed regarding relative deviations. The relation between the differences and possible influencing factors (elevation angles, pressure, temperature and relative humidity) are analyzed in this study. Besides the elevation, dependencies between the atmospheric humidity conditions, temperature and the differences in SIWV are found.

  3. Lagrangian Stochastic Dispersion Model IMS Model Suite and its Validation against Experimental Data

    International Nuclear Information System (INIS)

    Bartok, J.

    2010-01-01

    The dissertation presents IMS Lagrangian Dispersion Model, which is a 'new generation' Slovak dispersion model of long-range transport, developed by MicroStep-MIS. It solves trajectory equation for a vast number of Lagrangian 'particles' and stochastic equation that simulates the effects of turbulence. Model contains simulation of radioactive decay (full decay chains of more than 300 nuclides), and dry and wet deposition. Model was integrated into IMS Model Suite, a system in which several models and modules can run and cooperate, e.g. LAM model WRF preparing fine resolution meteorological data for dispersion. The main theme of the work is validation of dispersion model against large scale international campaigns CAPTEX and ETEX, which are two of the largest tracer experiments. Validation addressed treatment of missing data, data interpolation into comparable temporal and spatial representation. The best model results were observed for ETEX I, standard results for CAPTEXes and worst results for ETEX II, known in modelling community for its meteorological conditions that can be hardly resolved by models. The IMS Lagrangian Dispersion Model was identified as capable long range dispersion model for slowly- or nonreacting chemicals and radioactive matter. Influence of input data on simulation quality is discussed within the work. Additional modules were prepared according to praxis requirement: a) Recalculation of concentrations of radioactive pollutant into effective doses form inhalation, immersion in the plume and deposition. b) Dispersion of mineral dust was added and tested in desert locality, where wind and soil moisture were firstly analysed and forecast by WRF. The result was qualitatively verified in case study against satellite observations. (author)

  4. Validity of proxy data obtained by different psychological autopsy information reconstruction techniques.

    Science.gov (United States)

    Fang, L; Zhang, J

    2010-01-01

    Two informants were interviewed for each of 416 living controls (individuals sampled from the normal population) interviewed in a Chinese case-control psychological autopsy study. The validity of proxy data, obtained using seven psychological autopsy information reconstruction techniques (types 1, 2 and A - E), was evaluated, with living controls' self reports used as the gold-standard. Proxy data for reconstruction technique types 1, 2 and D on the Impulsivity Inventory Scale (total impulsivity score) were no different from the living controls' self report gold standard, whereas data for types A and E were smaller than data from living controls. On the 'acceptance or resignation' sub-scale of the avoidance coping dimension of the Moos Coping Response Inventory, information obtained by reconstruction technique types 1 and D was not significantly different from the living controls' self reports, whereas proxy data from types 2, A and E were smaller than those from the living controls. No statistically significant differences were identified for other proxy data obtained by reconstruction technique types 1, 2, A, D and E. These results indicate that using a second informant does not significantly enhance information reconstruction for the target.

  5. Generating Vegetation Leaf Area Index Earth System Data Record from Multiple Sensors. Part 2; Implementation, Analysis and Validation

    Science.gov (United States)

    Ganguly, Sangram; Samanta, Arindam; Schull, Mitchell A.; Shabanov, Nikolay V.; Milesi, Cristina; Nemani, Ramajrushna R,; Knyazikhin, Yuri; Myneni, Ranga B.

    2008-01-01

    The evaluation of a new global monthly leaf area index (LAI) data set for the period July 1981 to December 2006 derived from AVHRR Normalized Difference Vegetation Index (NDVI) data is described. The physically based algorithm is detailed in the first of the two part series. Here, the implementation, production and evaluation of the data set are described. The data set is evaluated both by direct comparisons to ground data and indirectly through inter-comparisons with similar data sets. This indirect validation showed satisfactory agreement with existing LAI products, importantly MODIS, at a range of spatial scales, and significant correlations with key climate variables in areas where temperature and precipitation limit plant growth. The data set successfully reproduced well-documented spatio-temporal trends and inter-annual variations in vegetation activity in the northern latitudes and semi-arid tropics. Comparison with plot scale field measurements over homogeneous vegetation patches indicated a 7% underestimation when all major vegetation types are taken into account. The error in mean values obtained from distributions of AVHRR LAI and high-resolution field LAI maps for different biomes is within 0.5 LAI for six out of the ten selected sites. These validation exercises though limited by the amount of field data, and thus less than comprehensive, indicated satisfactory agreement between the LAI product and field measurements. Overall, the intercomparison with short-term LAI data sets, evaluation of long term trends with known variations in climate variables, and validation with field measurements together build confidence in the utility of this new 26 year LAI record for long term vegetation monitoring and modeling studies.

  6. Model validation and error estimation of tsunami runup using high resolution data in Sadeng Port, Gunungkidul, Yogyakarta

    Science.gov (United States)

    Basith, Abdul; Prakoso, Yudhono; Kongko, Widjo

    2017-07-01

    A tsunami model using high resolution geometric data is indispensable in efforts to tsunami mitigation, especially in tsunami prone areas. It is one of the factors that affect the accuracy results of numerical modeling of tsunami. Sadeng Port is a new infrastructure in the Southern Coast of Java which could potentially hit by massive tsunami from seismic gap. This paper discusses validation and error estimation of tsunami model created using high resolution geometric data in Sadeng Port. Tsunami model validation uses the height wave of Tsunami Pangandaran 2006 recorded by Tide Gauge of Sadeng. Tsunami model will be used to accommodate the tsunami numerical modeling involves the parameters of earthquake-tsunami which is derived from the seismic gap. The validation results using t-test (student) shows that the height of the tsunami modeling results and observation in Tide Gauge of Sadeng are considered statistically equal at 95% confidence level and the value of the RMSE and NRMSE are 0.428 m and 22.12%, while the differences of tsunami wave travel time is 12 minutes.

  7. DBCG hypo trial validation of radiotherapy parameters from a national data bank versus manual reporting

    DEFF Research Database (Denmark)

    Brink, Carsten; Lorenzen, Ebbe L; Krogh, Simon Long

    2018-01-01

    of dose information, since the two patients had been treated with an electron boost based on a manual calculation, thus data was not exported to the data bank, and this was not detected prior to comparison with the manual data. For a few database fields in the manual data an ambiguity of the parameter...... definition of the specific field is seen in the data. This was not the case for the data bank, which extract all data consistently. CONCLUSIONS: In terms of data quality the data bank is superior to manually reported values. However, there is a need to allocate resources for checking the validity...... of the available data as well as ensuring that all relevant data is present. The data bank contains more detailed information, and thus facilitates research related to the actual dose distribution in the patients....

  8. Atmospheric correction at AERONET locations: A new science and validation data set

    Science.gov (United States)

    Wang, Y.; Lyapustin, A.I.; Privette, J.L.; Morisette, J.T.; Holben, B.

    2009-01-01

    This paper describes an Aerosol Robotic Network (AERONET)-based Surface Reflectance Validation Network (ASRVN) and its data set of spectral surface bidirectional reflectance and albedo based on Moderate Resolution Imaging Spectroradiometer (MODIS) TERRA and AQUA data. The ASRVN is an operational data collection and processing system. It receives 50 ?? 50 km2; subsets of MODIS level 1B (L1B) data from MODIS adaptive processing system and AERONET aerosol and water-vapor information. Then, it performs an atmospheric correction (AC) for about 100 AERONET sites based on accurate radiative-transfer theory with complex quality control of the input data. The ASRVN processing software consists of an L1B data gridding algorithm, a new cloud-mask (CM) algorithm based on a time-series analysis, and an AC algorithm using ancillary AERONET aerosol and water-vapor data. The AC is achieved by fitting the MODIS top-of-atmosphere measurements, accumulated for a 16-day interval, with theoretical reflectance parameterized in terms of the coefficients of the Li SparseRoss Thick (LSRT) model of the bidirectional reflectance factor (BRF). The ASRVN takes several steps to ensure high quality of results: 1) the filtering of opaque clouds by a CM algorithm; 2) the development of an aerosol filter to filter residual semitransparent and subpixel clouds, as well as cases with high inhomogeneity of aerosols in the processing area; 3) imposing the requirement of the consistency of the new solution with previously retrieved BRF and albedo; 4) rapid adjustment of the 16-day retrieval to the surface changes using the last day of measurements; and 5) development of a seasonal backup spectral BRF database to increase data coverage. The ASRVN provides a gapless or near-gapless coverage for the processing area. The gaps, caused by clouds, are filled most naturally with the latest solution for a given pixel. The ASRVN products include three parameters of the LSRT model (kL, kG, and kV), surface albedo

  9. The benchmark experiment on slab beryllium with D–T neutrons for validation of evaluated nuclear data

    Energy Technology Data Exchange (ETDEWEB)

    Nie, Y., E-mail: nieyb@ciae.ac.cn [Science and Technology on Nuclear Data Laboratory, China Institute of Atomic Energy, Beijing 102413 (China); Ren, J.; Ruan, X.; Bao, J. [Science and Technology on Nuclear Data Laboratory, China Institute of Atomic Energy, Beijing 102413 (China); Han, R. [Institute of Modern Physics, Chinese Academy of Sciences, Lanzhou 730000 (China); Zhang, S. [Institute of Modern Physics, Chinese Academy of Sciences, Lanzhou 730000 (China); Inner Mongolia University for the Nationalities, Inner Mongolia, Tongliao 028000 (China); Huang, H.; Li, X. [Science and Technology on Nuclear Data Laboratory, China Institute of Atomic Energy, Beijing 102413 (China); Ding, Y. [Science and Technology on Nuclear Data Laboratory, China Institute of Atomic Energy, Beijing 102413 (China); School of Nuclear Science and Technology, Lanzhou University, Lanzhou 730000 (China); Wu, H.; Liu, P.; Zhou, Z. [Science and Technology on Nuclear Data Laboratory, China Institute of Atomic Energy, Beijing 102413 (China)

    2016-04-15

    Highlights: • Evaluated data for beryllium are validated by a high precision benchmark experiment. • Leakage neutron spectra from pure beryllium slab are measured at 61° and 121° using time-of-flight method. • The experimental results are compared with the MCNP-4B calculations with the evaluated data from different libraries. - Abstract: Beryllium is the most favored neutron multiplier candidate for solid breeder blankets of future fusion power reactors. However, beryllium nuclear data are differently presented in modern nuclear data evaluations. In order to validate the evaluated nuclear data on beryllium, in the present study, a benchmark experiment has been performed at China Institution of Atomic Energy (CIAE). Neutron leakage spectra from pure beryllium slab samples were measured at 61° and 121° using time-of-flight method. The experimental results were compared with the calculated ones by MCNP-4B simulation, using the evaluated data of beryllium from the CENDL-3.1, ENDF/B-VII.1 and JENDL-4.0 libraries. From the comparison between the measured and the calculated spectra, it was found that the calculation results based on CENDL-3.1 caused overestimation in the energy range from about 3–12 MeV at 61°, while at 121°, all the libraries led to underestimation below 3 MeV.

  10. The benchmark experiment on slab beryllium with D–T neutrons for validation of evaluated nuclear data

    International Nuclear Information System (INIS)

    Nie, Y.; Ren, J.; Ruan, X.; Bao, J.; Han, R.; Zhang, S.; Huang, H.; Li, X.; Ding, Y.; Wu, H.; Liu, P.; Zhou, Z.

    2016-01-01

    Highlights: • Evaluated data for beryllium are validated by a high precision benchmark experiment. • Leakage neutron spectra from pure beryllium slab are measured at 61° and 121° using time-of-flight method. • The experimental results are compared with the MCNP-4B calculations with the evaluated data from different libraries. - Abstract: Beryllium is the most favored neutron multiplier candidate for solid breeder blankets of future fusion power reactors. However, beryllium nuclear data are differently presented in modern nuclear data evaluations. In order to validate the evaluated nuclear data on beryllium, in the present study, a benchmark experiment has been performed at China Institution of Atomic Energy (CIAE). Neutron leakage spectra from pure beryllium slab samples were measured at 61° and 121° using time-of-flight method. The experimental results were compared with the calculated ones by MCNP-4B simulation, using the evaluated data of beryllium from the CENDL-3.1, ENDF/B-VII.1 and JENDL-4.0 libraries. From the comparison between the measured and the calculated spectra, it was found that the calculation results based on CENDL-3.1 caused overestimation in the energy range from about 3–12 MeV at 61°, while at 121°, all the libraries led to underestimation below 3 MeV.

  11. Kepler Data Validation I: Architecture, Diagnostic Tests, and Data Products for Vetting Transiting Planet Candidates

    Science.gov (United States)

    Twicken, Joseph D.; Catanzarite, Joseph H.; Clarke, Bruce D.; Giroud, Forrest; Jenkins, Jon M.; Klaus, Todd C.; Li, Jie; McCauliff, Sean D.; Seader, Shawn E.; Tennenbaum, Peter; hide

    2018-01-01

    The Kepler Mission was designed to identify and characterize transiting planets in the Kepler Field of View and to determine their occurrence rates. Emphasis was placed on identification of Earth-size planets orbiting in the Habitable Zone of their host stars. Science data were acquired for a period of four years. Long-cadence data with 29.4 min sampling were obtained for approx. 200,000 individual stellar targets in at least one observing quarter in the primary Kepler Mission. Light curves for target stars are extracted in the Kepler Science Data Processing Pipeline, and are searched for transiting planet signatures. A Threshold Crossing Event is generated in the transit search for targets where the transit detection threshold is exceeded and transit consistency checks are satisfied. These targets are subjected to further scrutiny in the Data Validation (DV) component of the Pipeline. Transiting planet candidates are characterized in DV, and light curves are searched for additional planets after transit signatures are modeled and removed. A suite of diagnostic tests is performed on all candidates to aid in discrimination between genuine transiting planets and instrumental or astrophysical false positives. Data products are generated per target and planet candidate to document and display transiting planet model fit and diagnostic test results. These products are exported to the Exoplanet Archive at the NASA Exoplanet Science Institute, and are available to the community. We describe the DV architecture and diagnostic tests, and provide a brief overview of the data products. Transiting planet modeling and the search for multiple planets on individual targets are described in a companion paper. The final revision of the Kepler Pipeline code base is available to the general public through GitHub. The Kepler Pipeline has also been modified to support the Transiting Exoplanet Survey Satellite (TESS) Mission which is expected to commence in 2018.

  12. Kepler Data Validation I—Architecture, Diagnostic Tests, and Data Products for Vetting Transiting Planet Candidates

    Science.gov (United States)

    Twicken, Joseph D.; Catanzarite, Joseph H.; Clarke, Bruce D.; Girouard, Forrest; Jenkins, Jon M.; Klaus, Todd C.; Li, Jie; McCauliff, Sean D.; Seader, Shawn E.; Tenenbaum, Peter; Wohler, Bill; Bryson, Stephen T.; Burke, Christopher J.; Caldwell, Douglas A.; Haas, Michael R.; Henze, Christopher E.; Sanderfer, Dwight T.

    2018-06-01

    The Kepler Mission was designed to identify and characterize transiting planets in the Kepler Field of View and to determine their occurrence rates. Emphasis was placed on identification of Earth-size planets orbiting in the Habitable Zone of their host stars. Science data were acquired for a period of four years. Long-cadence data with 29.4 min sampling were obtained for ∼200,000 individual stellar targets in at least one observing quarter in the primary Kepler Mission. Light curves for target stars are extracted in the Kepler Science Data Processing Pipeline, and are searched for transiting planet signatures. A Threshold Crossing Event is generated in the transit search for targets where the transit detection threshold is exceeded and transit consistency checks are satisfied. These targets are subjected to further scrutiny in the Data Validation (DV) component of the Pipeline. Transiting planet candidates are characterized in DV, and light curves are searched for additional planets after transit signatures are modeled and removed. A suite of diagnostic tests is performed on all candidates to aid in discrimination between genuine transiting planets and instrumental or astrophysical false positives. Data products are generated per target and planet candidate to document and display transiting planet model fit and diagnostic test results. These products are exported to the Exoplanet Archive at the NASA Exoplanet Science Institute, and are available to the community. We describe the DV architecture and diagnostic tests, and provide a brief overview of the data products. Transiting planet modeling and the search for multiple planets on individual targets are described in a companion paper. The final revision of the Kepler Pipeline code base is available to the general public through GitHub. The Kepler Pipeline has also been modified to support the Transiting Exoplanet Survey Satellite (TESS) Mission which is expected to commence in 2018.

  13. Innovative use of soft data for the validation of a rainfall-runoff model forced by remote sensing data

    Science.gov (United States)

    van Emmerik, Tim; Eilander, Dirk; Piet, Marijn; Mulder, Gert

    2013-04-01

    The Chamcar Bei catchment in southern Cambodia is a typical ungauged basin. Neither meteorological data or discharge measurements are available. In this catchment, local farmers are highly dependent on the irrigation system. However, due to the unreliability of the water supply, it was required to make a hydrological model, with which further improvements of the irrigation system could be planned. First, we used knowledge generated in the IAHS decade on Predictions in Ungauged Basins (PUB) to estimate the annual water balance of the Chamcar Bei catchment. Next, using remotely sensed precipitation, vegetation, elevation and transpiration data, a monthly rainfall-runoff model has been developed. The rainfall-runoff model was linked to the irrigation system reservoir, which allowed to validate the model based on soft data such as historical knowledge of the reservoir water level and groundwater levels visible in wells. This study shows that combining existing remote sensing data and soft ground data can lead to useful modeling results. The approach presented in this study can be applied in other ungauged basins, which can be extremely helpful in managing water resources in developing countries.

  14. A study of existing experimental data and validation process for evaluated high energy nuclear data. Report of task force on integral test for JENDL High Energy File in Japanese Nuclear Data Committee

    International Nuclear Information System (INIS)

    Oyama, Yukio; Baba, Mamoru; Watanabe, Yukinobu

    1998-11-01

    JENDL High Energy File (JENDL-HE) is being produced by Japanese Nuclear Data Committee (JNDC) to provide common fundamental nuclear data in the intermediate energy region for many applications concerning a basic research, an accelerator-driven nuclear waste transmutation, a fusion material study, and medical applications like the radiation therapy. The first version of JENDL-HE, which contains the evaluated nuclear data up to 50 MeV, is planned to release in 1998. However, a method of integral test with which we can validate the high-energy nuclear data file has not been established. The validation of evaluated nuclear data through the integral tests is necessary to promote utilization of JENDL-HE. JNDC set up a task force in 1997 to discuss the problems concerning the integral tests of JENDL-HE. The task force members have surveyed and studied the current status of the problems for a year to obtain a guideline for development of the high-energy nuclear database. This report summarizes the results of the survey and study done by the task force for JNDC. (author)

  15. Protocol for Validation of the Land Surface Reflectance Fundamental Climate Data Record using AERONET: Application to the Global MODIS and VIIRS Data Records

    Science.gov (United States)

    Roger, J. C.; Vermote, E.; Holben, B. N.

    2014-12-01

    The land surface reflectance is a fundamental climate data record at the basis of the derivation of other climate data records (Albedo, LAI/Fpar, Vegetation indices) and a key parameter in the understanding of the land-surface-climate processes. It is essential that a careful validation of its uncertainties is performed on a global and continuous basis. One approach is the direct comparison of this product with ground measurements but that approach presents several issues related to scale, the episodic nature of ground measurements and the global representativeness. An alternative is to compare the surface reflectance product to reference reflectance determined from Top of atmosphere reflectance corrected using accurate radiative transfer code and very detailed measurements of the atmosphere obtained over the AERONET sites (Vermote and al, 2014, RSE) which allows to test for a large range of aerosol characteristics; formers being important inputs for atmospheric corrections. However, the application of this method necessitates the definition of a very detailed protocol for the use of AERONET data especially as far as size distribution and absorption are concerned, so that alternative validation methods or protocols could be compared. This paper describes the protocol we have been working on based on our experience with the AERONET data and its application to the MODIS and VIIRS record.

  16. Expert systems application to plant diagnosis and sensor data validation

    International Nuclear Information System (INIS)

    Hashemi, S.; Hajek, B.K.; Miller, D.W.; Chandrasekaran, B.; Josephson, J.R.

    1986-01-01

    In a nuclear power plant, over 2000 alarms and displays are available to the operator. For any given set of alarms and displays, the operator must be able to diagnose and correct the problem (s) quickly and accurately. At the same time, the operator is expected to distinguish the plant system faults from instrumentation channel failures and drifts. Needs for plant operator aids have been considered since the accident at TMI. Many of these aids are of the form of the Safety Parameter Display Systems and offer improved methods of displaying otherwise available data to the operator in a more concise and summarized format. diagnosis, however, remains a desirable objective of an operator aid. At The Ohio State University, faculty and students in nuclear engineering and computer science have evaluated this problem. The results of these studies have shown that plant diagnosis and sensor data validation must be considered as one integral problem and cannot be isolated from one another. Otherwise, an incorrect diagnosis based on faulty instrument information might be provided to the operator. In this study, the Knowlege Based System (KBS) technology is being incorporated to accomplish a final goal of an intelligent operator aid system

  17. Development and validity data of the Brazilian Internet Study on Temperament and Psychopathology (BRAINSTEP).

    Science.gov (United States)

    Lara, Diogo R; Ottoni, Gustavo L; Brunstein, Miriam G; Frozi, Julia; de Carvalho, Hudson W; Bisol, Luísa W

    2012-12-10

    The internet provides a research opportunity for psychiatry and psychology. This article presents the development and preliminary data of a large web-survey created to study how temperament relates to other psychological measures, behavior and psychiatric disorders. We used the Affective and Emotional Composite Temperament Scale (AFECTS) to evaluate temperament and we selected several self-report instruments to evaluate behavior, psychological constructs and mental disorders. The system provides anonymous psychological (phase 1) and psychiatric (phase 2) feedback and includes questions to assess the validity of the answers. Each phase has around 450 questions. This system was broadcast utilizing Brazilian media. After the exclusion of 21.5% of the volunteers (those who failed the validation questions), 41,427 participants concluded the first part of the system (mean age=31.2±10.5 yrs, 26.9% males), and 21,836 (mean age=32.5±10.9 yrs, 25.1% males) completed phase 2. Around 25% have received a psychiatric diagnosis from a mental health professional. Demographic and temperament profiles of those who completed either only 80 questions, only phase 1, or the whole system were similar. The rate of non-serious answers (e.g. on bizarre behaviors) was very low and congruency of answers was very high. The internal consistency of classical trait scales (TCI-R and PANAS) was high (Cronbach's alpha>0.80) for all dimensions. Relatively high dropout rate due to the length of the process and an overrepresentation of female, young and well-educated subjects. The BRAINSTEP provides valid and abundant data on psychological and psychiatric measures. Copyright © 2012 Elsevier B.V. All rights reserved.

  18. Support Vector Data Description Model to Map Specific Land Cover with Optimal Parameters Determined from a Window-Based Validation Set

    Directory of Open Access Journals (Sweden)

    Jinshui Zhang

    2017-04-01

    Full Text Available This paper developed an approach, the window-based validation set for support vector data description (WVS-SVDD, to determine optimal parameters for support vector data description (SVDD model to map specific land cover by integrating training and window-based validation sets. Compared to the conventional approach where the validation set included target and outlier pixels selected visually and randomly, the validation set derived from WVS-SVDD constructed a tightened hypersphere because of the compact constraint by the outlier pixels which were located neighboring to the target class in the spectral feature space. The overall accuracies for wheat and bare land achieved were as high as 89.25% and 83.65%, respectively. However, target class was underestimated because the validation set covers only a small fraction of the heterogeneous spectra of the target class. The different window sizes were then tested to acquire more wheat pixels for validation set. The results showed that classification accuracy increased with the increasing window size and the overall accuracies were higher than 88% at all window size scales. Moreover, WVS-SVDD showed much less sensitivity to the untrained classes than the multi-class support vector machine (SVM method. Therefore, the developed method showed its merits using the optimal parameters, tradeoff coefficient (C and kernel width (s, in mapping homogeneous specific land cover.

  19. External validation of a decision tree early warning score using only laboratory data

    DEFF Research Database (Denmark)

    Holm Atkins, Tara E; Öhman, Malin C; Brabrand, Mikkel

    2018-01-01

    INTRODUCTION: Early warning scores (EWS) have been developed to identify the degree of illness severity among acutely ill patients. One system, The Laboratory Decision Tree Early Warning Score (LDT-EWS) is wholly laboratory data based. Laboratory data was used in the development of a rare...... computerized method, developing a decision tree analysis. This article externally validates LDT-EWS, which is obligatory for an EWS before clinical use. METHOD: We conducted a retrospective review of prospectively collected data based on a time limited sample of all patients admitted through the medical......) and calibration (precision) as Hosmer-Lemeshow Goodness of fit test. RESULTS: A total of 5858 patients were admitted and 4902 included (83.7%). In-hospital mortality in our final dataset (n=4902) was 3.5%. Discriminatory power (95% CI), identifying in-hospital death was 0.809 (0.777-0.842). Calibration was good...

  20. Proceedings of the workshop on integral experiment covariance data for critical safety validation

    Energy Technology Data Exchange (ETDEWEB)

    Stuke, Maik (ed.)

    2016-04-15

    For some time, attempts to quantify the statistical dependencies of critical experiments and to account for them properly in validation procedures were discussed in the literature by various groups. Besides the development of suitable methods especially the quality and modeling issues of the freely available experimental data are in the focus of current discussions, carried out for example in the Expert Group on Uncertainty Analysis for Criticality Safety Assessment (UACSA) of the OECD-NEA Nuclear Science Committee. The same committee compiles and publishes also the freely available experimental data in the International Handbook of Evaluated Criticality Safety Benchmark Experiments. Most of these experiments were performed as series and might share parts of experimental setups leading to correlated results. The quality of the determination of these correlations and the underlying covariance data depend strongly on the quality of the documentation of experiments.

  1. Proceedings of the workshop on integral experiment covariance data for critical safety validation

    International Nuclear Information System (INIS)

    Stuke, Maik

    2016-04-01

    For some time, attempts to quantify the statistical dependencies of critical experiments and to account for them properly in validation procedures were discussed in the literature by various groups. Besides the development of suitable methods especially the quality and modeling issues of the freely available experimental data are in the focus of current discussions, carried out for example in the Expert Group on Uncertainty Analysis for Criticality Safety Assessment (UACSA) of the OECD-NEA Nuclear Science Committee. The same committee compiles and publishes also the freely available experimental data in the International Handbook of Evaluated Criticality Safety Benchmark Experiments. Most of these experiments were performed as series and might share parts of experimental setups leading to correlated results. The quality of the determination of these correlations and the underlying covariance data depend strongly on the quality of the documentation of experiments.

  2. Validation of self-reported erythema

    DEFF Research Database (Denmark)

    Petersen, B; Thieden, E; Lerche, C M

    2013-01-01

    Most epidemiological data of sunburn related to skin cancer have come from self-reporting in diaries and questionnaires. We thought it important to validate the reliability of such data.......Most epidemiological data of sunburn related to skin cancer have come from self-reporting in diaries and questionnaires. We thought it important to validate the reliability of such data....

  3. A mixed methods inquiry into the validity of data

    Directory of Open Access Journals (Sweden)

    Vaarst Mette

    2008-07-01

    Full Text Available Abstract Background Research in herd health management solely using a quantitative approach may present major challenges to the interpretation of the results, because the humans involved may have responded to their observations based on previous experiences and own beliefs. This challenge can be met through increased awareness and dialogue between researchers and farmers or other stakeholders about the background for data collection related to management and changes in management. By integrating quantitative and qualitative research methods in a mixed methods research approach, the researchers will improve their understanding of this potential bias of the observed data and farms, which will enable them to obtain more useful results of quantitative analyses. Case description An example is used to illustrate the potentials of combining quantitative and qualitative approaches to herd health related data analyses. The example is based on two studies on bovine metritis. The first study was a quantitative observational study of risk factors for metritis in Danish dairy cows based on data from the Danish Cattle Database. The other study was a semi-structured interview study involving 20 practicing veterinarians with the aim to gain insight into veterinarians' decision making when collecting and processing data related to metritis. Discussion and Evaluation The relations between risk factors and metritis in the first project supported the findings in several other quantitative observational studies; however, the herd incidence risk was highly skewed. There may be simple practical reasons for this, e.g. underreporting and differences in the veterinarians' decision making. Additionally, the interviews in the second project identified several problems with correctness and validity of data regarding the occurrence of metritis because of differences regarding case definitions and thresholds for treatments between veterinarians. Conclusion Studies where

  4. Using a mobile app and mobile workforce to validate data about emergency public health resources.

    Science.gov (United States)

    Chang, Anna Marie; Leung, Alison C; Saynisch, Olivia; Griffis, Heather; Hill, Shawndra; Hershey, John C; Becker, Lance B; Asch, David A; Seidman, Ariel; Merchant, Raina Martha

    2014-07-01

    Social media and mobile applications that allow people to work anywhere are changing the way people can contribute and collaborate. We sought to determine the feasibility of using mobile workforce technology to validate the locations of automated external defibrillators (AEDs), an emergency public health resource. We piloted the use of a mobile workforce application, to verify the location of 40 AEDs in Philadelphia county. AEDs were pre-identified in public locations for baseline data. The task of locating AEDs was posted online for a mobile workforce from October 2011 to January 2012. Participants were required to submit a mobile phone photo of AEDs and descriptions of the location. Thirty-five of the 40 AEDs were identified within the study period. Most, 91% (32/35) of the submitted AED photo information was confirmed project baseline data. Participants also provided additional data such as business hours and other nearby AEDs. It is feasible to engage a mobile workforce to complete health research-related tasks. Participants were able to validate information about emergency public health resources. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://group.bmj.com/group/rights-licensing/permissions.

  5. Validation of 18FDG biodistribution data in healthy mice obtained with G.E. LABPET4

    International Nuclear Information System (INIS)

    Rocha, Adriana Marcia Guimaraes; Mendes, Bruno Melo; Malamut, Carlos; Silva, Juliana Batista da; Campos, Danielle Cunha; Santos, Priscilla Figueiredo

    2013-01-01

    The aim of this study is to validate biodistribution data obtained with CDTN's MicroPET. To achieve this goal, correction and image acquisition procedures were established. 1 '8FDG dynamic images of 90 minutes were obtained following these procedures for Swiss healthy mice. Biodistribution data obtained after quantification of acquired images were compared with data available in literature. Considering the uptake time of 60 minutes and similar animal handling, data obtained in this work showed a satisfactory agreement with reference data. Some evaluated organs/tissues showed high interindividual variability. These findings are consistent with those observed in reference literature. However, improvements in VOI positioning VOI technique as well as increasing the number of animals (n) per group can minimize this problem. (author)

  6. Validation of RNAi Silencing Efficiency Using Gene Array Data shows 18.5% Failure Rate across 429 Independent Experiments

    Directory of Open Access Journals (Sweden)

    Gyöngyi Munkácsy

    2016-01-01

    Full Text Available No independent cross-validation of success rate for studies utilizing small interfering RNA (siRNA for gene silencing has been completed before. To assess the influence of experimental parameters like cell line, transfection technique, validation method, and type of control, we have to validate these in a large set of studies. We utilized gene chip data published for siRNA experiments to assess success rate and to compare methods used in these experiments. We searched NCBI GEO for samples with whole transcriptome analysis before and after gene silencing and evaluated the efficiency for the target and off-target genes using the array-based expression data. Wilcoxon signed-rank test was used to assess silencing efficacy and Kruskal–Wallis tests and Spearman rank correlation were used to evaluate study parameters. All together 1,643 samples representing 429 experiments published in 207 studies were evaluated. The fold change (FC of down-regulation of the target gene was above 0.7 in 18.5% and was above 0.5 in 38.7% of experiments. Silencing efficiency was lowest in MCF7 and highest in SW480 cells (FC = 0.59 and FC = 0.30, respectively, P = 9.3E−06. Studies utilizing Western blot for validation performed better than those with quantitative polymerase chain reaction (qPCR or microarray (FC = 0.43, FC = 0.47, and FC = 0.55, respectively, P = 2.8E−04. There was no correlation between type of control, transfection method, publication year, and silencing efficiency. Although gene silencing is a robust feature successfully cross-validated in the majority of experiments, efficiency remained insufficient in a significant proportion of studies. Selection of cell line model and validation method had the highest influence on silencing proficiency.

  7. Use of integral experiments in support to the validation of JEFF-3.2 nuclear data evaluation

    Science.gov (United States)

    Leclaire, Nicolas; Cochet, Bertrand; Jinaphanh, Alexis; Haeck, Wim

    2017-09-01

    For many years now, IRSN has developed its own Monte Carlo continuous energy capability, which allows testing various nuclear data libraries. In that prospect, a validation database of 1136 experiments was built from cases used for the validation of the APOLLO2-MORET 5 multigroup route of the CRISTAL V2.0 package. In this paper, the keff obtained for more than 200 benchmarks using the JEFF-3.1.1 and JEFF-3.2 libraries are compared to benchmark keff values and main discrepancies are analyzed regarding the neutron spectrum. Special attention is paid on benchmarks for which the results have been highly modified between both JEFF-3 versions.

  8. ValWorkBench: an open source Java library for cluster validation, with applications to microarray data analysis.

    Science.gov (United States)

    Giancarlo, R; Scaturro, D; Utro, F

    2015-02-01

    The prediction of the number of clusters in a dataset, in particular microarrays, is a fundamental task in biological data analysis, usually performed via validation measures. Unfortunately, it has received very little attention and in fact there is a growing need for software tools/libraries dedicated to it. Here we present ValWorkBench, a software library consisting of eleven well known validation measures, together with novel heuristic approximations for some of them. The main objective of this paper is to provide the interested researcher with the full software documentation of an open source cluster validation platform having the main features of being easily extendible in a homogeneous way and of offering software components that can be readily re-used. Consequently, the focus of the presentation is on the architecture of the library, since it provides an essential map that can be used to access the full software documentation, which is available at the supplementary material website [1]. The mentioned main features of ValWorkBench are also discussed and exemplified, with emphasis on software abstraction design and re-usability. A comparison with existing cluster validation software libraries, mainly in terms of the mentioned features, is also offered. It suggests that ValWorkBench is a much needed contribution to the microarray software development/algorithm engineering community. For completeness, it is important to mention that previous accurate algorithmic experimental analysis of the relative merits of each of the implemented measures [19,23,25], carried out specifically on microarray data, gives useful insights on the effectiveness of ValWorkBench for cluster validation to researchers in the microarray community interested in its use for the mentioned task. Copyright © 2014 Elsevier Ireland Ltd. All rights reserved.

  9. Validation of two case definitions to identify pressure ulcers using hospital administrative data.

    Science.gov (United States)

    Ho, Chester; Jiang, Jason; Eastwood, Cathy A; Wong, Holly; Weaver, Brittany; Quan, Hude

    2017-08-28

    Pressure ulcer development is a quality of care indicator, as pressure ulcers are potentially preventable. Yet pressure ulcer is a leading cause of morbidity, discomfort and additional healthcare costs for inpatients. Methods are lacking for accurate surveillance of pressure ulcer in hospitals to track occurrences and evaluate care improvement strategies. The main study aim was to validate hospital discharge abstract database (DAD) in recording pressure ulcers against nursing consult reports, and to calculate prevalence of pressure ulcers in Alberta, Canada in DAD. We hypothesised that a more inclusive case definition for pressure ulcers would enhance validity of cases identified in administrative data for research and quality improvement purposes. A cohort of patients with pressure ulcers were identified from enterostomal (ET) nursing consult documents at a large university hospital in 2011. There were 1217 patients with pressure ulcers in ET nursing documentation that were linked to a corresponding record in DAD to validate DAD for correct and accurate identification of pressure ulcer occurrence, using two case definitions for pressure ulcer. Using pressure ulcer definition 1 (7 codes), prevalence was 1.4%, and using definition 2 (29 codes), prevalence was 4.2% after adjusting for misclassifications. The results were lower than expected. Definition 1 sensitivity was 27.7% and specificity was 98.8%, while definition 2 sensitivity was 32.8% and specificity was 95.9%. Pressure ulcer in both DAD and ET consultation increased with age, number of comorbidities and length of stay. DAD underestimate pressure ulcer prevalence. Since various codes are used to record pressure ulcers in DAD, the case definition with more codes captures more pressure ulcer cases, and may be useful for monitoring facility trends. However, low sensitivity suggests that this data source may not be accurate for determining overall prevalence, and should be cautiously compared with other

  10. Validation of a CFD methodology for positive displacement LVAD analysis using PIV data.

    Science.gov (United States)

    Medvitz, Richard B; Reddy, Varun; Deutsch, Steve; Manning, Keefe B; Paterson, Eric G

    2009-11-01

    Computational fluid dynamics (CFD) is used to asses the hydrodynamic performance of a positive displacement left ventricular assist device. The computational model uses implicit large eddy simulation direct resolution of the chamber compression and modeled valve closure to reproduce the in vitro results. The computations are validated through comparisons with experimental particle image velocimetry (PIV) data. Qualitative comparisons of flow patterns, velocity fields, and wall-shear rates demonstrate a high level of agreement between the computations and experiments. Quantitatively, the PIV and CFD show similar probed velocity histories, closely matching jet velocities and comparable wall-strain rates. Overall, it has been shown that CFD can provide detailed flow field and wall-strain rate data, which is important in evaluating blood pump performance.

  11. Use of Sensitivity and Uncertainty Analysis to Select Benchmark Experiments for the Validation of Computer Codes and Data

    International Nuclear Information System (INIS)

    Elam, K.R.; Rearden, B.T.

    2003-01-01

    Sensitivity and uncertainty analysis methodologies under development at Oak Ridge National Laboratory were applied to determine whether existing benchmark experiments adequately cover the area of applicability for the criticality code and data validation of PuO 2 and mixed-oxide (MOX) powder systems. The study examined three PuO 2 powder systems and four MOX powder systems that would be useful for establishing mass limits for a MOX fuel fabrication facility. Using traditional methods to choose experiments for criticality analysis validation, 46 benchmark critical experiments were identified as applicable to the PuO 2 powder systems. However, only 14 experiments were thought to be within the area of applicability for dry MOX powder systems.The applicability of 318 benchmark critical experiments, including the 60 experiments initially identified, was assessed. Each benchmark and powder system was analyzed using the Tools for Sensitivity and UNcertainty Analysis Methodology Implementation (TSUNAMI) one-dimensional (TSUNAMI-1D) or TSUNAMI three-dimensional (TSUNAMI-3D) sensitivity analysis sequences, which will be included in the next release of the SCALE code system. This sensitivity data and cross-section uncertainty data were then processed with TSUNAMI-IP to determine the correlation of each application to each experiment in the benchmarking set. Correlation coefficients are used to assess the similarity between systems and determine the applicability of one system for the code and data validation of another.The applicability of most of the experiments identified using traditional methods was confirmed by the TSUNAMI analysis. In addition, some PuO 2 and MOX powder systems were determined to be within the area of applicability of several other benchmarks that would not have been considered using traditional methods. Therefore, the number of benchmark experiments useful for the validation of these systems exceeds the number previously expected. The TSUNAMI analysis

  12. Data collection, validation, and description for the Oak Ridge nuclear facilities mortality study

    International Nuclear Information System (INIS)

    Watkins, J.P.; Reagan, J.L.; Cragle, D.L.; West, C.M.; Tankersley, W.G.; Crawford-Brown, D.J.

    1995-01-01

    To investigate the long-term health effects of protracted occupational exposure to low levels of ionizing radiation, a mortality study was initiated by pooling data for 118,588 workers hired between 1943 and 1982, at three Department of Energy (DOE) facilities in Oak Ridge, Tennessee, with follow-up through 1984. Topics for this discussion will include issues involving the collection and validation of data for individuals in the study cohort, and characteristics of their demographic and radiation exposure data. Since the data were compiled between the late 1960s and the present under the direction of several principal investigators, it was essential to verify data precision and to understand how exposure data were generated prior to beginning any analysis. A stratified random sample of workers in the cohort was chosen for verification of their computerized data as it appeared in the database. Original source documents were reviewed to verify demographic data, as well as internal and external radiation exposure data. Extensive effort was expended to document the personal radiation monitoring policies and types of dosimeters used at each facility over the 42 years included in the study. Characteristics of internal and external exposure data by facility and year were examined by graphical methods with the intent of combining these monitoring data over time and across facilities

  13. CFD validation experiments for hypersonic flows

    Science.gov (United States)

    Marvin, Joseph G.

    1992-01-01

    A roadmap for CFD code validation is introduced. The elements of the roadmap are consistent with air-breathing vehicle design requirements and related to the important flow path components: forebody, inlet, combustor, and nozzle. Building block and benchmark validation experiments are identified along with their test conditions and measurements. Based on an evaluation criteria, recommendations for an initial CFD validation data base are given and gaps identified where future experiments could provide new validation data.

  14. Utilizing the social media data to validate 'climate change' indices

    Science.gov (United States)

    Molodtsova, T.; Kirilenko, A.; Stepchenkova, S.

    2013-12-01

    Reporting the observed and modeled changes in climate to public requires the measures understandable by the general audience. E.g., the NASA GISS Common Sense Climate Index (Hansen et al., 1998) reports the change in climate based on six practically observable parameters such as the air temperature exceeding the norm by one standard deviation. The utility of the constructed indices for reporting climate change depends, however, on an assumption that the selected parameters are felt and connected with the changing climate by a non-expert, which needs to be validated. Dynamic discussion of climate change issues in social media may provide data for this validation. We connected the intensity of public discussion of climate change in social networks with regional weather variations for the territory of the USA. We collected the entire 2012 population of Twitter microblogging activity on climate change topic, accumulating over 1.8 million separate records (tweets) globally. We identified the geographic location of the tweets and associated the daily and weekly intensity of twitting with the following parameters of weather for these locations: temperature anomalies, 'hot' temperature anomalies, 'cold' temperature anomalies, heavy rain/snow events. To account for non-weather related events we included the articles on climate change from the 'prestige press', a collection of major newspapers. We found that the regional changes in parameters of weather significantly affect the number of tweets published on climate change. This effect, however, is short-lived and varies throughout the country. We found that in different locations different weather parameters had the most significant effect on climate change microblogging activity. Overall 'hot' temperature anomalies had significant influence on climate change twitting intensity.

  15. The F4E programme on nuclear data validation and nuclear instrumentation techniques for TBM in ITER

    Czech Academy of Sciences Publication Activity Database

    Leichtle, D.; Angelone, M.; Batistoni, P.; Calderoni, P.; Fischer, U.; Izquierdo, J.; Klix, A.; Kodeli, I.; Kuc, T.; Lilley, S.; Majerle, Mitja; Packer, L.; Pillon, M.; Pohorecki, W.; Snoj, L.; Villari, R.

    2014-01-01

    Roč. 89, 9-10 (2014), s. 2169-2173 ISSN 0920-3796 Institutional support: RVO:61389005 Keywords : nuclear data * TBM * neutronic sensors * validation experiments Subject RIV: BG - Nuclear, Atomic and Molecular Physics, Colliders Impact factor: 1.152, year: 2014

  16. Validation of large-angle scattering data via shadow-bar experiment

    Energy Technology Data Exchange (ETDEWEB)

    Ohnishi, S., E-mail: ohnishi@nmri.go.jp [National Maritime Research Institute, 6-38-1, Shinkawa, Mitaka, Tokyo 181-0004 (Japan); Tamaki, S.; Murata, I. [Osaka University, 1-14-16-1, Yamadaoka, Suita-si, Osaka 565-0871 (Japan)

    2016-11-15

    Highlights: • An experiment to validate large-angle scattering cross section is conducted. • Pieces of Nb foil are set behind a shadow bar to obtain the {sup 92m}Nb production rates. • The results calculated using ENDF/B-VI library data exhibit a 57% overestimation. • The adjustment of cross section in large-angle region makes the C/E close to 1. - Abstract: An experiment emphasizing the influence of large-angle scattering on nuclear data was conducted, in which a Fe shadow bar and a Fe slab target were placed before a deuterium–tritium fusion (DT) neutron source. Two Nb foils were set on both sides of the shadow bar in order to monitor the neutron source intensity and to measure the neutrons scattered from the slab target. The {sup 93}Nb(n,2n){sup 92m}Nb reaction rate of the foil was measured following the DT neutron irradiation and calculated using the MCNP5 Monte Carlo radiation transportation code. The {sup 92m}Nb production rates calculated using data from the JEFF-3.1 and JENDL-4.0 libraries agreed with that measured in the experiment, while the result calculated using data from the ENDF/B-VI library exhibited a 57% overestimation. Because the sensitivity of the {sup 92m}Nb production rate to the scattering angular distribution was large in the angular region between scattering direction cosines of −0.9 and −0.4, the scattering angular distribution was adjusted in that region. This adjustment resulted in a calculation-to-experiment ratio close to 1, but had little influence on the existing integral benchmark experiment.

  17. Model-based clinical dose optimization for phenobarbital in neonates: An illustration of the importance of data sharing and external validation.

    Science.gov (United States)

    Völler, Swantje; Flint, Robert B; Stolk, Leo M; Degraeuwe, Pieter L J; Simons, Sinno H P; Pokorna, Paula; Burger, David M; de Groot, Ronald; Tibboel, Dick; Knibbe, Catherijne A J

    2017-11-15

    Particularly in the pediatric clinical pharmacology field, data-sharing offers the possibility of making the most of all available data. In this study, we utilize previously collected therapeutic drug monitoring (TDM) data of term and preterm newborns to develop a population pharmacokinetic model for phenobarbital. We externally validate the model using prospective phenobarbital data from an ongoing pharmacokinetic study in preterm neonates. TDM data from 53 neonates (gestational age (GA): 37 (24-42) weeks, bodyweight: 2.7 (0.45-4.5) kg; postnatal age (PNA): 4.5 (0-22) days) contained information on dosage histories, concentration and covariate data (including birth weight, actual weight, post-natal age (PNA), postmenstrual age, GA, sex, liver and kidney function, APGAR-score). Model development was carried out using NONMEM ® 7.3. After assessment of model fit, the model was validated using data of 17 neonates included in the DINO (Drug dosage Improvement in NeOnates)-study. Modelling of 229 plasma concentrations, ranging from 3.2 to 75.2mg/L, resulted in a one compartment model for phenobarbital. Clearance (CL) and volume (V d ) for a child with a birthweight of 2.6kg at PNA day 4.5 was 0.0091L/h (9%) and 2.38L (5%), respectively. Birthweight and PNA were the best predictors for CL maturation, increasing CL by 36.7% per kg birthweight and 5.3% per postnatal day of living, respectively. The best predictor for the increase in V d was actual bodyweight (0.31L/kg). External validation showed that the model can adequately predict the pharmacokinetics in a prospective study. Data-sharing can help to successfully develop and validate population pharmacokinetic models in neonates. From the results it seems that both PNA and bodyweight are required to guide dosing of phenobarbital in term and preterm neonates. Copyright © 2017 The Authors. Published by Elsevier B.V. All rights reserved.

  18. Validation of COMMIX with Westinghouse AP-600 PCCS test data

    International Nuclear Information System (INIS)

    Sun, J.G.; Chien, T.H.; Ding, J.; Sha, W.T.

    1993-01-01

    Small-scale test data for the Westinghouse AP-600 Passive Containment Cooling System (PCCS) have been used to validate the COMMIX computer code. To evaluate the performance of the PCCS, two transient liquid-film tracking models have been developed and implemented in the CO code. A set of heat transfer models and a mass transfer model based on heat and mass transfer analogy were used for the analysis of the AP-600 PCCS. It was found that the flow of the air stream in the annulus is a highly turbulent forced convection and that the flow of the air/steam mixture in the containment vessel is a mixed convection. Accordingly, a turbulent-forced-convection heat transfer model is used on the outside of the steel containment vessel wall and a mixed-convection heat transfer model is used on the inside of the steel containment vessel wall. The results from the CO calculations are compared with the experimental data from Westinghouse PCCS small-scale tests for average wall heat flux, evaporation rate, containment vessel pressure, and vessel wall temperature and heat flux distributions; agreement is good. The CO calculations also provide detailed distributions of velocity, temperature, and steam and air concentrations

  19. Principles of Proper Validation

    DEFF Research Database (Denmark)

    Esbensen, Kim; Geladi, Paul

    2010-01-01

    to suffer from the same deficiencies. The PPV are universal and can be applied to all situations in which the assessment of performance is desired: prediction-, classification-, time series forecasting-, modeling validation. The key element of PPV is the Theory of Sampling (TOS), which allow insight......) is critically necessary for the inclusion of the sampling errors incurred in all 'future' situations in which the validated model must perform. Logically, therefore, all one data set re-sampling approaches for validation, especially cross-validation and leverage-corrected validation, should be terminated...

  20. Validation of Satellite AOD Data with the Ground PM10 Data over Islamabad Pakistan

    Science.gov (United States)

    Bulbul, Gufran; Shahid, Imran

    2016-07-01

    health. In this study, concentrations of PM10 will be monitored at different sites in H-12 sector and Kashmir Highway Islamabad using High volume air sampler and its chemical characterization will be done using Energy Dispersive XRF. The first application of satellite remote sensing for aerosol monitoring began in the mid-1970s to detect the desert particles above the ocean using data from Landsat, GOES, and AVHRR remote sensing satellites. Maps of Aerosol Optical Depth (AOD) over the ocean were produced using the 0.63 µm channel of Advanced Very High Resolution Radiometer (AVHRR) . Aerosols properties were retrieved using AVHRR. The useable range of wavelengths of spectrum (shorter wavelengths and the longer wavelengths) for the remote sensing of the aerosols particles is mostly restricted due to ozone and gaseous absorptions. The purpose of the study is to validate the satellite Aerosol Optical Depth (AOD) data for the regional and local scale for Pakistan Objectives • To quantify the concentration of PM10 • To investigate their elemental composition • To find out their possible sources • Validation with MODIS satellite AOD Methodology: PM10 concentration will be measured at different sites of NUST Islamabad, Pakistan using High volume air sampler an Air sampling equipment capable of sampling high volumes of air (typically 57,000 ft3 or 1,600 m3) at high flow rates (typically 1.13 m3/min or 40 ft3/min) over an extended sampling duration (typically 24 hrs). The sampling period will be of 24 hours. Particles in the PM10 size range are then collected on the filter(s) during the specified 24-h sampling period. Each sample filter will be weighed before and after sampling to determine the net weight (mass) gain of the collected PM10 sample (40 CFR Part 50, Appendix M, US EPA). Next step will be the chemical characterization. Element concentrations will be determined by energy dispersive X-ray fluorescence (ED-XRF) technique. The ED-XRF system uses an X-ray tube to

  1. Validation Process Methods

    Energy Technology Data Exchange (ETDEWEB)

    Lewis, John E. [National Renewable Energy Lab. (NREL), Golden, CO (United States); English, Christine M. [National Renewable Energy Lab. (NREL), Golden, CO (United States); Gesick, Joshua C. [National Renewable Energy Lab. (NREL), Golden, CO (United States); Mukkamala, Saikrishna [National Renewable Energy Lab. (NREL), Golden, CO (United States)

    2018-01-04

    This report documents the validation process as applied to projects awarded through Funding Opportunity Announcements (FOAs) within the U.S. Department of Energy Bioenergy Technologies Office (DOE-BETO). It describes the procedures used to protect and verify project data, as well as the systematic framework used to evaluate and track performance metrics throughout the life of the project. This report also describes the procedures used to validate the proposed process design, cost data, analysis methodologies, and supporting documentation provided by the recipients.

  2. Development and validation of a nuclear data and calculation system for Superphenix with steel reflectors

    International Nuclear Information System (INIS)

    Bosq, J.Ch.

    1998-01-01

    This thesis concerns the definition and the validation of the ERANOS neutronic calculation system for steel reflected fast reactors. The calculation system uses JEF2.2 evaluated nuclear data, the ECCO cell code and the BISTRO and VARIANT transport codes. After a description of the physical phenomena induced by the existence of the these sub-critical media, an inventory of the past studies related to steel reflectors is reported. A calculational scheme taking into account the important physical phenomena (strong neutronic slowing-down, presence of broad resonances of the structural materials and spatial variation of the spectrum in the reflector) is defined. This method is validated with the TRIPOLI4 reference Monte-Carlo code. The use of this upgraded calculation method for the analysis of the part of the CIRANO experimental program devoted to the study of steel reflected configurations leads to discrepancies between the calculated and measured values. These remaining discrepancies obtained for the reactivity and the fission rate traverses are due to inaccurate nuclear data for the structural materials. The adjustment of these nuclear data in order to reduce these discrepancies id demonstrated. The additional uncertainty associated to the integral parameters of interest for a nuclear reactor (reactivity and power distribution) induced by the replacement of a fertile blanket by a steel reflector is determined for the Superphenix reactor and is proved to be small. (author)

  3. Validation and upgrading of the recommended cross-section data of charged particle reactions: Gamma emitter radioisotopes

    International Nuclear Information System (INIS)

    Takacs, S.; Tarkanyi, F.; Hermanne, A.

    2005-01-01

    An upgrade and validation test of the recommended cross-section database for production of gamma emitter radioisotopes by charged particle induced reactions, published by the IAEA in 2001, was performed. Experimental microscopic cross-section data published earlier or measured recently and not yet included in the evaluation work were collected and added to the primary database in order to improve the quality of the recommended data. The newly compiled experimental data in general supported the previous recommended data, but in a few cases they influenced the decision and resulted in different selected cross-section data sets. A Spline fitting method was used to calculate the recommended data from the selected data sets. Integral thick target yields were deduced from the newly calculated recommended cross-sections and were critically compared with the available experimental yield data

  4. Use of integral experiments in support to the validation of JEFF-3.2 nuclear data evaluation

    Directory of Open Access Journals (Sweden)

    Leclaire Nicolas

    2017-01-01

    Full Text Available For many years now, IRSN has developed its own Monte Carlo continuous energy capability, which allows testing various nuclear data libraries. In that prospect, a validation database of 1136 experiments was built from cases used for the validation of the APOLLO2-MORET 5 multigroup route of the CRISTAL V2.0 package. In this paper, the keff obtained for more than 200 benchmarks using the JEFF-3.1.1 and JEFF-3.2 libraries are compared to benchmark keff values and main discrepancies are analyzed regarding the neutron spectrum. Special attention is paid on benchmarks for which the results have been highly modified between both JEFF-3 versions.

  5. Texas Panhandle soil-crop-beef food chain for uranium: a dynamic model validated by experimental data

    International Nuclear Information System (INIS)

    Wenzel, W.J.; Wallwork-Barber, K.M.; Rodgers, J.C.; Gallegos, A.F.

    1982-01-01

    Long-term simulations of uranium transport in the soil-crop-beef food chain were performed using the BIOTRAN model. Experimental data means from an extensive Pantex beef cattle study are presented. Experimental data were used to validate the computer model. Measurements of uranium in air, soil, water, range grasses, feed, and cattle tissues are compared to simulated uranium output values in these matrices when the BIOTRAN model was set at the measured soil and air values. The simulations agreed well with experimental data even though metabolic details for ruminants and uranium chemical form in the environment remain to be studied

  6. The use of absolute gravity data for the validation of Global Geopotential Models and for improving quasigeoid heights determined from satellite-only Global Geopotential Models

    Science.gov (United States)

    Godah, Walyeldeen; Krynski, Jan; Szelachowska, Malgorzata

    2018-05-01

    The objective of this paper is to demonstrate the usefulness of absolute gravity data for the validation of Global Geopotential Models (GGMs). It is also aimed at improving quasigeoid heights determined from satellite-only GGMs using absolute gravity data. The area of Poland, as a unique one, covered with a homogeneously distributed set of absolute gravity data, has been selected as a study area. The gravity anomalies obtained from GGMs were validated using the corresponding ones determined from absolute gravity data. The spectral enhancement method was implemented to overcome the spectral inconsistency in data being validated. The quasigeoid heights obtained from the satellite-only GGM as well as from the satellite-only GGM in combination with absolute gravity data were evaluated with high accuracy GNSS/levelling data. Estimated accuracy of gravity anomalies obtained from GGMs investigated is of 1.7 mGal. Considering omitted gravity signal, e.g. from degree and order 101 to 2190, satellite-only GGMs can be validated at the accuracy level of 1 mGal using absolute gravity data. An improvement up to 59% in the accuracy of quasigeoid heights obtained from the satellite-only GGM can be observed when combining the satellite-only GGM with absolute gravity data.

  7. Estimating uncertainty of inference for validation

    Energy Technology Data Exchange (ETDEWEB)

    Booker, Jane M [Los Alamos National Laboratory; Langenbrunner, James R [Los Alamos National Laboratory; Hemez, Francois M [Los Alamos National Laboratory; Ross, Timothy J [UNM

    2010-09-30

    We present a validation process based upon the concept that validation is an inference-making activity. This has always been true, but the association has not been as important before as it is now. Previously, theory had been confirmed by more data, and predictions were possible based on data. The process today is to infer from theory to code and from code to prediction, making the role of prediction somewhat automatic, and a machine function. Validation is defined as determining the degree to which a model and code is an accurate representation of experimental test data. Imbedded in validation is the intention to use the computer code to predict. To predict is to accept the conclusion that an observable final state will manifest; therefore, prediction is an inference whose goodness relies on the validity of the code. Quantifying the uncertainty of a prediction amounts to quantifying the uncertainty of validation, and this involves the characterization of uncertainties inherent in theory/models/codes and the corresponding data. An introduction to inference making and its associated uncertainty is provided as a foundation for the validation problem. A mathematical construction for estimating the uncertainty in the validation inference is then presented, including a possibility distribution constructed to represent the inference uncertainty for validation under uncertainty. The estimation of inference uncertainty for validation is illustrated using data and calculations from Inertial Confinement Fusion (ICF). The ICF measurements of neutron yield and ion temperature were obtained for direct-drive inertial fusion capsules at the Omega laser facility. The glass capsules, containing the fusion gas, were systematically selected with the intent of establishing a reproducible baseline of high-yield 10{sup 13}-10{sup 14} neutron output. The deuterium-tritium ratio in these experiments was varied to study its influence upon yield. This paper on validation inference is the

  8. Online data validation

    DEFF Research Database (Denmark)

    Nielsen, Jan Nygaard; Madsen, Henrik

    1998-01-01

    The objective of this report is to discuss methods for identifying, rejecting, making allowances for, or minimizing the influence of, outlyer observations.The literature provides a variety of vague definitions of the concept of an outlying observation, see (Barnett and Lewis, 1994), yet consensus...... seems to be within reach with the following definition:An Outlier is an observation (or a subset of observations) which appear to be inconsistent with a model of the majority of the data....

  9. Parametric adaptive filtering and data validation in the bar GW detector AURIGA

    CERN Document Server

    Ortolan, A; Cerdonio, M; Prodi, G A; Vedovato, G; Vitale, S

    2002-01-01

    We report on our experience gained in the signal processing of the resonant GW detector AURIGA. Signal amplitude and arrival time are estimated by means of a matched-adaptive Wiener filter. The detector noise, entering in the filter set-up, is modelled as a parametric ARMA process; to account for slow non-stationarity of the noise, the ARMA parameters are estimated on an hourly basis. A requirement of the set-up of an unbiased Wiener filter is the separation of time spans with 'almost Gaussian' noise from non-Gaussian and/or strongly non-stationary time spans. The separation algorithm consists basically of a variance estimate with the Chauvenet convergence method and a threshold on the Curtosis index. The subsequent validation of data is strictly connected with the separation procedure: in fact, by injecting a large number of artificial GW signals into the 'almost Gaussian' part of the AURIGA data stream, we have demonstrated that the effective probability distributions of the signal-to-noise ratio chi sup 2 ...

  10. Validation of CFD predictions using process data obtained from flow through an industrial control valve

    International Nuclear Information System (INIS)

    Green, J; Mishra, R; Charlton, M; Owen, R

    2012-01-01

    This study uses the experimental flow test data to validate CFD simulations for a complex control valve trim. In both the simulation and the experimental flow test the capacity of the trim (Cv) is calculated in order to test the ability of CFD software to provide a design tool for these trims. While CFD tests produced results for the capacity which were consistent across a series of five different simulations, it differed from the experimental flow data by nearly 25%. This indicates that CFD simulations need to be properly calibrated before being used in designing complex valve trims.

  11. Application of aerosol speciation data as an in situ dust proxy for validation of the Dust Regional Atmospheric Model (DREAM)

    Science.gov (United States)

    Shaw, Patrick

    The Dust REgional Atmospheric Model (DREAM) predicts concentrations of mineral dust aerosols in time and space, but validation is challenging with current in situ particulate matter (PM) concentration measurements. Measured levels of ambient PM often contain anthropogenic components as well as windblown mineral dust. In this study, two approaches to model validation were performed with data from preexisting air quality monitoring networks: using hourly concentrations of total PM with aerodynamic diameter less than 2.5 μm (PM 2.5); and using a daily averaged speciation-derived soil component. Validation analyses were performed for point locations within the cities of El Paso (TX), Austin (TX), Phoenix (AZ), Salt Lake City (UT) and Bakersfield (CA) for most of 2006. Hourly modeled PM 2.5 did not validate at all with hourly observations among the sites (combined R hourly values). Aerosol chemical speciation data distinguished between mineral (soil) dust from anthropogenic ambient PM. As expected, statistically significant improvements in correlation among all stations (combined R = 0.16, N = 343 daily values) were found when the soil component alone was used to validate DREAM. The validation biases that result from anthropogenic aerosols were also reduced using the soil component. This is seen in the reduction of the root mean square error between hourly in situ versus hourly modeled (RMSE hourly = 18.6 μg m -3) and 24-h in situ speciation values versus daily averaged observed (RMSE soil = 12.0 μg m -3). However, the lack of a total reduction in RMSE indicates there is still room for improvement in the model. While the soil component is the theoretical proxy of choice for a dust transport model, the current sparse and infrequent sampling is not ideal for routine hourly air quality forecast validation.

  12. Validation of a Climate-Data Record of the "Clear-Kky" Surface Temperature of the Greenland Ice Sheet

    Science.gov (United States)

    Hall, Dorothy K.; Box, Jason E.; Koenig, Lora S.; DiGirolamo, Nicolo E.; Comiso, Josefino C.; Shuman, Christopher A.

    2011-01-01

    Surface temperatures on the Greenland Ice Sheet have been studied on the ground, using automatic weather station (AWS) data from the Greenland-Climate Network (GC-Net), and from analysis of satellite sensor data. Using Advanced Very High Frequency Radiometer (AVHRR) weekly surface temperature maps, warming of the surface of the Greenland Ice Sheet has been documented since 1981. We extended and refined this record using higher-resolution Moderate-Resolution Imaging Spectroradiometer (MODIS) data from March 2000 to the present. We developed a daily and monthly climate-data record (CDR) of the "clear-sky" surface temperature of the Greenland Ice Sheet using an ice-surface temperature (1ST) algorithm developed for use with MODIS data. Validation of this CDR is ongoing. MODIS Terra swath data are projected onto a polar stereographic grid at 6.25-km resolution to develop binary, gridded daily and mean-monthly 1ST maps. Each monthly map also has a color-coded image map that is available to download. Also included with the monthly maps is an accompanying map showing number of days in the month that were used to calculate the mean-monthly 1ST. This is important because no 1ST decision is made by the algorithm for cells that are considered cloudy by the internal cloud mask, so a sufficient number of days must be available to produce a mean 1ST for each grid cell. Validation of the CDR consists of several facets: 1) comparisons between ISTs and in-situ measurements; 2) comparisons between ISTs and AWS data; and 3) comparisons of ISTs with surface temperatures derived from other satellite instruments such as the Thermal Emission and Reflection Radiometer (ASTER) and Enhanced Thematic Mapper Plus (ETM+). Previous work shows that Terra MODIS ISTs are about 3 C lower than in-situ temperatures measured at Summit Camp, during the winter of 2008-09 under clear skies. In this work we begin to compare surface temperatures derived from AWS data with ISTs from the MODIS CDR. The

  13. EOS Terra Validation Program

    Science.gov (United States)

    Starr, David

    2000-01-01

    The EOS Terra mission will be launched in July 1999. This mission has great relevance to the atmospheric radiation community and global change issues. Terra instruments include Advanced Spaceborne Thermal Emission and Reflection Radiometer (ASTER), Clouds and Earth's Radiant Energy System (CERES), Multi-Angle Imaging Spectroradiometer (MISR), Moderate Resolution Imaging Spectroradiometer (MODIS) and Measurements of Pollution in the Troposphere (MOPITT). In addition to the fundamental radiance data sets, numerous global science data products will be generated, including various Earth radiation budget, cloud and aerosol parameters, as well as land surface, terrestrial ecology, ocean color, and atmospheric chemistry parameters. Significant investments have been made in on-board calibration to ensure the quality of the radiance observations. A key component of the Terra mission is the validation of the science data products. This is essential for a mission focused on global change issues and the underlying processes. The Terra algorithms have been subject to extensive pre-launch testing with field data whenever possible. Intensive efforts will be made to validate the Terra data products after launch. These include validation of instrument calibration (vicarious calibration) experiments, instrument and cross-platform comparisons, routine collection of high quality correlative data from ground-based networks, such as AERONET, and intensive sites, such as the SGP ARM site, as well as a variety field experiments, cruises, etc. Airborne simulator instruments have been developed for the field experiment and underflight activities including the MODIS Airborne Simulator (MAS) AirMISR, MASTER (MODIS-ASTER), and MOPITT-A. All are integrated on the NASA ER-2 though low altitude platforms are more typically used for MASTER. MATR is an additional sensor used for MOPITT algorithm development and validation. The intensive validation activities planned for the first year of the Terra

  14. Using Office Discipline Referral Data for Decision Making about Student Behavior in Elementary and Middle Schools: An Empirical Evaluation of Validity

    Science.gov (United States)

    Irvin, Larry K.; Horner, Robert H.; Ingram, Kimberly; Todd, Anne W.; Sugai, George; Sampson, Nadia Katul; Boland, Joseph B.

    2006-01-01

    In this evaluation we used Messick's construct validity as a conceptual framework for an empirical study assessing the validity of use, utility, and impact of office discipline referral (ODR) measures for data-based decision making about student behavior in schools. The Messick approach provided a rubric for testing the fit of our theory of use of…

  15. Melt pond fraction and spectral sea ice albedo retrieval from MERIS data – Part 1: Validation against in situ, aerial, and ship cruise data

    OpenAIRE

    Istomina, L.; Heygster, G.; Huntemann, M.; Schwarz, P.; Birnbaum, G.; Scharien, R.; Polashenski, C.; Perovich, D.; Zege, E.; Malinka, A.; Prikhach, A.; Katsev, I.

    2015-01-01

    The presence of melt ponds on the Arctic sea ice strongly affects the energy balance of the Arctic Ocean in summer. It affects albedo as well as transmittance through the sea ice, which has consequences for the heat balance and mass balance of sea ice. An algorithm to retrieve melt pond fraction and sea ice albedo from Medium Resolution Imaging Spectrometer (MERIS) data is validated against aerial, shipborne and in situ campaign data. The results show the best correlation fo...

  16. Probabilistic Matching of Deidentified Data From a Trauma Registry and a Traumatic Brain Injury Model System Center: A Follow-up Validation Study.

    Science.gov (United States)

    Kumar, Raj G; Wang, Zhensheng; Kesinger, Matthew R; Newman, Mark; Huynh, Toan T; Niemeier, Janet P; Sperry, Jason L; Wagner, Amy K

    2018-04-01

    In a previous study, individuals from a single Traumatic Brain Injury Model Systems and trauma center were matched using a novel probabilistic matching algorithm. The Traumatic Brain Injury Model Systems is a multicenter prospective cohort study containing more than 14,000 participants with traumatic brain injury, following them from inpatient rehabilitation to the community over the remainder of their lifetime. The National Trauma Databank is the largest aggregation of trauma data in the United States, including more than 6 million records. Linking these two databases offers a broad range of opportunities to explore research questions not otherwise possible. Our objective was to refine and validate the previous protocol at another independent center. An algorithm generation and validation data set were created, and potential matches were blocked by age, sex, and year of injury; total probabilistic weight was calculated based on of 12 common data fields. Validity metrics were calculated using a minimum probabilistic weight of 3. The positive predictive value was 98.2% and 97.4% and sensitivity was 74.1% and 76.3%, in the algorithm generation and validation set, respectively. These metrics were similar to the previous study. Future work will apply the refined probabilistic matching algorithm to the Traumatic Brain Injury Model Systems and the National Trauma Databank to generate a merged data set for clinical traumatic brain injury research use.

  17. Thai SF-36 health survey: tests of data quality, scaling assumptions, reliability and validity in healthy men and women

    Directory of Open Access Journals (Sweden)

    Sleigh Adrian

    2008-07-01

    Full Text Available Abstract Background Since its translation to Thai in 2000, the SF-36 Health Survey has been used extensively in many different clinical settings in Thailand. Its popularity has increased despite the absence of published evidence that the translated instrument satisfies scoring assumptions, the psychometric properties required for valid interpretation of the SF-36 summated ratings scales. The purpose of this paper was to examine these properties and to report on the reliability and validity of the Thai SF-36 in a non-clinical general population. Methods 1345 distance-education university students who live in all areas of Thailand completed a questionnaire comprising the Thai SF-36 (Version 1. Median age was 31 years. Psychometric tests recommended by the International Quality of Life Assessment Project were used. Results Data quality was satisfactory: questionnaire completion rate was high (97.5% and missing data rates were low ( Conclusion The summated ratings method can be used for scoring the Thai SF-36. The instrument was found to be reliable and valid for use in a general non-clinical population. Version 2 of the SF-36 could improve ceiling and floor effects in the role functioning scales. Further work is warranted to refine items that measure the concepts of social functioning, vitality and mental health to improve the reliability and discriminant validity of these scales.

  18. The Potential of Autonomous Ship-Borne Hyperspectral Radiometers for the Validation of Ocean Color Radiometry Data

    Directory of Open Access Journals (Sweden)

    Vittorio E. Brando

    2016-02-01

    Full Text Available Calibration and validation of satellite observations are essential and on-going tasks to ensure compliance with mission accuracy requirements. An automated above water hyperspectral radiometer significantly augmented Australia’s ability to contribute to global and regional ocean color validation and algorithm design activities. The hyperspectral data can be re-sampled for comparison with current and future sensor wavebands. The continuous spectral acquisition along the ship track enables spatial resampling to match satellite footprint. This study reports spectral comparisons of the radiometer data with Visible Infrared Imaging Radiometer Suite (VIIRS and Moderate Resolution Imaging Spectroradiometer (MODIS-Aqua for contrasting water types in tropical waters off northern Australia based on the standard NIR atmospheric correction implemented in SeaDAS. Consistent match-ups are shown for transects of up to 50 km over a range of reflectance values. The MODIS and VIIRS satellite reflectance data consistently underestimated the in situ spectra in the blue with a bias relative to the “dynamic above water radiance and irradiance collector” (DALEC at 443 nm ranging from 9.8 × 10−4 to 3.1 × 10−3 sr−1. Automated acquisition has produced good quality data under standard operating and maintenance procedures. A sensitivity analysis explored the effects of some assumptions in the data reduction methods, indicating the need for a comprehensive investigation and quantification of each source of uncertainty in the estimate of the DALEC reflectances. Deployment on a Research Vessel provides the potential for the radiometric data to be combined with other sampling and observational activities to contribute to algorithm development in the wider bio-optical research community.

  19. Validation of Case Finding Algorithms for Hepatocellular Cancer From Administrative Data and Electronic Health Records Using Natural Language Processing.

    Science.gov (United States)

    Sada, Yvonne; Hou, Jason; Richardson, Peter; El-Serag, Hashem; Davila, Jessica

    2016-02-01

    Accurate identification of hepatocellular cancer (HCC) cases from automated data is needed for efficient and valid quality improvement initiatives and research. We validated HCC International Classification of Diseases, 9th Revision (ICD-9) codes, and evaluated whether natural language processing by the Automated Retrieval Console (ARC) for document classification improves HCC identification. We identified a cohort of patients with ICD-9 codes for HCC during 2005-2010 from Veterans Affairs administrative data. Pathology and radiology reports were reviewed to confirm HCC. The positive predictive value (PPV), sensitivity, and specificity of ICD-9 codes were calculated. A split validation study of pathology and radiology reports was performed to develop and validate ARC algorithms. Reports were manually classified as diagnostic of HCC or not. ARC generated document classification algorithms using the Clinical Text Analysis and Knowledge Extraction System. ARC performance was compared with manual classification. PPV, sensitivity, and specificity of ARC were calculated. A total of 1138 patients with HCC were identified by ICD-9 codes. On the basis of manual review, 773 had HCC. The HCC ICD-9 code algorithm had a PPV of 0.67, sensitivity of 0.95, and specificity of 0.93. For a random subset of 619 patients, we identified 471 pathology reports for 323 patients and 943 radiology reports for 557 patients. The pathology ARC algorithm had PPV of 0.96, sensitivity of 0.96, and specificity of 0.97. The radiology ARC algorithm had PPV of 0.75, sensitivity of 0.94, and specificity of 0.68. A combined approach of ICD-9 codes and natural language processing of pathology and radiology reports improves HCC case identification in automated data.

  20. The COSIMA-experiments, a data base for validation of two-phase flow computer codes

    International Nuclear Information System (INIS)

    Class, G.; Meyder, R.; Stratmanns, E.

    1985-12-01

    The report presents an overview on the large data base generated with COSIMA. The data base is to be used to validate and develop computer codes for two-phase flow. In terms of fuel rod behavior it was found that during blowdown under realistic conditions only small strains are reached. For clad rupture extremely high rod internal pressure is necessary. Additionally important results were found in the behavior of a fuel rod simulator and on the effect of thermocouples attached on the cladding outer surface. Post-test calculations, performed with the codes RELAP and DRUFAN show a good agreement with the experiments. This however can be improved if the phase separation models in the codes would be updated. (orig./HP) [de

  1. Melt pond fraction and spectral sea ice albedo retrieval from MERIS data - Part 1: Validation against in situ, aerial, and ship cruise data

    Science.gov (United States)

    Istomina, L.; Heygster, G.; Huntemann, M.; Schwarz, P.; Birnbaum, G.; Scharien, R.; Polashenski, C.; Perovich, D.; Zege, E.; Malinka, A.; Prikhach, A.; Katsev, I.

    2015-08-01

    The presence of melt ponds on the Arctic sea ice strongly affects the energy balance of the Arctic Ocean in summer. It affects albedo as well as transmittance through the sea ice, which has consequences for the heat balance and mass balance of sea ice. An algorithm to retrieve melt pond fraction and sea ice albedo from Medium Resolution Imaging Spectrometer (MERIS) data is validated against aerial, shipborne and in situ campaign data. The results show the best correlation for landfast and multiyear ice of high ice concentrations. For broadband albedo, R2 is equal to 0.85, with the RMS (root mean square) being equal to 0.068; for the melt pond fraction, R2 is equal to 0.36, with the RMS being equal to 0.065. The correlation for lower ice concentrations, subpixel ice floes, blue ice and wet ice is lower due to ice drift and challenging for the retrieval surface conditions. Combining all aerial observations gives a mean albedo RMS of 0.089 and a mean melt pond fraction RMS of 0.22. The in situ melt pond fraction correlation is R2 = 0.52 with an RMS = 0.14. Ship cruise data might be affected by documentation of varying accuracy within the Antarctic Sea Ice Processes and Climate (ASPeCt) protocol, which may contribute to the discrepancy between the satellite value and the observed value: mean R2 = 0.044, mean RMS = 0.16. An additional dynamic spatial cloud filter for MERIS over snow and ice has been developed to assist with the validation on swath data.

  2. Data supporting the validation of a simulation model for multi-component gas separation in polymeric membranes.

    Science.gov (United States)

    Giordano, Lorena; Roizard, Denis; Bounaceur, Roda; Favre, Eric

    2016-12-01

    The article describes data concerning the separation performances of polymeric hollow-fiber membranes. The data were obtained using a model for simulating gas separation, described in the research article entitled "Interplay of inlet temperature and humidity on energy penalty for CO 2 post-combustion capture: rigorous analysis and simulation of a single stage gas permeation process" (L. Giordano, D. Roizard, R. Bounaceur, E. Favre, 2016) [1]. The data were used to validate the model by comparison with literature results. Considering a membrane system based on feed compression only, data from the model proposed and that from literature were compared with respect to the molar composition of permeate stream, the membrane area and specific energy requirement, varying the feed pressure and the CO 2 separation degree.

  3. A Global Lake Ecological Observatory Network (GLEON) for synthesising high-frequency sensor data for validation of deterministic ecological models

    Science.gov (United States)

    David, Hamilton P; Carey, Cayelan C.; Arvola, Lauri; Arzberger, Peter; Brewer, Carol A.; Cole, Jon J; Gaiser, Evelyn; Hanson, Paul C.; Ibelings, Bas W; Jennings, Eleanor; Kratz, Tim K; Lin, Fang-Pang; McBride, Christopher G.; de Motta Marques, David; Muraoka, Kohji; Nishri, Ami; Qin, Boqiang; Read, Jordan S.; Rose, Kevin C.; Ryder, Elizabeth; Weathers, Kathleen C.; Zhu, Guangwei; Trolle, Dennis; Brookes, Justin D

    2014-01-01

    A Global Lake Ecological Observatory Network (GLEON; www.gleon.org) has formed to provide a coordinated response to the need for scientific understanding of lake processes, utilising technological advances available from autonomous sensors. The organisation embraces a grassroots approach to engage researchers from varying disciplines, sites spanning geographic and ecological gradients, and novel sensor and cyberinfrastructure to synthesise high-frequency lake data at scales ranging from local to global. The high-frequency data provide a platform to rigorously validate process- based ecological models because model simulation time steps are better aligned with sensor measurements than with lower-frequency, manual samples. Two case studies from Trout Bog, Wisconsin, USA, and Lake Rotoehu, North Island, New Zealand, are presented to demonstrate that in the past, ecological model outputs (e.g., temperature, chlorophyll) have been relatively poorly validated based on a limited number of directly comparable measurements, both in time and space. The case studies demonstrate some of the difficulties of mapping sensor measurements directly to model state variable outputs as well as the opportunities to use deviations between sensor measurements and model simulations to better inform process understanding. Well-validated ecological models provide a mechanism to extrapolate high-frequency sensor data in space and time, thereby potentially creating a fully 3-dimensional simulation of key variables of interest.

  4. Accuracy of postpartum haemorrhage data in the 2011 Victorian Perinatal Data Collection: Results of a validation study.

    Science.gov (United States)

    Flood, Margaret; Pollock, Wendy; McDonald, Susan J; Davey, Mary-Ann

    2018-04-01

    The postpartum haemorrhage (PPH) rate in Victoria in 2009 for women having their first birth, based on information reported to the Victorian Perinatal Data Collection (VPDC), was 23.6% (primiparas). Prior to 2009 PPH was collected via a tick box item on the perinatal form. Estimated blood loss (EBL) volume is now collected and it is from this item the PPH rate is calculated. Periodic assessment of data accuracy is essential to inform clinicians and others who rely on these data of their quality and limitations. This paper describes the results of a state-wide validation study of the accuracy of EBL volume and EBL-related data items reported to VPDC. PPH data from a random sample of 1% of births in Victoria in 2011 were extracted from source medical records and compared with information submitted to the VPDC. Accuracy was determined, together with sensitivity, specificity, positive predictive value and negative predictive value for dichotomous items. Accuracy of reporting for EBL ≥ 500 mL was 97.2% and for EBL ≥ 1500 mL was 99.7%. Sensitivity for EBL ≥ 500 mL was 89.0% (CI 83.1-93.0) and for EBL ≥ 1500 mL was 71.4% (CI 35.9-91.8). Blood product transfusion, peripartum hysterectomy and procedures to control bleeding were all accurately reported in >99% of cases. Most PPH-related data items in the 2011 VPDC may be considered reliable. Our results suggest EBL ≥ 1500 mL is likely to be under-reported. Changes to policies and practices of recording blood loss could further increase accuracy of reporting. © 2017 The Royal Australian and New Zealand College of Obstetricians and Gynaecologists.

  5. A CFD validation roadmap for hypersonic flows

    Science.gov (United States)

    Marvin, Joseph G.

    1993-01-01

    A roadmap for computational fluid dynamics (CFD) code validation is developed. The elements of the roadmap are consistent with air-breathing vehicle design requirements and related to the important flow path components: forebody, inlet, combustor, and nozzle. Building block and benchmark validation experiments are identified along with their test conditions and measurements. Based on an evaluation criteria, recommendations for an initial CFD validation data base are given and gaps identified where future experiments would provide the needed validation data.

  6. Parametric adaptive filtering and data validation in the bar GW detector AURIGA

    Science.gov (United States)

    Ortolan, A.; Baggio, L.; Cerdonio, M.; Prodi, G. A.; Vedovato, G.; Vitale, S.

    2002-04-01

    We report on our experience gained in the signal processing of the resonant GW detector AURIGA. Signal amplitude and arrival time are estimated by means of a matched-adaptive Wiener filter. The detector noise, entering in the filter set-up, is modelled as a parametric ARMA process; to account for slow non-stationarity of the noise, the ARMA parameters are estimated on an hourly basis. A requirement of the set-up of an unbiased Wiener filter is the separation of time spans with 'almost Gaussian' noise from non-Gaussian and/or strongly non-stationary time spans. The separation algorithm consists basically of a variance estimate with the Chauvenet convergence method and a threshold on the Curtosis index. The subsequent validation of data is strictly connected with the separation procedure: in fact, by injecting a large number of artificial GW signals into the 'almost Gaussian' part of the AURIGA data stream, we have demonstrated that the effective probability distributions of the signal-to-noise ratio χ2 and the time of arrival are those that are expected.

  7. Parametric adaptive filtering and data validation in the bar GW detector AURIGA

    International Nuclear Information System (INIS)

    Ortolan, A; Baggio, L; Cerdonio, M; Prodi, G A; Vedovato, G; Vitale, S

    2002-01-01

    We report on our experience gained in the signal processing of the resonant GW detector AURIGA. Signal amplitude and arrival time are estimated by means of a matched-adaptive Wiener filter. The detector noise, entering in the filter set-up, is modelled as a parametric ARMA process; to account for slow non-stationarity of the noise, the ARMA parameters are estimated on an hourly basis. A requirement of the set-up of an unbiased Wiener filter is the separation of time spans with 'almost Gaussian' noise from non-Gaussian and/or strongly non-stationary time spans. The separation algorithm consists basically of a variance estimate with the Chauvenet convergence method and a threshold on the Curtosis index. The subsequent validation of data is strictly connected with the separation procedure: in fact, by injecting a large number of artificial GW signals into the 'almost Gaussian' part of the AURIGA data stream, we have demonstrated that the effective probability distributions of the signal-to-noise ratio χ 2 and the time of arrival are those that are expected

  8. Parametric adaptive filtering and data validation in the bar GW detector AURIGA

    Energy Technology Data Exchange (ETDEWEB)

    Ortolan, A [INFN - Laboratori Nazionali di Legnaro, Via Romea, 4 I-35020 Legnaro, Padova (Italy); Baggio, L [Department of Physics, University of Trento and INFN Gruppo Collegato di Trento, I-38050 Povo, Trento (Italy); Cerdonio, M [Department of Physics, University of Padova and INFN Sezione di Padova, Via Marzolo 8, I-35131 Padova (Italy); Prodi, G A [Department of Physics, University of Trento and INFN Gruppo Collegato di Trento, I-38050 Povo, Trento (Italy); Vedovato, G [INFN - Laboratori Nazionali di Legnaro, Via Romea, 4 I-35020 Legnaro, Padova (Italy); Vitale, S [Department of Physics, University of Trento and INFN Gruppo Collegato di Trento, I-38050 Povo, Trento (Italy)

    2002-04-07

    We report on our experience gained in the signal processing of the resonant GW detector AURIGA. Signal amplitude and arrival time are estimated by means of a matched-adaptive Wiener filter. The detector noise, entering in the filter set-up, is modelled as a parametric ARMA process; to account for slow non-stationarity of the noise, the ARMA parameters are estimated on an hourly basis. A requirement of the set-up of an unbiased Wiener filter is the separation of time spans with 'almost Gaussian' noise from non-Gaussian and/or strongly non-stationary time spans. The separation algorithm consists basically of a variance estimate with the Chauvenet convergence method and a threshold on the Curtosis index. The subsequent validation of data is strictly connected with the separation procedure: in fact, by injecting a large number of artificial GW signals into the 'almost Gaussian' part of the AURIGA data stream, we have demonstrated that the effective probability distributions of the signal-to-noise ratio {chi}{sup 2} and the time of arrival are those that are expected.

  9. Validation and Refinement of a Pain Information Model from EHR Flowsheet Data.

    Science.gov (United States)

    Westra, Bonnie L; Johnson, Steven G; Ali, Samira; Bavuso, Karen M; Cruz, Christopher A; Collins, Sarah; Furukawa, Meg; Hook, Mary L; LaFlamme, Anne; Lytle, Kay; Pruinelli, Lisiane; Rajchel, Tari; Settergren, Theresa Tess; Westman, Kathryn F; Whittenburg, Luann

    2018-01-01

    Secondary use of electronic health record (EHR) data can reduce costs of research and quality reporting. However, EHR data must be consistent within and across organizations. Flowsheet data provide a rich source of interprofessional data and represents a high volume of documentation; however, content is not standardized. Health care organizations design and implement customized content for different care areas creating duplicative data that is noncomparable. In a prior study, 10 information models (IMs) were derived from an EHR that included 2.4 million patients. There was a need to evaluate the generalizability of the models across organizations. The pain IM was selected for evaluation and refinement because pain is a commonly occurring problem associated with high costs for pain management. The purpose of our study was to validate and further refine a pain IM from EHR flowsheet data that standardizes pain concepts, definitions, and associated value sets for assessments, goals, interventions, and outcomes. A retrospective observational study was conducted using an iterative consensus-based approach to map, analyze, and evaluate data from 10 organizations. The aggregated metadata from the EHRs of 8 large health care organizations and the design build in 2 additional organizations represented flowsheet data from 6.6 million patients, 27 million encounters, and 683 million observations. The final pain IM has 30 concepts, 4 panels (classes), and 396 value set items. Results are built on Logical Observation Identifiers Names and Codes (LOINC) pain assessment terms and extend the need for additional terms to support interoperability. The resulting pain IM is a consensus model based on actual EHR documentation in the participating health systems. The IM captures the most important concepts related to pain. Schattauer GmbH Stuttgart.

  10. Approaches to ascertaining comorbidity information: validation of routine hospital episode data with clinician-based case note review.

    Science.gov (United States)

    Soo, Martin; Robertson, Lynn M; Ali, Tariq; Clark, Laura E; Fluck, Nicholas; Johnston, Marjorie; Marks, Angharad; Prescott, Gordon J; Smith, William Cairns S; Black, Corri

    2014-04-21

    In clinical practice, research, and increasingly health surveillance, planning and costing, there is a need for high quality information to determine comorbidity information about patients. Electronic, routinely collected healthcare data is capturing increasing amounts of clinical information as part of routine care. The aim of this study was to assess the validity of routine hospital administrative data to determine comorbidity, as compared with clinician-based case note review, in a large cohort of patients with chronic kidney disease. A validation study using record linkage. Routine hospital administrative data were compared with clinician-based case note review comorbidity data in a cohort of 3219 patients with chronic kidney disease. To assess agreement, we calculated prevalence, kappa statistic, sensitivity, specificity, positive predictive value and negative predictive value. Subgroup analyses were also performed. Median age at index date was 76.3 years, 44% were male, 67% had stage 3 chronic kidney disease and 31% had at least three comorbidities. For most comorbidities, we found a higher prevalence recorded from case notes compared with administrative data. The best agreement was found for cerebrovascular disease (κ = 0.80) ischaemic heart disease (κ = 0.63) and diabetes (κ = 0.65). Hypertension, peripheral vascular disease and dementia showed only fair agreement (κ = 0.28, 0.39, 0.38 respectively) and smoking status was found to be poorly recorded in administrative data. The patterns of prevalence across subgroups were as expected and for most comorbidities, agreement between case note and administrative data was similar. Agreement was less, however, in older ages and for those with three or more comorbidities for some conditions. This study demonstrates that hospital administrative comorbidity data compared moderately well with case note review data for cerebrovascular disease, ischaemic heart disease and diabetes, however there was

  11. Estimating mortality from external causes using data from retrospective surveys: A validation study in Niakhar (Senegal

    Directory of Open Access Journals (Sweden)

    Gilles Pison

    2018-03-01

    Full Text Available Background: In low- and middle-income countries (LMICs, data on causes of death is often inaccurate or incomplete. In this paper, we test whether adding a few questions about injuries and accidents to mortality questionnaires used in representative household surveys would yield accurate estimates of the extent of mortality due to external causes (accidents, homicides, or suicides. Methods: We conduct a validation study in Niakhar (Senegal, during which we compare reported survey data to high-quality prospective records of deaths collected by a health and demographic surveillance system (HDSS. Results: Survey respondents more frequently list the deaths of their adult siblings who die of external causes than the deaths of those who die from other causes. The specificity of survey data is high, but sensitivity is low. Among reported deaths, less than 60Š of the deaths classified as due to external causes by the HDSS are also classified as such by survey respondents. Survey respondents better report deaths due to road-traffic accidents than deaths from suicides and homicides. Conclusions: Asking questions about deaths resulting from injuries and accidents during surveys might help measure mortality from external causes in LMICs, but the resulting data displays systematic bias in a rural population of Senegal. Future studies should 1 investigate whether similar biases also apply in other settings and 2 test new methods to further improve the accuracy of survey data on mortality from external causes. Contribution: This study helps strengthen the monitoring of sustainable development targets in LMICs by validating a simple approach for the measurement of mortality from external causes.

  12. Kepler Data Validation Time Series File: Description of File Format and Content

    Science.gov (United States)

    Mullally, Susan E.

    2016-01-01

    The Kepler space mission searches its time series data for periodic, transit-like signatures. The ephemerides of these events, called Threshold Crossing Events (TCEs), are reported in the TCE tables at the NASA Exoplanet Archive (NExScI). Those TCEs are then further evaluated to create planet candidates and populate the Kepler Objects of Interest (KOI) table, also hosted at the Exoplanet Archive. The search, evaluation and export of TCEs is performed by two pipeline modules, TPS (Transit Planet Search) and DV (Data Validation). TPS searches for the strongest, believable signal and then sends that information to DV to fit a transit model, compute various statistics, and remove the transit events so that the light curve can be searched for other TCEs. More on how this search is done and on the creation of the TCE table can be found in Tenenbaum et al. (2012), Seader et al. (2015), Jenkins (2002). For each star with at least one TCE, the pipeline exports a file that contains the light curves used by TPS and DV to find and evaluate the TCE(s). This document describes the content of these DV time series files, and this introduction provides a bit of context for how the data in these files are used by the pipeline.

  13. Predicting DMS-IV cluster B personality disorder criteria from MMPI-2 and Rorschach data: a test of incremental validity.

    Science.gov (United States)

    Blais, M A; Hilsenroth, M J; Castlebury, F; Fowler, J C; Baity, M R

    2001-02-01

    Despite their frequent conjoint clinical use, the incremental validity of Rorschach (Rorschach, 1921/1942) and MMPI (Hathaway & McKinley, 1943) data has not been adequately established, nor has any study to date explored the incremental validity of these tests for predicting Diagnostic and Statistical Manual of Mental Disorders (4th ed. [DSM-IV]; American Psychiatric Association, 1994) personality disorders (PDs). In a reanalysis of existing data, we used select Rorschach variables and the MMPI PD scales to predict DSM-IV antisocial, borderline, histrionic, and narcissistic PD criteria in a sample of treatment-seeking outpatients. The correlational findings revealed alimited relation between Rorschach and MMPI-2 (Butcher, Dahlstrom, Graham, Tellegen, & Kaemmer, 1989) variables, with only 5 of 30 correlations reaching significance (p psychological characteristics of the DSM-IV Cluster B PDs.

  14. Thai SF-36 health survey: tests of data quality, scaling assumptions, reliability and validity in healthy men and women.

    Science.gov (United States)

    Lim, Lynette L-Y; Seubsman, Sam-Ang; Sleigh, Adrian

    2008-07-18

    Since its translation to Thai in 2000, the SF-36 Health Survey has been used extensively in many different clinical settings in Thailand. Its popularity has increased despite the absence of published evidence that the translated instrument satisfies scoring assumptions, the psychometric properties required for valid interpretation of the SF-36 summated ratings scales. The purpose of this paper was to examine these properties and to report on the reliability and validity of the Thai SF-36 in a non-clinical general population. 1345 distance-education university students who live in all areas of Thailand completed a questionnaire comprising the Thai SF-36 (Version 1). Median age was 31 years. Psychometric tests recommended by the International Quality of Life Assessment Project were used. Data quality was satisfactory: questionnaire completion rate was high (97.5%) and missing data rates were low (Vitality scale correlated better with the Mental Health scale than with itself, possibly because a healthy mental state is central to the concept of vitality in Thailand. The summated ratings method can be used for scoring the Thai SF-36. The instrument was found to be reliable and valid for use in a general non-clinical population. Version 2 of the SF-36 could improve ceiling and floor effects in the role functioning scales. Further work is warranted to refine items that measure the concepts of social functioning, vitality and mental health to improve the reliability and discriminant validity of these scales.

  15. Development of the knee quality of life (KQoL-26) 26-item questionnaire: data quality, reliability, validity and responsiveness.

    Science.gov (United States)

    Garratt, Andrew M; Brealey, Stephen; Robling, Michael; Atwell, Chris; Russell, Ian; Gillespie, William; King, David

    2008-07-10

    This article describes the development and validation of a self-reported questionnaire, the KQoL-26, that is based on the views of patients with a suspected ligamentous or meniscal injury of the knee that assesses the impact of their knee problem on the quality of their lives. Patient interviews and focus groups were used to derive questionnaire content. The instrument was assessed for data quality, reliability, validity, and responsiveness using data from a randomised trial and patient survey about general practitioners' use of Magnetic Resonance Imaging for patients with a suspected ligamentous or meniscal injury. Interview and focus group data produced a 40-item questionnaire designed for self-completion. 559 trial patients and 323 survey patients responded to the questionnaire. Following principal components analysis and Rasch analysis, 26 items were found to contribute to three scales of knee-related quality of life: physical functioning, activity limitations, and emotional functioning. Item-total correlations ranged from 0.60-0.82. Cronbach's alpha and test retest reliability estimates were 0.91-0.94 and 0.80-0.93 respectively. Hypothesised correlations with the Lysholm Knee Scale, EQ-5D, SF-36 and knee symptom questions were evidence for construct validity. The instrument produced highly significant change scores for 65 trial patients indicating that their knee was a little or somewhat better at six months. The new instrument had higher effect sizes (range 0.86-1.13) and responsiveness statistics (range 1.50-2.13) than the EQ-5D and SF-36. The KQoL-26 has good evidence for internal reliability, test-retest reliability, validity and responsiveness, and is recommended for use in randomised trials and other evaluative studies of patients with a suspected ligamentous or meniscal injury.

  16. Development of the Knee Quality of Life (KQoL-26 26-item questionnaire: data quality, reliability, validity and responsiveness

    Directory of Open Access Journals (Sweden)

    Atwell Chris

    2008-07-01

    Full Text Available Abstract Background This article describes the development and validation of a self-reported questionnaire, the KQoL-26, that is based on the views of patients with a suspected ligamentous or meniscal injury of the knee that assesses the impact of their knee problem on the quality of their lives. Methods Patient interviews and focus groups were used to derive questionnaire content. The instrument was assessed for data quality, reliability, validity, and responsiveness using data from a randomised trial and patient survey about general practitioners' use of Magnetic Resonance Imaging for patients with a suspected ligamentous or meniscal injury. Results Interview and focus group data produced a 40-item questionnaire designed for self-completion. 559 trial patients and 323 survey patients responded to the questionnaire. Following principal components analysis and Rasch analysis, 26 items were found to contribute to three scales of knee-related quality of life: physical functioning, activity limitations, and emotional functioning. Item-total correlations ranged from 0.60–0.82. Cronbach's alpha and test retest reliability estimates were 0.91–0.94 and 0.80–0.93 respectively. Hypothesised correlations with the Lysholm Knee Scale, EQ-5D, SF-36 and knee symptom questions were evidence for construct validity. The instrument produced highly significant change scores for 65 trial patients indicating that their knee was a little or somewhat better at six months. The new instrument had higher effect sizes (range 0.86–1.13 and responsiveness statistics (range 1.50–2.13 than the EQ-5D and SF-36. Conclusion The KQoL-26 has good evidence for internal reliability, test-retest reliability, validity and responsiveness, and is recommended for use in randomised trials and other evaluative studies of patients with a suspected ligamentous or meniscal injury.

  17. Validation and discovery of genotype-phenotype associations in chronic diseases using linked data.

    Science.gov (United States)

    Pathak, Jyotishman; Kiefer, Richard; Freimuth, Robert; Chute, Christopher

    2012-01-01

    This study investigates federated SPARQL queries over Linked Open Data (LOD) in the Semantic Web to validate existing, and potentially discover new genotype-phenotype associations from public datasets. In particular, we report our preliminary findings for identifying such associations for commonly occurring chronic diseases using the Online Mendelian Inheritance in Man (OMIM) and Database for SNPs (dbSNP) within the LOD knowledgebase and compare them with Gene Wiki for coverage and completeness. Our results indicate that Semantic Web technologies can play an important role for in-silico identification of novel disease-gene-SNP associations, although additional verification is required before such information can be applied and used effectively.

  18. Obtaining Valid Safety Data for Software Safety Measurement and Process Improvement

    Science.gov (United States)

    Basili, Victor r.; Zelkowitz, Marvin V.; Layman, Lucas; Dangle, Kathleen; Diep, Madeline

    2010-01-01

    We report on a preliminary case study to examine software safety risk in the early design phase of the NASA Constellation spaceflight program. Our goal is to provide NASA quality assurance managers with information regarding the ongoing state of software safety across the program. We examined 154 hazard reports created during the preliminary design phase of three major flight hardware systems within the Constellation program. Our purpose was two-fold: 1) to quantify the relative importance of software with respect to system safety; and 2) to identify potential risks due to incorrect application of the safety process, deficiencies in the safety process, or the lack of a defined process. One early outcome of this work was to show that there are structural deficiencies in collecting valid safety data that make software safety different from hardware safety. In our conclusions we present some of these deficiencies.

  19. Development of the Galaxy Chronic Obstructive Pulmonary Disease (COPD) Model Using Data from ECLIPSE: Internal Validation of a Linked-Equations Cohort Model.

    Science.gov (United States)

    Briggs, Andrew H; Baker, Timothy; Risebrough, Nancy A; Chambers, Mike; Gonzalez-McQuire, Sebastian; Ismaila, Afisi S; Exuzides, Alex; Colby, Chris; Tabberer, Maggie; Muellerova, Hana; Locantore, Nicholas; Rutten van Mölken, Maureen P M H; Lomas, David A

    2017-05-01

    The recent joint International Society for Pharmacoeconomics and Outcomes Research / Society for Medical Decision Making Modeling Good Research Practices Task Force emphasized the importance of conceptualizing and validating models. We report a new model of chronic obstructive pulmonary disease (COPD) (part of the Galaxy project) founded on a conceptual model, implemented using a novel linked-equation approach, and internally validated. An expert panel developed a conceptual model including causal relationships between disease attributes, progression, and final outcomes. Risk equations describing these relationships were estimated using data from the Evaluation of COPD Longitudinally to Identify Predictive Surrogate Endpoints (ECLIPSE) study, with costs estimated from the TOwards a Revolution in COPD Health (TORCH) study. Implementation as a linked-equation model enabled direct estimation of health service costs and quality-adjusted life years (QALYs) for COPD patients over their lifetimes. Internal validation compared 3 years of predicted cohort experience with ECLIPSE results. At 3 years, the Galaxy COPD model predictions of annual exacerbation rate and annual decline in forced expiratory volume in 1 second fell within the ECLIPSE data confidence limits, although 3-year overall survival was outside the observed confidence limits. Projections of the risk equations over time permitted extrapolation to patient lifetimes. Averaging the predicted cost/QALY outcomes for the different patients within the ECLIPSE cohort gives an estimated lifetime cost of £25,214 (undiscounted)/£20,318 (discounted) and lifetime QALYs of 6.45 (undiscounted/5.24 [discounted]) per ECLIPSE patient. A new form of model for COPD was conceptualized, implemented, and internally validated, based on a series of linked equations using epidemiological data (ECLIPSE) and cost data (TORCH). This Galaxy model predicts COPD outcomes from treatment effects on disease attributes such as lung function

  20. Data supporting the validation of a simulation model for multi-component gas separation in polymeric membranes

    Directory of Open Access Journals (Sweden)

    Lorena Giordano

    2016-12-01

    The data were obtained using a model for simulating gas separation, described in the research article entitled “Interplay of inlet temperature and humidity on energy penalty for CO2 post-combustion capture: rigorous analysis and simulation of a single stage gas permeation process” (L. Giordano, D. Roizard, R. Bounaceur, E. Favre, 2016 [1]. The data were used to validate the model by comparison with literature results. Considering a membrane system based on feed compression only, data from the model proposed and that from literature were compared with respect to the molar composition of permeate stream, the membrane area and specific energy requirement, varying the feed pressure and the CO2 separation degree.

  1. ODM Data Analysis-A tool for the automatic validation, monitoring and generation of generic descriptive statistics of patient data.

    Science.gov (United States)

    Brix, Tobias Johannes; Bruland, Philipp; Sarfraz, Saad; Ernsting, Jan; Neuhaus, Philipp; Storck, Michael; Doods, Justin; Ständer, Sonja; Dugas, Martin

    2018-01-01

    A required step for presenting results of clinical studies is the declaration of participants demographic and baseline characteristics as claimed by the FDAAA 801. The common workflow to accomplish this task is to export the clinical data from the used electronic data capture system and import it into statistical software like SAS software or IBM SPSS. This software requires trained users, who have to implement the analysis individually for each item. These expenditures may become an obstacle for small studies. Objective of this work is to design, implement and evaluate an open source application, called ODM Data Analysis, for the semi-automatic analysis of clinical study data. The system requires clinical data in the CDISC Operational Data Model format. After uploading the file, its syntax and data type conformity of the collected data is validated. The completeness of the study data is determined and basic statistics, including illustrative charts for each item, are generated. Datasets from four clinical studies have been used to evaluate the application's performance and functionality. The system is implemented as an open source web application (available at https://odmanalysis.uni-muenster.de) and also provided as Docker image which enables an easy distribution and installation on local systems. Study data is only stored in the application as long as the calculations are performed which is compliant with data protection endeavors. Analysis times are below half an hour, even for larger studies with over 6000 subjects. Medical experts have ensured the usefulness of this application to grant an overview of their collected study data for monitoring purposes and to generate descriptive statistics without further user interaction. The semi-automatic analysis has its limitations and cannot replace the complex analysis of statisticians, but it can be used as a starting point for their examination and reporting.

  2. Validation of satellite derived LHF using coare_3.0 scheme and time series data over north-east Indian Ocean

    Digital Repository Service at National Institute of Oceanography (India)

    Muraleedharan, P.M.; Pankajakshan, T.; Sathe, P.V.

    to the scientific community as it call for near perfect observational platforms and sensors to Page 1 of 10Gayana (Concepción) - VALIDATION OF SATELLITE DERIVED LHF USING C... 8/11/2006http://www.scielo.cl/scielo.php?script=sci_arttext&pid=S0717...>VALIDATION OF SATELLITE DERIVED LHF USING C... 8/11/2006http://www.scielo.cl/scielo.php?script=sci_arttext&pid=S0717-65382004000300019&lng=... Day and night passes of SSMI (wind speed and columnar water vapor) and TMI (sea surface temperature) data for the period July...

  3. Quality assessment of the Ozone_cci Climate Research Data Package (release 2017) - Part 1: Ground-based validation of total ozone column data products

    Science.gov (United States)

    Garane, Katerina; Lerot, Christophe; Coldewey-Egbers, Melanie; Verhoelst, Tijl; Elissavet Koukouli, Maria; Zyrichidou, Irene; Balis, Dimitris S.; Danckaert, Thomas; Goutail, Florence; Granville, Jose; Hubert, Daan; Keppens, Arno; Lambert, Jean-Christopher; Loyola, Diego; Pommereau, Jean-Pierre; Van Roozendael, Michel; Zehner, Claus

    2018-03-01

    The GOME-type Total Ozone Essential Climate Variable (GTO-ECV) is a level-3 data record, which combines individual sensor products into one single cohesive record covering the 22-year period from 1995 to 2016, generated in the frame of the European Space Agency's Climate Change Initiative Phase II. It is based on level-2 total ozone data produced by the GODFIT (GOME-type Direct FITting) v4 algorithm as applied to the GOME/ERS-2, OMI/Aura, SCIAMACHY/Envisat and GOME-2/Metop-A and Metop-B observations. In this paper we examine whether GTO-ECV meets the specific requirements set by the international climate-chemistry modelling community for decadal stability long-term and short-term accuracy. In the following, we present the validation of the 2017 release of the Climate Research Data Package Total Ozone Column (CRDP TOC) at both level 2 and level 3. The inter-sensor consistency of the individual level-2 data sets has mean differences generally within 0.5 % at moderate latitudes (±50°), whereas the level-3 data sets show mean differences with respect to the OMI reference data record that span between -0.2 ± 0.9 % (for GOME-2B) and 1.0 ± 1.4 % (for SCIAMACHY). Very similar findings are reported for the level-2 validation against independent ground-based TOC observations reported by Brewer, Dobson and SAOZ instruments: the mean bias between GODFIT v4 satellite TOC and the ground instrument is well within 1.0 ± 1.0 % for all sensors, the drift per decade spans between -0.5 % and 1.0 ± 1.0 % depending on the sensor, and the peak-to-peak seasonality of the differences ranges from ˜ 1 % for GOME and OMI to ˜ 2 % for SCIAMACHY. For the level-3 validation, our first goal was to show that the level-3 CRDP produces findings consistent with the level-2 individual sensor comparisons. We show a very good agreement with 0.5 to 2 % peak-to-peak amplitude for the monthly mean difference time series and a negligible drift per decade of the differences in the Northern Hemisphere

  4. Experimental validation of a linear model for data reduction in chirp-pulse microwave CT.

    Science.gov (United States)

    Miyakawa, M; Orikasa, K; Bertero, M; Boccacci, P; Conte, F; Piana, M

    2002-04-01

    Chirp-pulse microwave computerized tomography (CP-MCT) is an imaging modality developed at the Department of Biocybernetics, University of Niigata (Niigata, Japan), which intends to reduce the microwave-tomography problem to an X-ray-like situation. We have recently shown that data acquisition in CP-MCT can be described in terms of a linear model derived from scattering theory. In this paper, we validate this model by showing that the theoretically computed response function is in good agreement with the one obtained from a regularized multiple deconvolution of three data sets measured with the prototype of CP-MCT. Furthermore, the reliability of the model as far as image restoration in concerned, is tested in the case of space-invariant conditions by considering the reconstruction of simple on-axis cylindrical phantoms.

  5. Validity of Self-reported Healthcare Utilization Data in the Community Health Survey in Korea

    Science.gov (United States)

    Rim, Hwayoung; Lee, Kunsei; Chang, Sounghoon; Hovell, Melbourne F; Kim, Young-Taek; Kim, Yuna; Kang, Gilwon; Tak, Yangju; Im, Jeehye

    2011-01-01

    To evaluate the sensitivity and specificity of Community Health Survey (CHS), we analyzed data from 11,217 participants aged ≥ 19 yr, in 13 cities and counties in 2008. Three healthcare utilization indices (admission, outpatient visits, dental visits) as comparative variables and the insurance benefit claim data of the Health Insurance Review & Assessment Service as the gold-standard were used. The sensitivities of admission, outpatient visits, and dental visits in CHS were 54.8%, 52.1%, and 61.0%, respectively. The specificities were 96.4%, 85.6%, and 82.7%, respectively. This is the first study to evaluate the validity of nationwide health statistics resulting from questionnaire surveys and shows that CHS needs a lot of efforts to reflect the true health status, health behavior, and healthcare utilization of the population. PMID:22065895

  6. Validity and validation of expert (Q)SAR systems.

    Science.gov (United States)

    Hulzebos, E; Sijm, D; Traas, T; Posthumus, R; Maslankiewicz, L

    2005-08-01

    At a recent workshop in Setubal (Portugal) principles were drafted to assess the suitability of (quantitative) structure-activity relationships ((Q)SARs) for assessing the hazards and risks of chemicals. In the present study we applied some of the Setubal principles to test the validity of three (Q)SAR expert systems and validate the results. These principles include a mechanistic basis, the availability of a training set and validation. ECOSAR, BIOWIN and DEREK for Windows have a mechanistic or empirical basis. ECOSAR has a training set for each QSAR. For half of the structural fragments the number of chemicals in the training set is >4. Based on structural fragments and log Kow, ECOSAR uses linear regression to predict ecotoxicity. Validating ECOSAR for three 'valid' classes results in predictivity of > or = 64%. BIOWIN uses (non-)linear regressions to predict the probability of biodegradability based on fragments and molecular weight. It has a large training set and predicts non-ready biodegradability well. DEREK for Windows predictions are supported by a mechanistic rationale and literature references. The structural alerts in this program have been developed with a training set of positive and negative toxicity data. However, to support the prediction only a limited number of chemicals in the training set is presented to the user. DEREK for Windows predicts effects by 'if-then' reasoning. The program predicts best for mutagenicity and carcinogenicity. Each structural fragment in ECOSAR and DEREK for Windows needs to be evaluated and validated separately.

  7. Validation of abridged mini-mental state examination scales using population-based data from Sweden and USA.

    Science.gov (United States)

    Ericsson, Malin Christina; Gatz, Margaret; Kåreholt, Ingemar; Parker, Marti G; Fors, Stefan

    2017-06-01

    The objective of this study is to validate two abridged versions of the mini-mental state examination (MMSE): one intended for use in face-to-face interviews, and the other developed for telephonic interviews, using data from Sweden and the US to validate the abridged scales against dementia diagnoses as well as to compare their performance to that of the full MMSE scale. The abridged versions were based on eight domains from the original MMSE scale. The domains included in the MMSE-SF were registration, orientation, delayed recall, attention, and visual spatial ability. In the MMSE-SF-C, the visual spatial ability item was excluded, and instead, one additional orientation item was added. There were 794 participants from the Swedish HARMONY study [mean age 81.8 (4.8); the proportion of cognitively impaired was 51 %] and 576 participants from the US ADAMS study [mean age 83.2 (5.7); the proportion of cognitively impaired was 65 %] where it was possible to compare abridged MMSE scales to dementia diagnoses and to the full MMSE scale. We estimated the sensitivity and specificity levels of the abridged tests, using clinical diagnoses as reference. Analyses with both the HARMONY and the ADAMS data indicated comparable levels of sensitivity and specificity in detecting cognitive impairment for the two abridged scales relative to the full MMSE. Receiver operating characteristic curves indicated that the two abridged scales corresponded well to those of the full MMSE. The two abridged tests have adequate validity and correspond well with the full MMSE. The abridged versions could therefore be alternatives to consider in larger population studies where interview length is restricted, and the respondent burden is high.

  8. Earth Observation for Citizen Science Validation, or Citizen Science for Earth Observation Validation? The Role of Quality Assurance of Volunteered Observations

    Directory of Open Access Journals (Sweden)

    Didier G. Leibovici

    2017-10-01

    Full Text Available Environmental policy involving citizen science (CS is of growing interest. In support of this open data stream of information, validation or quality assessment of the CS geo-located data to their appropriate usage for evidence-based policy making needs a flexible and easily adaptable data curation process ensuring transparency. Addressing these needs, this paper describes an approach for automatic quality assurance as proposed by the Citizen OBservatory WEB (COBWEB FP7 project. This approach is based upon a workflow composition that combines different quality controls, each belonging to seven categories or “pillars”. Each pillar focuses on a specific dimension in the types of reasoning algorithms for CS data qualification. These pillars attribute values to a range of quality elements belonging to three complementary quality models. Additional data from various sources, such as Earth Observation (EO data, are often included as part of the inputs of quality controls within the pillars. However, qualified CS data can also contribute to the validation of EO data. Therefore, the question of validation can be considered as “two sides of the same coin”. Based on an invasive species CS study, concerning Fallopia japonica (Japanese knotweed, the paper discusses the flexibility and usefulness of qualifying CS data, either when using an EO data product for the validation within the quality assurance process, or validating an EO data product that describes the risk of occurrence of the plant. Both validation paths are found to be improved by quality assurance of the CS data. Addressing the reliability of CS open data, issues and limitations of the role of quality assurance for validation, due to the quality of secondary data used within the automatic workflow, are described, e.g., error propagation, paving the route to improvements in the approach.

  9. HZETRN radiation transport validation using balloon-based experimental data

    Science.gov (United States)

    Warner, James E.; Norman, Ryan B.; Blattnig, Steve R.

    2018-05-01

    The deterministic radiation transport code HZETRN (High charge (Z) and Energy TRaNsport) was developed by NASA to study the effects of cosmic radiation on astronauts and instrumentation shielded by various materials. This work presents an analysis of computed differential flux from HZETRN compared with measurement data from three balloon-based experiments over a range of atmospheric depths, particle types, and energies. Model uncertainties were quantified using an interval-based validation metric that takes into account measurement uncertainty both in the flux and the energy at which it was measured. Average uncertainty metrics were computed for the entire dataset as well as subsets of the measurements (by experiment, particle type, energy, etc.) to reveal any specific trends of systematic over- or under-prediction by HZETRN. The distribution of individual model uncertainties was also investigated to study the range and dispersion of errors beyond just single scalar and interval metrics. The differential fluxes from HZETRN were generally well-correlated with balloon-based measurements; the median relative model difference across the entire dataset was determined to be 30%. The distribution of model uncertainties, however, revealed that the range of errors was relatively broad, with approximately 30% of the uncertainties exceeding ± 40%. The distribution also indicated that HZETRN systematically under-predicts the measurement dataset as a whole, with approximately 80% of the relative uncertainties having negative values. Instances of systematic bias for subsets of the data were also observed, including a significant underestimation of alpha particles and protons for energies below 2.5 GeV/u. Muons were found to be systematically over-predicted at atmospheric depths deeper than 50 g/cm2 but under-predicted for shallower depths. Furthermore, a systematic under-prediction of alpha particles and protons was observed below the geomagnetic cutoff, suggesting that

  10. Pre-Launch Phase 1 Calibration and Validation Rehearsal of Geophysical Data Products of Soil Moisture Active Passive (SMAP) Mission

    Science.gov (United States)

    Colliander, A.; Jackson, T. J.; Chan, S.; Dunbar, R.; Das, N. N.; Kim, S.; Reichle, R. H.; De Lannoy, G. J.; Liu, Q.; Kimball, J. S.; Yi, Y.; Cosh, M. H.; Bindlish, R.; Crow, W. T.; Dang, L.; Yueh, S. H.; Njoku, E. G.

    2013-12-01

    NASA's Soil Moisture Active Passive (SMAP) Mission is scheduled for launch in October 2014. The objective of the mission is global mapping of soil moisture and freeze/thaw state. SMAP utilizes an L-band radar and radiometer sharing a rotating 6-meter mesh reflector antenna. The instruments will operate onboard the SMAP spacecraft in a 685-km Sun-synchronous near-polar orbit, viewing the surface at a constant 40-degree incidence angle with a 1000-km swath width. Merging of active and passive L-band observations of the mission will enable an unprecedented combination of accuracy, resolution, coverage and revisit-time for soil moisture and freeze/thaw state retrieval. SMAP measurements will enable significantly improved estimates of water, energy and carbon transfers between the land and atmosphere. The SMAP science data product suite of geophysical parameters will include estimates of surface (top 5 cm) and root-zone (down to 1-m depth) soil moisture, net ecosystem exchange, and classification of the frozen/non-frozen state of the landscape. The primary validation reference of the data products will be ground-based measurements. Other remote sensing and model-based products will be used as additional resources. The post-launch timeline of the mission requires that the geophysical data products are validated (with respect to the mission requirements) within 12 months after a 3-month in-orbit check-out phase. SMAP is taking several preparatory steps in order to meet this schedule. One of the main steps consists of running a rehearsal to exercise calibration and validation procedures planned for the Cal/Val Phase. The rehearsal is divided into two stages. Phase 1, which was conducted in June-August 2013, focused on validation methodologies for the geophysical data products. Phase 2, which will be conducted in May-June 2014, includes operational aspects including a fully functioning SMAP Science Data System. (Note that the rehearsals do not include an airborne field

  11. Injury surveillance in community sport: Can we obtain valid data from sports trainers?

    Science.gov (United States)

    Ekegren, C L; Gabbe, B J; Finch, C F

    2015-06-01

    A lack of available injury data on community sports participants has hampered the development of informed preventive strategies for the broad-base of sports participation. In community sports settings, sports trainers or first-aiders are well-placed to carry out injury surveillance, but few studies have evaluated their ability to do so. The aim of this study was to investigate the reporting rate and completeness of sports trainers' injury records and agreement between sports trainers' and players' reports of injury in community Australian football. Throughout the football season, one sports trainer from each of four clubs recorded players' injuries. To validate these data, we collected self-reported injury data from players via short message service (SMS). In total, 210 discrete injuries were recorded for 139 players, 21% by sports trainers only, 59% by players via SMS only, and 21% by both. Completeness of injury records ranged from 95% to 100%. Agreement between sports trainers and players ranged from K = 0.32 (95% confidence interval: 0.27, 0.37) for date of return to football to K = 1.00 for activity when injured. Injury data collected by sports trainers may be of adequate quality for providing an understanding of the profile of injuries. However, data are likely to underestimate injury rates and should be interpreted with caution. © 2014 John Wiley & Sons A/S. Published by John Wiley & Sons Ltd.

  12. From patient care to research: a validation study examining the factors contributing to data quality in a primary care electronic medical record database.

    Science.gov (United States)

    Coleman, Nathan; Halas, Gayle; Peeler, William; Casaclang, Natalie; Williamson, Tyler; Katz, Alan

    2015-02-05

    Electronic Medical Records (EMRs) are increasingly used in the provision of primary care and have been compiled into databases which can be utilized for surveillance, research and informing practice. The primary purpose of these records is for the provision of individual patient care; validation and examination of underlying limitations is crucial for use for research and data quality improvement. This study examines and describes the validity of chronic disease case definition algorithms and factors affecting data quality in a primary care EMR database. A retrospective chart audit of an age stratified random sample was used to validate and examine diagnostic algorithms applied to EMR data from the Manitoba Primary Care Research Network (MaPCReN), part of the Canadian Primary Care Sentinel Surveillance Network (CPCSSN). The presence of diabetes, hypertension, depression, osteoarthritis and chronic obstructive pulmonary disease (COPD) was determined by review of the medical record and compared to algorithm identified cases to identify discrepancies and describe the underlying contributing factors. The algorithm for diabetes had high sensitivity, specificity and positive predictive value (PPV) with all scores being over 90%. Specificities of the algorithms were greater than 90% for all conditions except for hypertension at 79.2%. The largest deficits in algorithm performance included poor PPV for COPD at 36.7% and limited sensitivity for COPD, depression and osteoarthritis at 72.0%, 73.3% and 63.2% respectively. Main sources of discrepancy included missing coding, alternative coding, inappropriate diagnosis detection based on medications used for alternate indications, inappropriate exclusion due to comorbidity and loss of data. Comparison to medical chart review shows that at MaPCReN the CPCSSN case finding algorithms are valid with a few limitations. This study provides the basis for the validated data to be utilized for research and informs users of its

  13. Validation of Copernicus Height-resolved Ozone data Products from Sentinel-5P TROPOMI using global sonde and lidar networks (CHEOPS-5P)

    Science.gov (United States)

    Keppens, Arno; Lambert, Jean-Christopher; Hubert, Daan; Verhoelst, Tijl; Granville, José; Ancellet, Gérard; Balis, Dimitris; Delcloo, Andy; Duflot, Valentin; Godin-Beekmann, Sophie; Koukouli, Marilisa; Leblanc, Thierry; Stavrakou, Trissevgeni; Steinbrecht, Wolfgang; Stübi, Réné; Thompson, Anne

    2017-04-01

    Monitoring of and research on air quality, stratospheric ozone and climate change require global and long-term observation of the vertical distribution of atmospheric ozone, at ever-improving resolution and accuracy. Global tropospheric and stratospheric ozone profile measurement capabilities from space have therefore improved substantially over the last decades. Being a part of the space segment of the Copernicus Atmosphere and Climate Services that is currently under implementation, the upcoming Sentinel-5 Precursor (S5P) mission with its imaging spectrometer TROPOMI (Tropospheric Monitoring Instrument) is dedicated to the measurement of nadir atmospheric radiance and solar irradiance in the UV-VIS-NIR-SWIR spectral range. Ozone profile and tropospheric ozone column data will be retrieved from these measurements by use of several complementary retrieval methods. The geophysical validation of the enhanced height-resolved ozone data products, as well as support to the continuous evolution of the associated retrieval algorithms, is a key objective of the CHEOPS-5P project, a contributor to the ESA-led S5P Validation Team (S5PVT). This work describes the principles and implementation of the CHEOPS-5P quality assessment (QA) and validation system. The QA/validation methodology relies on the analysis of S5P retrieval diagnostics and on comparisons of S5P data with reference ozone profile measurements. The latter are collected from ozonesonde, stratospheric lidar and tropospheric lidar stations performing network operation in the context of WMO's Global Atmosphere Watch, including the NDACC global and SHADOZ tropical networks. After adaptation of the Multi-TASTE versatile satellite validation environment currently operational in the context of ESA's CCI, EUMETSAT O3M-SAF, and CEOS and SPARC initiatives, a list of S5P data Quality Indicators (QI) will be derived from complementary investigations: (1) data content and information content studies of the S5P data retrievals

  14. Validation of models with multivariate output

    International Nuclear Information System (INIS)

    Rebba, Ramesh; Mahadevan, Sankaran

    2006-01-01

    This paper develops metrics for validating computational models with experimental data, considering uncertainties in both. A computational model may generate multiple response quantities and the validation experiment might yield corresponding measured values. Alternatively, a single response quantity may be predicted and observed at different spatial and temporal points. Model validation in such cases involves comparison of multiple correlated quantities. Multiple univariate comparisons may give conflicting inferences. Therefore, aggregate validation metrics are developed in this paper. Both classical and Bayesian hypothesis testing are investigated for this purpose, using multivariate analysis. Since, commonly used statistical significance tests are based on normality assumptions, appropriate transformations are investigated in the case of non-normal data. The methodology is implemented to validate an empirical model for energy dissipation in lap joints under dynamic loading

  15. Application of validation data for assessing spatial interpolation methods for 8-h ozone or other sparsely monitored constituents.

    Science.gov (United States)

    Joseph, John; Sharif, Hatim O; Sunil, Thankam; Alamgir, Hasanat

    2013-07-01

    The adverse health effects of high concentrations of ground-level ozone are well-known, but estimating exposure is difficult due to the sparseness of urban monitoring networks. This sparseness discourages the reservation of a portion of the monitoring stations for validation of interpolation techniques precisely when the risk of overfitting is greatest. In this study, we test a variety of simple spatial interpolation techniques for 8-h ozone with thousands of randomly selected subsets of data from two urban areas with monitoring stations sufficiently numerous to allow for true validation. Results indicate that ordinary kriging with only the range parameter calibrated in an exponential variogram is the generally superior method, and yields reliable confidence intervals. Sparse data sets may contain sufficient information for calibration of the range parameter even if the Moran I p-value is close to unity. R script is made available to apply the methodology to other sparsely monitored constituents. Copyright © 2013 Elsevier Ltd. All rights reserved.

  16. Validation of MOPITT carbon monoxide using ground-based Fourier transform infrared spectrometer data from NDACC

    Science.gov (United States)

    Buchholz, Rebecca R.; Deeter, Merritt N.; Worden, Helen M.; Gille, John; Edwards, David P.; Hannigan, James W.; Jones, Nicholas B.; Paton-Walsh, Clare; Griffith, David W. T.; Smale, Dan; Robinson, John; Strong, Kimberly; Conway, Stephanie; Sussmann, Ralf; Hase, Frank; Blumenstock, Thomas; Mahieu, Emmanuel; Langerock, Bavo

    2017-06-01

    The Measurements of Pollution in the Troposphere (MOPITT) satellite instrument provides the longest continuous dataset of carbon monoxide (CO) from space. We perform the first validation of MOPITT version 6 retrievals using total column CO measurements from ground-based remote-sensing Fourier transform infrared spectrometers (FTSs). Validation uses data recorded at 14 stations, that span a wide range of latitudes (80° N to 78° S), in the Network for the Detection of Atmospheric Composition Change (NDACC). MOPITT measurements are spatially co-located with each station, and different vertical sensitivities between instruments are accounted for by using MOPITT averaging kernels (AKs). All three MOPITT retrieval types are analyzed: thermal infrared (TIR-only), joint thermal and near infrared (TIR-NIR), and near infrared (NIR-only). Generally, MOPITT measurements overestimate CO relative to FTS measurements, but the bias is typically less than 10 %. Mean bias is 2.4 % for TIR-only, 5.1 % for TIR-NIR, and 6.5 % for NIR-only. The TIR-NIR and NIR-only products consistently produce a larger bias and lower correlation than the TIR-only. Validation performance of MOPITT for TIR-only and TIR-NIR retrievals over land or water scenes is equivalent. The four MOPITT detector element pixels are validated separately to account for their different uncertainty characteristics. Pixel 1 produces the highest standard deviation and lowest correlation for all three MOPITT products. However, for TIR-only and TIR-NIR, the error-weighted average that includes all four pixels often provides the best correlation, indicating compensating pixel biases and well-captured error characteristics. We find that MOPITT bias does not depend on latitude but rather is influenced by the proximity to rapidly changing atmospheric CO. MOPITT bias drift has been bound geographically to within ±0.5 % yr-1 or lower at almost all locations.

  17. Data Quality in Institutional Arthroplasty Registries: Description of a Model of Validation and Report of Preliminary Results.

    Science.gov (United States)

    Bautista, Maria P; Bonilla, Guillermo A; Mieth, Klaus W; Llinás, Adolfo M; Rodríguez, Fernanda; Cárdenas, Laura L

    2017-07-01

    Arthroplasty registries are a relevant source of information for research and quality improvement in patient care and its value depends on the quality of the recorded data. The purpose of this study is to describe a model of validation and present the findings of validation of an Institutional Arthroplasty Registry (IAR). Information from 209 primary arthroplasties and revision surgeries of the hip, knee, and shoulder recorded in the IAR between March and September 2015 were analyzed in the following domains. Adherence is defined as the proportion of patients included in the registry, completeness is defined as the proportion of data effectively recorded, and accuracy is defined as the proportion of data consistent with medical records. A random sample of 53 patients (25.4%) was selected to assess the latest 2 domains. A direct comparison between the registry's database and medical records was performed. In total, 324 variables containing information on demographic data, surgical procedure, clinical outcomes, and key performance indicators were analyzed. Two hundred nine of 212 patients who underwent surgery during the study period were included in the registry, accounting for an adherence of 98.6%. Completeness was 91.7% and accuracy was 85.8%. Most errors were found in the preoperative range of motion and timely administration of prophylactic antibiotics and thromboprophylaxis. This model provides useful information regarding the quality of the recorded data since it identified deficient areas within the IAR. We recommend that institutional arthroplasty registries be constantly monitored for data quality before using their information for research or quality improvement purposes. Copyright © 2017 Elsevier Inc. All rights reserved.

  18. Validation of meter-scale surface faulting offset measurements from high-resolution topographic data

    Science.gov (United States)

    Salisbury, Barrett; Haddad, D.E.; Rockwell, T.K.; Arrowsmith, R.; Madugo, C.; Zielke, O.; Scharer, Katherine M.

    2015-01-01

    Studies of active fault zones have flourished with the availability of high-resolution topographic data, particularly where airborne light detection and ranging (lidar) and structure from motion (SfM) data sets provide a means to remotely analyze submeter-scale fault geomorphology. To determine surface offset at a point along a strike-slip earthquake rupture, geomorphic features (e.g., stream channels) are measured days to centuries after the event. Analysis of these and cumulatively offset features produces offset distributions for successive earthquakes that are used to understand earthquake rupture behavior. As researchers expand studies to more varied terrain types, climates, and vegetation regimes, there is an increasing need to standardize and uniformly validate measurements of tectonically displaced geomorphic features. A recently compiled catalog of nearly 5000 earthquake offsets across a range of measurement and reporting styles provides insight into quality rating and uncertainty trends from which we formulate best-practice and reporting recommendations for remote studies. In addition, a series of public and beginner-level studies validate the remote methodology for a number of tools and emphasize considerations to enhance measurement accuracy and precision for beginners and professionals. Our investigation revealed that (1) standardizing remote measurement methods and reporting quality rating schemes is essential for the utility and repeatability of fault-offset measurements; (2) measurement discrepancies often involve misinterpretation of the offset geomorphic feature and are a function of the investigator’s experience; (3) comparison of measurements made by a single investigator in different climatic regions reveals systematic differences in measurement uncertainties attributable to variation in feature preservation; (4) measuring more components of a displaced geomorphic landform produces more consistently repeatable estimates of offset; and (5

  19. Validation of meter-scale surface faulting offset measurements from high-resolution topographic data

    KAUST Repository

    Salisbury, J. Barrett

    2015-10-24

    Studies of active fault zones have flourished with the availability of high-resolution topographic data, particularly where airborne light detection and ranging (lidar) and structure from motion (SfM) data sets provide a means to remotely analyze submeter- scale fault geomorphology. To determine surface offset at a point along a strike-slip earthquake rupture, geomorphic features (e.g., stream channels) are measured days to centuries after the event. Analysis of these and cumulatively offset features produces offset distributions for successive earthquakes that are used to understand earthquake rupture behavior. As researchers expand studies to more varied terrain types, climates, and vegetation regimes, there is an increasing need to standardize and uniformly validate measurements of tectonically displaced geomorphic features. A recently compiled catalog of nearly 5000 earthquake offsets across a range of measurement and reporting styles provides insight into quality rating and uncertainty trends from which we formulate best-practice and reporting recommendations for remote studies. In addition, a series of public and beginner-level studies validate the remote methodology for a number of tools and emphasize considerations to enhance measurement accuracy and precision for beginners and professionals. Our investigation revealed that (1) standardizing remote measurement methods and reporting quality rating schemes is essential for the utility and repeatability of fault-offset measurements; (2) measurement discrepancies often involve misinterpretation of the offset geomorphic feature and are a function of the investigator\\'s experience; (3) comparison of measurements made by a single investigator in different climatic regions reveals systematic differences in measurement uncertainties attributable to variation in feature preservation; (4) measuring more components of a displaced geomorphic landform produces more consistently repeatable estimates of offset; and (5

  20. Preliminary data on validity of the Drug Addiction Treatment Efficacy Questionnaire.

    Science.gov (United States)

    Kastelic, Andrej; Mlakar, Janez; Pregelj, Peter

    2013-09-01

    This study describes the validation process for the Slovenian version of the Drug Addiction Treatment Efficacy Questionnaire (DATEQ). DATEQ was constructed from the questionnaires used at the Centre for the Treatment of Drug Addiction, Ljubljana University Psychiatric Hospital, and within the network of Centres for the Prevention and Treatment of Drug Addiction in Slovenia during the past 14 years. The Slovenian version of the DATEQ was translated to English using the 'forward-backward' procedure by its authors and their co-workers. The validation process included 100 male and female patients with established addiction to illicit drugs who had been prescribed opioid substitution therapy. The DATEQ questionnaire was used in the study, together with clinical evaluation to measure psychological state and to evaluate the efficacy of treatment in the last year. To determinate the validity of DATEQ the correlation with the clinical assessments of the outcome was calculated using one-way ANOVA. The F value was 44.4, p<0.001 (sum of squares: between groups 210.4, df=2, within groups 229.7, df=97, total 440.1, df=99). At the cut-off 4 the sensitivity is 81% and specificity 83%. The validation process for the Slovenian DATEQ version shows metric properties similar to those found in international studies of similar questionnaires, suggesting that it measures the same constructs, in the same way and as similar questionnaires. However, the relatively low sensitivity and specificity suggests caution when using DATEQ as the only measure of outcome.

  1. Sound quality indicators for urban places in Paris cross-validated by Milan data.

    Science.gov (United States)

    Ricciardi, Paola; Delaitre, Pauline; Lavandier, Catherine; Torchia, Francesca; Aumond, Pierre

    2015-10-01

    A specific smartphone application was developed to collect perceptive and acoustic data in Paris. About 3400 questionnaires were analyzed, regarding the global sound environment characterization, the perceived loudness of some emergent sources and the presence time ratio of sources that do not emerge from the background. Sound pressure level was recorded each second from the mobile phone's microphone during a 10-min period. The aim of this study is to propose indicators of urban sound quality based on linear regressions with perceptive variables. A cross validation of the quality models extracted from Paris data was carried out by conducting the same survey in Milan. The proposed sound quality general model is correlated with the real perceived sound quality (72%). Another model without visual amenity and familiarity is 58% correlated with perceived sound quality. In order to improve the sound quality indicator, a site classification was performed by Kohonen's Artificial Neural Network algorithm, and seven specific class models were developed. These specific models attribute more importance on source events and are slightly closer to the individual data than the global model. In general, the Parisian models underestimate the sound quality of Milan environments assessed by Italian people.

  2. The ICR96 exon CNV validation series: a resource for orthogonal assessment of exon CNV calling in NGS data [version 1; referees: 2 approved

    Directory of Open Access Journals (Sweden)

    Shazia Mahamdallie

    2017-05-01

    Full Text Available Detection of deletions and duplications of whole exons (exon CNVs is a key requirement of genetic testing. Accurate detection of this variant type has proved very challenging in targeted next-generation sequencing (NGS data, particularly if only a single exon is involved. Many different NGS exon CNV calling methods have been developed over the last five years. Such methods are usually evaluated using simulated and/or in-house data due to a lack of publicly-available datasets with orthogonally generated results. This hinders tool comparisons, transparency and reproducibility. To provide a community resource for assessment of exon CNV calling methods in targeted NGS data, we here present the ICR96 exon CNV validation series. The dataset includes high-quality sequencing data from a targeted NGS assay (the TruSight Cancer Panel together with Multiplex Ligation-dependent Probe Amplification (MLPA results for 96 independent samples. 66 samples contain at least one validated exon CNV and 30 samples have validated negative results for exon CNVs in 26 genes. The dataset includes 46 exon CNVs in BRCA1, BRCA2, TP53, MLH1, MSH2, MSH6, PMS2, EPCAM or PTEN, giving excellent representation of the cancer predisposition genes most frequently tested in clinical practice. Moreover, the validated exon CNVs include 25 single exon CNVs, the most difficult type of exon CNV to detect. The FASTQ files for the ICR96 exon CNV validation series can be accessed through the European-Genome phenome Archive (EGA under the accession number EGAS00001002428.

  3. CRED Optical Validation Data at the island of Ta'u in American Samoa, 2004 to Support Benthic Habitat Mapping (TOAD)

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — Optical validation data were collected using a Tethered Optical Assessment Device (TOAD), an underwater sled equipped with an underwater digital video camera and...

  4. CRED Optical Validation Data at the island of Ta'u in American Samoa, 2006 to support Benthic Habitat Mapping (TOAD)

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — Optical validation data were collected using a Tethered Optical Assessment Device (TOAD), an underwater sled equipped with an underwater digital video camera and...

  5. Reliability, Validity, Comparability and Practical Utility of Cybercrime-Related Data, Metrics, and Information

    Directory of Open Access Journals (Sweden)

    Nir Kshetri

    2013-02-01

    Full Text Available With an increasing pervasiveness, prevalence and severity of cybercrimes, various metrics, measures and statistics have been developed and used to measure various aspects of this phenomenon. Cybercrime-related data, metrics, and information, however, pose important and difficult dilemmas regarding the issues of reliability, validity, comparability and practical utility. While many of the issues of the cybercrime economy are similar to other underground and underworld industries, this economy also has various unique aspects. For one thing, this industry also suffers from a problem partly rooted in the incredibly broad definition of the term “cybercrime”. This article seeks to provide insights and analysis into this phenomenon, which is expected to advance our understanding into cybercrime-related information.

  6. Validation of the mathematical model of the NAPS PHT system flow with test data

    International Nuclear Information System (INIS)

    Rajesh Kumar, K.; Vani, K.; Chakraborty, G.; Venkat Raj, V.

    1994-01-01

    A dynamic analysis code to predict the time dependent behaviour of the reactor coolant system flow following the tripping and starting of Primary Circulating Pumps in the different operating modes has been developed for Indian Pressurised Heavy Water Reactor (PHWR) type power plants. The model is comprised of reactor coolant momentum equation, Primary Heat Transport (PHT) pump dynamic equation and pump characteristics. This model forms one of the modules of the integrated system code being developed for transient analysis of 220 MWe PHWR power plants. The Narora Atomic Power Station (NAPS) PHT system flow transient results for different combinations of pump operation predicted by the model have been compared with the experimental data obtained from a test carried out in NAPS-2 for validation of the model. The predicted results are in good agreement with the experimental data. (author). 3 refs., 5 figs

  7. Validation of dengue infection severity score

    Directory of Open Access Journals (Sweden)

    Pongpan S

    2014-03-01

    Full Text Available Surangrat Pongpan,1,2 Jayanton Patumanond,3 Apichart Wisitwong,4 Chamaiporn Tawichasri,5 Sirianong Namwongprom1,6 1Clinical Epidemiology Program, Faculty of Medicine, Chiang Mai University, Chiang Mai, Thailand; 2Department of Occupational Medicine, Phrae Hospital, Phrae, Thailand; 3Clinical Epidemiology Program, Faculty of Medicine, Thammasat University, Bangkok, Thailand; 4Department of Social Medicine, Sawanpracharak Hospital, Nakorn Sawan, Thailand; 5Clinical Epidemiology Society at Chiang Mai, Chiang Mai, Thailand; 6Department of Radiology, Faculty of Medicine, Chiang Mai University, Chiang Mai, Thailand Objective: To validate a simple scoring system to classify dengue viral infection severity to patients in different settings. Methods: The developed scoring system derived from 777 patients from three tertiary-care hospitals was applied to 400 patients in the validation data obtained from another three tertiary-care hospitals. Percentage of correct classification, underestimation, and overestimation was compared. The score discriminative performance in the two datasets was compared by analysis of areas under the receiver operating characteristic curves. Results: Patients in the validation data were different from those in the development data in some aspects. In the validation data, classifying patients into three severity levels (dengue fever, dengue hemorrhagic fever, and dengue shock syndrome yielded 50.8% correct prediction (versus 60.7% in the development data, with clinically acceptable underestimation (18.6% versus 25.7% and overestimation (30.8% versus 13.5%. Despite the difference in predictive performances between the validation and the development data, the overall prediction of the scoring system is considered high. Conclusion: The developed severity score may be applied to classify patients with dengue viral infection into three severity levels with clinically acceptable under- or overestimation. Its impact when used in routine

  8. Numerical Validation of a Vortex Model against ExperimentalData on a Straight-Bladed Vertical Axis Wind Turbine

    Directory of Open Access Journals (Sweden)

    Eduard Dyachuk

    2015-10-01

    Full Text Available Cyclic blade motion during operation of vertical axis wind turbines (VAWTs imposes challenges on the simulations models of the aerodynamics of VAWTs. A two-dimensional vortex model is validated against the new experimental data on a 12-kW straight-bladed VAWT, which is operated at an open site. The results on the normal force on one blade are analyzed. The model is assessed against the measured data in the wide range of tip speed ratios: from 1.8 to 4.6. The predicted results within one revolution have a similar shape and magnitude as the measured data, though the model does not reproduce every detail of the experimental data. The present model can be used when dimensioning the turbine for maximum loads.

  9. Validation of Temperament and Character Inventory (TCI in Iranian Sample: Normative Data

    Directory of Open Access Journals (Sweden)

    M Poor Naseh

    2005-05-01

    Full Text Available Background: This Study aimed to validate the temperament and character inventory (TCI in an Iranian sample of men and women with different ages. TCI contains subscales designed to measure seven different personality traits and characteristics. Materials and Methods: In the first step, subjects (n=1212 completed the questionnaire. In the second step, to examine the reliability of the questionnaire, 101 randomly chosen subjects were re-tested one to two months after the first test. Also, in order to examine the validity of the questionnaire, 100 subjects were interviewed by two psychologists using a checklist based on the Cloninger's biological theory of personality. The interviewers, who were blind to the subjects scores on the seven subscales, rated each subject for the seven traits and characteristics on a 10-point rating scale (from 1 to 10. Results & Conclusion: The results showed normative data for the subscales novelty seeking (NS, harm avoidance (HA, reward dependence (RD, persistence (Per, self directiveness (SD, cooperation (Co and self transcendence (ST for different gender and age classes. Correlations between the scores and ratings of the test and re-test revealed significant coefficients, confirming reliability for all subscales. A good internal consistency was found for each subscale. The results also showed no significant correlations higher than 0.40 among NS, HA, Per and RD; the temperament subscales were independent from each other. The only significant correlation, higher than 0.40, among the character subscales was between SD and Co. Applied and clinical implication of the present findings will be discussed.

  10. Cross-Validation of a Glucose-Insulin-Glucagon Pharmacodynamics Model for Simulation using Data from Patients with Type 1 Diabetes

    DEFF Research Database (Denmark)

    Wendt, Sabrina Lyngbye; Ranjan, Ajenthen; Møller, Jan Kloppenborg

    2017-01-01

    three PD model test fits in each of the seven subjects. Thus, we successfully validated the PD model by leave-one-out cross-validation in seven out of eight T1D patients. Conclusions: The PD model accurately simulates glucose excursions based on plasma insulin and glucagon concentrations. The reported...... for concentrations of glucagon, insulin, and glucose. We fitted pharmacokinetic (PK) models to insulin and glucagon data using maximum likelihood and maximum a posteriori estimation methods. Similarly, we fitted a pharmacodynamic (PD) model to glucose data. The PD model included multiplicative effects of insulin...... and glucagon on EGP. Bias and precision of PD model test fits were assessed by mean predictive error (MPE) and mean absolute predictive error (MAPE). Results: Assuming constant variables in a subject across nonoutlier visits and using thresholds of ±15% MPE and 20% MAPE, we accepted at least one and at most...

  11. Satellite Validation: A Project to Create a Data-Logging System to Monitor Lake Tahoe

    Science.gov (United States)

    Roy, Rudy A.

    2005-01-01

    Flying aboard the satellite Terra, the Advanced Spaceborne Thermal Emission and Reflection Radiometer (ASTER) is an imaging instrument used to acquire detailed maps of Earth's surface temperature, elevation, emissivity, and reflectance. An automated site consisting of four buoys was established 6 years ago at Lake Tahoe for the validation of ASTERS thermal infrared data. Using Campbell CR23X Dataloggers, a replacement system to be deployed on a buoy was designed and constructed for the measurement of the lake's temperature profile, surrounding air temperature, humidity, wind direction and speed, net radiation, and surface skin temperature. Each Campbell Datalogger has been programmed to control, power, and monitor 14 different temperature sensors, a JPL-built radiometer, and an RM Young 32500 meteorological station. The logger communicates with the radiometer and meteorological station through a Campbell SDM-SIO4 RS232 serial interface, sending polling commands, and receiving filtered data back from the sensors. This data is then cataloged and sent back across a cellular modem network every hour to JPL. Each instrument is wired via a panel constructed with 18 individual plugs that allow for simple installation and expansion. Data sent back from the system are analyzed at JPL, where they are used to calibrate ASTER data.

  12. Measurement of leakage neutron spectra for Tungsten with D-T neutrons and validation of evaluated nuclear data

    International Nuclear Information System (INIS)

    Zhang, S.; Chen, Z.; Nie, Y.; Wada, R.; Ruan, X.; Han, R.; Liu, X.; Lin, W.; Liu, J.; Shi, F.; Ren, P.; Tian, G.; Luo, F.; Ren, J.; Bao, J.

    2015-01-01

    Highlights: • Evaluated data for Tungsten are validated by integral experiment. • Leakage neutron spectra from the irradiation of D-T neutrons on Tungsten are measured at 60° and 120° by using a time-of-flight method. • The measured results are compared to the MCNP-4C calculated ones with evaluated data of the different libraries. - Abstract: Integral neutronics experiments have been investigated at Institute of Modern Physics, Chinese Academy of Sciences (IMP, CAS) in order to validate evaluated nuclear data related to the design of Chinese Initiative Accelerator Driven Systems (CIADS). In the present paper, the accuracy of evaluated nuclear data for Tungsten has been examined by comparing measured leakage neutron spectra with calculated ones. Leakage neutron spectra from the irradiation of D-T neutrons on Tungsten slab sample were experimentally measured at 60° and 120° by using a time-of-flight method. Theoretical calculations are carried out by Monte Carlo neutron transport code MCNP-4C with evaluated nuclear data of the ADS-2.0, ENDF/B-VII.0, ENDF/B-VII.1, JENDL-4.0 and CENDL-3.1 libraries. From the comparisons, it is found that the calculations with ADS-2.0 and ENDF/B-VII.1 give good agreements with the experiments in the whole energy regions at 60°, while a large discrepancy is observed at 120° in the elastic scattering peak, caused by a slight difference in the oscillation pattern of the elastic angular distribution at angles larger than 20°. However, the calculated spectra using data from ENDF/B-VII.0, JENDL-4.0 and CENDL-3.1 libraries showed larger discrepancies with the measured ones, especially around 8.5–13.5 MeV. Further studies are presented for these disagreements

  13. MARS Validation Plan and Status

    International Nuclear Information System (INIS)

    Ahn, Seung-hoon; Cho, Yong-jin

    2008-01-01

    The KINS Reactor Thermal-hydraulic Analysis System (KINS-RETAS) under development is directed toward a realistic analysis approach of best-estimate (BE) codes and realistic assumptions. In this system, MARS is pivoted to provide the BE Thermal-Hydraulic (T-H) response in core and reactor coolant system to various operational transients and accidental conditions. As required for other BE codes, the qualification is essential to ensure reliable and reasonable accuracy for a targeted MARS application. Validation is a key element of the code qualification, and determines the capability of a computer code in predicting the major phenomena expected to occur. The MARS validation was made by its developer KAERI, on basic premise that its backbone code RELAP5/MOD3.2 is well qualified against analytical solutions, test or operational data. A screening was made to select the test data for MARS validation; some models transplanted from RELAP5, if already validated and found to be acceptable, were screened out from assessment. It seems to be reasonable, but does not demonstrate whether code adequacy complies with the software QA guidelines. Especially there may be much difficulty in validating the life-cycle products such as code updates or modifications. This paper presents the plan for MARS validation, and the current implementation status

  14. Model Validation Status Review

    International Nuclear Information System (INIS)

    E.L. Hardin

    2001-01-01

    The primary objective for the Model Validation Status Review was to perform a one-time evaluation of model validation associated with the analysis/model reports (AMRs) containing model input to total-system performance assessment (TSPA) for the Yucca Mountain site recommendation (SR). This review was performed in response to Corrective Action Request BSC-01-C-01 (Clark 2001, Krisha 2001) pursuant to Quality Assurance review findings of an adverse trend in model validation deficiency. The review findings in this report provide the following information which defines the extent of model validation deficiency and the corrective action needed: (1) AMRs that contain or support models are identified, and conversely, for each model the supporting documentation is identified. (2) The use for each model is determined based on whether the output is used directly for TSPA-SR, or for screening (exclusion) of features, events, and processes (FEPs), and the nature of the model output. (3) Two approaches are used to evaluate the extent to which the validation for each model is compliant with AP-3.10Q (Analyses and Models). The approaches differ in regard to whether model validation is achieved within individual AMRs as originally intended, or whether model validation could be readily achieved by incorporating information from other sources. (4) Recommendations are presented for changes to the AMRs, and additional model development activities or data collection, that will remedy model validation review findings, in support of licensing activities. The Model Validation Status Review emphasized those AMRs that support TSPA-SR (CRWMS M and O 2000bl and 2000bm). A series of workshops and teleconferences was held to discuss and integrate the review findings. The review encompassed 125 AMRs (Table 1) plus certain other supporting documents and data needed to assess model validity. The AMRs were grouped in 21 model areas representing the modeling of processes affecting the natural and

  15. Model Validation Status Review

    Energy Technology Data Exchange (ETDEWEB)

    E.L. Hardin

    2001-11-28

    The primary objective for the Model Validation Status Review was to perform a one-time evaluation of model validation associated with the analysis/model reports (AMRs) containing model input to total-system performance assessment (TSPA) for the Yucca Mountain site recommendation (SR). This review was performed in response to Corrective Action Request BSC-01-C-01 (Clark 2001, Krisha 2001) pursuant to Quality Assurance review findings of an adverse trend in model validation deficiency. The review findings in this report provide the following information which defines the extent of model validation deficiency and the corrective action needed: (1) AMRs that contain or support models are identified, and conversely, for each model the supporting documentation is identified. (2) The use for each model is determined based on whether the output is used directly for TSPA-SR, or for screening (exclusion) of features, events, and processes (FEPs), and the nature of the model output. (3) Two approaches are used to evaluate the extent to which the validation for each model is compliant with AP-3.10Q (Analyses and Models). The approaches differ in regard to whether model validation is achieved within individual AMRs as originally intended, or whether model validation could be readily achieved by incorporating information from other sources. (4) Recommendations are presented for changes to the AMRs, and additional model development activities or data collection, that will remedy model validation review findings, in support of licensing activities. The Model Validation Status Review emphasized those AMRs that support TSPA-SR (CRWMS M&O 2000bl and 2000bm). A series of workshops and teleconferences was held to discuss and integrate the review findings. The review encompassed 125 AMRs (Table 1) plus certain other supporting documents and data needed to assess model validity. The AMRs were grouped in 21 model areas representing the modeling of processes affecting the natural and

  16. Methodology for Validating Building Energy Analysis Simulations

    Energy Technology Data Exchange (ETDEWEB)

    Judkoff, R.; Wortman, D.; O' Doherty, B.; Burch, J.

    2008-04-01

    The objective of this report was to develop a validation methodology for building energy analysis simulations, collect high-quality, unambiguous empirical data for validation, and apply the validation methodology to the DOE-2.1, BLAST-2MRT, BLAST-3.0, DEROB-3, DEROB-4, and SUNCAT 2.4 computer programs. This report covers background information, literature survey, validation methodology, comparative studies, analytical verification, empirical validation, comparative evaluation of codes, and conclusions.

  17. 1:50 Scale Testing of Three Floating Wind Turbines at MARIN and Numerical Model Validation Against Test Data

    Energy Technology Data Exchange (ETDEWEB)

    Dagher, Habib [Univ. of Maine, Orno, ME (United States); Viselli, Anthony [Univ. of Maine, Orno, ME (United States); Goupee, Andrew [Univ. of Maine, Orno, ME (United States); Allen, Christopher [Univ. of Maine, Orno, ME (United States)

    2017-08-15

    The primary goal of the basin model test program discussed herein is to properly scale and accurately capture physical data of the rigid body motions, accelerations and loads for different floating wind turbine platform technologies. The intended use for this data is for performing comparisons with predictions from various aero-hydro-servo-elastic floating wind turbine simulators for calibration and validation. Of particular interest is validating the floating offshore wind turbine simulation capabilities of NREL’s FAST open-source simulation tool. Once the validation process is complete, coupled simulators such as FAST can be used with a much greater degree of confidence in design processes for commercial development of floating offshore wind turbines. The test program subsequently described in this report was performed at MARIN (Maritime Research Institute Netherlands) in Wageningen, the Netherlands. The models considered consisted of the horizontal axis, NREL 5 MW Reference Wind Turbine (Jonkman et al., 2009) with a flexible tower affixed atop three distinct platforms: a tension leg platform (TLP), a spar-buoy modeled after the OC3 Hywind (Jonkman, 2010) and a semi-submersible. The three generic platform designs were intended to cover the spectrum of currently investigated concepts, each based on proven floating offshore structure technology. The models were tested under Froude scale wind and wave loads. The high-quality wind environments, unique to these tests, were realized in the offshore basin via a novel wind machine which exhibits negligible swirl and low turbulence intensity in the flow field. Recorded data from the floating wind turbine models included rotor torque and position, tower top and base forces and moments, mooring line tensions, six-axis platform motions and accelerations at key locations on the nacelle, tower, and platform. A large number of tests were performed ranging from simple free-decay tests to complex operating conditions with

  18. Validation of a Climate-Data Record of the "Clear-Sky" Surface Temperature of the Greenland Ice Sheet

    Science.gov (United States)

    Hall, Dorothy K.; Box, Jason E.; Koenig, Lora S.; DiGirolamo, Nicolo E.; Comiso, Josefino C.; Shuman, Christopher A.

    2011-01-01

    Surface temperatures on the Greenland Ice Sheet have been studied on the ground, using automatic weather station (AWS) data from the Greenland-Climate Network (GC-Net), and from analysis of satellite sensor data. Using Advanced Very High Frequency Radiometer (AVHRR) weekly surface temperature maps, warming of the surface of the Greenland Ice Sheet has been documented since 1981. We extended and refined this record using higher-resolution Moderate-Resolution Imaging Spectroradiometer (MODIS) data from March 2000 to the present. We developed a daily and monthly climate-data record (CDR) of the "clear-sky" surface temperature of the Greenland Ice Sheet using an ice-surface temperature (1ST) algorithm developed for use with MODIS data. Validation of this CDR is ongoing. MODIS Terra swath data are projected onto a polar stereographic grid at 6.25-km resolution to develop binary, gridded daily and mean-monthly 1ST maps. Each monthly map also has a color-coded image map that is available to download. Also included with the monthly maps is an accompanying map showing number of days in the month that were used to calculate the mean-monthly 1ST. This is important because no 1ST decision is made by the algorithm for cells that are considered cloudy by the internal cloud mask, so a sufficient number of days must be available to produce a mean 1ST for each grid cell. Validation of the CDR consists of several facets: 1) comparisons between ISTs and in-situ measurements; 2) comparisons between ISTs and AWS data; and 3) comparisons of ISTs with surface temperatures derived from other satellite instruments such as the Thermal Emission and Reflection Radiometer (ASTER) and Enhanced Thematic Mapper Plus (ETM+). Previous work shows that Terra MODIS ISTs are about 3 C lower than in-situ temperatures measured at Summit Camp, during the winter of 2008-09 under clear skies. In this work we begin to compare surface temperatures derived from AWS data with ISTs from the MODIS CDR.

  19. Data Validation Package - July 2016 Groundwater Sampling at the Gunnison, Colorado, Disposal Site

    Energy Technology Data Exchange (ETDEWEB)

    Linard, Joshua [USDOE Office of Legacy Management, Washington, DC (United States); Campbell, Sam [Navarro Research and Engineering, Inc., Las Vegas, NV (United States)

    2016-10-25

    Groundwater sampling at the Gunnison, Colorado, Disposal Site is conducted every 5 years to monitor disposal cell performance. During this event, samples were collected from eight monitoring wells as specified in the 1997 Long-Term Surveillance Plan for the Gunnison, Colorado, Disposal Site. Sampling and analyses were conducted as specified in the Sampling and Analysis Plan for US Department of Energy Office of Legacy Management Sites (LMS/PRO/S04351, continually updated, http://energy.gov/lm/downloads/sampling-and­ analysis-plan-us-department-energy-office-legacy-management-sites). Planned monitoring locations are shown in Attachment 1, Sampling and Analysis Work Order. A duplicate sample was collected from location 0723. Water levels were measured at all monitoring wells that were sampled and seven additional wells. The analytical data and associated qualifiers can be viewed in environmental database reports and are also available for viewing with dynamic mapping via the GEMS (Geospatial Environmental Mapping System) website at http://gems.lm.doe.gov/#. No issues were identified during the data validation process that require additional action or follow-up.

  20. CellProfiler Tracer: exploring and validating high-throughput, time-lapse microscopy image data.

    Science.gov (United States)

    Bray, Mark-Anthony; Carpenter, Anne E

    2015-11-04

    Time-lapse analysis of cellular images is an important and growing need in biology. Algorithms for cell tracking are widely available; what researchers have been missing is a single open-source software package to visualize standard tracking output (from software like CellProfiler) in a way that allows convenient assessment of track quality, especially for researchers tuning tracking parameters for high-content time-lapse experiments. This makes quality assessment and algorithm adjustment a substantial challenge, particularly when dealing with hundreds of time-lapse movies collected in a high-throughput manner. We present CellProfiler Tracer, a free and open-source tool that complements the object tracking functionality of the CellProfiler biological image analysis package. Tracer allows multi-parametric morphological data to be visualized on object tracks, providing visualizations that have already been validated within the scientific community for time-lapse experiments, and combining them with simple graph-based measures for highlighting possible tracking artifacts. CellProfiler Tracer is a useful, free tool for inspection and quality control of object tracking data, available from http://www.cellprofiler.org/tracer/.

  1. Predicting dementia risk in primary care: development and validation of the Dementia Risk Score using routinely collected data.

    Science.gov (United States)

    Walters, K; Hardoon, S; Petersen, I; Iliffe, S; Omar, R Z; Nazareth, I; Rait, G

    2016-01-21

    Existing dementia risk scores require collection of additional data from patients, limiting their use in practice. Routinely collected healthcare data have the potential to assess dementia risk without the need to collect further information. Our objective was to develop and validate a 5-year dementia risk score derived from primary healthcare data. We used data from general practices in The Health Improvement Network (THIN) database from across the UK, randomly selecting 377 practices for a development cohort and identifying 930,395 patients aged 60-95 years without a recording of dementia, cognitive impairment or memory symptoms at baseline. We developed risk algorithm models for two age groups (60-79 and 80-95 years). An external validation was conducted by validating the model on a separate cohort of 264,224 patients from 95 randomly chosen THIN practices that did not contribute to the development cohort. Our main outcome was 5-year risk of first recorded dementia diagnosis. Potential predictors included sociodemographic, cardiovascular, lifestyle and mental health variables. Dementia incidence was 1.88 (95% CI, 1.83-1.93) and 16.53 (95% CI, 16.15-16.92) per 1000 PYAR for those aged 60-79 (n = 6017) and 80-95 years (n = 7104), respectively. Predictors for those aged 60-79 included age, sex, social deprivation, smoking, BMI, heavy alcohol use, anti-hypertensive drugs, diabetes, stroke/TIA, atrial fibrillation, aspirin, depression. The discrimination and calibration of the risk algorithm were good for the 60-79 years model; D statistic 2.03 (95% CI, 1.95-2.11), C index 0.84 (95% CI, 0.81-0.87), and calibration slope 0.98 (95% CI, 0.93-1.02). The algorithm had a high negative predictive value, but lower positive predictive value at most risk thresholds. Discrimination and calibration were poor for the 80-95 years model. Routinely collected data predicts 5-year risk of recorded diagnosis of dementia for those aged 60-79, but not those aged 80+. This

  2. The Validation of a Beta-Binomial Model for Overdispersed Binomial Data.

    Science.gov (United States)

    Kim, Jongphil; Lee, Ji-Hyun

    2017-01-01

    The beta-binomial model has been widely used as an analytically tractable alternative that captures the overdispersion of an intra-correlated, binomial random variable, X . However, the model validation for X has been rarely investigated. As a beta-binomial mass function takes on a few different shapes, the model validation is examined for each of the classified shapes in this paper. Further, the mean square error (MSE) is illustrated for each shape by the maximum likelihood estimator (MLE) based on a beta-binomial model approach and the method of moments estimator (MME) in order to gauge when and how much the MLE is biased.

  3. Validation of MIPAS HNO3 operational data

    Directory of Open Access Journals (Sweden)

    C. D. Boone

    2007-09-01

    Full Text Available Nitric acid (HNO3 is one of the key products that are operationally retrieved by the European Space Agency (ESA from the emission spectra measured by the Michelson Interferometer for Passive Atmospheric Sounding (MIPAS onboard ENVISAT. The product version 4.61/4.62 for the observation period between July 2002 and March 2004 is validated by comparisons with a number of independent observations from ground-based stations, aircraft/balloon campaigns, and satellites. Individual HNO3 profiles of the ESA MIPAS level-2 product show good agreement with those of MIPAS-B and MIPAS-STR (the balloon and aircraft version of MIPAS, respectively, and the balloon-borne infrared spectrometers MkIV and SPIRALE, mostly matching the reference data within the combined instrument error bars. In most cases differences between the correlative measurement pairs are less than 1 ppbv (5–10% throughout the entire altitude range up to about 38 km (~6 hPa, and below 0.5 ppbv (15–20% or more above 30 km (~17 hPa. However, differences up to 4 ppbv compared to MkIV have been found at high latitudes in December 2002 in the presence of polar stratospheric clouds. The degree of consistency is further largely affected by the temporal and spatial coincidence, and differences of 2 ppbv may be observed between 22 and 26 km (~50 and 30 hPa at high latitudes near the vortex boundary, due to large horizontal inhomogeneity of HNO3. Similar features are also observed in the mean differences of the MIPAS ESA HNO3 VMRs with respect to the ground-based FTIR measurements at five stations, aircraft-based SAFIRE-A and ASUR, and the balloon campaign IBEX. The mean relative differences between the MIPAS and FTIR HNO3 partial columns are within ±2%, comparable to the MIPAS systematic error of ~2%. For the vertical profiles, the biases between the MIPAS and FTIR data are generally below 10% in the altitudes of 10 to 30 km. The MIPAS and SAFIRE HNO3 data generally match within their total error

  4. Evidence for validity of five secondary data sources for enumerating retail food outlets in seven American Indian Communities in North Carolina

    Science.gov (United States)

    2012-01-01

    Background Most studies on the local food environment have used secondary sources to describe the food environment, such as government food registries or commercial listings (e.g., Reference USA). Most of the studies exploring evidence for validity of secondary retail food data have used on-site verification and have not conducted analysis by data source (e.g., sensitivity of Reference USA) or by food outlet type (e.g., sensitivity of Reference USA for convenience stores). Few studies have explored the food environment in American Indian communities. To advance the science on measuring the food environment, we conducted direct, on-site observations of a wide range of food outlets in multiple American Indian communities, without a list guiding the field observations, and then compared our findings to several types of secondary data. Methods Food outlets located within seven State Designated Tribal Statistical Areas in North Carolina (NC) were gathered from online Yellow Pages, Reference USA, Dun & Bradstreet, local health departments, and the NC Department of Agriculture and Consumer Services. All TIGER/Line 2009 roads (>1,500 miles) were driven in six of the more rural tribal areas and, for the largest tribe, all roads in two of its cities were driven. Sensitivity, positive predictive value, concordance, and kappa statistics were calculated to compare secondary data sources to primary data. Results 699 food outlets were identified during primary data collection. Match rate for primary data and secondary data differed by type of food outlet observed, with the highest match rates found for grocery stores (97%), general merchandise stores (96%), and restaurants (91%). Reference USA exhibited almost perfect sensitivity (0.89). Local health department data had substantial sensitivity (0.66) and was almost perfect when focusing only on restaurants (0.91). Positive predictive value was substantial for Reference USA (0.67) and moderate for local health department data (0

  5. Production and validation of nuclear data for reactor and fuel cycle applications; Production et validation des donnees nucleaires pour les applications reacteurs et cycle du combustible

    Energy Technology Data Exchange (ETDEWEB)

    Trakas, C [Framatome ANP GmbH NBTT, Erlangen (Germany); Verwaerde, D [Electricite de France EDF, 75 - Paris (France); Toubon, H [Cogema, 78 - Velizy Villacoublay (France); and others

    2002-07-01

    The aim of this technical meeting is the improvement of the existing nuclear data and the production of new data of interest for the upstream and downstream of the fuel cycle (enrichment, fabrication, management, storage, transport, reprocessing), for the industrial reactors, the research reactors and the new reactor concepts (criticality, dimensioning, exploitation), for the instrumentation systems (external and internal sensors), the radioprotection, the residual power, the structures (neutron bombardment effect on vessels, rods etc..), and for the activation of steel structures (Fr, Ni, Co). The expected result is the collection of more reliable and accurate data in a wider spectrum of energies and temperatures thanks to more precise computer codes and measurement techniques. This document brings together the communications presented at this meeting and dealing with: the process of production and validation of nuclear data; the measurement facilities and the big international programs; the users needs and the industrial priorities; the basic nuclear data (BND) needs at Cogema; the expression and evaluation of BND; the evaluation work: the efficient cross-sections; the processing of data and the creation of activation libraries; from the integral measurement to the qualification and the feedback on nuclear data. (J.S.)

  6. The cross-validated AUC for MCP-logistic regression with high-dimensional data.

    Science.gov (United States)

    Jiang, Dingfeng; Huang, Jian; Zhang, Ying

    2013-10-01

    We propose a cross-validated area under the receiving operator characteristic (ROC) curve (CV-AUC) criterion for tuning parameter selection for penalized methods in sparse, high-dimensional logistic regression models. We use this criterion in combination with the minimax concave penalty (MCP) method for variable selection. The CV-AUC criterion is specifically designed for optimizing the classification performance for binary outcome data. To implement the proposed approach, we derive an efficient coordinate descent algorithm to compute the MCP-logistic regression solution surface. Simulation studies are conducted to evaluate the finite sample performance of the proposed method and its comparison with the existing methods including the Akaike information criterion (AIC), Bayesian information criterion (BIC) or Extended BIC (EBIC). The model selected based on the CV-AUC criterion tends to have a larger predictive AUC and smaller classification error than those with tuning parameters selected using the AIC, BIC or EBIC. We illustrate the application of the MCP-logistic regression with the CV-AUC criterion on three microarray datasets from the studies that attempt to identify genes related to cancers. Our simulation studies and data examples demonstrate that the CV-AUC is an attractive method for tuning parameter selection for penalized methods in high-dimensional logistic regression models.

  7. Absorber and regenerator models for liquid desiccant air conditioning systems. Validation and comparison using experimental data

    Energy Technology Data Exchange (ETDEWEB)

    Krause, M.; Heinzen, R.; Jordan, U.; Vajen, K. [Kassel Univ., Inst. of Thermal Engineering, Kassel (Germany); Saman, W.; Halawa, E. [Sustainable Energy Centre, Univ. of South Australia, Mawson Lakes, Adelaide (Australia)

    2008-07-01

    Solar assisted air conditioning systems using liquid desiccants represent a promising option to decrease high summer energy demand caused by electrically driven vapor compression machines. The main components of liquid desiccant systems are absorbers for dehumidifying and cooling of supply air and regenerators for concentrating the desiccant. However, high efficient and validated reliable components are required and the design and operation have to be adjusted to each respective building design, location, and user demand. Simulation tools can help to optimize component and system design. The present paper presents new developed numerical models for absorbers and regenerators, as well as experimental data of a regenerator prototype. The models have been compared with a finite-difference method model as well as experimental data. The data are gained from the regenerator prototype presented and an absorber presented in the literature. (orig.)

  8. Site characterization and validation - drift and borehole fracture data. Stage 1

    International Nuclear Information System (INIS)

    Gale, J.; Straahle, A.

    1988-09-01

    This report describes the procedures used in mapping fractures intersecting seven scanlines along the southern and eastern boundaries of the Site Characterization and Validation (SCV) site and the procedures used in logging and orienting the fractures intersecting the core from six 'boundary boreholes' that were drilled as part of the site characterization program for the SCV site at the 360 m level in the Stripa mine. Scanline mapping along the mine drifts provided a detailed description of the fracture geomentry on the boundaries of the SCV site. The cores from the boundary boreholes have been logged, reconstructed and oriented using a borehole Televiewer and a borehole TV camera and the true fracture orientations calcilated. This has provide additional data on the fracture geometry within the SCV site. The fractire data from both the scanlines and the core logging are presented in the appendices. In addition, an initial analysis has been completed of the fracture orientations, trace lengths and spacings. Based on the variation in fracture orientations over the SCV site, there are two strong subvertical fracture sets or or clusters and a poorly represented sub-horizontal fracture set. An empirical approach, based on the 'blind zone' concept has been used to correct for orientation bias and to predict the orientations of the fracture system that will be intersected by the C and D boreholes in stage III. (33 figs., 6 tabl., 19 refs.)

  9. Experimental validation of decay heat calculation codes and associated nuclear data libraries for fusion energy

    International Nuclear Information System (INIS)

    Maekawa, Fujio; Wada, Masayuki; Ikeda, Yujiro

    2001-01-01

    Validity of decay heat calculations for safety designs of fusion reactors was investigated by using decay heat experimental data on thirty-two fusion reactor relevant materials obtained at the 14-MeV neutron source facility of FNS in JAERI. Calculation codes developed in Japan, ACT4 and CINAC version 4, and nuclear data bases such as JENDL/Act-96, FENDL/A-2.0 and Lib90 were used for the calculation. Although several corrections in algorithms for both the calculation codes were needed, it was shown by comparing calculated results with the experimental data that most of activation cross sections and decay data were adequate. In cases of type 316 stainless steel and copper which were important for ITER, prediction accuracy of decay heat within ±10% was confirmed. However, it was pointed out that there were some problems in parts of data such as improper activation cross sections, e,g., the 92 Mo(n, 2n) 91g Mo reaction in FENDL, and lack of activation cross section data, e.g., the 138 Ba(n, 2n) 137m Ba reaction in JENDL. Modifications of cross section data were recommended for 19 reactions in JENDL and FENDL. It was also pointed out that X-ray and conversion electron energies should be included in decay data. (author)

  10. Experimental validation of decay heat calculation codes and associated nuclear data libraries for fusion energy

    Energy Technology Data Exchange (ETDEWEB)

    Maekawa, Fujio; Wada, Masayuki; Ikeda, Yujiro [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment

    2001-01-01

    Validity of decay heat calculations for safety designs of fusion reactors was investigated by using decay heat experimental data on thirty-two fusion reactor relevant materials obtained at the 14-MeV neutron source facility of FNS in JAERI. Calculation codes developed in Japan, ACT4 and CINAC version 4, and nuclear data bases such as JENDL/Act-96, FENDL/A-2.0 and Lib90 were used for the calculation. Although several corrections in algorithms for both the calculation codes were needed, it was shown by comparing calculated results with the experimental data that most of activation cross sections and decay data were adequate. In cases of type 316 stainless steel and copper which were important for ITER, prediction accuracy of decay heat within {+-}10% was confirmed. However, it was pointed out that there were some problems in parts of data such as improper activation cross sections, e,g., the {sup 92}Mo(n, 2n){sup 91g}Mo reaction in FENDL, and lack of activation cross section data, e.g., the {sup 138}Ba(n, 2n){sup 137m}Ba reaction in JENDL. Modifications of cross section data were recommended for 19 reactions in JENDL and FENDL. It was also pointed out that X-ray and conversion electron energies should be included in decay data. (author)

  11. DBCG hypo trial validation of radiotherapy parameters from a national data bank versus manual reporting.

    Science.gov (United States)

    Brink, Carsten; Lorenzen, Ebbe L; Krogh, Simon Long; Westberg, Jonas; Berg, Martin; Jensen, Ingelise; Thomsen, Mette Skovhus; Yates, Esben Svitzer; Offersen, Birgitte Vrou

    2018-01-01

    The current study evaluates the data quality achievable using a national data bank for reporting radiotherapy parameters relative to the classical manual reporting method of selected parameters. The data comparison is based on 1522 Danish patients of the DBCG hypo trial with data stored in the Danish national radiotherapy data bank. In line with standard DBCG trial practice selected parameters were also reported manually to the DBCG database. Categorical variables are compared using contingency tables, and comparison of continuous parameters is presented in scatter plots. For categorical variables 25 differences between the data bank and manual values were located. Of these 23 were related to mistakes in the manual reported value whilst the remaining two were a wrong classification in the data bank. The wrong classification in the data bank was related to lack of dose information, since the two patients had been treated with an electron boost based on a manual calculation, thus data was not exported to the data bank, and this was not detected prior to comparison with the manual data. For a few database fields in the manual data an ambiguity of the parameter definition of the specific field is seen in the data. This was not the case for the data bank, which extract all data consistently. In terms of data quality the data bank is superior to manually reported values. However, there is a need to allocate resources for checking the validity of the available data as well as ensuring that all relevant data is present. The data bank contains more detailed information, and thus facilitates research related to the actual dose distribution in the patients.

  12. Calibration, validation, and sensitivity analysis: What's what

    International Nuclear Information System (INIS)

    Trucano, T.G.; Swiler, L.P.; Igusa, T.; Oberkampf, W.L.; Pilch, M.

    2006-01-01

    One very simple interpretation of calibration is to adjust a set of parameters associated with a computational science and engineering code so that the model agreement is maximized with respect to a set of experimental data. One very simple interpretation of validation is to quantify our belief in the predictive capability of a computational code through comparison with a set of experimental data. Uncertainty in both the data and the code are important and must be mathematically understood to correctly perform both calibration and validation. Sensitivity analysis, being an important methodology in uncertainty analysis, is thus important to both calibration and validation. In this paper, we intend to clarify the language just used and express some opinions on the associated issues. We will endeavor to identify some technical challenges that must be resolved for successful validation of a predictive modeling capability. One of these challenges is a formal description of a 'model discrepancy' term. Another challenge revolves around the general adaptation of abstract learning theory as a formalism that potentially encompasses both calibration and validation in the face of model uncertainty

  13. Implementing an online tool for genome-wide validation of survival-associated biomarkers in ovarian-cancer using microarray data from 1287 patients

    DEFF Research Database (Denmark)

    Győrffy, Balázs; Lánczky, András; Szállási, Zoltán

    2012-01-01

    was set up using gene expression data and survival information of 1287 ovarian cancer patients downloaded from Gene Expression Omnibus and The Cancer Genome Atlas (Affymetrix HG-U133A, HG-U133A 2.0, and HG-U133 Plus 2.0 microarrays). After quality control and normalization, only probes present on all......). A Kaplan–Meier survival plot was generated and significance was computed. The tool can be accessed online at www.kmplot.com/ovar. We used this integrative data analysis tool to validate the prognostic power of 37 biomarkers identified in the literature. Of these, CA125 (MUC16; P=3.7x10–5, hazard ratio (HR...... biomarker validation platform that mines all available microarray data to assess the prognostic power of 22 277 genes in 1287 ovarian cancer patients. We specifically used this tool to evaluate the effect of 37 previously published biomarkers on ovarian cancer prognosis....

  14. Political Representation and Gender Inequalities Testing the Validity of Model Developed for Pakistan using a Data Set of Malaysia

    OpenAIRE

    Najeebullah Khan; Adnan Hussein; Zahid Awan; Bakhtiar Khan

    2012-01-01

    This study measured the impacts of six independent variables (political rights, election system type, political quota, literacy rate, labor force participation and GDP per capita at current price in US dollar) on the dependent variable (percentage of women representation in national legislature) using multiple linear regression models. At a first step we developed and tested the model without of sample data of Pakistan. For model construction and validation ten years data from the year 1999 a...

  15. Estimation and Validation of Land Surface Temperatures from Chinese Second-Generation Polar-Orbit FY-3A VIRR Data

    Directory of Open Access Journals (Sweden)

    Bo-Hui Tang

    2015-03-01

    Full Text Available This work estimated and validated the land surface temperature (LST from thermal-infrared Channels 4 (10.8 µm and 5 (12.0 µm of the Visible and Infrared Radiometer (VIRR onboard the second-generation Chinese polar-orbiting FengYun-3A (FY-3A meteorological satellite. The LST, mean emissivity and atmospheric water vapor content (WVC were divided into several tractable sub-ranges with little overlap to improve the fitting accuracy. The experimental results showed that the root mean square errors (RMSEs were proportional to the viewing zenith angles (VZAs and WVC. The RMSEs were below 1.0 K for VZA sub-ranges less than 30° or for VZA sub-ranges less than 60° and WVC less than 3.5 g/cm2, provided that the land surface emissivities were known. A preliminary validation using independently simulated data showed that the estimated LSTs were quite consistent with the actual inputs, with a maximum RMSE below 1 K for all VZAs. An inter-comparison using the Moderate Resolution Imaging Spectroradiometer (MODIS-derived LST product MOD11_L2 showed that the minimum RMSE was 1.68 K for grass, and the maximum RMSE was 3.59 K for barren or sparsely vegetated surfaces. In situ measurements at the Hailar field site in northeastern China from October, 2013, to September, 2014, were used to validate the proposed method. The result showed that the RMSE between the LSTs calculated from the ground measurements and derived from the VIRR data was 1.82 K.

  16. Color Trails Test: normative data and criterion validity for the greek adult population.

    Science.gov (United States)

    Messinis, Lambros; Malegiannaki, Amaryllis-Chryssi; Christodoulou, Tessa; Panagiotopoulos, Vassillis; Papathanasopoulos, Panagiotis

    2011-06-01

    The Color Trails Test (CTT) was developed as a culturally fair analog of the Trail Making Test. In the present study, normative data for the CTT were developed for the Greek adult population and further the criterion validity of the CTT was examined in two clinical groups (29 Parkinson's disease [PD] and 25 acute stroke patients). The instrument was applied to 163 healthy participants, aged 19-75. Stepwise linear regression analyses revealed a significant influence of age and education level on completion time in both parts of the CTT (increased age and decreased educational level contributed to slower completion times for both parts), whereas gender did not influence time to completion of part B. Further, the CTT appears to discriminate adequately between the performance of PD and acute stroke patients and matched healthy controls.

  17. Validity analysis on merged and averaged data using within and between analysis: focus on effect of qualitative social capital on self-rated health.

    Science.gov (United States)

    Shin, Sang Soo; Shin, Young-Jeon

    2016-01-01

    With an increasing number of studies highlighting regional social capital (SC) as a determinant of health, many studies are using multi-level analysis with merged and averaged scores of community residents' survey responses calculated from community SC data. Sufficient examination is required to validate if the merged and averaged data can represent the community. Therefore, this study analyzes the validity of the selected indicators and their applicability in multi-level analysis. Within and between analysis (WABA) was performed after creating community variables using merged and averaged data of community residents' responses from the 2013 Community Health Survey in Korea, using subjective self-rated health assessment as a dependent variable. Further analysis was performed following the model suggested by WABA result. Both E-test results (1) and WABA results (2) revealed that single-level analysis needs to be performed using qualitative SC variable with cluster mean centering. Through single-level multivariate regression analysis, qualitative SC with cluster mean centering showed positive effect on self-rated health (0.054, panalysis using SC variables without cluster mean centering or multi-level analysis. As modification in qualitative SC was larger within the community than between communities, we validate that relational analysis of individual self-rated health can be performed within the group, using cluster mean centering. Other tests besides the WABA can be performed in the future to confirm the validity of using community variables and their applicability in multi-level analysis.

  18. Development and validation of MIX: comprehensive free software for meta-analysis of causal research data

    Directory of Open Access Journals (Sweden)

    Ikeda Noriaki

    2006-10-01

    Full Text Available Abstract Background Meta-analysis has become a well-known method for synthesis of quantitative data from previously conducted research in applied health sciences. So far, meta-analysis has been particularly useful in evaluating and comparing therapies and in assessing causes of disease. Consequently, the number of software packages that can perform meta-analysis has increased over the years. Unfortunately, it can take a substantial amount of time to get acquainted with some of these programs and most contain little or no interactive educational material. We set out to create and validate an easy-to-use and comprehensive meta-analysis package that would be simple enough programming-wise to remain available as a free download. We specifically aimed at students and researchers who are new to meta-analysis, with important parts of the development oriented towards creating internal interactive tutoring tools and designing features that would facilitate usage of the software as a companion to existing books on meta-analysis. Results We took an unconventional approach and created a program that uses Excel as a calculation and programming platform. The main programming language was Visual Basic, as implemented in Visual Basic 6 and Visual Basic for Applications in Excel 2000 and higher. The development took approximately two years and resulted in the 'MIX' program, which can be downloaded from the program's website free of charge. Next, we set out to validate the MIX output with two major software packages as reference standards, namely STATA (metan, metabias, and metatrim and Comprehensive Meta-Analysis Version 2. Eight meta-analyses that had been published in major journals were used as data sources. All numerical and graphical results from analyses with MIX were identical to their counterparts in STATA and CMA. The MIX program distinguishes itself from most other programs by the extensive graphical output, the click-and-go (Excel interface, and the

  19. Development and validation of a dissolution test for lodenafil carbonate based on in vivo data.

    Science.gov (United States)

    Codevilla, Cristiane Franco; Castilhos, Tamara dos Santos; Cirne, Carolina Araújo; Froehlich, Pedro Eduardo; Bergold, Ana Maria

    2014-04-01

    Lodenafil carbonate is a phosphodiesterase type 5 inhibitor used for the treatment of erectile dysfunction. Currently, there is no dissolution test reported for lodenafil carbonate and this drug is not listed in any pharmacopoeia. The present study focused on the development and validation of a dissolution test for lodenafil carbonate tablets, using a simulated absorption profile based on in vivo data. The appropriate conditions were determined after testing sink conditions. Different conditions as medium, surfactant concentration and rotation speed were evaluated. The percentage of dose absorbed was calculated by deconvolution, using the Wagner-Nelson method. According to the obtained results, the use of 0.1 M HCl + 1.5% SLS (900 mL, at 37 + 0.5 °C) as the dissolution medium, paddles at 25 rpm were considered adequate. The samples were quantified by UV spectroscopy at 295 nm and the validation was performed according to international guidelines. The method showed specificity, linearity, accuracy and precision, within the acceptable range. Kinetics of drug release was better described by the first-order model. The proposed dissolution test can be used for the routine quality control of lodenafil carbonate in tablets.

  20. Nuclear power plant performance monitoring using Data Validation and Reconciliation (DVR). Application at the Brazilian Angra 2 PWR plant

    International Nuclear Information System (INIS)

    Tran Quang, Anh Tho; Closon, Hervé; Chares, Robert; Azola, Edson

    2011-01-01

    Operational decisions related to plant performance monitoring mainly result from raw process measurement analysis. When first signs of sub-optimal behavior occur, traditional methods mainly focus on the observation of selective pieces of information. The main disadvantages of these methods are: Investigation efforts are required to localize the problems, entailing time losses and costs; Validity and reliability of the pieces of information cannot be checked as long as the measurements are observed individually. The problem is not the lack of methods and techniques but rather a lack of reliable and consistent data and information across the entire plant. To overcome drawbacks of traditional methods, measurements are considered as interacting with one another. When related to the other measurements of the plant, the observed information becomes of an interest: its incoherency to the others identifies and localizes a problem. The Data Validation and Reconciliation technology (DVR) is based on an advanced process data coherency treatment. By using all available plant information and by closing the plant heat and mass balances based on rigorous thermodynamics, the method generates: A single set of reliable and most accurate plant process data; Alarms for incoherent measurements, highlighting potential problems; Alarms for equipment and process performance degradation; Alarms for faulty and drifting measurements. The use of the advanced DVR software package VALI offers various benefits as it allows to base operational decisions on reliable and accurate data. (author)

  1. Validation and optimisation of an ICD-10-coded case definition for sepsis using administrative health data

    Science.gov (United States)

    Jolley, Rachel J; Jetté, Nathalie; Sawka, Keri Jo; Diep, Lucy; Goliath, Jade; Roberts, Derek J; Yipp, Bryan G; Doig, Christopher J

    2015-01-01

    Objective Administrative health data are important for health services and outcomes research. We optimised and validated in intensive care unit (ICU) patients an International Classification of Disease (ICD)-coded case definition for sepsis, and compared this with an existing definition. We also assessed the definition's performance in non-ICU (ward) patients. Setting and participants All adults (aged ≥18 years) admitted to a multisystem ICU with general medicosurgical ICU care from one of three tertiary care centres in the Calgary region in Alberta, Canada, between 1 January 2009 and 31 December 2012 were included. Research design Patient medical records were randomly selected and linked to the discharge abstract database. In ICU patients, we validated the Canadian Institute for Health Information (CIHI) ICD-10-CA (Canadian Revision)-coded definition for sepsis and severe sepsis against a reference standard medical chart review, and optimised this algorithm through examination of other conditions apparent in sepsis. Measures Sensitivity (Sn), specificity (Sp), positive predictive value (PPV) and negative predictive value (NPV) were calculated. Results Sepsis was present in 604 of 1001 ICU patients (60.4%). The CIHI ICD-10-CA-coded definition for sepsis had Sn (46.4%), Sp (98.7%), PPV (98.2%) and NPV (54.7%); and for severe sepsis had Sn (47.2%), Sp (97.5%), PPV (95.3%) and NPV (63.2%). The optimised ICD-coded algorithm for sepsis increased Sn by 25.5% and NPV by 11.9% with slightly lowered Sp (85.4%) and PPV (88.2%). For severe sepsis both Sn (65.1%) and NPV (70.1%) increased, while Sp (88.2%) and PPV (85.6%) decreased slightly. Conclusions This study demonstrates that sepsis is highly undercoded in administrative data, thus under-ascertaining the true incidence of sepsis. The optimised ICD-coded definition has a higher validity with higher Sn and should be preferentially considered if used for surveillance purposes. PMID:26700284

  2. Validation of the Organizational Culture Assessment Instrument

    Science.gov (United States)

    Heritage, Brody; Pollock, Clare; Roberts, Lynne

    2014-01-01

    Organizational culture is a commonly studied area in industrial/organizational psychology due to its important role in workplace behaviour, cognitions, and outcomes. Jung et al.'s [1] review of the psychometric properties of organizational culture measurement instruments noted many instruments have limited validation data despite frequent use in both theoretical and applied situations. The Organizational Culture Assessment Instrument (OCAI) has had conflicting data regarding its psychometric properties, particularly regarding its factor structure. Our study examined the factor structure and criterion validity of the OCAI using robust analysis methods on data gathered from 328 (females = 226, males = 102) Australian employees. Confirmatory factor analysis supported a four factor structure of the OCAI for both ideal and current organizational culture perspectives. Current organizational culture data demonstrated expected reciprocally-opposed relationships between three of the four OCAI factors and the outcome variable of job satisfaction but ideal culture data did not, thus indicating possible weak criterion validity when the OCAI is used to assess ideal culture. Based on the mixed evidence regarding the measure's properties, further examination of the factor structure and broad validity of the measure is encouraged. PMID:24667839

  3. Validation of the organizational culture assessment instrument.

    Directory of Open Access Journals (Sweden)

    Brody Heritage

    Full Text Available Organizational culture is a commonly studied area in industrial/organizational psychology due to its important role in workplace behaviour, cognitions, and outcomes. Jung et al.'s [1] review of the psychometric properties of organizational culture measurement instruments noted many instruments have limited validation data despite frequent use in both theoretical and applied situations. The Organizational Culture Assessment Instrument (OCAI has had conflicting data regarding its psychometric properties, particularly regarding its factor structure. Our study examined the factor structure and criterion validity of the OCAI using robust analysis methods on data gathered from 328 (females = 226, males = 102 Australian employees. Confirmatory factor analysis supported a four factor structure of the OCAI for both ideal and current organizational culture perspectives. Current organizational culture data demonstrated expected reciprocally-opposed relationships between three of the four OCAI factors and the outcome variable of job satisfaction but ideal culture data did not, thus indicating possible weak criterion validity when the OCAI is used to assess ideal culture. Based on the mixed evidence regarding the measure's properties, further examination of the factor structure and broad validity of the measure is encouraged.

  4. Ocean Optics Protocols for Satellite Ocean Color Sensor Validation, Revision 4, Volume IV: Inherent Optical Properties: Instruments, Characterizations, Field Measurements and Data Analysis Protocols

    Science.gov (United States)

    Mueller, J. L.; Fargion, G. S.; McClain, C. R. (Editor); Pegau, S.; Zanefeld, J. R. V.; Mitchell, B. G.; Kahru, M.; Wieland, J.; Stramska, M.

    2003-01-01

    This document stipulates protocols for measuring bio-optical and radiometric data for the Sensor Intercomparision and Merger for Biological and Interdisciplinary Oceanic Studies (SIMBIOS) Project activities and algorithm development. The document is organized into 6 separate volumes as Ocean Optics Protocols for Satellite Ocean Color Sensor Validation, Revision 4. Volume I: Introduction, Background, and Conventions; Volume II: Instrument Specifications, Characterization and Calibration; Volume III: Radiometric Measurements and Data Analysis Methods; Volume IV: Inherent Optical Properties: Instruments, Characterization, Field Measurements and Data Analysis Protocols; Volume V: Biogeochemical and Bio-Optical Measurements and Data Analysis Methods; Volume VI: Special Topics in Ocean Optics Protocols and Appendices. The earlier version of Ocean Optics Protocols for Satellite Ocean Color Sensor Validation, Revision 3 is entirely superseded by the six volumes of Revision 4 listed above.

  5. Validity of Danish Breast Cancer Group (DBCG) registry data used in the predictors of breast cancer recurrence (ProBeCaRe) premenopausal breast cancer cohort study

    DEFF Research Database (Denmark)

    Cronin-Fenton, Deirdre P; Kjærsgaard, Anders; Ahern, Thomas P

    2017-01-01

    BACKGROUND: Validation studies of the Danish Breast Cancer Group (DBCG) registry show good agreement with medical records for adjuvant treatment data, but inconsistent recurrence information. No studies have validated changes in menopausal status or endocrine therapy during follow-up. In a longit...

  6. A new dataset validation system for the Planetary Science Archive

    Science.gov (United States)

    Manaud, N.; Zender, J.; Heather, D.; Martinez, S.

    2007-08-01

    The Planetary Science Archive is the official archive for the Mars Express mission. It has received its first data by the end of 2004. These data are delivered by the PI teams to the PSA team as datasets, which are formatted conform to the Planetary Data System (PDS). The PI teams are responsible for analyzing and calibrating the instrument data as well as the production of reduced and calibrated data. They are also responsible of the scientific validation of these data. ESA is responsible of the long-term data archiving and distribution to the scientific community and must ensure, in this regard, that all archived products meet quality. To do so, an archive peer-review is used to control the quality of the Mars Express science data archiving process. However a full validation of its content is missing. An independent review board recently recommended that the completeness of the archive as well as the consistency of the delivered data should be validated following well-defined procedures. A new validation software tool is being developed to complete the overall data quality control system functionality. This new tool aims to improve the quality of data and services provided to the scientific community through the PSA, and shall allow to track anomalies in and to control the completeness of datasets. It shall ensure that the PSA end-users: (1) can rely on the result of their queries, (2) will get data products that are suitable for scientific analysis, (3) can find all science data acquired during a mission. We defined dataset validation as the verification and assessment process to check the dataset content against pre-defined top-level criteria, which represent the general characteristics of good quality datasets. The dataset content that is checked includes the data and all types of information that are essential in the process of deriving scientific results and those interfacing with the PSA database. The validation software tool is a multi-mission tool that

  7. Validation of putative reference genes for normalization of Q-RT-PCR data from paraffin-embedded lymphoid tissue

    DEFF Research Database (Denmark)

    Green, Tina Marie; de Stricker, Karin; Møller, Michael Boe

    2009-01-01

    Normalization of quantitative reverse transcription-PCR (Q-RT-PCR) data to appropriate tissue-specific reference genes is an essential part of interpreting the results. This study aimed to determine the most appropriate reference genes for normalizing gene expressions in lymphatic tissue...... was 0.93 (Pnormalization with the appropriate reference genes. Thus, we show that formalin-fixed, paraffin-embedded lymphoid samples are suitable for Q-RT-PCR when using thoroughly validated reference genes....

  8. Development and Validation of a Data-Based Food Frequency Questionnaire for Adults in Eastern Rural Area of Rwanda

    Directory of Open Access Journals (Sweden)

    Ayumi Yanagisawa

    2016-01-01

    Full Text Available This study aimed to develop and evaluate the validity of a food frequency questionnaire (FFQ for rural Rwandans. Since our FFQ was developed to assess malnutrition, it measured energy, protein, vitamin A, and iron intakes only. We collected 260 weighed food records (WFRs from a total of 162 Rwandans. Based on the WFR data, we developed a tentative FFQ and examined the food list by percent contribution to energy and nutrient intakes. To assess the validity, nutrient intakes estimated from the FFQ were compared with those calculated from three-day WFRs by correlation coefficient and cross-classification for 17 adults. Cumulative contributions of the 18-item FFQ to the total intakes of energy and nutrients reached nearly 100%. Crude and energy-adjusted correlation coefficients ranged from -0.09 (vitamin A to 0.58 (protein and from -0.19 (vitamin A to 0.68 (iron, respectively. About 50%-60% of the participants were classified into the same tertile. Our FFQ provided acceptable validity for energy and iron intakes and could rank Rwandan adults in eastern rural area correctly according to their energy and iron intakes.

  9. Assessing the Validity of Automated Webcrawlers as Data Collection Tools to Investigate Online Child Sexual Exploitation.

    Science.gov (United States)

    Westlake, Bryce; Bouchard, Martin; Frank, Richard

    2017-10-01

    The distribution of child sexual exploitation (CE) material has been aided by the growth of the Internet. The graphic nature and prevalence of the material has made researching and combating difficult. Although used to study online CE distribution, automated data collection tools (e.g., webcrawlers) have yet to be shown effective at targeting only relevant data. Using CE-related image and keyword criteria, we compare networks starting from CE websites to those from similar non-CE sexuality websites and dissimilar sports websites. Our results provide evidence that (a) webcrawlers have the potential to provide valid CE data, if the appropriate criterion is selected; (b) CE distribution is still heavily image-based suggesting images as an effective criterion; (c) CE-seeded networks are more hub-based and differ from non-CE-seeded networks on several website characteristics. Recommendations for improvements to reliable criteria selection are discussed.

  10. The validation and assessment of machine learning: a game of prediction from high-dimensional data.

    Directory of Open Access Journals (Sweden)

    Tune H Pers

    Full Text Available In applied statistics, tools from machine learning are popular for analyzing complex and high-dimensional data. However, few theoretical results are available that could guide to the appropriate machine learning tool in a new application. Initial development of an overall strategy thus often implies that multiple methods are tested and compared on the same set of data. This is particularly difficult in situations that are prone to over-fitting where the number of subjects is low compared to the number of potential predictors. The article presents a game which provides some grounds for conducting a fair model comparison. Each player selects a modeling strategy for predicting individual response from potential predictors. A strictly proper scoring rule, bootstrap cross-validation, and a set of rules are used to make the results obtained with different strategies comparable. To illustrate the ideas, the game is applied to data from the Nugenob Study where the aim is to predict the fat oxidation capacity based on conventional factors and high-dimensional metabolomics data. Three players have chosen to use support vector machines, LASSO, and random forests, respectively.

  11. Validation and Analysis of SRTM and VCL Data Over Tropical Volcanoes

    Science.gov (United States)

    Mouginis-Mark, Peter J.

    2004-01-01

    The focus of our investigation was on the application of digital topographic data in conducting first-order volcanological and structural studies of tropical volcanoes, focusing on the Java, the Philippines and the Galapagos Islands. Kilauea volcano, Hawaii, served as our test site for SRTM data validation. Volcanoes in humid tropical environments are frequently cloud covered, typically densely vegetated and erode rapidly, so that it was expected that new insights into the styles of eruption of these volcanoes could be obtained from analysis of topographic data. For instance, in certain parts of the world, such as Indonesia, even the regional structural context of volcanic centers is poorly known, and the distribution of volcanic products (e.g., lava flows, pyroclastic flows, and lahars) are not well mapped. SRTM and Vegetation Canopy Lidar (VCL) data were expected to provide new information on these volcanoes. Due to the cancellation of the VCL mission, we did not conduct any lidar studies during the duration of this project. Digital elevation models (DEMs) such as those collected by SRTM provide quantitative information about the time-integrated typical activity on a volcano and allow an assessment of the spatial and temporal contributions of various constructional and destructional processes to each volcano's present morphology. For basaltic volcanoes, P_c?w!m-d and Garbed (2000) have shown that gradual slopes (less than 5 deg.) occur where lava and tephra pond within calderas or in the saddles between adjacent volcanoes, as well as where lava deltas coalesce to form coastal plains. Vent concentration zones (axes of rift zones) have slopes ranging from 10 deg. to 12 deg. Differential vertical growth rates between vent concentration zones and adjacent mostly-lava flanks produce steep constructional slopes up to 40". The steepest slopes (locally approaching 90 deg.) are produced by fluvial erosion, caldera collapse, faulting, and catastrophic avalanches, all of

  12. Translation, data quality, reliability, validity and responsiveness of the Norwegian version of the Effective Musculoskeletal Consumer Scale (EC-17

    Directory of Open Access Journals (Sweden)

    Kristjansson Elizabeth

    2010-01-01

    Full Text Available Abstract Background The Effective Musculoskeletal Consumer Scale (EC-17 is a self-administered questionnaire for evaluating self-management interventions that empower and educate people with rheumatic conditions. The aim of the study was to translate and evaluate the Norwegian version of EC-17 against the necessary criteria for a patient-reported outcome measure, including responsiveness to change. Methods Data quality, reliability, validity and responsiveness were assessed in two groups. One group comprising 103 patients received a questionnaire before and at the end of a self-management programme. The second group comprising 96 patients' received the questionnaire two weeks before and on arrival of the program. Internal consistency and test-retest reliability were assessed. Construct validity was assessed through comparisons with the Brief Approach/Avoidance Coping Questionnaire, (BACQ, the Emotional Approach Coping Scale (EAC and the General Health Questionnaire (GHQ-20. Responsiveness was assessed with the Standardised Response Mean (SRM. Results Respondents included 66 (64% and 52 (54% patients from the first and second groups respectively. Levels of missing data were low for all items. There was good evidence for unidimensionality, item-total correlations ranged from 0.59 to 0.82 and Cronbach's Alpha and test-retest correlations were over 0.90. As hypothesised EC-17 scores had statistically significant low to moderate correlations with the BACQ, EAC and GHQ-20 in the range 0.26 to 0.42. Following the self-management program, EC-17 scores showed a significant improvement with an SRM of 0.48. Conclusion The Norwegian version of the EC-17 has evidence for data quality, internal consistency and test-retest reliability, construct validity and responsiveness to change. The EC-17 seems promising as an outcome measure for evaluating self-management interventions for people with rheumatic conditions, but further studies are needed.

  13. Translation, data quality, reliability, validity and responsiveness of the Norwegian version of the Effective Musculoskeletal Consumer Scale (EC-17).

    Science.gov (United States)

    Hamnes, Bente; Garratt, Andrew; Kjeken, Ingvild; Kristjansson, Elizabeth; Hagen, Kåre B

    2010-01-29

    The Effective Musculoskeletal Consumer Scale (EC-17) is a self-administered questionnaire for evaluating self-management interventions that empower and educate people with rheumatic conditions. The aim of the study was to translate and evaluate the Norwegian version of EC-17 against the necessary criteria for a patient-reported outcome measure, including responsiveness to change. Data quality, reliability, validity and responsiveness were assessed in two groups. One group comprising 103 patients received a questionnaire before and at the end of a self-management programme. The second group comprising 96 patients' received the questionnaire two weeks before and on arrival of the program. Internal consistency and test-retest reliability were assessed. Construct validity was assessed through comparisons with the Brief Approach/Avoidance Coping Questionnaire, (BACQ), the Emotional Approach Coping Scale (EAC) and the General Health Questionnaire (GHQ-20). Responsiveness was assessed with the Standardised Response Mean (SRM). Respondents included 66 (64%) and 52 (54%) patients from the first and second groups respectively. Levels of missing data were low for all items. There was good evidence for unidimensionality, item-total correlations ranged from 0.59 to 0.82 and Cronbach's Alpha and test-retest correlations were over 0.90. As hypothesised EC-17 scores had statistically significant low to moderate correlations with the BACQ, EAC and GHQ-20 in the range 0.26 to 0.42. Following the self-management program, EC-17 scores showed a significant improvement with an SRM of 0.48. The Norwegian version of the EC-17 has evidence for data quality, internal consistency and test-retest reliability, construct validity and responsiveness to change. The EC-17 seems promising as an outcome measure for evaluating self-management interventions for people with rheumatic conditions, but further studies are needed.

  14. 3D structure tensor analysis of light microscopy data for validating diffusion MRI.

    Science.gov (United States)

    Khan, Ahmad Raza; Cornea, Anda; Leigland, Lindsey A; Kohama, Steven G; Jespersen, Sune Nørhøj; Kroenke, Christopher D

    2015-05-01

    Diffusion magnetic resonance imaging (d-MRI) is a powerful non-invasive and non-destructive technique for characterizing brain tissue on the microscopic scale. However, the lack of validation of d-MRI by independent experimental means poses an obstacle to accurate interpretation of data acquired using this method. Recently, structure tensor analysis has been applied to light microscopy images, and this technique holds promise to be a powerful validation strategy for d-MRI. Advantages of this approach include its similarity to d-MRI in terms of averaging the effects of a large number of cellular structures, and its simplicity, which enables it to be implemented in a high-throughput manner. However, a drawback of previous implementations of this technique arises from it being restricted to 2D. As a result, structure tensor analyses have been limited to tissue sectioned in a direction orthogonal to the direction of interest. Here we describe the analytical framework for extending structure tensor analysis to 3D, and utilize the results to analyze serial image "stacks" acquired with confocal microscopy of rhesus macaque hippocampal tissue. Implementation of 3D structure tensor procedures requires removal of sources of anisotropy introduced in tissue preparation and confocal imaging. This is accomplished with image processing steps to mitigate the effects of anisotropic tissue shrinkage, and the effects of anisotropy in the point spread function (PSF). In order to address the latter confound, we describe procedures for measuring the dependence of PSF anisotropy on distance from the microscope objective within tissue. Prior to microscopy, ex vivo d-MRI measurements performed on the hippocampal tissue revealed three regions of tissue with mutually orthogonal directions of least restricted diffusion that correspond to CA1, alveus and inferior longitudinal fasciculus. We demonstrate the ability of 3D structure tensor analysis to identify structure tensor orientations that

  15. Application of validation data for assessing spatial interpolation methods for 8-h ozone or other sparsely monitored constituents

    International Nuclear Information System (INIS)

    Joseph, John; Sharif, Hatim O.; Sunil, Thankam; Alamgir, Hasanat

    2013-01-01

    The adverse health effects of high concentrations of ground-level ozone are well-known, but estimating exposure is difficult due to the sparseness of urban monitoring networks. This sparseness discourages the reservation of a portion of the monitoring stations for validation of interpolation techniques precisely when the risk of overfitting is greatest. In this study, we test a variety of simple spatial interpolation techniques for 8-h ozone with thousands of randomly selected subsets of data from two urban areas with monitoring stations sufficiently numerous to allow for true validation. Results indicate that ordinary kriging with only the range parameter calibrated in an exponential variogram is the generally superior method, and yields reliable confidence intervals. Sparse data sets may contain sufficient information for calibration of the range parameter even if the Moran I p-value is close to unity. R script is made available to apply the methodology to other sparsely monitored constituents. -- Highlights: •Spatial interpolation methods were tested for thousands of sparse ozone data sets. •A particular single-parameter ordinary kriging was found to be generally superior. •A Moran I p-value in the training set is not helpful in selecting the method. •The sum of the squares of the residuals is helpful in selecting the method. •R script is available for application to other sites and constituents. -- Spatial interpolation methods were compared for thousands of subsets of data for 8-h ozone using R script applicable to other constituents as well, and available from the authors

  16. Systematic review of validated case definitions for diabetes in ICD-9-coded and ICD-10-coded data in adult populations.

    Science.gov (United States)

    Khokhar, Bushra; Jette, Nathalie; Metcalfe, Amy; Cunningham, Ceara Tess; Quan, Hude; Kaplan, Gilaad G; Butalia, Sonia; Rabi, Doreen

    2016-08-05

    With steady increases in 'big data' and data analytics over the past two decades, administrative health databases have become more accessible and are now used regularly for diabetes surveillance. The objective of this study is to systematically review validated International Classification of Diseases (ICD)-based case definitions for diabetes in the adult population. Electronic databases, MEDLINE and Embase, were searched for validation studies where an administrative case definition (using ICD codes) for diabetes in adults was validated against a reference and statistical measures of the performance reported. The search yielded 2895 abstracts, and of the 193 potentially relevant studies, 16 met criteria. Diabetes definition for adults varied by data source, including physician claims (sensitivity ranged from 26.9% to 97%, specificity ranged from 94.3% to 99.4%, positive predictive value (PPV) ranged from 71.4% to 96.2%, negative predictive value (NPV) ranged from 95% to 99.6% and κ ranged from 0.8 to 0.9), hospital discharge data (sensitivity ranged from 59.1% to 92.6%, specificity ranged from 95.5% to 99%, PPV ranged from 62.5% to 96%, NPV ranged from 90.8% to 99% and κ ranged from 0.6 to 0.9) and a combination of both (sensitivity ranged from 57% to 95.6%, specificity ranged from 88% to 98.5%, PPV ranged from 54% to 80%, NPV ranged from 98% to 99.6% and κ ranged from 0.7 to 0.8). Overall, administrative health databases are useful for undertaking diabetes surveillance, but an awareness of the variation in performance being affected by case definition is essential. The performance characteristics of these case definitions depend on the variations in the definition of primary diagnosis in ICD-coded discharge data and/or the methodology adopted by the healthcare facility to extract information from patient records. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/

  17. SMAP RADAR Calibration and Validation

    Science.gov (United States)

    West, R. D.; Jaruwatanadilok, S.; Chaubel, M. J.; Spencer, M.; Chan, S. F.; Chen, C. W.; Fore, A.

    2015-12-01

    The Soil Moisture Active Passive (SMAP) mission launched on Jan 31, 2015. The mission employs L-band radar and radiometer measurements to estimate soil moisture with 4% volumetric accuracy at a resolution of 10 km, and freeze-thaw state at a resolution of 1-3 km. Immediately following launch, there was a three month instrument checkout period, followed by six months of level 1 (L1) calibration and validation. In this presentation, we will discuss the calibration and validation activities and results for the L1 radar data. Early SMAP radar data were used to check commanded timing parameters, and to work out issues in the low- and high-resolution radar processors. From April 3-13 the radar collected receive only mode data to conduct a survey of RFI sources. Analysis of the RFI environment led to a preferred operating frequency. The RFI survey data were also used to validate noise subtraction and scaling operations in the radar processors. Normal radar operations resumed on April 13. All radar data were examined closely for image quality and calibration issues which led to improvements in the radar data products for the beta release at the end of July. Radar data were used to determine and correct for small biases in the reported spacecraft attitude. Geo-location was validated against coastline positions and the known positions of corner reflectors. Residual errors at the time of the beta release are about 350 m. Intra-swath biases in the high-resolution backscatter images are reduced to less than 0.3 dB for all polarizations. Radiometric cross-calibration with Aquarius was performed using areas of the Amazon rain forest. Cross-calibration was also examined using ocean data from the low-resolution processor and comparing with the Aquarius wind model function. Using all a-priori calibration constants provided good results with co-polarized measurements matching to better than 1 dB, and cross-polarized measurements matching to about 1 dB in the beta release. During the

  18. FAST Model Calibration and Validation of the OC5-DeepCwind Floating Offshore Wind System Against Wave Tank Test Data

    Energy Technology Data Exchange (ETDEWEB)

    Wendt, Fabian F [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Robertson, Amy N [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Jonkman, Jason [National Renewable Energy Laboratory (NREL), Golden, CO (United States)

    2017-06-03

    During the course of the Offshore Code Comparison Collaboration, Continued, with Correlation (OC5) project, which focused on the validation of numerical methods through comparison against tank test data, the authors created a numerical FAST model of the 1:50-scale DeepCwind semisubmersible system that was tested at the Maritime Research Institute Netherlands ocean basin in 2013. This paper discusses several model calibration studies that were conducted to identify model adjustments that improve the agreement between the numerical simulations and the experimental test data. These calibration studies cover wind-field-specific parameters (coherence, turbulence), hydrodynamic and aerodynamic modeling approaches, as well as rotor model (blade-pitch and blade-mass imbalances) and tower model (structural tower damping coefficient) adjustments. These calibration studies were conducted based on relatively simple calibration load cases (wave only/wind only). The agreement between the final FAST model and experimental measurements is then assessed based on more-complex combined wind and wave validation cases.

  19. Initialization of the Euler model MODIS with field data from the 'EPRI plume model validation project'

    International Nuclear Information System (INIS)

    Petersen, G.; Eppel, D.; Lautenschlager, M.; Mueller, A.

    1985-01-01

    The program deck MODIS (''MOment DIStribution'') is designed to be used as operational tool for modelling the dispersion of a point source under general atmospheric conditions. The concentration distribution is determined by calculating its cross-wind moments on a vertical grid oriented in the main wind direction. The model contains a parametrization for horizontal and vertical coefficients based on a second order closure model. The Eulerian time scales, preliminary determined by fitting measured plume cross sections, are confirmed by comparison with data from the EPRI plume model validation project. (orig.) [de

  20. The ALICE Software Release Validation cluster

    International Nuclear Information System (INIS)

    Berzano, D; Krzewicki, M

    2015-01-01

    One of the most important steps of software lifecycle is Quality Assurance: this process comprehends both automatic tests and manual reviews, and all of them must pass successfully before the software is approved for production. Some tests, such as source code static analysis, are executed on a single dedicated service: in High Energy Physics, a full simulation and reconstruction chain on a distributed computing environment, backed with a sample “golden” dataset, is also necessary for the quality sign off. The ALICE experiment uses dedicated and virtualized computing infrastructures for the Release Validation in order not to taint the production environment (i.e. CVMFS and the Grid) with non-validated software and validation jobs: the ALICE Release Validation cluster is a disposable virtual cluster appliance based on CernVM and the Virtual Analysis Facility, capable of deploying on demand, and with a single command, a dedicated virtual HTCondor cluster with an automatically scalable number of virtual workers on any cloud supporting the standard EC2 interface. Input and output data are externally stored on EOS, and a dedicated CVMFS service is used to provide the software to be validated. We will show how the Release Validation Cluster deployment and disposal are completely transparent for the Release Manager, who simply triggers the validation from the ALICE build system's web interface. CernVM 3, based entirely on CVMFS, permits to boot any snapshot of the operating system in time: we will show how this allows us to certify each ALICE software release for an exact CernVM snapshot, addressing the problem of Long Term Data Preservation by ensuring a consistent environment for software execution and data reprocessing in the future. (paper)

  1. Screening for postdeployment conditions: development and cross-validation of an embedded validity scale in the neurobehavioral symptom inventory.

    Science.gov (United States)

    Vanderploeg, Rodney D; Cooper, Douglas B; Belanger, Heather G; Donnell, Alison J; Kennedy, Jan E; Hopewell, Clifford A; Scott, Steven G

    2014-01-01

    To develop and cross-validate internal validity scales for the Neurobehavioral Symptom Inventory (NSI). Four existing data sets were used: (1) outpatient clinical traumatic brain injury (TBI)/neurorehabilitation database from a military site (n = 403), (2) National Department of Veterans Affairs TBI evaluation database (n = 48 175), (3) Florida National Guard nonclinical TBI survey database (n = 3098), and (4) a cross-validation outpatient clinical TBI/neurorehabilitation database combined across 2 military medical centers (n = 206). Secondary analysis of existing cohort data to develop (study 1) and cross-validate (study 2) internal validity scales for the NSI. The NSI, Mild Brain Injury Atypical Symptoms, and Personality Assessment Inventory scores. Study 1: Three NSI validity scales were developed, composed of 5 unusual items (Negative Impression Management [NIM5]), 6 low-frequency items (LOW6), and the combination of 10 nonoverlapping items (Validity-10). Cut scores maximizing sensitivity and specificity on these measures were determined, using a Mild Brain Injury Atypical Symptoms score of 8 or more as the criterion for invalidity. Study 2: The same validity scale cut scores again resulted in the highest classification accuracy and optimal balance between sensitivity and specificity in the cross-validation sample, using a Personality Assessment Inventory Negative Impression Management scale with a T score of 75 or higher as the criterion for invalidity. The NSI is widely used in the Department of Defense and Veterans Affairs as a symptom-severity assessment following TBI, but is subject to symptom overreporting or exaggeration. This study developed embedded NSI validity scales to facilitate the detection of invalid response styles. The NSI Validity-10 scale appears to hold considerable promise for validity assessment when the NSI is used as a population-screening tool.

  2. Quality assessment of the Ozone_cci Climate Research Data Package (release 2017 – Part 1: Ground-based validation of total ozone column data products

    Directory of Open Access Journals (Sweden)

    K. Garane

    2018-03-01

    Full Text Available The GOME-type Total Ozone Essential Climate Variable (GTO-ECV is a level-3 data record, which combines individual sensor products into one single cohesive record covering the 22-year period from 1995 to 2016, generated in the frame of the European Space Agency's Climate Change Initiative Phase II. It is based on level-2 total ozone data produced by the GODFIT (GOME-type Direct FITting v4 algorithm as applied to the GOME/ERS-2, OMI/Aura, SCIAMACHY/Envisat and GOME-2/Metop-A and Metop-B observations. In this paper we examine whether GTO-ECV meets the specific requirements set by the international climate–chemistry modelling community for decadal stability long-term and short-term accuracy. In the following, we present the validation of the 2017 release of the Climate Research Data Package Total Ozone Column (CRDP TOC at both level 2 and level 3. The inter-sensor consistency of the individual level-2 data sets has mean differences generally within 0.5 % at moderate latitudes (±50°, whereas the level-3 data sets show mean differences with respect to the OMI reference data record that span between −0.2 ± 0.9 % (for GOME-2B and 1.0 ± 1.4 % (for SCIAMACHY. Very similar findings are reported for the level-2 validation against independent ground-based TOC observations reported by Brewer, Dobson and SAOZ instruments: the mean bias between GODFIT v4 satellite TOC and the ground instrument is well within 1.0 ± 1.0 % for all sensors, the drift per decade spans between −0.5 % and 1.0 ± 1.0 % depending on the sensor, and the peak-to-peak seasonality of the differences ranges from ∼ 1 % for GOME and OMI to  ∼ 2 % for SCIAMACHY. For the level-3 validation, our first goal was to show that the level-3 CRDP produces findings consistent with the level-2 individual sensor comparisons. We show a very good agreement with 0.5 to 2 % peak-to-peak amplitude for the monthly mean difference time series and a

  3. Ensemble Kalman filter regularization using leave-one-out data cross-validation

    KAUST Repository

    Rayo Schiappacasse, Lautaro Jeró nimo; Hoteit, Ibrahim

    2012-01-01

    In this work, the classical leave-one-out cross-validation method for selecting a regularization parameter for the Tikhonov problem is implemented within the EnKF framework. Following the original concept, the regularization parameter is selected

  4. Selected ICAR Data from the SAPA-Project: Development and Initial Validation of a Public-Domain Measure

    Directory of Open Access Journals (Sweden)

    David M. Condon

    2016-01-01

    Full Text Available These data were collected during the initial evaluation of the International Cognitive Ability Resource (ICAR project. ICAR is an international collaborative effort to develop open-source public-domain tools for cognitive ability assessment, including tools that can be administered in non-proctored environments (e.g., online administration and those which are based on automatic item generation algorithms. These data provide initial validation of the first four ICAR item types as reported in Condon & Revelle [1]. The 4 item types contain a total of 60 items: 9 Letter and Number Series items, 11 Matrix Reasoning items, 16 Verbal Reasoning items and 24 Three-dimensional Rotation items. Approximately 97,000 individuals were administered random subsets of these 60 items using the Synthetic Aperture Personality Assessment method between August 18, 2010 and May 20, 2013. The data are available in rdata and csv formats and are accompanied by documentation stored as a text file. Re-use potential includes a wide range of structural and item-level analyses.

  5. Creation of a Human Secretome: A Novel Composite Library of Human Secreted Proteins: Validation Using Ovarian Cancer Gene Expression Data and a Virtual Secretome Array.

    Science.gov (United States)

    Vathipadiekal, Vinod; Wang, Victoria; Wei, Wei; Waldron, Levi; Drapkin, Ronny; Gillette, Michael; Skates, Steven; Birrer, Michael

    2015-11-01

    To generate a comprehensive "Secretome" of proteins potentially found in the blood and derive a virtual Affymetrix array. To validate the utility of this database for the discovery of novel serum-based biomarkers using ovarian cancer transcriptomic data. The secretome was constructed by aggregating the data from databases of known secreted proteins, transmembrane or membrane proteins, signal peptides, G-protein coupled receptors, or proteins existing in the extracellular region, and the virtual array was generated by mapping them to Affymetrix probeset identifiers. Whole-genome microarray data from ovarian cancer, normal ovarian surface epithelium, and fallopian tube epithelium were used to identify transcripts upregulated in ovarian cancer. We established the secretome from eight public databases and a virtual array consisting of 16,521 Affymetrix U133 Plus 2.0 probesets. Using ovarian cancer transcriptomic data, we identified candidate blood-based biomarkers for ovarian cancer and performed bioinformatic validation by demonstrating rediscovery of known biomarkers including CA125 and HE4. Two novel top biomarkers (FGF18 and GPR172A) were validated in serum samples from an independent patient cohort. We present the secretome, comprising the most comprehensive resource available for protein products that are potentially found in the blood. The associated virtual array can be used to translate gene-expression data into cancer biomarker discovery. A list of blood-based biomarkers for ovarian cancer detection is reported and includes CA125 and HE4. FGF18 and GPR172A were identified and validated by ELISA as being differentially expressed in the serum of ovarian cancer patients compared with controls. ©2015 American Association for Cancer Research.

  6. Validation of the Social Appearance Anxiety Scale: factor, convergent, and divergent validity.

    Science.gov (United States)

    Levinson, Cheri A; Rodebaugh, Thomas L

    2011-09-01

    The Social Appearance Anxiety Scale (SAAS) was created to assess fear of overall appearance evaluation. Initial psychometric work indicated that the measure had a single-factor structure and exhibited excellent internal consistency, test-retest reliability, and convergent validity. In the current study, the authors further examined the factor, convergent, and divergent validity of the SAAS in two samples of undergraduates. In Study 1 (N = 323), the authors tested the factor structure, convergent, and divergent validity of the SAAS with measures of the Big Five personality traits, negative affect, fear of negative evaluation, and social interaction anxiety. In Study 2 (N = 118), participants completed a body evaluation that included measurements of height, weight, and body fat content. The SAAS exhibited excellent convergent and divergent validity with self-report measures (i.e., self-esteem, trait anxiety, ethnic identity, and sympathy), predicted state anxiety experienced during the body evaluation, and predicted body fat content. In both studies, results confirmed a single-factor structure as the best fit to the data. These results lend additional support for the use of the SAAS as a valid measure of social appearance anxiety.

  7. Validation of MIPAS-ENVISAT H2O operational data collected between July 2002 and March 2004

    Directory of Open Access Journals (Sweden)

    G. Wetzel

    2013-06-01

    Full Text Available Water vapour (H2O is one of the operationally retrieved key species of the Michelson Interferometer for Passive Atmospheric Sounding (MIPAS instrument aboard the Environmental Satellite (ENVISAT which was launched into its sun-synchronous orbit on 1 March 2002 and operated until April 2012. Within the MIPAS validation activities, independent observations from balloons, aircraft, satellites, and ground-based stations have been compared to European Space Agency (ESA version 4.61 operational H2O data comprising the time period from July 2002 until March 2004 where MIPAS measured with full spectral resolution. No significant bias in the MIPAS H2O data is seen in the lower stratosphere (above the hygropause between about 15 and 30 km. Differences of H2O quantities observed by MIPAS and the validation instruments are mostly well within the combined total errors in this altitude region. In the upper stratosphere (above about 30 km, a tendency towards a small positive bias (up to about 10% is present in the MIPAS data when compared to its balloon-borne counterpart MIPAS-B, to the satellite instruments HALOE (Halogen Occultation Experiment and ACE-FTS (Atmospheric Chemistry Experiment, Fourier Transform Spectrometer, and to the millimeter-wave airborne sensor AMSOS (Airborne Microwave Stratospheric Observing System. In the mesosphere the situation is unclear due to the occurrence of different biases when comparing HALOE and ACE-FTS data. Pronounced deviations between MIPAS and the correlative instruments occur in the lowermost stratosphere and upper troposphere, a region where retrievals of H2O are most challenging. Altogether it can be concluded that MIPAS H2O profiles yield valuable information on the vertical distribution of H2O in the stratosphere with an overall accuracy of about 10 to 30% and a precision of typically 5 to 15% – well within the predicted error budget, showing that these global and continuous data are very valuable for scientific

  8. Analysis of progressive distorsion. Validation of the method based on effective primary stress. Discussion of Anderson's experimental data

    International Nuclear Information System (INIS)

    Moulin, Didier.

    1981-02-01

    An empirical rule usable for design by analysis against progressive distorsion has been set up from experiments conducted in C.E.N. Saclay. This rule is checked with experimental data obtained by W.F. ANDERSON, this experiment is sufficiently different from the Saclay one to evaluate the merits of the rule. The satisfactory results achieved, are another validation of the efficiency diagram on which the method is based [fr

  9. Validation of MIPAS-ENVISAT NO2 operational data

    Directory of Open Access Journals (Sweden)

    R. Ruhnke

    2007-06-01

    Full Text Available The Michelson Interferometer for Passive Atmospheric Sounding (MIPAS instrument was launched aboard the environmental satellite ENVISAT into its sun-synchronous orbit on 1 March 2002. The short-lived species NO2 is one of the key target products of MIPAS that are operationally retrieved from limb emission spectra measured in the stratosphere and mesosphere. Within the MIPAS validation activities, a large number of independent observations from balloons, satellites and ground-based stations have been compared to European Space Agency (ESA version 4.61 operational NO2 data comprising the time period from July 2002 until March 2004 where MIPAS measured with full spectral resolution. Comparisons between MIPAS and balloon-borne observations carried out in 2002 and 2003 in the Arctic, at mid-latitudes, and in the tropics show a very good agreement below 40 km altitude with a mean deviation of roughly 3%, virtually without any significant bias. The comparison to ACE satellite observations exhibits only a small negative bias of MIPAS which appears not to be significant. The independent satellite instruments HALOE, SAGE II, and POAM III confirm in common for the spring-summer time period a negative bias of MIPAS in the Arctic and a positive bias in the Antarctic middle and upper stratosphere exceeding frequently the combined systematic error limits. In contrast to the ESA operational processor, the IMK/IAA retrieval code allows accurate inference of NO2 volume mixing ratios under consideration of all important non-LTE processes. Large differences between both retrieval results appear especially at higher altitudes, above about 50 to 55 km. These differences might be explained at least partly by non-LTE under polar winter conditions but not at mid-latitudes. Below this altitude region mean differences between both processors remain within 5% (during night and up to 10% (during day under undisturbed (September 2002 conditions and up to 40% under perturbed

  10. Autism detection in early childhood (ADEC): reliability and validity data for a Level 2 screening tool for autistic disorder.

    Science.gov (United States)

    Nah, Yong-Hwee; Young, Robyn L; Brewer, Neil; Berlingeri, Genna

    2014-03-01

    The Autism Detection in Early Childhood (ADEC; Young, 2007) was developed as a Level 2 clinician-administered autistic disorder (AD) screening tool that was time-efficient, suitable for children under 3 years, easy to administer, and suitable for persons with minimal training and experience with AD. A best estimate clinical Diagnostic and Statistical Manual of Mental Disorders (4th ed., text rev.; DSM-IV-TR; American Psychiatric Association, 2000) diagnosis of AD was made for 70 children using all available information and assessment results, except for the ADEC data. A screening study compared these children on the ADEC with 57 children with other developmental disorders and 64 typically developing children. Results indicated high internal consistency (α = .91). Interrater reliability and test-retest reliability of the ADEC were also adequate. ADEC scores reliably discriminated different diagnostic groups after controlling for nonverbal IQ and Vineland Adaptive Behavior Composite scores. Construct validity (using exploratory factor analysis) and concurrent validity using performance on the Autism Diagnostic Observation Schedule (Lord et al., 2000), the Autism Diagnostic Interview-Revised (Le Couteur, Lord, & Rutter, 2003), and DSM-IV-TR criteria were also demonstrated. Signal detection analysis identified the optimal ADEC cutoff score, with the ADEC identifying all children who had an AD (N = 70, sensitivity = 1.0) but overincluding children with other disabilities (N = 13, specificity ranging from .74 to .90). Together, the reliability and validity data indicate that the ADEC has potential to be established as a suitable and efficient screening tool for infants with AD. 2014 APA

  11. Small Scale Variability and the Problem of Data Validation

    Science.gov (United States)

    Sparling, L. C.; Avallone, L.; Einaudi, Franco (Technical Monitor)

    2000-01-01

    Numerous measurements taken with a variety of airborne, balloon borne and ground based instruments over the past decade have revealed a complex multiscaled 3D structure in both chemical and dynamical fields in the upper troposphere/lower stratosphere. The variability occurs on scales that are well below the resolution of satellite measurements, leading to problems in measurement validation. We discuss some statistical ideas that can shed some light on the contribution of the natural variability to the inevitable differences in correlative measurements that are not strictly colocated, or that have different spatial resolution.

  12. Blockchain Technology: A Data Framework to Improve Validity, Trust, and Accountability of Information Exchange in Health Professions Education.

    Science.gov (United States)

    Funk, Eric; Riddell, Jeff; Ankel, Felix; Cabrera, Daniel

    2018-06-12

    Health professions educators face multiple challenges, among them the need to adapt educational methods to new technologies. In the last decades multiple new digital platforms have appeared in the learning arena, including massive open online courses and social media-based education. The major critique of these novel methods is the lack of the ability to ascertain the origin, validity, and accountability of the knowledge that is created, shared, and acquired. Recently, a novel technology based on secured data storage and transmission, called blockchain, has emerged as a way to generate networks where validity, trust, and accountability can be created. Conceptually blockchain is an open, public, distributed, and secure digital registry where information transactions are secured and have a clear origin, explicit pathways, and concrete value. Health professions education based on the blockchain will potentially allow improved tracking of content and the individuals who create it, quantify educational impact on multiple generations of learners, and build a relative value of educational interventions. Furthermore, institutions adopting blockchain technology would be able to provide certification and credentialing of healthcare professionals with no intermediaries. There is potential for blockchain to significantly change the future of health professions education and radically transform how patients, professionals, educators, and learners interact around safe, valid, and accountable information.

  13. The Moon Mineralogy Mapper (M3) imaging spectrometerfor lunar science: Instrument description, calibration, on‐orbit measurements, science data calibration and on‐orbit validation

    Science.gov (United States)

    C. Pieters,; P. Mouroulis,; M. Eastwood,; J. Boardman,; Green, R.O.; Glavich, T.; Isaacson, P.; Annadurai, M.; Besse, S.; Cate, D.; Chatterjee, A.; Clark, R.; Barr, D.; Cheek, L.; Combe, J.; Dhingra, D.; Essandoh, V.; Geier, S.; Goswami, J.N.; Green, R.; Haemmerle, V.; Head, J.; Hovland, L.; Hyman, S.; Klima, R.; Koch, T.; Kramer, G.; Kumar, A.S.K.; Lee, K.; Lundeen, S.; Malaret, E.; McCord, T.; McLaughlin, S.; Mustard, J.; Nettles, J.; Petro, N.; Plourde, K.; Racho, C.; Rodriguez, J.; Runyon, C.; Sellar, G.; Smith, C.; Sobel, H.; Staid, M.; Sunshine, J.; Taylor, L.; Thaisen, K.; Tompkins, S.; Tseng, H.; Vane, G.; Varanasi, P.; White, M.; Wilson, D.

    2011-01-01

    The NASA Discovery Moon Mineralogy Mapper imaging spectrometer was selected to pursue a wide range of science objectives requiring measurement of composition at fine spatial scales over the full lunar surface. To pursue these objectives, a broad spectral range imaging spectrometer with high uniformity and high signal-to-noise ratio capable of measuring compositionally diagnostic spectral absorption features from a wide variety of known and possible lunar materials was required. For this purpose the Moon Mineralogy Mapper imaging spectrometer was designed and developed that measures the spectral range from 430 to 3000 nm with 10 nm spectral sampling through a 24 degree field of view with 0.7 milliradian spatial sampling. The instrument has a signal-to-noise ratio of greater than 400 for the specified equatorial reference radiance and greater than 100 for the polar reference radiance. The spectral cross-track uniformity is >90% and spectral instantaneous field-of-view uniformity is >90%. The Moon Mineralogy Mapper was launched on Chandrayaan-1 on the 22nd of October. On the 18th of November 2008 the Moon Mineralogy Mapper was turned on and collected a first light data set within 24 h. During this early checkout period and throughout the mission the spacecraft thermal environment and orbital parameters varied more than expected and placed operational and data quality constraints on the measurements. On the 29th of August 2009, spacecraft communication was lost. Over the course of the flight mission 1542 downlinked data sets were acquired that provide coverage of more than 95% of the lunar surface. An end-to-end science data calibration system was developed and all measurements have been passed through this system and delivered to the Planetary Data System (PDS.NASA.GOV). An extensive effort has been undertaken by the science team to validate the Moon Mineralogy Mapper science measurements in the context of the mission objectives. A focused spectral, radiometric

  14. Dual Rotating Rake Measurements of Higher-Order Duct Modes: Validation Using Experimental and Numerical Data

    Science.gov (United States)

    Dahl, Milo D.; Hixon, Duane R.; Sutliff, Daniel L.

    2018-01-01

    A rotating rake mode measurement system was designed to measure acoustic duct modes generated by a fan stage. After analysis of the measured data, the mode coefficient amplitudes and phases were quantified. Early studies using this system found that mode power levels computed from rotating rake measured data would agree with the far-field power levels. However, this agreement required that the sound from the noise sources within the duct propagated outward from the duct exit without reflection and previous studies suggested conditions could exist where significant reflections could occur. This paper shows that mounting a second rake to the rotating system, with an offset in both the axial and the azimuthal directions, measures the data necessary to determine the modes propagating in both directions within a duct. The rotating rake data analysis technique was extended to include the data measured by the second rake. The analysis resulted in a set of circumferential mode coefficients at each of the two rake microphone locations. Radial basis functions were then least-squares fit to this data to obtain the radial mode coefficients for the modes propagating in both directions within the duct while accounting for the presence of evanescent modes. The validation of the dual-rotating-rake measurements was conducted using data from a combination of experiments and numerical calculations to compute reflection coefficients and other mode coefficient ratios. Compared to results from analytical and numerical computations, the results from dual-rotating-rake measured data followed the expected trends when frequency, mode number, and duct termination geometry were changed.

  15. Validation of Land Cover Products Using Reliability Evaluation Methods

    Directory of Open Access Journals (Sweden)

    Wenzhong Shi

    2015-06-01

    Full Text Available Validation of land cover products is a fundamental task prior to data applications. Current validation schemes and methods are, however, suited only for assessing classification accuracy and disregard the reliability of land cover products. The reliability evaluation of land cover products should be undertaken to provide reliable land cover information. In addition, the lack of high-quality reference data often constrains validation and affects the reliability results of land cover products. This study proposes a validation schema to evaluate the reliability of land cover products, including two methods, namely, result reliability evaluation and process reliability evaluation. Result reliability evaluation computes the reliability of land cover products using seven reliability indicators. Process reliability evaluation analyzes the reliability propagation in the data production process to obtain the reliability of land cover products. Fuzzy fault tree analysis is introduced and improved in the reliability analysis of a data production process. Research results show that the proposed reliability evaluation scheme is reasonable and can be applied to validate land cover products. Through the analysis of the seven indicators of result reliability evaluation, more information on land cover can be obtained for strategic decision-making and planning, compared with traditional accuracy assessment methods. Process reliability evaluation without the need for reference data can facilitate the validation and reflect the change trends of reliabilities to some extent.

  16. Underway Sampling of Marine Inherent Optical Properties on the Tara Oceans Expedition as a Novel Resource for Ocean Color Satellite Data Product Validation

    Science.gov (United States)

    Werdell, P. Jeremy; Proctor, Christopher W.; Boss, Emmanuel; Leeuw, Thomas; Ouhssain, Mustapha

    2013-01-01

    Developing and validating data records from operational ocean color satellite instruments requires substantial volumes of high quality in situ data. In the absence of broad, institutionally supported field programs, organizations such as the NASA Ocean Biology Processing Group seek opportunistic datasets for use in their operational satellite calibration and validation activities. The publicly available, global biogeochemical dataset collected as part of the two and a half year Tara Oceans expedition provides one such opportunity. We showed how the inline measurements of hyperspectral absorption and attenuation coefficients collected onboard the R/V Tara can be used to evaluate near-surface estimates of chlorophyll-a, spectral particulate backscattering coefficients, particulate organic carbon, and particle size classes derived from the NASA Moderate Resolution Imaging Spectroradiometer onboard Aqua (MODISA). The predominant strength of such flow-through measurements is their sampling rate-the 375 days of measurements resulted in 165 viable MODISA-to-in situ match-ups, compared to 13 from discrete water sampling. While the need to apply bio-optical models to estimate biogeochemical quantities of interest from spectroscopy remains a weakness, we demonstrated how discrete samples can be used in combination with flow-through measurements to create data records of sufficient quality to conduct first order evaluations of satellite-derived data products. Given an emerging agency desire to rapidly evaluate new satellite missions, our results have significant implications on how calibration and validation teams for these missions will be constructed.

  17. Comparison of results from the MCNP criticality validation suite using ENDF/B-VI and preliminary ENDF/B-VII nuclear data

    Energy Technology Data Exchange (ETDEWEB)

    Mosteller, R. D. (Russell D.)

    2004-01-01

    The MCNP Criticality Validation Suite is a collection of 31 benchmarks taken from the International Handbook of Evaluated Criticality Safety Benchmark Experiments. MCNP5 calculations clearly demonstrate that, overall, nuclear data for a preliminary version of ENDFB-VII produce better agreement with the benchmarks in the suite than do corresponding data from ENDF/B-VI. Additional calculations identify areas where improvements in the data still are needed. Based on results for the MCNP Criticality Validation Suite, the Pre-ENDF/B-VII nuclear data produce substantially better overall results than do their ENDF/B-VI counterparts. The calculated values for k{sub eff} for bare metal spheres and for an IEU cylinder reflected by normal uranium are in much better agreement with the benchmark values. In addition, the values of k{sub eff} for the bare metal spheres are much more consistent with those for corresponding metal spheres reflected by normal uranium or water. In addition, a long-standing controversy about the need for an ad hoc adjustment to the {sup 238}U resonance integral for thermal systems may finally be resolved. On the other hand, improvements still are needed in a number of areas. Those areas include intermediate-energy cross sections for {sup 235}U, angular distributions for elastic scattering in deuterium, and fast cross sections for {sup 237}Np.

  18. Are cannabis prevalence estimates comparable across countries and regions? A cross-cultural validation using search engine query data.

    Science.gov (United States)

    Steppan, Martin; Kraus, Ludwig; Piontek, Daniela; Siciliano, Valeria

    2013-01-01

    Prevalence estimation of cannabis use is usually based on self-report data. Although there is evidence on the reliability of this data source, its cross-cultural validity is still a major concern. External objective criteria are needed for this purpose. In this study, cannabis-related search engine query data are used as an external criterion. Data on cannabis use were taken from the 2007 European School Survey Project on Alcohol and Other Drugs (ESPAD). Provincial data came from three Italian nation-wide studies using the same methodology (2006-2008; ESPAD-Italia). Information on cannabis-related search engine query data was based on Google search volume indices (GSI). (1) Reliability analysis was conducted for GSI. (2) Latent measurement models of "true" cannabis prevalence were tested using perceived availability, web-based cannabis searches and self-reported prevalence as indicators. (3) Structure models were set up to test the influences of response tendencies and geographical position (latitude, longitude). In order to test the stability of the models, analyses were conducted on country level (Europe, US) and on provincial level in Italy. Cannabis-related GSI were found to be highly reliable and constant over time. The overall measurement model was highly significant in both data sets. On country level, no significant effects of response bias indicators and geographical position on perceived availability, web-based cannabis searches and self-reported prevalence were found. On provincial level, latitude had a significant positive effect on availability indicating that perceived availability of cannabis in northern Italy was higher than expected from the other indicators. Although GSI showed weaker associations with cannabis use than perceived availability, the findings underline the external validity and usefulness of search engine query data as external criteria. The findings suggest an acceptable relative comparability of national (provincial) prevalence

  19. Assessing the Relative Performance of Microwave-Based Satellite Rain Rate Retrievals Using TRMM Ground Validation Data

    Science.gov (United States)

    Wolff, David B.; Fisher, Brad L.

    2011-01-01

    Space-borne microwave sensors provide critical rain information used in several global multi-satellite rain products, which in turn are used for a variety of important studies, including landslide forecasting, flash flood warning, data assimilation, climate studies, and validation of model forecasts of precipitation. This study employs four years (2003-2006) of satellite data to assess the relative performance and skill of SSM/I (F13, F14 and F15), AMSU-B (N15, N16 and N17), AMSR-E (Aqua) and the TRMM Microwave Imager (TMI) in estimating surface rainfall based on direct instantaneous comparisons with ground-based rain estimates from Tropical Rainfall Measuring Mission (TRMM) Ground Validation (GV) sites at Kwajalein, Republic of the Marshall Islands (KWAJ) and Melbourne, Florida (MELB). The relative performance of each of these satellite estimates is examined via comparisons with space- and time-coincident GV radar-based rain rate estimates. Because underlying surface terrain is known to affect the relative performance of the satellite algorithms, the data for MELB was further stratified into ocean, land and coast categories using a 0.25deg terrain mask. Of all the satellite estimates compared in this study, TMI and AMSR-E exhibited considerably higher correlations and skills in estimating/observing surface precipitation. While SSM/I and AMSU-B exhibited lower correlations and skills for each of the different terrain categories, the SSM/I absolute biases trended slightly lower than AMSR-E over ocean, where the observations from both emission and scattering channels were used in the retrievals. AMSU-B exhibited the least skill relative to GV in all of the relevant statistical categories, and an anomalous spike was observed in the probability distribution functions near 1.0 mm/hr. This statistical artifact appears to be related to attempts by algorithm developers to include some lighter rain rates, not easily detectable by its scatter-only frequencies. AMSU

  20. Validating FMEA output against incident learning data: A study in stereotactic body radiation therapy.

    Science.gov (United States)

    Yang, F; Cao, N; Young, L; Howard, J; Logan, W; Arbuckle, T; Sponseller, P; Korssjoen, T; Meyer, J; Ford, E

    2015-06-01

    Though failure mode and effects analysis (FMEA) is becoming more widely adopted for risk assessment in radiation therapy, to our knowledge, its output has never been validated against data on errors that actually occur. The objective of this study was to perform FMEA of a stereotactic body radiation therapy (SBRT) treatment planning process and validate the results against data recorded within an incident learning system. FMEA on the SBRT treatment planning process was carried out by a multidisciplinary group including radiation oncologists, medical physicists, dosimetrists, and IT technologists. Potential failure modes were identified through a systematic review of the process map. Failure modes were rated for severity, occurrence, and detectability on a scale of one to ten and risk priority number (RPN) was computed. Failure modes were then compared with historical reports identified as relevant to SBRT planning within a departmental incident learning system that has been active for two and a half years. Differences between FMEA anticipated failure modes and existing incidents were identified. FMEA identified 63 failure modes. RPN values for the top 25% of failure modes ranged from 60 to 336. Analysis of the incident learning database identified 33 reported near-miss events related to SBRT planning. Combining both methods yielded a total of 76 possible process failures, of which 13 (17%) were missed by FMEA while 43 (57%) identified by FMEA only. When scored for RPN, the 13 events missed by FMEA ranked within the lower half of all failure modes and exhibited significantly lower severity relative to those identified by FMEA (p = 0.02). FMEA, though valuable, is subject to certain limitations. In this study, FMEA failed to identify 17% of actual failure modes, though these were of lower risk. Similarly, an incident learning system alone fails to identify a large number of potentially high-severity process errors. Using FMEA in combination with incident learning may

  1. Shift Verification and Validation

    Energy Technology Data Exchange (ETDEWEB)

    Pandya, Tara M. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Evans, Thomas M. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Davidson, Gregory G [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Johnson, Seth R. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Godfrey, Andrew T. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)

    2016-09-07

    This documentation outlines the verification and validation of Shift for the Consortium for Advanced Simulation of Light Water Reactors (CASL). Five main types of problems were used for validation: small criticality benchmark problems; full-core reactor benchmarks for light water reactors; fixed-source coupled neutron-photon dosimetry benchmarks; depletion/burnup benchmarks; and full-core reactor performance benchmarks. We compared Shift results to measured data and other simulated Monte Carlo radiation transport code results, and found very good agreement in a variety of comparison measures. These include prediction of critical eigenvalue, radial and axial pin power distributions, rod worth, leakage spectra, and nuclide inventories over a burn cycle. Based on this validation of Shift, we are confident in Shift to provide reference results for CASL benchmarking.

  2. Statistically validated mobile communication networks: the evolution of motifs in European and Chinese data

    International Nuclear Information System (INIS)

    Li, Ming-Xia; Jiang, Zhi-Qiang; Zhou, Wei-Xing; Palchykov, Vasyl; Kaski, Kimmo; Kertész, János; Miccichè, Salvatore; Tumminello, Michele; N Mantegna, Rosario

    2014-01-01

    Big data open up unprecedented opportunities for investigating complex systems, including society. In particular, communication data serve as major sources for computational social sciences, but they have to be cleaned and filtered as they may contain spurious information due to recording errors as well as interactions, like commercial and marketing activities, not directly related to the social network. The network constructed from communication data can only be considered as a proxy for the network of social relationships. Here we apply a systematic method, based on multiple-hypothesis testing, to statistically validate the links and then construct the corresponding Bonferroni network, generalized to the directed case. We study two large datasets of mobile phone records, one from Europe and the other from China. For both datasets we compare the raw data networks with the corresponding Bonferroni networks and point out significant differences in the structures and in the basic network measures. We show evidence that the Bonferroni network provides a better proxy for the network of social interactions than the original one. Using the filtered networks, we investigated the statistics and temporal evolution of small directed 3-motifs and concluded that closed communication triads have a formation time scale, which is quite fast and typically intraday. We also find that open communication triads preferentially evolve into other open triads with a higher fraction of reciprocated calls. These stylized facts were observed for both datasets. (paper)

  3. Statistically validated mobile communication networks: the evolution of motifs in European and Chinese data

    Science.gov (United States)

    Li, Ming-Xia; Palchykov, Vasyl; Jiang, Zhi-Qiang; Kaski, Kimmo; Kertész, János; Miccichè, Salvatore; Tumminello, Michele; Zhou, Wei-Xing; Mantegna, Rosario N.

    2014-08-01

    Big data open up unprecedented opportunities for investigating complex systems, including society. In particular, communication data serve as major sources for computational social sciences, but they have to be cleaned and filtered as they may contain spurious information due to recording errors as well as interactions, like commercial and marketing activities, not directly related to the social network. The network constructed from communication data can only be considered as a proxy for the network of social relationships. Here we apply a systematic method, based on multiple-hypothesis testing, to statistically validate the links and then construct the corresponding Bonferroni network, generalized to the directed case. We study two large datasets of mobile phone records, one from Europe and the other from China. For both datasets we compare the raw data networks with the corresponding Bonferroni networks and point out significant differences in the structures and in the basic network measures. We show evidence that the Bonferroni network provides a better proxy for the network of social interactions than the original one. Using the filtered networks, we investigated the statistics and temporal evolution of small directed 3-motifs and concluded that closed communication triads have a formation time scale, which is quite fast and typically intraday. We also find that open communication triads preferentially evolve into other open triads with a higher fraction of reciprocated calls. These stylized facts were observed for both datasets.

  4. A comparison of accuracy validation methods for genomic and pedigree-based predictions of swine litter size traits using Large White and simulated data.

    Science.gov (United States)

    Putz, A M; Tiezzi, F; Maltecca, C; Gray, K A; Knauer, M T

    2018-02-01

    The objective of this study was to compare and determine the optimal validation method when comparing accuracy from single-step GBLUP (ssGBLUP) to traditional pedigree-based BLUP. Field data included six litter size traits. Simulated data included ten replicates designed to mimic the field data in order to determine the method that was closest to the true accuracy. Data were split into training and validation sets. The methods used were as follows: (i) theoretical accuracy derived from the prediction error variance (PEV) of the direct inverse (iLHS), (ii) approximated accuracies from the accf90(GS) program in the BLUPF90 family of programs (Approx), (iii) correlation between predictions and the single-step GEBVs from the full data set (GEBV Full ), (iv) correlation between predictions and the corrected phenotypes of females from the full data set (Y c ), (v) correlation from method iv divided by the square root of the heritability (Y ch ) and (vi) correlation between sire predictions and the average of their daughters' corrected phenotypes (Y cs ). Accuracies from iLHS increased from 0.27 to 0.37 (37%) in the Large White. Approximation accuracies were very consistent and close in absolute value (0.41 to 0.43). Both iLHS and Approx were much less variable than the corrected phenotype methods (ranging from 0.04 to 0.27). On average, simulated data showed an increase in accuracy from 0.34 to 0.44 (29%) using ssGBLUP. Both iLHS and Y ch approximated the increase well, 0.30 to 0.46 and 0.36 to 0.45, respectively. GEBV Full performed poorly in both data sets and is not recommended. Results suggest that for within-breed selection, theoretical accuracy using PEV was consistent and accurate. When direct inversion is infeasible to get the PEV, correlating predictions to the corrected phenotypes divided by the square root of heritability is adequate given a large enough validation data set. © 2017 Blackwell Verlag GmbH.

  5. Designing the Social Context for Easier Verification, Validation, and Uncertainty Quantification of Earth Science Data

    Science.gov (United States)

    Barkstrom, B. R.; Loeb, N. G.; Wielicki, B. A.

    2017-12-01

    Verification, Validation, and Uncertainty Quantification (VVUQ) are key actions that support conclusions based on Earth science data. Communities of data producers and users must undertake VVUQ when they create and use their data. The strategies [S] and tools [T] suggested below come from successful use on two large NASA projects. The first was the Earth Radiation Budget Experiment (ERBE). The second is the investigation of Clouds and the Earth's Radiant Energy System (CERES). [S] 1. Partition the production system into subsystems that deal with data transformations confined to limited space and time scales. Simplify the subsystems to minimize the number of data transformations in each subsystem. [S] 2. Derive algorithms from the fundamental physics and chemistry governing the parameters in each subsystem including those for instrument calibration. [S] 3. Use preliminary uncertainty estimates to detect unexpected discrepancies. Removing these requires diagnostic work as well as development and testing of fixes. [S] 4. Make sure there are adequate resources to support multiple end-to-end reprocessing of all data products. [T] 1. Create file identifiers that accommodate temporal and spatial sequences of data files and subsystem version changes. [T] 2. Create libraries of parameters used in common by different subsystems to reduce errors due to inconsistent values. [T] 3. Maintain a list of action items to record progress on resolving discrepancies. [T] 4. Plan on VVUQ activities that use independent data sources and peer review before distributing and archiving data. The goal of VVUQ is to provide a transparent link between the data and the physics and chemistry governing the measured quantities. The VVUQ effort also involves specialized domain experience and nomenclature. It often requires as much effort as the original system development. ERBE and CERES demonstrated that these strategies and tools can reduce the cost of VVUQ for Earth science data products.

  6. Validation of secondary commercial data sources for physical activity facilities in urban and nonurban settings.

    Science.gov (United States)

    Han, Euna; Powell, Lisa; Slater, Sandy; Quinn, Christopher

    2012-11-01

    Secondary data are often necessary to assess the availability of commercial physical activity (PA) facilities and examine its association with individual behaviors and outcomes, yet the validity of such sources has been explored only in a limited number of studies. Field data were collected on the presence and attributes of commercial PA facilities in a random sample of 30 urban, 15 suburban, and 15 rural Census tracts in the Chicago metropolitan statistical area and surrounding area. Approximately 40% of PA establishments in the field data were listed for both urban and nonurban tracts in both lists except for nonurban tracts in D&B (35%), which was significantly improved in the combined list of D&B and InfoUSA. Approximately one-quarter of the PA facilities listed in D&B were found on the ground, whereas 40% to 50% of PA facilities listed in InfoUSA were found on the ground. PA establishments that offered instruction programs or lessons or that had a court or pool were less likely to be listed, particularly in the nonurban tracts. Secondary commercial business lists on PA facilities should be used with caution in assessing the built environment.

  7. Overview of SCIAMACHY validation: 2002 2004

    Science.gov (United States)

    Piters, A. J. M.; Bramstedt, K.; Lambert, J.-C.; Kirchhoff, B.

    2005-08-01

    SCIAMACHY, on board Envisat, is now in operation for almost three years. This UV/visible/NIR spectrometer measures the solar irradiance, the earthshine radiance scattered at nadir and from the limb, and the attenuation of solar radiation by the atmosphere during sunrise and sunset, from 240 to 2380 nm and at moderate spectral resolution. Vertical columns and profiles of a variety of atmospheric constituents are inferred from the SCIAMACHY radiometric measurements by dedicated retrieval algorithms. With the support of ESA and several international partners, a methodical SCIAMACHY validation programme has been developed jointly by Germany, the Netherlands and Belgium (the three instrument providing countries) to face complex requirements in terms of measured species, altitude range, spatial and temporal scales, geophysical states and intended scientific applications. This summary paper describes the approach adopted to address those requirements. The actual validation of the operational SCIAMACHY processors established at DLR on behalf of ESA has been hampered by data distribution and processor problems. Since first data releases in summer 2002, operational processors were upgraded regularly and some data products - level-1b spectra, level-2 O3, NO2, BrO and clouds data - have improved significantly. Validation results summarised in this paper conclude that for limited periods and geographical domains they can already be used for atmospheric research. Nevertheless, remaining processor problems cause major errors preventing from scientific usability in other periods and domains. Untied to the constraints of operational processing, seven scientific institutes (BIRA-IASB, IFE, IUP-Heidelberg, KNMI, MPI, SAO and SRON) have developed their own retrieval algorithms and generated SCIAMACHY data products, together addressing nearly all targeted constituents. Most of the UV-visible data products (both columns and profiles) already have acceptable, if not excellent, quality

  8. FY2017 Pilot Project Plan for the Nuclear Energy Knowledge and Validation Center Initiative

    Energy Technology Data Exchange (ETDEWEB)

    Ren, Weiju [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)

    2016-10-30

    To prepare for technical development of computational code validation under the Nuclear Energy Knowledge and Validation Center (NEKVAC) initiative, several meetings were held by a group of experts of the Idaho National Laboratory (INL) and the Oak Ridge National Laboratory (ORNL) to develop requirements of, and formulate a structure for, a transient fuel database through leveraging existing resources. It was concluded in discussions of these meetings that a pilot project is needed to address the most fundamental issues that can generate immediate stimulus to near-future validation developments as well as long-lasting benefits to NEKVAC operation. The present project is proposed based on the consensus of these discussions. Analysis of common scenarios in code validation indicates that the incapability of acquiring satisfactory validation data is often a showstopper that must first be tackled before any confident validation developments can be carried out. Validation data are usually found scattered in different places most likely with interrelationships among the data not well documented, incomplete with information for some parameters missing, nonexistent, or unrealistic to experimentally generate. Furthermore, with very different technical backgrounds, the modeler, the experimentalist, and the knowledgebase developer that must be involved in validation data development often cannot communicate effectively without a data package template that is representative of the data structure for the information domain of interest to the desired code validation. This pilot project is proposed to use the legendary TREAT Experiments Database to provide core elements for creating an ideal validation data package. Data gaps and missing data interrelationships will be identified from these core elements. All the identified missing elements will then be filled in with experimental data if available from other existing sources or with dummy data if nonexistent. The resulting hybrid

  9. Development of Indian cross section data files for Th-232 and U-233 and integral validation studies

    International Nuclear Information System (INIS)

    Ganesan, S.

    1988-01-01

    This paper presents an overview of the tasks performed towards the development of Indian cross section data files for Th-232 and U-233. Discrepancies in various neutron induced reaction cross sections in various available evaluated data files have been obtained by processing the basic data into multigroup form and intercomparison of the latter. Interesting results of integral validation studies for capture, fission and (n,2n) cross sections for Th-232 by analyses of selected integral measurements are presented. In the resonance range, energy regions where significant differences in the calculated self-shielding factors for Th-232 occur have been identified by a comparison of self-shielded multigroup cross sections derived from two recent evaluated data files, viz., ENDF/B-V (Rev.2) and JENDL-2, for several dilutions and temperatures. For U-233, the three different basic data files ENDF/B-IV, JENDL-2 and ENDL-84 were intercompared. Interesting observations on the predictional capability of these files for the criticality of the spherical metal U-233 system are given. The current status of Indian data file is presented. (author) 62 ref

  10. Measurement error correction in the least absolute shrinkage and selection operator model when validation data are available.

    Science.gov (United States)

    Vasquez, Monica M; Hu, Chengcheng; Roe, Denise J; Halonen, Marilyn; Guerra, Stefano

    2017-01-01

    Measurement of serum biomarkers by multiplex assays may be more variable as compared to single biomarker assays. Measurement error in these data may bias parameter estimates in regression analysis, which could mask true associations of serum biomarkers with an outcome. The Least Absolute Shrinkage and Selection Operator (LASSO) can be used for variable selection in these high-dimensional data. Furthermore, when the distribution of measurement error is assumed to be known or estimated with replication data, a simple measurement error correction method can be applied to the LASSO method. However, in practice the distribution of the measurement error is unknown and is expensive to estimate through replication both in monetary cost and need for greater amount of sample which is often limited in quantity. We adapt an existing bias correction approach by estimating the measurement error using validation data in which a subset of serum biomarkers are re-measured on a random subset of the study sample. We evaluate this method using simulated data and data from the Tucson Epidemiological Study of Airway Obstructive Disease (TESAOD). We show that the bias in parameter estimation is reduced and variable selection is improved.

  11. FAST Model Calibration and Validation of the OC5- DeepCwind Floating Offshore Wind System Against Wave Tank Test Data: Preprint

    Energy Technology Data Exchange (ETDEWEB)

    Wendt, Fabian F [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Robertson, Amy N [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Jonkman, Jason [National Renewable Energy Laboratory (NREL), Golden, CO (United States)

    2017-08-09

    During the course of the Offshore Code Comparison Collaboration, Continued, with Correlation (OC5) project, which focused on the validation of numerical methods through comparison against tank test data, the authors created a numerical FAST model of the 1:50-scale DeepCwind semisubmersible system that was tested at the Maritime Research Institute Netherlands ocean basin in 2013. This paper discusses several model calibration studies that were conducted to identify model adjustments that improve the agreement between the numerical simulations and the experimental test data. These calibration studies cover wind-field-specific parameters (coherence, turbulence), hydrodynamic and aerodynamic modeling approaches, as well as rotor model (blade-pitch and blade-mass imbalances) and tower model (structural tower damping coefficient) adjustments. These calibration studies were conducted based on relatively simple calibration load cases (wave only/wind only). The agreement between the final FAST model and experimental measurements is then assessed based on more-complex combined wind and wave validation cases.

  12. CVTresh: R Package for Level-Dependent Cross-Validation Thresholding

    Directory of Open Access Journals (Sweden)

    Donghoh Kim

    2006-04-01

    Full Text Available The core of the wavelet approach to nonparametric regression is thresholding of wavelet coefficients. This paper reviews a cross-validation method for the selection of the thresholding value in wavelet shrinkage of Oh, Kim, and Lee (2006, and introduces the R package CVThresh implementing details of the calculations for the procedures. This procedure is implemented by coupling a conventional cross-validation with a fast imputation method, so that it overcomes a limitation of data length, a power of 2. It can be easily applied to the classical leave-one-out cross-validation and K-fold cross-validation. Since the procedure is computationally fast, a level-dependent cross-validation can be developed for wavelet shrinkage of data with various sparseness according to levels.

  13. CVTresh: R Package for Level-Dependent Cross-Validation Thresholding

    Directory of Open Access Journals (Sweden)

    Donghoh Kim

    2006-04-01

    Full Text Available The core of the wavelet approach to nonparametric regression is thresholding of wavelet coefficients. This paper reviews a cross-validation method for the selection of the thresholding value in wavelet shrinkage of Oh, Kim, and Lee (2006, and introduces the R package CVThresh implementing details of the calculations for the procedures.This procedure is implemented by coupling a conventional cross-validation with a fast imputation method, so that it overcomes a limitation of data length, a power of 2. It can be easily applied to the classical leave-one-out cross-validation and K-fold cross-validation. Since the procedure is computationally fast, a level-dependent cross-validation can be developed for wavelet shrinkage of data with various sparseness according to levels.

  14. Excellent cross-cultural validity, intra-test reliability and construct validity of the dutch rivermead mobility index in patients after stroke undergoing rehabilitation

    NARCIS (Netherlands)

    Roorda, Leo D.; Green, John; De Kluis, Kiki R. A.; Molenaar, Ivo W.; Bagley, Pam; Smith, Jane; Geurts, Alexander C. H.

    2008-01-01

    Objective: To investigate the cross-cultural validity of international Dutch-English comparisons when using the Dutch Rivermead Mobility Index (RMI), and the intra-test reliability and construct validity of the Dutch RMI. Methods: Cross-cultural validity was studied in a combined data-set of Dutch

  15. IMPLEMENTATION AND VALIDATION OF STATISTICAL TESTS IN RESEARCH'S SOFTWARE HELPING DATA COLLECTION AND PROTOCOLS ANALYSIS IN SURGERY.

    Science.gov (United States)

    Kuretzki, Carlos Henrique; Campos, Antônio Carlos Ligocki; Malafaia, Osvaldo; Soares, Sandramara Scandelari Kusano de Paula; Tenório, Sérgio Bernardo; Timi, Jorge Rufino Ribas

    2016-03-01

    The use of information technology is often applied in healthcare. With regard to scientific research, the SINPE(c) - Integrated Electronic Protocols was created as a tool to support researchers, offering clinical data standardization. By the time, SINPE(c) lacked statistical tests obtained by automatic analysis. Add to SINPE(c) features for automatic realization of the main statistical methods used in medicine . The study was divided into four topics: check the interest of users towards the implementation of the tests; search the frequency of their use in health care; carry out the implementation; and validate the results with researchers and their protocols. It was applied in a group of users of this software in their thesis in the strict sensu master and doctorate degrees in one postgraduate program in surgery. To assess the reliability of the statistics was compared the data obtained both automatically by SINPE(c) as manually held by a professional in statistics with experience with this type of study. There was concern for the use of automatic statistical tests, with good acceptance. The chi-square, Mann-Whitney, Fisher and t-Student were considered as tests frequently used by participants in medical studies. These methods have been implemented and thereafter approved as expected. The incorporation of the automatic SINPE (c) Statistical Analysis was shown to be reliable and equal to the manually done, validating its use as a research tool for medical research.

  16. A cross-validation package driving Netica with python

    Science.gov (United States)

    Fienen, Michael N.; Plant, Nathaniel G.

    2014-01-01

    Bayesian networks (BNs) are powerful tools for probabilistically simulating natural systems and emulating process models. Cross validation is a technique to avoid overfitting resulting from overly complex BNs. Overfitting reduces predictive skill. Cross-validation for BNs is known but rarely implemented due partly to a lack of software tools designed to work with available BN packages. CVNetica is open-source, written in Python, and extends the Netica software package to perform cross-validation and read, rebuild, and learn BNs from data. Insights gained from cross-validation and implications on prediction versus description are illustrated with: a data-driven oceanographic application; and a model-emulation application. These examples show that overfitting occurs when BNs become more complex than allowed by supporting data and overfitting incurs computational costs as well as causing a reduction in prediction skill. CVNetica evaluates overfitting using several complexity metrics (we used level of discretization) and its impact on performance metrics (we used skill).

  17. Valid methods: the quality assurance of test method development, validation, approval, and transfer for veterinary testing laboratories.

    Science.gov (United States)

    Wiegers, Ann L

    2003-07-01

    Third-party accreditation is a valuable tool to demonstrate a laboratory's competence to conduct testing. Accreditation, internationally and in the United States, has been discussed previously. However, accreditation is only I part of establishing data credibility. A validated test method is the first component of a valid measurement system. Validation is defined as confirmation by examination and the provision of objective evidence that the particular requirements for a specific intended use are fulfilled. The international and national standard ISO/IEC 17025 recognizes the importance of validated methods and requires that laboratory-developed methods or methods adopted by the laboratory be appropriate for the intended use. Validated methods are therefore required and their use agreed to by the client (i.e., end users of the test results such as veterinarians, animal health programs, and owners). ISO/IEC 17025 also requires that the introduction of methods developed by the laboratory for its own use be a planned activity conducted by qualified personnel with adequate resources. This article discusses considerations and recommendations for the conduct of veterinary diagnostic test method development, validation, evaluation, approval, and transfer to the user laboratory in the ISO/IEC 17025 environment. These recommendations are based on those of nationally and internationally accepted standards and guidelines, as well as those of reputable and experienced technical bodies. They are also based on the author's experience in the evaluation of method development and transfer projects, validation data, and the implementation of quality management systems in the area of method development.

  18. Imputation by the mean score should be avoided when validating a Patient Reported Outcomes questionnaire by a Rasch model in presence of informative missing data

    LENUS (Irish Health Repository)

    Hardouin, Jean-Benoit

    2011-07-14

    Abstract Background Nowadays, more and more clinical scales consisting in responses given by the patients to some items (Patient Reported Outcomes - PRO), are validated with models based on Item Response Theory, and more specifically, with a Rasch model. In the validation sample, presence of missing data is frequent. The aim of this paper is to compare sixteen methods for handling the missing data (mainly based on simple imputation) in the context of psychometric validation of PRO by a Rasch model. The main indexes used for validation by a Rasch model are compared. Methods A simulation study was performed allowing to consider several cases, notably the possibility for the missing values to be informative or not and the rate of missing data. Results Several imputations methods produce bias on psychometrical indexes (generally, the imputation methods artificially improve the psychometric qualities of the scale). In particular, this is the case with the method based on the Personal Mean Score (PMS) which is the most commonly used imputation method in practice. Conclusions Several imputation methods should be avoided, in particular PMS imputation. From a general point of view, it is important to use an imputation method that considers both the ability of the patient (measured for example by his\\/her score), and the difficulty of the item (measured for example by its rate of favourable responses). Another recommendation is to always consider the addition of a random process in the imputation method, because such a process allows reducing the bias. Last, the analysis realized without imputation of the missing data (available case analyses) is an interesting alternative to the simple imputation in this context.

  19. Results of a monitoring programme in the environs of Berkeley aimed at collecting Chernobyl data for foodchain model validation

    International Nuclear Information System (INIS)

    Nair, S.; Darley, P.J.; Shaer, J.

    1989-03-01

    The results of a fallout measurement programme which was carried out in the environs of Berkeley Nuclear Laboratory in the United Kingdom following the Chernobyl reactor accident in April 1986 are presented in this report. The programme was aimed at establishing a time-dependent data base of concentrations of Chernobyl fallout radionuclides in selected agricultural products. Results were obtained for milk, grass, silage, soil and wheat over an eighteen month period from May 1986. It is intended to use the data to validate the CEGB's dynamic foodchain model, which is incorporated in the FOODWEB module of the NECTAR environmental code. (author)

  20. 78 FR 32255 - HHS-Operated Risk Adjustment Data Validation Stakeholder Meeting

    Science.gov (United States)

    2013-05-29

    ... States'') is assigned a host (in accordance with the Department Foreign Visitor Management Policy... general public. Visitors to the complex are required to show a valid U.S. Government issued photo... lobby, and the cafeteria. If a visitor is found outside of those areas without proper escort, they may...

  1. Collocation mismatch uncertainties in satellite aerosol retrieval validation

    Science.gov (United States)

    Virtanen, Timo H.; Kolmonen, Pekka; Sogacheva, Larisa; Rodríguez, Edith; Saponaro, Giulia; de Leeuw, Gerrit

    2018-02-01

    Satellite-based aerosol products are routinely validated against ground-based reference data, usually obtained from sun photometer networks such as AERONET (AEROsol RObotic NETwork). In a typical validation exercise a spatial sample of the instantaneous satellite data is compared against a temporal sample of the point-like ground-based data. The observations do not correspond to exactly the same column of the atmosphere at the same time, and the representativeness of the reference data depends on the spatiotemporal variability of the aerosol properties in the samples. The associated uncertainty is known as the collocation mismatch uncertainty (CMU). The validation results depend on the sampling parameters. While small samples involve less variability, they are more sensitive to the inevitable noise in the measurement data. In this paper we study systematically the effect of the sampling parameters in the validation of AATSR (Advanced Along-Track Scanning Radiometer) aerosol optical depth (AOD) product against AERONET data and the associated collocation mismatch uncertainty. To this end, we study the spatial AOD variability in the satellite data, compare it against the corresponding values obtained from densely located AERONET sites, and assess the possible reasons for observed differences. We find that the spatial AOD variability in the satellite data is approximately 2 times larger than in the ground-based data, and the spatial variability correlates only weakly with that of AERONET for short distances. We interpreted that only half of the variability in the satellite data is due to the natural variability in the AOD, and the rest is noise due to retrieval errors. However, for larger distances (˜ 0.5°) the correlation is improved as the noise is averaged out, and the day-to-day changes in regional AOD variability are well captured. Furthermore, we assess the usefulness of the spatial variability of the satellite AOD data as an estimate of CMU by comparing the

  2. Reactivity worth measurements on the CALIBAN reactor: interpretation of integral experiments for the nuclear data validation

    International Nuclear Information System (INIS)

    Richard, B.

    2012-01-01

    The good knowledge of nuclear data, input parameters for the neutron transport calculation codes, is necessary to support the advances of the nuclear industry. The purpose of this work is to bring pertinent information regarding the nuclear data integral validation process. Reactivity worth measurements have been performed on the Caliban reactor, they concern four materials of interest for the nuclear industry: gold, lutetium, plutonium and uranium 238. Experiments which have been conducted in order to improve the characterization of the core are also described and discussed, the latter are necessary to the good interpretation of reactivity worth measurements. The experimental procedures are described with their associated uncertainties, measurements are then compared to numerical results. The methods used in numerical calculations are reported, especially the multigroup cross sections generation for deterministic codes. The modeling of the experiments is presented along with the associated uncertainties. This comparison led to an interpretation concerning the qualification of nuclear data libraries. Discrepancies are reported, discussed and justify the need of such experiments. (author) [fr

  3. Validation of virtual instrument for data analysis in metrology of time and frequency; Validacao do instrumento virtual para analise de dados em metrologia de tempo e frequencia

    Energy Technology Data Exchange (ETDEWEB)

    Jordao, Bruno; Quaresma, Daniel; Rocha, Pedro; Carvalho, Ricardo, E-mail: bjordan@on.br [Observatorio Nacional (ON), Rio de Janeiro, RJ (Brazil). Laboratorio Primario de Tempo e Frequencia; Peixoto, Jose Guilherme [Instituto de Radioprotecao e Dosimetria (LNMRI/IRD/CNEN-RJ), Rio de Janeiro, RJ (Brazil). Laboratorio Nacional de Metrologia das Radiacoes Ionizantes

    2016-07-01

    Commercial Software (CS) for collection, analysis and plot time and frequency data plots are being increasingly used in reference laboratories worldwide. With this, it has greatly improved the results of calculations of uncertainty for these values. We propose the creation of a collection of software and data analysis using Virtual Instruments (VI) developed the Primary Laboratory Time and frequency of the National Observatory - ON and validation of this instrument. To validate the instrument developed, it made a comparative analysis between the results obtained (VI) with the results obtained by (CS) widely used in many metrology laboratories. From these results we can conclude that there was equivalence between the analyzed data. (author)

  4. Preliminary Assessment of ATR-C Capabilities to Provide Integral Benchmark Data for Key Structural/Matrix Materials that May be Used for Nuclear Data Testing and Analytical Methods Validation

    Energy Technology Data Exchange (ETDEWEB)

    John D. Bess

    2009-03-01

    The purpose of this research is to provide a fundamental computational investigation into the possible integration of experimental activities with the Advanced Test Reactor Critical (ATR-C) facility with the development of benchmark experiments. Criticality benchmarks performed in the ATR-C could provide integral data for key matrix and structural materials used in nuclear systems. Results would then be utilized in the improvement of nuclear data libraries and as a means for analytical methods validation. It is proposed that experiments consisting of well-characterized quantities of materials be placed in the Northwest flux trap position of the ATR-C. The reactivity worth of the material could be determined and computationally analyzed through comprehensive benchmark activities including uncertainty analyses. Experiments were modeled in the available benchmark model of the ATR using MCNP5 with the ENDF/B-VII.0 cross section library. A single bar (9.5 cm long, 0.5 cm wide, and 121.92 cm high) of each material could provide sufficient reactivity difference in the core geometry for computational modeling and analysis. However, to provide increased opportunity for the validation of computational models, additional bars of material placed in the flux trap would increase the effective reactivity up to a limit of 1$ insertion. For simplicity in assembly manufacture, approximately four bars of material could provide a means for additional experimental benchmark configurations, except in the case of strong neutron absorbers and many materials providing positive reactivity. Future tasks include the cost analysis and development of the experimental assemblies, including means for the characterization of the neutron flux and spectral indices. Oscillation techniques may also serve to provide additional means for experimentation and validation of computational methods and acquisition of integral data for improving neutron cross sections. Further assessment of oscillation

  5. MO-G-BRE-09: Validating FMEA Against Incident Learning Data: A Study in Stereotactic Body Radiation Therapy

    International Nuclear Information System (INIS)

    Yang, F; Cao, N; Young, L; Howard, J; Sponseller, P; Logan, W; Arbuckle, T; Korssjoen, T; Meyer, J; Ford, E

    2014-01-01

    Purpose: Though FMEA (Failure Mode and Effects Analysis) is becoming more widely adopted for risk assessment in radiation therapy, to our knowledge it has never been validated against actual incident learning data. The objective of this study was to perform an FMEA analysis of an SBRT (Stereotactic Body Radiation Therapy) treatment planning process and validate this against data recorded within an incident learning system. Methods: FMEA on the SBRT treatment planning process was carried out by a multidisciplinary group including radiation oncologists, medical physicists, and dosimetrists. Potential failure modes were identified through a systematic review of the workflow process. Failure modes were rated for severity, occurrence, and detectability on a scale of 1 to 10 and RPN (Risk Priority Number) was computed. Failure modes were then compared with historical reports identified as relevant to SBRT planning within a departmental incident learning system that had been active for two years. Differences were identified. Results: FMEA identified 63 failure modes. RPN values for the top 25% of failure modes ranged from 60 to 336. Analysis of the incident learning database identified 33 reported near-miss events related to SBRT planning. FMEA failed to anticipate 13 of these events, among which 3 were registered with severity ratings of severe or critical in the incident learning system. Combining both methods yielded a total of 76 failure modes, and when scored for RPN the 13 events missed by FMEA ranked within the middle half of all failure modes. Conclusion: FMEA, though valuable, is subject to certain limitations, among them the limited ability to anticipate all potential errors for a given process. This FMEA exercise failed to identify a significant number of possible errors (17%). Integration of FMEA with retrospective incident data may be able to render an improved overview of risks within a process

  6. Physics validation of detector simulation tools for LHC

    International Nuclear Information System (INIS)

    Beringer, J.

    2004-01-01

    Extensive studies aimed at validating the physics processes built into the detector simulation tools Geant4 and Fluka are in progress within all Large Hardon Collider (LHC) experiments, within the collaborations developing these tools, and within the LHC Computing Grid (LCG) Simulation Physics Validation Project, which has become the primary forum for these activities. This work includes detailed comparisons with test beam data, as well as benchmark studies of simple geometries and materials with single incident particles of various energies for which experimental data is available. We give an overview of these validation activities with emphasis on the latest results

  7. A broad view of model validation

    International Nuclear Information System (INIS)

    Tsang, C.F.

    1989-10-01

    The safety assessment of a nuclear waste repository requires the use of models. Such models need to be validated to ensure, as much as possible, that they are a good representation of the actual processes occurring in the real system. In this paper we attempt to take a broad view by reviewing step by step the modeling process and bringing out the need to validating every step of this process. This model validation includes not only comparison of modeling results with data from selected experiments, but also evaluation of procedures for the construction of conceptual models and calculational models as well as methodologies for studying data and parameter correlation. The need for advancing basic scientific knowledge in related fields, for multiple assessment groups, and for presenting our modeling efforts in open literature to public scrutiny is also emphasized. 16 refs

  8. Validation of Long-Term Global Aerosol Climatology Project Optical Thickness Retrievals Using AERONET and MODIS Data

    Science.gov (United States)

    Geogdzhayev, Igor V.; Mishchenko, Michael I.

    2015-01-01

    A comprehensive set of monthly mean aerosol optical thickness (AOT) data from coastal and island AErosol RObotic NETwork (AERONET) stations is used to evaluate Global Aerosol Climatology Project (GACP) retrievals for the period 1995-2009 during which contemporaneous GACP and AERONET data were available. To put the GACP performance in broader perspective, we also compare AERONET and MODerate resolution Imaging Spectroradiometer (MODIS) Aqua level-2 data for 2003-2009 using the same methodology. We find that a large mismatch in geographic coverage exists between the satellite and ground-based datasets, with very limited AERONET coverage of open-ocean areas. This is especially true of GACP because of the smaller number of AERONET stations at the early stages of the network development. Monthly mean AOTs from the two over-the-ocean satellite datasets are well-correlated with the ground-based values, the correlation coefficients being 0.81-0.85 for GACP and 0.74-0.79 for MODIS. Regression analyses demonstrate that the GACP mean AOTs are approximately 17%-27% lower than the AERONET values on average, while the MODIS mean AOTs are 5%-25% higher. The regression coefficients are highly dependent on the weighting assumptions (e.g., on the measure of aerosol variability) as well as on the set of AERONET stations used for comparison. Comparison of over-the-land and over-the-ocean MODIS monthly mean AOTs in the vicinity of coastal AERONET stations reveals a significant bias. This may indicate that aerosol amounts in coastal locations can differ significantly from those in adjacent open-ocean areas. Furthermore, the color of coastal waters and peculiarities of coastline meteorological conditions may introduce biases in the GACP AOT retrievals. We conclude that the GACP and MODIS over-the-ocean retrieval algorithms show similar ranges of discrepancy when compared to available coastal and island AERONET stations. The factors mentioned above may limit the performance of the

  9. Signal-to-noise assessment for diffusion tensor imaging with single data set and validation using a difference image method with data from a multicenter study

    Energy Technology Data Exchange (ETDEWEB)

    Wang, Zhiyue J., E-mail: jerry.wang@childrens.com [Department of Radiology, Children' s Medical Center, Dallas, Texas 75235 and Department of Radiology, University of Texas Southwestern Medical Center, Dallas, Texas 75390 (United States); Chia, Jonathan M. [Clinical Science, Philips Healthcare, Cleveland, Ohio 44143 (United States); Ahmed, Shaheen; Rollins, Nancy K. [Department of Radiology, Children' s Medical Center, Dallas, TX 75235 and Department of Radiology, University of Texas Southwestern Medical Center, Dallas, TX 75390 (United States)

    2014-09-15

    Purpose: To describe a quantitative method for determination of SNR that extracts the local noise level using a single diffusion data set. Methods: Brain data sets came from a multicenter study (eight sites; three MR vendors). Data acquisition protocol required b = 0, 700 s/mm{sup 2}, fov = 256 × 256 mm{sup 2}, acquisition matrix size 128 × 128, reconstruction matrix size 256 × 256; 30 gradient encoding directions and voxel size 2 × 2 × 2 mm{sup 3}. Regions-of-interest (ROI) were placed manually on the b = 0 image volume on transverse slices, and signal was recorded as the mean value of the ROI. The noise level from the ROI was evaluated using Fourier Transform based Butterworth high-pass filtering. Patients were divided into two groups, one for filter parameter optimization (N = 17) and one for validation (N = 10). Six white matter areas (the genu and splenium of corpus callosum, right and left centrum semiovale, right and left anterior corona radiata) were analyzed. The Bland–Altman method was used to compare the resulting SNR with that from the difference image method. The filter parameters were optimized for each brain area, and a set of “global” parameters was also obtained, which represent an average of all regions. Results: The Bland–Altman analysis on the validation group using “global” filter parameters revealed that the 95% limits of agreement of percent bias between the SNR obtained with the new and the reference methods were −15.5% (median of the lower limit, range [−24.1%, −8.9%]) and 14.5% (median of the higher limits, range [12.7%, 18.0%]) for the 6 brain areas. Conclusions: An FT-based high-pass filtering method can be used for local area SNR assessment using only one DTI data set. This method could be used to evaluate SNR for patient studies in a multicenter setting.

  10. Validation of an explanatory tool for data-fused displays for high-technology future aircraft

    Science.gov (United States)

    Fletcher, Georgina C. L.; Shanks, Craig R.; Selcon, Stephen J.

    1996-05-01

    As the number of sensor and data sources in the military cockpit increases, pilots will suffer high levels of workload which could result in reduced performance and the loss of situational awareness. A DRA research program has been investigating the use of data-fused displays in decision support and has developed and laboratory-tested an explanatory tool for displaying information in air combat scenarios. The tool has been designed to provide pictorial explanations of data that maintain situational awareness by involving the pilot in the hostile aircraft threat assessment task. This paper reports a study carried out to validate the success of the explanatory tool in a realistic flight simulation facility. Aircrew were asked to perform a threat assessment task, either with or without the explanatory tool providing information in the form of missile launch success zone envelopes, while concurrently flying a waypoint course within set flight parameters. The results showed that there was a significant improvement (p less than 0.01) in threat assessment accuracy of 30% when using the explanatory tool. This threat assessment performance advantage was achieved without a trade-off with flying task performance. Situational awareness measures showed no general differences between the explanatory and control conditions, but significant learning effects suggested that the explanatory tool makes the task initially more intuitive and hence less demanding on the pilots' attentional resources. The paper concludes that DRA's data-fused explanatory tool is successful at improving threat assessment accuracy in a realistic simulated flying environment, and briefly discusses the requirements for further research in the area.

  11. Use of In-Flight Data to Validate Mars Samle Return Autonomous RvD GNC

    DEFF Research Database (Denmark)

    Barrena, V.; Colmenarejo, P.; Suatoni, M.

    D is based on RF, camera and LIDAR measurements. It includes design, prototyping and verification at three different levels: algorithms design and verification in a Functional Engineering Simulator, SW demonstrator verified in Real Time Avionics Test Benching and Dynamic Test Benching. Moreover...... and testing of a vision based optical stimulator (ViSOS by DTU) to enhance the on-ground validation capabilities. After checking different alternatives for the proposed HARvD-GNC experiment with PRISMA resources, an efficient but cost-effective approach was chosen. The approach is based on designing MSR......-like dedicated manoeuvres sequencing using the already existing on-board PRISMA GNC/AOCS system (based on relative GPS measurements for the closed-loop execution of the manoeuvres sequencing and acquiring RF and camera images as part of the HARvD-GNC experiment data). This option allows downloading the sensor...

  12. Application of validity theory and methodology to patient-reported outcome measures (PROMs): building an argument for validity.

    Science.gov (United States)

    Hawkins, Melanie; Elsworth, Gerald R; Osborne, Richard H

    2018-07-01

    Data from subjective patient-reported outcome measures (PROMs) are now being used in the health sector to make or support decisions about individuals, groups and populations. Contemporary validity theorists define validity not as a statistical property of the test but as the extent to which empirical evidence supports the interpretation of test scores for an intended use. However, validity testing theory and methodology are rarely evident in the PROM validation literature. Application of this theory and methodology would provide structure for comprehensive validation planning to support improved PROM development and sound arguments for the validity of PROM score interpretation and use in each new context. This paper proposes the application of contemporary validity theory and methodology to PROM validity testing. The validity testing principles will be applied to a hypothetical case study with a focus on the interpretation and use of scores from a translated PROM that measures health literacy (the Health Literacy Questionnaire or HLQ). Although robust psychometric properties of a PROM are a pre-condition to its use, a PROM's validity lies in the sound argument that a network of empirical evidence supports the intended interpretation and use of PROM scores for decision making in a particular context. The health sector is yet to apply contemporary theory and methodology to PROM development and validation. The theoretical and methodological processes in this paper are offered as an advancement of the theory and practice of PROM validity testing in the health sector.

  13. Validation of the Danish PAROLE lexicon (upubliceret)

    DEFF Research Database (Denmark)

    Møller, Margrethe; Christoffersen, Ellen

    2000-01-01

    This validation is based on the Danish PAROLE lexicon dated June 20, 1998, downloaded on March 16, 1999. Subsequently, the developers of the lexicon have informed us that they have been revising the lexicon, in particular the morphological level. Morphological entries were originally generated...... automatically from a machine-readable version of the Official Danish Spelling Dictionary (Retskrivningsordbogen 1986, in the following RO86), and this resulted in some overgeneration, which the developers started eliminating after submitting the Danish PAROLE lexicon for validation. The present validation is......, however, based on the January 1997 version of the lexicon. The validation as such complies with the specifications described in ELRA validation manuals for lexical data, i.e. Underwood and Navaretta: "A Draft Manual for the Validation of Lexica, Final Report" [Underwood & Navaretta1997] and Braasch: "A...

  14. User's guide for signal validation software: Final report

    International Nuclear Information System (INIS)

    Swisher, V.I.

    1987-09-01

    Northeast Utilities has implemented a real-time signal validation program into the safety parameter display systems (SPDS) at Millstone Units 2 and 3. Signal validation has been incorporated to improve the reliability of the information being used in the SPDS. Signal validation uses Parity Space Vector Analysis to process SPDS sensor data. The Parity Space algorithm determines consistency among independent, redundant input measurements. This information is then used to calculate a validated estimate of that parameter. Additional logic is incorporated to compare partially redundant measurement data. In both plants the SPDS has been designed to monitor the status of critical safety functions (CSFs) and provide information that can be used with plant-specific emergency operating procedures (EOPs). However the CSF logic, EOPs, and complement of plant sensors vary for these plants due to their different design characteristics (MP2 - 870 MWe Combustion Engineering PWR, MP3 - 1150 MWe Westinghouse PWR). These differences in plant design and information requirements result in a variety of signal validation applications

  15. A Generic Approach for Inversion of Surface Reflectance over Land: Overview, Application and Validation Using MODIS and LANDSAT8 Data

    Science.gov (United States)

    Vermote, E.; Roger, J. C.; Justice, C. O.; Franch, B.; Claverie, M.

    2016-01-01

    This paper presents a generic approach developed to derive surface reflectance over land from a variety of sensors. This technique builds on the extensive dataset acquired by the Terra platform by combining MODIS and MISR to derive an explicit and dynamic map of band ratio's between blue and red channels and is a refinement of the operational approach used for MODIS and LANDSAT over the past 15 years. We will present the generic approach and the application to MODIS and LANDSAT data and its validation using the AERONET data.

  16. Use of RTIGS data streams for validating the performance of the IGS Ultra-Rapid products

    Science.gov (United States)

    Thaler, Gottfried; Weber, Robert

    2010-05-01

    The IGS (International GNSS Service) Real-Time Working Group (RTIGS) disseminates for several years raw observation data of a globally distributed steady growing station network in real-time via the internet. This observation data can be used for validating the performance of the IGS predicted orbits and clocks (Ultra-Rapid (IGU)). Therefore, based on pre-processed ITRF- station coordinates, clock corrections w.r.t GPS-Time for GPS-satellites and site-receivers as well as satellite orbits are calculated in quasi real-time and compared to the IGU solutions. The Institute for "Geodesy and Geophysics" of the Technical University of Vienna develops based on the software RTIGS Multicast Receive (RTIGSMR) provided by National Resources Canada (NRCan) the software RTIGU-Control. Using Code-smoothed observations RTIGU-Control calculates in a first step by means of a linear KALMAN-Filter and based on the orbit information of the IGUs real-time clock corrections and clock drifts w.r.t GPS-Time for the GPS-satellites and stations. The second extended KALMAN-Filter (kinematic approach) uses again the Code-smoothed observations corrected for the clock corrections of step 1 to calculate the positions and velocities of the satellites. The calculation interval is set to 30 seconds. The results and comparisons to IGU-products are displayed online but also stored as clock-RINEX- and SP3-files on the ftp-server of the institute, e.g. for validation of the performance of the IGU predicted products. A comparison to the more precise but delayed issued IGS Rapid products (IGR) allows also to validate the performance of RTIGU-Control. To carry out these comparisons the MatLab routine RTIGU-Analyse was established. This routine is for example able to import and process standard clock-RINEX-files of several sources and delivers a variety of comparisons both in graphical or numerical form. Results will become part of this presentation. Another way to analyse the quality and consistency of

  17. Report on workshop "Study of the polar atmosphere and cryosphere using satellite data with surface validation observations including unmanned one"

    Directory of Open Access Journals (Sweden)

    Hiroshi Kanzawa

    1993-07-01

    Full Text Available The workshop was organized to discuss algorithms to derive parameters of the polar atmosphere and cryosphere using satellite data received mainly at Syowa Station (69°S, 40°E, Antarctica, i.e., the data from NOAA, MOS (Marine Observation Satellite-1,ERS (European Remote Sensing Satellite-1,JERS (Japanese Earth Resources Satellite-1 with validation data at the surface. It was held on 16 March 1993 at the National Institute of Polar Research (NIPR, total number of participants being about 40. The contents of the workshop are as follows : The present status of receipt and utilization of the satellite data of NOAA, MOS-1,ERS-1,JERS-1; The Atmosphere; Sea ice; The Cryosphere; Introduction to the satellite data analysis system at the Information Science Center at NIPR.

  18. Development and validation of factor analysis for dynamic in-vivo imaging data sets

    Science.gov (United States)

    Goldschmied, Lukas; Knoll, Peter; Mirzaei, Siroos; Kalchenko, Vyacheslav

    2018-02-01

    In-vivo optical imaging method provides information about the anatomical structures and function of tissues ranging from single cell to entire organisms. Dynamic Fluorescent Imaging (DFI) is used to examine dynamic events related to normal physiology or disease progression in real time. In this work we improve this method by using factor analysis (FA) to automatically separate overlying structures.The proposed method is based on a previously introduced Transcranial Optical Vascular Imaging (TOVI), which employs natural and sufficient transparency through the intact cranial bones of a mouse. Fluorescent image acquisition is performed after intravenous fluorescent tracer administration. Afterwards FA is used to extract structures with different temporal characteristics from dynamic contrast enhanced studies without making any a priori assumptions about physiology. The method was validated by a dynamic light phantom based on the Arduino hardware platform and dynamic fluorescent cerebral hemodynamics data sets. Using the phantom data FA can separate various light channels without user intervention. FA applied on an image sequence obtained after fluorescent tracer administration is allowing extracting valuable information about cerebral blood vessels anatomy and functionality without a-priory assumptions of their anatomy or physiology while keeping the mouse cranium intact. Unsupervised color-coding based on FA enhances visibility and distinguishing of blood vessels belonging to different compartments. DFI based on FA especially in case of transcranial imaging can be used to separate dynamic structures.

  19. Current Status of the Validation of the Atmospheric Chemistry Instruments on Envisat

    Science.gov (United States)

    Lecomte, P.; Koopman, R.; Zehner, C.; Laur, H.; Attema, E.; Wursteisen, P.; Snoeij, P.

    2003-04-01

    Envisat is ESA's advanced Earth observing satellite launched in March 2002 and is designed to provide measurements of the atmosphere, ocean, land and ice over a five-year period. After the launch and the switch-on period, a six-month commissioning phase has taken place for instrument calibration and geophysical validation, concluded with the Envisat Calibration Review held in September 2002. In addition to ESA and its industrial partners in the Envisat consortium, many other companies and research institutes have contributed to the calibration and validation programme under ESA contract as expert support laboratories (ESLs). A major contribution has also been made by the Principal Investigators of approved proposals submitted to ESA in response to a worldwide "Announcement of Opportunity for the Exploitation of the Envisat Data Products" in 1998. Working teams have been formed in which the different participants worked side by side to achieve the objectives of the calibration and validation programme. Validation is a comparison of Envisat level-2 data products and estimates of the different geophysical variables obtained by independent means, the validation instruments. Validation is closely linked to calibration because inconsistencies discovered in the comparison of Envisat Level 2 data products to well-known external instruments can have many different sources, including inaccuracies of the Envisat instrument calibration and the data calibration algorithms. Therefore, initial validation of the geophysical variables has provided feedback to calibration, de-bugging and algorithm improvement. The initial validation phase ended in December 2002 with the Envisat Validation Workshop at which, for a number of products, a final quality statement was given. Full validation of all data products available from the Atmospheric Chemistry Instruments on Envisat (MIPAS, GOMOS and SCIAMACHY) is quite a challenge and therefore it has been decided to adopt a step-wise approach

  20. Temperature Data Assimilation with Salinity Corrections: Validation for the NSIPP Ocean Data Assimilation System in the Tropical Pacific Ocean, 1993-1998

    Science.gov (United States)

    Troccoli, Alberto; Rienecker, Michele M.; Keppenne, Christian L.; Johnson, Gregory C.

    2003-01-01

    The NASA Seasonal-to-Interannual Prediction Project (NSIPP) has developed an Ocean data assimilation system to initialize the quasi-isopycnal ocean model used in our experimental coupled-model forecast system. Initial tests of the system have focused on the assimilation of temperature profiles in an optimal interpolation framework. It is now recognized that correction of temperature only often introduces spurious water masses. The resulting density distribution can be statically unstable and also have a detrimental impact on the velocity distribution. Several simple schemes have been developed to try to correct these deficiencies. Here the salinity field is corrected by using a scheme which assumes that the temperature-salinity relationship of the model background is preserved during the assimilation. The scheme was first introduced for a zlevel model by Troccoli and Haines (1999). A large set of subsurface observations of salinity and temperature is used to cross-validate two data assimilation experiments run for the 6-year period 1993-1998. In these two experiments only subsurface temperature observations are used, but in one case the salinity field is also updated whenever temperature observations are available.

  1. Solution Validation for a Double Façade Prototype

    Directory of Open Access Journals (Sweden)

    Pau Fonseca i Casas

    2017-12-01

    Full Text Available A Solution Validation involves comparing the data obtained from the system that are implemented following the model recommendations, as well as the model results. This paper presents a Solution Validation that has been performed with the aim of certifying that a set of computer-optimized designs, for a double façade, are consistent with reality. To validate the results obtained through simulation models, based on dynamic thermal calculation and using Computational Fluid Dynamic techniques, a comparison with the data obtained by monitoring a real implemented prototype has been carried out. The new validated model can be used to describe the system thermal behavior in different climatic zones without having to build a new prototype. The good performance of the proposed double façade solution is confirmed since the validation assures there is a considerable energy saving, preserving and even improving interior comfort. This work shows all the processes in the Solution Validation depicting some of the problems we faced and represents an example of this kind of validation that often is not considered in a simulation project.

  2. Validation of the VTT's reactor physics code system

    International Nuclear Information System (INIS)

    Tanskanen, A.

    1998-01-01

    At VTT Energy several international reactor physics codes and nuclear data libraries are used in a variety of applications. The codes and libraries are under constant development and every now and then new updated versions are released, which are taken in use as soon as they have been validated at VTT Energy. The primary aim of the validation is to ensure that the code works properly, and that it can be used correctly. Moreover, the applicability of the codes and libraries are studied in order to establish their advantages and weak points. The capability of generating program-specific nuclear data for different reactor physics codes starting from the same evaluated data is sometimes of great benefit. VTT Energy has acquired a nuclear data processing system based on the NJOY-94.105 and TRANSX-2.15 processing codes. The validity of the processing system has been demonstrated by generating pointwise (MCNP) and groupwise (ANISN) temperature-dependent cross section sets for the benchmark calculations of the Doppler coefficient of reactivity. At VTT Energy the KENO-VI three-dimensional Monte Carlo code is used in criticality safety analyses. The KENO-VI code and the 44GROUPNDF5 data library have been validated at VTT Energy against the ZR-6 and LR-0 critical experiments. Burnup Credit refers to the reduction in reactivity of burned nuclear fuel due to the change in composition during irradiation. VTT Energy has participated in the calculational VVER-440 burnup credit benchmark in order to validate criticality safety calculation tools. (orig.)

  3. Determination of methylmercury in marine sediment samples: Method validation and occurrence data

    Energy Technology Data Exchange (ETDEWEB)

    Carrasco, Luis; Vassileva, Emilia, E-mail: e.vasileva-veleva@iaea.org

    2015-01-01

    Highlights: • A method for MeHg determination at trace level in marine sediments is completely validated. • Validation is performed according to ISO-17025 and Eurachem guidelines. • The extraction efficiency of four sample preparation procedures is evaluated. • The uncertainty budget is used as a tool for evaluation of main uncertainty contributors. • Comparison with independent methods yields good agreement within stated uncertainty. - Abstract: The determination of methylmercury (MeHg) in sediment samples is a difficult task due to the extremely low MeHg/THg (total mercury) ratio and species interconversion. Here, we present the method validation of a cost-effective fit-for-purpose analytical procedure for the measurement of MeHg in sediments, which is based on aqueous phase ethylation, followed by purge and trap and hyphenated gas chromatography–pyrolysis–atomic fluorescence spectrometry (GC–Py–AFS) separation and detection. Four different extraction techniques, namely acid and alkaline leaching followed by solvent extraction and evaporation, microwave-assisted extraction with 2-mercaptoethanol, and acid leaching, solvent extraction and back extraction into sodium thiosulfate, were examined regarding their potential to selectively extract MeHg from estuarine sediment IAEA-405 certified reference material (CRM). The procedure based on acid leaching with HNO{sub 3}/CuSO{sub 4}, solvent extraction and back extraction into Na{sub 2}S{sub 2}O{sub 3} yielded the highest extraction recovery, i.e., 94 ± 3% and offered the possibility to perform the extraction of a large number of samples in a short time, by eliminating the evaporation step. The artifact formation of MeHg was evaluated by high performance liquid chromatography coupled to inductively coupled plasma mass spectrometry (HPLC–ICP–MS), using isotopically enriched Me{sup 201}Hg and {sup 202}Hg and it was found to be nonexistent. A full validation approach in line with ISO 17025 and

  4. Determination of methylmercury in marine sediment samples: Method validation and occurrence data

    International Nuclear Information System (INIS)

    Carrasco, Luis; Vassileva, Emilia

    2015-01-01

    Highlights: • A method for MeHg determination at trace level in marine sediments is completely validated. • Validation is performed according to ISO-17025 and Eurachem guidelines. • The extraction efficiency of four sample preparation procedures is evaluated. • The uncertainty budget is used as a tool for evaluation of main uncertainty contributors. • Comparison with independent methods yields good agreement within stated uncertainty. - Abstract: The determination of methylmercury (MeHg) in sediment samples is a difficult task due to the extremely low MeHg/THg (total mercury) ratio and species interconversion. Here, we present the method validation of a cost-effective fit-for-purpose analytical procedure for the measurement of MeHg in sediments, which is based on aqueous phase ethylation, followed by purge and trap and hyphenated gas chromatography–pyrolysis–atomic fluorescence spectrometry (GC–Py–AFS) separation and detection. Four different extraction techniques, namely acid and alkaline leaching followed by solvent extraction and evaporation, microwave-assisted extraction with 2-mercaptoethanol, and acid leaching, solvent extraction and back extraction into sodium thiosulfate, were examined regarding their potential to selectively extract MeHg from estuarine sediment IAEA-405 certified reference material (CRM). The procedure based on acid leaching with HNO 3 /CuSO 4 , solvent extraction and back extraction into Na 2 S 2 O 3 yielded the highest extraction recovery, i.e., 94 ± 3% and offered the possibility to perform the extraction of a large number of samples in a short time, by eliminating the evaporation step. The artifact formation of MeHg was evaluated by high performance liquid chromatography coupled to inductively coupled plasma mass spectrometry (HPLC–ICP–MS), using isotopically enriched Me 201 Hg and 202 Hg and it was found to be nonexistent. A full validation approach in line with ISO 17025 and Eurachem guidelines was followed

  5. Dry/Wet Conditions Monitoring Based on TRMM Rainfall Data and Its Reliability Validation over Poyang Lake Basin, China

    Directory of Open Access Journals (Sweden)

    Xianghu Li

    2013-11-01

    Full Text Available Local dry/wet conditions are of great concern in regional water resource and floods/droughts disaster risk management. Satellite-based precipitation products have greatly improved their accuracy and applicability and are expected to offer an alternative to ground rain gauges data. This paper investigated the capability of Tropical Rainfall Measuring Mission (TRMM rainfall data for monitoring the temporal and spatial variation of dry/wet conditions in Poyang Lake basin during 1998–2010, and validated its reliability with rain gauges data from 14 national meteorological stations in the basin. The results show that: (1 the daily TRMM rainfall data does not describe the occurrence and contribution rates of precipitation accurately, but monthly TRMM data have a good linear relationship with rain gauges rainfall data; (2 both the Z index and Standardized Precipitation Index (SPI based on monthly TRMM rainfall data oscillate around zero and show a consistent interannual variability as compared with rain gauges data; (3 the spatial pattern of moisture status, either in dry months or wet months, based on both the Z index and SPI using TRMM data, agree with the observed rainfall. In conclusion, the monthly TRMM rainfall data can be used for monitoring the variation and spatial distribution of dry/wet conditions in Poyang Lake basin.

  6. Representation of occupational information across resources and validation of the occupational data for health model.

    Science.gov (United States)

    Rajamani, Sripriya; Chen, Elizabeth S; Lindemann, Elizabeth; Aldekhyyel, Ranyah; Wang, Yan; Melton, Genevieve B

    2018-02-01

    Reports by the National Academy of Medicine and leading public health organizations advocate including occupational information as part of an individual's social context. Given recent National Academy of Medicine recommendations on occupation-related data in the electronic health record, there is a critical need for improved representation. The National Institute for Occupational Safety and Health has developed an Occupational Data for Health (ODH) model, currently in draft format. This study aimed to validate the ODH model by mapping occupation-related elements from resources representing recommendations, standards, public health reports and surveys, and research measures, along with preliminary evaluation of associated value sets. All 247 occupation-related items across 20 resources mapped to the ODH model. Recommended value sets had high variability across the evaluated resources. This study demonstrates the ODH model's value, the multifaceted nature of occupation information, and the critical need for occupation value sets to support clinical care, population health, and research. © The Author 2017. Published by Oxford University Press on behalf of the American Medical Informatics Association. All rights reserved. For Permissions, please email: journals.permissions@oup.com.

  7. A New Method for Analyzing Content Validity Data Using Multidimensional Scaling

    Science.gov (United States)

    Li, Xueming; Sireci, Stephen G.

    2013-01-01

    Validity evidence based on test content is of essential importance in educational testing. One source for such evidence is an alignment study, which helps evaluate the congruence between tested objectives and those specified in the curriculum. However, the results of an alignment study do not always sufficiently capture the degree to which a test…

  8. Verification and validation benchmarks.

    Energy Technology Data Exchange (ETDEWEB)

    Oberkampf, William Louis; Trucano, Timothy Guy

    2007-02-01

    Verification and validation (V&V) are the primary means to assess the accuracy and reliability of computational simulations. V&V methods and procedures have fundamentally improved the credibility of simulations in several high-consequence fields, such as nuclear reactor safety, underground nuclear waste storage, and nuclear weapon safety. Although the terminology is not uniform across engineering disciplines, code verification deals with assessing the reliability of the software coding, and solution verification deals with assessing the numerical accuracy of the solution to a computational model. Validation addresses the physics modeling accuracy of a computational simulation by comparing the computational results with experimental data. Code verification benchmarks and validation benchmarks have been constructed for a number of years in every field of computational simulation. However, no comprehensive guidelines have been proposed for the construction and use of V&V benchmarks. For example, the field of nuclear reactor safety has not focused on code verification benchmarks, but it has placed great emphasis on developing validation benchmarks. Many of these validation benchmarks are closely related to the operations of actual reactors at near-safety-critical conditions, as opposed to being more fundamental-physics benchmarks. This paper presents recommendations for the effective design and use of code verification benchmarks based on manufactured solutions, classical analytical solutions, and highly accurate numerical solutions. In addition, this paper presents recommendations for the design and use of validation benchmarks, highlighting the careful design of building-block experiments, the estimation of experimental measurement uncertainty for both inputs and outputs to the code, validation metrics, and the role of model calibration in validation. It is argued that the understanding of predictive capability of a computational model is built on the level of

  9. Identification of valid reference genes for the normalization of RT qPCR gene expression data in human brain tissue

    Directory of Open Access Journals (Sweden)

    Ravid Rivka

    2008-05-01

    Full Text Available Abstract Background Studies of gene expression in post mortem human brain can contribute to understanding of the pathophysiology of neurodegenerative diseases, including Alzheimer's disease (AD, Parkinson's disease (PD and dementia with Lewy bodies (DLB. Quantitative real-time PCR (RT qPCR is often used to analyse gene expression. The validity of results obtained using RT qPCR is reliant on accurate data normalization. Reference genes are generally used to normalize RT qPCR data. Given that expression of some commonly used reference genes is altered in certain conditions, this study aimed to establish which reference genes were stably expressed in post mortem brain tissue from individuals with AD, PD or DLB. Results The present study investigated the expression stability of 8 candidate reference genes, (ubiquitin C [UBC], tyrosine-3-monooxygenase [YWHAZ], RNA polymerase II polypeptide [RP II], hydroxymethylbilane synthase [HMBS], TATA box binding protein [TBP], β-2-microglobulin [B2M], glyceraldehyde-3-phosphate dehydrogenase [GAPDH], and succinate dehydrogenase complex-subunit A, [SDHA] in cerebellum and medial temporal gyrus of 6 AD, 6 PD, 6 DLB subjects, along with 5 matched controls using RT qPCR (TaqMan® Gene Expression Assays. Gene expression stability was analysed using geNorm to rank the candidate genes in order of decreasing stability in each disease group. The optimal number of genes recommended for accurate data normalization in each disease state was determined by pairwise variation analysis. Conclusion This study identified validated sets of mRNAs which would be appropriate for the normalization of RT qPCR data when studying gene expression in brain tissue of AD, PD, DLB and control subjects.

  10. Reprocessing input data validation

    International Nuclear Information System (INIS)

    Persiani, P.J.; Bucher, R.G.; Pond, R.B.; Cornella, R.J.

    1990-01-01

    The Isotope Correlation Technique (ICT), in conjunction with the gravimetric (Pu/U ratio) method for mass determination, provides an independent verification of the input accountancy at the dissolver or accountancy stage of the reprocessing plant. The Isotope Correlation Technique has been applied to many classes of domestic and international reactor systems (light-water, heavy-water, graphite, and liquid-metal) operating in a variety of modes (power, research, production, and breeder), and for a variety of reprocessing fuel cycle management strategies. Analysis of reprocessing operations data based on isotopic correlations derived for assemblies in a PWR environment and fuel management scheme, yielded differences between the measurement-derived and ICT-derived plutonium mass determinations of (-0.02 ± 0.23)% for the measured U-235 and (+0.50 ± 0.31)% for the measured Pu-239, for a core campaign. The ICT analyses has been implemented for the plutonium isotopics in a depleted uranium assembly in a heavy-water, enriched uranium system and for the uranium isotopes in the fuel assemblies in light-water, highly-enriched systems. 7 refs., 5 figs., 4 tabs

  11. The Treatment Validity of Autism Screening Instruments

    Science.gov (United States)

    Livanis, Andrew; Mouzakitis, Angela

    2010-01-01

    Treatment validity is a frequently neglected topic of screening instruments used to identify autism spectrum disorders. Treatment validity, however, should represent an important aspect of these instruments to link the resulting data to the selection of interventions as well as make decisions about treatment length and intensity. Research…

  12. SeaWiFS data from oceanic waters around New Zealand: Validation and applications

    Science.gov (United States)

    Richardson, K.; Boyd, P.; Gall, M.; Pinkerton, M.

    Satellite observations of ocean colour are the only realistic way to measure phytoplankton abundance at regional and global scales. NASA's Sea-viewing Wide Field -o f-view Sensor (SeaWiFS) began operation in September 1997 and is still providing data today. The data are of particular value to New Zealand, which has the fourth largest Exclusive Economic Zone (EEZ) in the world (some 4 million km2 ). Analysis of moderate resolution (9 km) SeaWiFS monthly Standard Mapped Images has substantially increased knowledge of the dynamics of chlorophyll concentrations around New Zealand. SeaWiFS data over nearly three years shows that northern New Zealand Subtropical and Tasman Sea waters follow a classical cycle of spring and autumn chlorophyll blooms consistent with production being co-limited by nitrate and light. Subantarctic Waters south of New Zealand had a low-magnitude annual cycle of chlorophyll abundance that peaked in early autumn, consistent with production being principally iron-limited. Chlorophyll was generally highest in the Subtropical Front either side of New Zealand where Subtropical and Subantarctic waters mix. NIWA (National Institute of Water and Atmospheric Research) has been receiving and processing high resolution (1.1 km) SeaWiFS data for the NZ region since May 2000. In addition to this, extensive bio-optical data from a number of NIWA cruises are being used to validate the satellite data and assess the accuracy of the ocean products in New Zealand open ocean and coastal waters. The performance of the SeaWiFS chlorophyll-a algorithm (OC4v4) has been investigated by comparing high-precision in situ measurements of the underwater radiation field with measurements of phytoplankton pigment concentration. Analyses of these results suggest that the algorithm may be performing well in the open ocean for chlorophyll- a concentrations below 0.3-0.4 mg m-3 but overestimating by a factor of two or more at higher concentrations. NIWA believes that ocean colour

  13. Validation of POLDER/ADEOS data using a ground-based lidar network: Preliminary results for semi-transparent and cirrus clouds

    Science.gov (United States)

    Chepfer, H.; Sauvage, L.; Flamant, P. H.; Pelon, J.; Goloub, P.; Brogniez, G.; spinhirne, J.; Lavorato, M.; Sugimoto, N.

    1998-01-01

    At mid and tropical latitudes, cirrus clouds are present more than 50% of the time in satellites observations. Due to their large spatial and temporal coverage, and associated low temperatures, cirrus clouds have a major influence on the Earth-Ocean-Atmosphere energy balance through their effects on the incoming solar radiation and outgoing infrared radiation. At present the impact of cirrus clouds on climate is well recognized but remains to be asserted more precisely, for their optical and radiative properties are not very well known. In order to understand the effects of cirrus clouds on climate, their optical and radiative characteristics of these clouds need to be determined accurately at different scales in different locations i.e. latitude. Lidars are well suited to observe cirrus clouds, they can detect very thin and semi-transparent layers, and retrieve the clouds geometrical properties i.e. altitude and multilayers, as well as radiative properties i.e. optical depth, backscattering phase functions of ice crystals. Moreover the linear depolarization ratio can give information on the ice crystal shape. In addition, the data collected with an airborne version of POLDER (POLarization and Directionality of Earth Reflectances) instrument have shown that bidirectional polarized measurements can provide information on cirrus cloud microphysical properties (crystal shapes, preferred orientation in space). The spaceborne version of POLDER-1 has been flown on ADEOS-1 platform during 8 months (October 96 - June 97), and the next POLDER-2 instrument will be launched in 2000 on ADEOS-2. The POLDER-1 cloud inversion algorithms are currently under validation. For cirrus clouds, a validation based on comparisons between cloud properties retrieved from POLDER-1 data and cloud properties inferred from a ground-based lidar network is currently under consideration. We present the first results of the validation.

  14. On the Need for Quality Control in Validation Research.

    Science.gov (United States)

    Maier, Milton H.

    1988-01-01

    Validated aptitude tests used to help make personnel decisions about military recruits against hands-on tests of job performance in radio repairers and automotive mechanics. Data were filled with errors, reducing accuracy of validity coefficients. Discusses how validity coefficients can be made more accurate by exercising quality control during…

  15. Validation of MCNP and WIMS-AECL/DRAGON/RFSP for ACR-1000 applications

    International Nuclear Information System (INIS)

    Bromley, Blair P.; Adams, Fred P.; Zeller, Michael B.; Watts, David G.; Shukhman, Boris V.; Pencer, Jeremy

    2008-01-01

    This paper gives a summary of the validation of the reactor physics codes WIMS-AECL, DRAGON, RFSP and MCNP5, which are being used in the design, operation, and safety analysis of the ACR-1000 R . The standards and guidelines being followed for code validation of the suite are established in CSA Standard N286.7-99 and ANS Standard ANS-19.3-2005. These codes are being validated for the calculation of key output parameters associated with various reactor physics phenomena of importance during normal operations and postulated accident conditions in an ACR-1000 reactor. Experimental data from a variety of sources are being used for validation. The bulk of the validation data is from critical experiments in the ZED-2 research reactor with ACR-type lattices. To supplement and complement ZED-2 data, qualified and applicable data are being taken from other power and research reactors, such as existing CANDU R units, FUGEN, NRU and SPERT research reactors, and the DCA critical facility. MCNP simulations of the ACR-1000 are also being used for validating WIMS-AECL/ DRAGON/RFSP, which involves extending the validation results for MCNP through the assistance of TSUNAMI analyses. Code validation against commissioning data in the first-build ACR-1000 will be confirmatory. The code validation is establishing the biases and uncertainties in the calculations of the WIMS-AECL/DRAGON/RFSP suite for the evaluation of various key parameters of importance in the reactor physics analysis of the ACR-1000. (authors)

  16. Optical Tracking Data Validation and Orbit Estimation for Sparse Observations of Satellites by the OWL-Net.

    Science.gov (United States)

    Choi, Jin; Jo, Jung Hyun; Yim, Hong-Suh; Choi, Eun-Jung; Cho, Sungki; Park, Jang-Hyun

    2018-06-07

    An Optical Wide-field patroL-Network (OWL-Net) has been developed for maintaining Korean low Earth orbit (LEO) satellites' orbital ephemeris. The OWL-Net consists of five optical tracking stations. Brightness signals of reflected sunlight of the targets were detected by a charged coupled device (CCD). A chopper system was adopted for fast astrometric data sampling, maximum 50 Hz, within a short observation time. The astrometric accuracy of the optical observation data was validated with precise orbital ephemeris such as Consolidated Prediction File (CPF) data and precise orbit determination result with onboard Global Positioning System (GPS) data from the target satellite. In the optical observation simulation of the OWL-Net for 2017, an average observation span for a single arc of 11 LEO observation targets was about 5 min, while an average optical observation separation time was 5 h. We estimated the position and velocity with an atmospheric drag coefficient of LEO observation targets using a sequential-batch orbit estimation technique after multi-arc batch orbit estimation. Post-fit residuals for the multi-arc batch orbit estimation and sequential-batch orbit estimation were analyzed for the optical measurements and reference orbit (CPF and GPS data). The post-fit residuals with reference show few tens-of-meters errors for in-track direction for multi-arc batch and sequential-batch orbit estimation results.

  17. Optical Tracking Data Validation and Orbit Estimation for Sparse Observations of Satellites by the OWL-Net

    Directory of Open Access Journals (Sweden)

    Jin Choi

    2018-06-01

    Full Text Available An Optical Wide-field patroL-Network (OWL-Net has been developed for maintaining Korean low Earth orbit (LEO satellites’ orbital ephemeris. The OWL-Net consists of five optical tracking stations. Brightness signals of reflected sunlight of the targets were detected by a charged coupled device (CCD. A chopper system was adopted for fast astrometric data sampling, maximum 50 Hz, within a short observation time. The astrometric accuracy of the optical observation data was validated with precise orbital ephemeris such as Consolidated Prediction File (CPF data and precise orbit determination result with onboard Global Positioning System (GPS data from the target satellite. In the optical observation simulation of the OWL-Net for 2017, an average observation span for a single arc of 11 LEO observation targets was about 5 min, while an average optical observation separation time was 5 h. We estimated the position and velocity with an atmospheric drag coefficient of LEO observation targets using a sequential-batch orbit estimation technique after multi-arc batch orbit estimation. Post-fit residuals for the multi-arc batch orbit estimation and sequential-batch orbit estimation were analyzed for the optical measurements and reference orbit (CPF and GPS data. The post-fit residuals with reference show few tens-of-meters errors for in-track direction for multi-arc batch and sequential-batch orbit estimation results.

  18. Validity evidence based on test content.

    Science.gov (United States)

    Sireci, Stephen; Faulkner-Bond, Molly

    2014-01-01

    Validity evidence based on test content is one of the five forms of validity evidence stipulated in the Standards for Educational and Psychological Testing developed by the American Educational Research Association, American Psychological Association, and National Council on Measurement in Education. In this paper, we describe the logic and theory underlying such evidence and describe traditional and modern methods for gathering and analyzing content validity data. A comprehensive review of the literature and of the aforementioned Standards is presented. For educational tests and other assessments targeting knowledge and skill possessed by examinees, validity evidence based on test content is necessary for building a validity argument to support the use of a test for a particular purpose. By following the methods described in this article, practitioners have a wide arsenal of tools available for determining how well the content of an assessment is congruent with and appropriate for the specific testing purposes.

  19. Validation of moderator-level reactivity coefficient using station data

    Energy Technology Data Exchange (ETDEWEB)

    Younis, M.; Martchouk, I., E-mail: mohamed.younis@amecfw.com, E-mail: iouri.martchouk@amecfw.com [Amec Foster Wheeler, Toronto, ON (Canada); Buchan, P.D., E-mail: david.buchan@opg.com [Ontario Power Generation, Pickering, ON (Canada)

    2015-07-01

    The reactivity effect due to variations in the moderator level has been recognized as a reactor physics phenomenon of importance during normal operation and accident analysis. The moderator-level reactivity coefficient is an important parameter in safety analysis of CANDU reactors, e.g., during Loss of Moderator Heat Sink as well as in the simulation of Reactor Regulating System action in CANDU reactors that use moderator level for reactivity control. This paper presents the results of the validation exercise of the reactor-physics toolset using the measurements performed in Pickering Unit 4 in 2003. The capability of the code suite of predicting moderator-level reactivity effect was tested by comparing measured and predicted reactor-physics parameters. (author)

  20. Ageing nuclear plants inspection, maintenance and performance monitoring eased by Data Validation and Reconciliation (DVR) approach

    International Nuclear Information System (INIS)

    Quang, A. Tran; Chares, R.

    2012-01-01

    With the long term operation of nuclear power plants, decisions to replace or repair should be optimized from an economical, human resource and safety point of view. In that frame, the operation requires a continuous inspection and testing of systems in order to detect degradations and failures at an early stage. Nevertheless, in spite of the fact that such assiduous vigilance ensures safety, it is not optimized in terms of man power and maintenance costs. Preventive strategies show the same types of drawbacks. On top of that, these maintenance procedures rely on process measurements which accuracy, availability and reliability cannot always be ensured. The present paper introduces a new approach to the maintenance management of ageing installations and suggests a method that overcomes the actual drawbacks. The Reconciliation and Validation of data (DVR) is an advanced and innovative technology that relies on process data statistics, thermodynamic and on-site process knowledge. Benefits of real-time applications are also presented. (author)

  1. Empirical validation of the S-Score algorithm in the analysis of gene expression data

    Directory of Open Access Journals (Sweden)

    Archer Kellie J

    2006-03-01

    Full Text Available Abstract Background Current methods of analyzing Affymetrix GeneChip® microarray data require the estimation of probe set expression summaries, followed by application of statistical tests to determine which genes are differentially expressed. The S-Score algorithm described by Zhang and colleagues is an alternative method that allows tests of hypotheses directly from probe level data. It is based on an error model in which the detected signal is proportional to the probe pair signal for highly expressed genes, but approaches a background level (rather than 0 for genes with low levels of expression. This model is used to calculate relative change in probe pair intensities that converts probe signals into multiple measurements with equalized errors, which are summed over a probe set to form the S-Score. Assuming no expression differences between chips, the S-Score follows a standard normal distribution, allowing direct tests of hypotheses to be made. Using spike-in and dilution datasets, we validated the S-Score method against comparisons of gene expression utilizing the more recently developed methods RMA, dChip, and MAS5. Results The S-score showed excellent sensitivity and specificity in detecting low-level gene expression changes. Rank ordering of S-Score values more accurately reflected known fold-change values compared to other algorithms. Conclusion The S-score method, utilizing probe level data directly, offers significant advantages over comparisons using only probe set expression summaries.

  2. Validation and correction of rainfall data from the WegenerNet high density network in southeast Austria

    Science.gov (United States)

    O, Sungmin; Foelsche, U.; Kirchengast, G.; Fuchsberger, J.

    2018-01-01

    Eight years of daily rainfall data from WegenerNet were analyzed by comparison with data from Austrian national weather stations. WegenerNet includes 153 ground level weather stations in an area of about 15 km × 20 km in the Feldbach region in southeast Austria. Rainfall has been measured by tipping bucket gauges at 150 stations of the network since the beginning of 2007. Since rain gauge measurements are considered close to true rainfall, there are increasing needs for WegenerNet data for the validation of rainfall data products such as remote sensing based estimates or model outputs. Serving these needs, this paper aims at providing a clearer interpretation on WegenerNet rainfall data for users in hydro-meteorological communities. Five clusters - a cluster consists of one national weather station and its four closest WegenerNet stations - allowed us close comparison of datasets between the stations. Linear regression analysis and error estimation with statistical indices were conducted to quantitatively evaluate the WegenerNet daily rainfall data. It was found that rainfall data between the stations show good linear relationships with an average correlation coefficient (r) of 0.97 , while WegenerNet sensors tend to underestimate rainfall according to the regression slope (0.87). For the five clusters investigated, the bias and relative bias were - 0.97 mm d-1 and - 11.5 % on average (except data from new sensors). The average of bias and relative bias, however, could be reduced by about 80 % through a simple linear regression-slope correction, with the assumption that the underestimation in WegenerNet data was caused by systematic errors. The results from the study have been employed to improve WegenerNet data for user applications so that a new version of the data (v5) is now available at the WegenerNet data portal (www.wegenernet.org).

  3. MODIS Observation of Aerosols over Southern Africa During SAFARI 2000: Data, Validation, and Estimation of Aerosol Radiative Forcing

    Science.gov (United States)

    Ichoku, Charles; Kaufman, Yoram; Remer, Lorraine; Chu, D. Allen; Mattoo, Shana; Tanre, Didier; Levy, Robert; Li, Rong-Rong; Kleidman, Richard; Lau, William K. M. (Technical Monitor)

    2001-01-01

    Aerosol properties, including optical thickness and size parameters, are retrieved operationally from the MODIS sensor onboard the Terra satellite launched on 18 December 1999. The predominant aerosol type over the Southern African region is smoke, which is generated from biomass burning on land and transported over the southern Atlantic Ocean. The SAFARI-2000 period experienced smoke aerosol emissions from the regular biomass burning activities as well as from the prescribed burns administered on the auspices of the experiment. The MODIS Aerosol Science Team (MAST) formulates and implements strategies for the retrieval of aerosol products from MODIS, as well as for validating and analyzing them in order to estimate aerosol effects in the radiative forcing of climate as accurately as possible. These activities are carried out not only from a global perspective, but also with a focus on specific regions identified as having interesting characteristics, such as the biomass burning phenomenon in southern Africa and the associated smoke aerosol, particulate, and trace gas emissions. Indeed, the SAFARI-2000 aerosol measurements from the ground and from aircraft, along with MODIS, provide excellent data sources for a more intensive validation and a closer study of the aerosol characteristics over Southern Africa. The SAFARI-2000 ground-based measurements of aerosol optical thickness (AOT) from both the automatic Aerosol Robotic Network (AERONET) and handheld Sun photometers have been used to validate MODIS retrievals, based on a sophisticated spatio-temporal technique. The average global monthly distribution of aerosol from MODIS has been combined with other data to calculate the southern African aerosol daily averaged (24 hr) radiative forcing over the ocean for September 2000. It is estimated that on the average, for cloud free conditions over an area of 9 million square kin, this predominantly smoke aerosol exerts a forcing of -30 W/square m C lose to the terrestrial

  4. Investigation of a two-phase nozzle flow and validation of several computer codes by the experimental data

    International Nuclear Information System (INIS)

    Kedziur, F.

    1980-03-01

    Stationary experiments with a convergent nozzle are performed in order to validate advanced two-phase computer codes, which find application in the blowdown-phase of a loss-of-coolant accident (LOCA). The steam/water flow presents a broad variety of initial conditions: The pressure varies between 2 and 13 MPa, the void fraction between 0 (subcooled) and about 80%, a great number of subcritical as well as critical experiments with different flow pattern is investigated. Additional air/water experiments serve for the separation of phase transition effects. The transient acceleration of the fluid in the LOCA-case is simulated by a local acceleration in the experiments. The layout of the nozzle and the applied measurement technique allow for a separate testing of physical models and the determination of empirical model parameters, respectively: In the four codes DUESE, DRIX-20, RELAP4/MOD6 and STRUYA the models - if they exist - for slip between the phases, thermodynamic non-equilibrium, pipe friction and critical mass flow rate are validated and criticised in comparison with the experimental data, and the corresponding model parameters are determined. The parameters essentially are a function of the void fraction. (orig.) [de

  5. Experimental validation of Monte Carlo calculations for organ dose

    International Nuclear Information System (INIS)

    Yalcintas, M.G.; Eckerman, K.F.; Warner, G.G.

    1980-01-01

    The problem of validating estimates of absorbed dose due to photon energy deposition is examined. The computational approaches used for the estimation of the photon energy deposition is examined. The limited data for validation of these approaches is discussed and suggestions made as to how better validation information might be obtained

  6. Copper benchmark experiment at the Frascati Neutron Generator for nuclear data validation

    Energy Technology Data Exchange (ETDEWEB)

    Angelone, M., E-mail: maurizio.angelone@enea.it; Flammini, D.; Loreti, S.; Moro, F.; Pillon, M.; Villari, R.

    2016-11-01

    Highlights: • A benchmark experiment was performed using pure copper with 14 MeV neutrons. • The experiment was performed at the Frascati Neutron Generator (FNG). • Activation foils, thermoluminescent dosimeters and scintillators were used to measure reactions rates (RR), nuclear heating and neutron spectra. • The paper presents the RR measurements and the post analysis using MCNP5 and JEFF-3.1.1, JEFF-3.2 and FENDL-3.1 libraries. • C/Es are presented showing the need for deep revision of Cu cross sections. - Abstract: A neutronics benchmark experiment on a pure Copper block (dimensions 60 × 70 × 60 cm{sup 3}), aimed at testing and validating the recent nuclear data libraries for fusion applications, was performed at the 14-MeV Frascati Neutron Generator (FNG) as part of a F4E specific grant (F4E-FPA-395-01) assigned to the European Consortium on Nuclear Data and Experimental Techniques. The relevant neutronics quantities (e.g., reaction rates, neutron flux spectra, doses, etc.) were measured using different experimental techniques and the results were compared to the calculated quantities using fusion relevant nuclear data libraries. This paper focuses on the analyses carried-out by ENEA through the activation foils techniques. {sup 197}Au(n,γ){sup 198}Au, {sup 186}W(n,γ){sup 187}W, {sup 115}In(n,n′){sup 115}In, {sup 58}Ni(n,p){sup 58}Co, {sup 27}Al(n,α){sup 24}Na, {sup 93}Nb(n,2n){sup 92}Nb{sup m} activation reactions were used. The foils were placed at eight different positions along the Cu block and irradiated with 14 MeV neutrons. Activation measurements were performed by means of High Purity Germanium (HPGe) detector. Detailed simulation of the experiment was carried-out using MCNP5 Monte Carlo code and the European JEFF-3.1.1 and 3.2 nuclear cross-sections data files for neutron transport and IRDFF-v1.05 library for the reaction rates in activation foils. The calculated reaction rates (C) were compared to the experimental quantities (E) and

  7. Validity of Linder Hypothesis in Bric Countries

    Directory of Open Access Journals (Sweden)

    Rana Atabay

    2016-03-01

    Full Text Available In this study, the theory of similarity in preferences (Linder hypothesis has been introduced and trade in BRIC countries has been examined whether the trade between these countries was valid for this hypothesis. Using the data for the period 1996 – 2010, the study applies to panel data analysis in order to provide evidence regarding the empirical validity of the Linder hypothesis for BRIC countries’ international trade. Empirical findings show that the trade between BRIC countries is in support of Linder hypothesis.

  8. Evaluation of radiochemical data usability

    International Nuclear Information System (INIS)

    Paar, J.G.; Porterfield, D.R.

    1997-04-01

    This procedure provides a framework for implementation of radiochemical data verification and validation for environmental remediation activities. It has been developed through participation of many individuals currently involved in analytical radiochemistry, radiochemical validation, and validation program development throughout the DOE complex. It should be regarded as a guidance to use in developing an implementable radiochemical validation strategy. This procedure provides specifications for developing and implementing a radiochemical validation methodology flexible enough to allow evaluation of data useability for project-specific Data Quality Objectives (DQO). Data produced by analytical methods for which this procedure provides limited guidance are classified as open-quotes non-routineclose quotes radionuclides and methods, and analyses by these methods may necessitate adoption of modified criteria from this procedure

  9. Application of Kramers-Kronig relationships for titanium impedance data validation in a Ringer's solution

    Directory of Open Access Journals (Sweden)

    Bastidas, D. M.

    2004-08-01

    Full Text Available This paper studies the applicability of Kramers-Kronig (KK relationships to assess the validity of real and imaginary impedance measurements for titanium in a Ringer's solution using the electrochemical impedance spectroscopy (EIS method. Nyquist and Bode plots showed a capacitive behaviour with a high impedance modulus including two time constants. Two procedures were employed in the implementation of the KK integral relationships based on an equivalent circuit which satisfies KK relationships and an ohmic resistance shunted to the measured EIS data. The titanium's EIS data satisfied the KK relationships.

    Este artículo estudia la aplicabilidad de las relaciones de Kramers-Kronig (KK al estudio de la validez de las medidas de impedancia (EIS, parte real y parte imaginaria, del titanio en contacto con la solución de Ringer. Los diagramas de Nyquist y Bode muestran un comportamiento capacitivo, con un módulo de impedancia elevado y con dos constantes de tiempo. En la implementación de las integrales de KK se emplearon dos procedimientos, que se basan en un circuito equivalente que cumple las relaciones de KK y en una resistencia óhmica en paralelo añadida a los datos de impedancia medidos. Los resultados de impedancia del titanio satisfacen las relaciones de KK.

  10. Flight code validation simulator

    Science.gov (United States)

    Sims, Brent A.

    1996-05-01

    An End-To-End Simulation capability for software development and validation of missile flight software on the actual embedded computer has been developed utilizing a 486 PC, i860 DSP coprocessor, embedded flight computer and custom dual port memory interface hardware. This system allows real-time interrupt driven embedded flight software development and checkout. The flight software runs in a Sandia Digital Airborne Computer and reads and writes actual hardware sensor locations in which Inertial Measurement Unit data resides. The simulator provides six degree of freedom real-time dynamic simulation, accurate real-time discrete sensor data and acts on commands and discretes from the flight computer. This system was utilized in the development and validation of the successful premier flight of the Digital Miniature Attitude Reference System in January of 1995 at the White Sands Missile Range on a two stage attitude controlled sounding rocket.

  11. Validity of Sensory Systems as Distinct Constructs

    OpenAIRE

    Su, Chia-Ting; Parham, L. Diane

    2014-01-01

    Confirmatory factor analysis testing whether sensory questionnaire items represented distinct sensory system constructs found, using data from two age groups, that such constructs can be measured validly using questionnaire data.

  12. 40 CFR 152.93 - Citation of a previously submitted valid study.

    Science.gov (United States)

    2010-07-01

    ... Data Submitters' Rights § 152.93 Citation of a previously submitted valid study. An applicant may demonstrate compliance for a data requirement by citing a valid study previously submitted to the Agency. The... the original data submitter, the applicant may cite the study only in accordance with paragraphs (b...

  13. The Moon Mineralogy Mapper (M3) imaging spectrometer for lunar science: Instrument description, calibration, on-orbit measurements, science data calibration and on-orbit validation

    Science.gov (United States)

    Green, R.O.; Pieters, C.; Mouroulis, P.; Eastwood, M.; Boardman, J.; Glavich, T.; Isaacson, P.; Annadurai, M.; Besse, S.; Barr, D.; Buratti, B.; Cate, D.; Chatterjee, A.; Clark, R.; Cheek, L.; Combe, J.; Dhingra, D.; Essandoh, V.; Geier, S.; Goswami, J.N.; Green, R.; Haemmerle, V.; Head, J.; Hovland, L.; Hyman, S.; Klima, R.; Koch, T.; Kramer, G.; Kumar, A.S.K.; Lee, Kenneth; Lundeen, S.; Malaret, E.; McCord, T.; McLaughlin, S.; Mustard, J.; Nettles, J.; Petro, N.; Plourde, K.; Racho, C.; Rodriquez, J.; Runyon, C.; Sellar, G.; Smith, C.; Sobel, H.; Staid, M.; Sunshine, J.; Taylor, L.; Thaisen, K.; Tompkins, S.; Tseng, H.; Vane, G.; Varanasi, P.; White, M.; Wilson, D.

    2011-01-01

    The NASA Discovery Moon Mineralogy Mapper imaging spectrometer was selected to pursue a wide range of science objectives requiring measurement of composition at fine spatial scales over the full lunar surface. To pursue these objectives, a broad spectral range imaging spectrometer with high uniformity and high signal-to-noise ratio capable of measuring compositionally diagnostic spectral absorption features from a wide variety of known and possible lunar materials was required. For this purpose the Moon Mineralogy Mapper imaging spectrometer was designed and developed that measures the spectral range from 430 to 3000 nm with 10 nm spectral sampling through a 24 degree field of view with 0.7 milliradian spatial sampling. The instrument has a signal-to-noise ratio of greater than 400 for the specified equatorial reference radiance and greater than 100 for the polar reference radiance. The spectral cross-track uniformity is >90% and spectral instantaneous field-of-view uniformity is >90%. The Moon Mineralogy Mapper was launched on Chandrayaan-1 on the 22nd of October. On the 18th of November 2008 the Moon Mineralogy Mapper was turned on and collected a first light data set within 24 h. During this early checkout period and throughout the mission the spacecraft thermal environment and orbital parameters varied more than expected and placed operational and data quality constraints on the measurements. On the 29th of August 2009, spacecraft communication was lost. Over the course of the flight mission 1542 downlinked data sets were acquired that provide coverage of more than 95% of the lunar surface. An end-to-end science data calibration system was developed and all measurements have been passed through this system and delivered to the Planetary Data System (PDS.NASA.GOV). An extensive effort has been undertaken by the science team to validate the Moon Mineralogy Mapper science measurements in the context of the mission objectives. A focused spectral, radiometric

  14. GPM Ground Validation: Pre to Post-Launch Era

    Science.gov (United States)

    Petersen, Walt; Skofronick-Jackson, Gail; Huffman, George

    2015-04-01

    NASA GPM Ground Validation (GV) activities have transitioned from the pre to post-launch era. Prior to launch direct validation networks and associated partner institutions were identified world-wide, covering a plethora of precipitation regimes. In the U.S. direct GV efforts focused on use of new operational products such as the NOAA Multi-Radar Multi-Sensor suite (MRMS) for TRMM validation and GPM radiometer algorithm database development. In the post-launch, MRMS products including precipitation rate, accumulation, types and data quality are being routinely generated to facilitate statistical GV of instantaneous (e.g., Level II orbit) and merged (e.g., IMERG) GPM products. Toward assessing precipitation column impacts on product uncertainties, range-gate to pixel-level validation of both Dual-Frequency Precipitation Radar (DPR) and GPM microwave imager data are performed using GPM Validation Network (VN) ground radar and satellite data processing software. VN software ingests quality-controlled volumetric radar datasets and geo-matches those data to coincident DPR and radiometer level-II data. When combined MRMS and VN datasets enable more comprehensive interpretation of both ground and satellite-based estimation uncertainties. To support physical validation efforts eight (one) field campaigns have been conducted in the pre (post) launch era. The campaigns span regimes from northern latitude cold-season snow to warm tropical rain. Most recently the Integrated Precipitation and Hydrology Experiment (IPHEx) took place in the mountains of North Carolina and involved combined airborne and ground-based measurements of orographic precipitation and hydrologic processes underneath the GPM Core satellite. One more U.S. GV field campaign (OLYMPEX) is planned for late 2015 and will address cold-season precipitation estimation, process and hydrology in the orographic and oceanic domains of western Washington State. Finally, continuous direct and physical validation

  15. Validation in the Absence of Observed Events.

    Science.gov (United States)

    Lathrop, John; Ezell, Barry

    2016-04-01

    This article addresses the problem of validating models in the absence of observed events, in the area of weapons of mass destruction terrorism risk assessment. We address that problem with a broadened definition of "validation," based on stepping "up" a level to considering the reason why decisionmakers seek validation, and from that basis redefine validation as testing how well the model can advise decisionmakers in terrorism risk management decisions. We develop that into two conditions: validation must be based on cues available in the observable world; and it must focus on what can be done to affect that observable world, i.e., risk management. That leads to two foci: (1) the real-world risk generating process, and (2) best use of available data. Based on our experience with nine WMD terrorism risk assessment models, we then describe three best use of available data pitfalls: SME confidence bias, lack of SME cross-referencing, and problematic initiation rates. Those two foci and three pitfalls provide a basis from which we define validation in this context in terms of four tests--Does the model: … capture initiation? … capture the sequence of events by which attack scenarios unfold? … consider unanticipated scenarios? … consider alternative causal chains? Finally, we corroborate our approach against three validation tests from the DOD literature: Is the model a correct representation of the process to be simulated? To what degree are the model results comparable to the real world? Over what range of inputs are the model results useful? © 2015 Society for Risk Analysis.

  16. Nuclear data for fusion: Validation of typical pre-processing methods for radiation transport calculations

    International Nuclear Information System (INIS)

    Hutton, T.; Sublet, J.C.; Morgan, L.; Leadbeater, T.W.

    2015-01-01

    Highlights: • We quantify the effect of processing nuclear data from ENDF to ACE format. • We consider the differences between fission and fusion angular distributions. • C-nat(n,el) at 2.0 MeV has a 0.6% deviation between original and processed data. • Fe-56(n,el) at 14.1 MeV has a 11.0% deviation between original and processed data. • Processed data do not accurately depict ENDF distributions for fusion energies. - Abstract: Nuclear data form the basis of the radiation transport codes used to design and simulate the behaviour of nuclear facilities, such as the ITER and DEMO fusion reactors. Typically these data and codes are biased towards fission and high-energy physics applications yet are still applied to fusion problems. With increasing interest in fusion applications, the lack of fusion specific codes and relevant data libraries is becoming increasingly apparent. Industry standard radiation transport codes require pre-processing of the evaluated data libraries prior to use in simulation. Historically these methods focus on speed of simulation at the cost of accurate data representation. For legacy applications this has not been a major concern, but current fusion needs differ significantly. Pre-processing reconstructs the differential and double differential interaction cross sections with a coarse binned structure, or more recently as a tabulated cumulative distribution function. This work looks at the validity of applying these processing methods to data used in fusion specific calculations in comparison to fission. The relative effects of applying this pre-processing mechanism, to both fission and fusion relevant reaction channels are demonstrated, and as such the poor representation of these distributions for the fusion energy regime. For the nat C(n,el) reaction at 2.0 MeV, the binned differential cross section deviates from the original data by 0.6% on average. For the 56 Fe(n,el) reaction at 14.1 MeV, the deviation increases to 11.0%. We

  17. [Risk Prediction Using Routine Data: Development and Validation of Multivariable Models Predicting 30- and 90-day Mortality after Surgical Treatment of Colorectal Cancer].

    Science.gov (United States)

    Crispin, Alexander; Strahwald, Brigitte; Cheney, Catherine; Mansmann, Ulrich

    2018-06-04

    Quality control, benchmarking, and pay for performance (P4P) require valid indicators and statistical models allowing adjustment for differences in risk profiles of the patient populations of the respective institutions. Using hospital remuneration data for measuring quality and modelling patient risks has been criticized by clinicians. Here we explore the potential of prediction models for 30- and 90-day mortality after colorectal cancer surgery based on routine data. Full census of a major statutory health insurer. Surgical departments throughout the Federal Republic of Germany. 4283 and 4124 insurants with major surgery for treatment of colorectal cancer during 2013 and 2014, respectively. Age, sex, primary and secondary diagnoses as well as tumor locations as recorded in the hospital remuneration data according to §301 SGB V. 30- and 90-day mortality. Elixhauser comorbidities, Charlson conditions, and Charlson scores were generated from the ICD-10 diagnoses. Multivariable prediction models were developed using a penalized logistic regression approach (logistic ridge regression) in a derivation set (patients treated in 2013). Calibration and discrimination of the models were assessed in an internal validation sample (patients treated in 2014) using calibration curves, Brier scores, receiver operating characteristic curves (ROC curves) and the areas under the ROC curves (AUC). 30- and 90-day mortality rates in the learning-sample were 5.7 and 8.4%, respectively. The corresponding values in the validation sample were 5.9% and once more 8.4%. Models based on Elixhauser comorbidities exhibited the highest discriminatory power with AUC values of 0.804 (95% CI: 0.776 -0.832) and 0.805 (95% CI: 0.782-0.828) for 30- and 90-day mortality. The Brier scores for these models were 0.050 (95% CI: 0.044-0.056) and 0.067 (95% CI: 0.060-0.074) and similar to the models based on Charlson conditions. Regardless of the model, low predicted probabilities were well calibrated, while

  18. Can we rely on cancer mortality data? Checking the validity of cervical cancer mortality data for Slovenia

    International Nuclear Information System (INIS)

    Primic Zakelj, M.; Pompe Kirn, V.; Skrlec, F.; Selb, J.

    2001-01-01

    Background. Valid inference on cervical cancer mortality is very difficult since - on the basis of death certificates - it is not always possible to distinguish between cervix, corpus and unspecified uterine cancer deaths. Our aim was to estimate the extent to which cervical cancer as the official cause of death reflects the true mortality from cervical cancer in Slovenia. Material and methods. The data on 2245 deaths from cervix, corpus uteri, and uterus-unspecified cancers for the period 1985-1999 were linked to the Cancer Registry of Slovenia database from the mortality database of Slovenia. Results. Officially, in the period 1985-1999, there were 878 cervical cancer deaths. The comparison of these causes of death with the cancer sites registered in the Cancer Registry revealed that they include only 87.7% patients with a previous diagnosis of cervical cancer. Of 650 corpus uteri cancer deaths, 17. 1 % of patients were registered to have cervical cancer, and of 717 unspecified uterine cancer deaths, 31.4% were registered. Taking into account the correctly identified cervical cancer cases among cervical cancer deaths and misclassified cervical cancer deaths as corpus uteri and unspecified uterine, the corrected number of deaths would be 1106. Conclusions. When evaluating the impact of cervical cancer mortality from national mortality rates, the stated underestimation should be taken into account. However, this does not hold for some other cancers. (author)

  19. Design for validation: An approach to systems validation

    Science.gov (United States)

    Carter, William C.; Dunham, Janet R.; Laprie, Jean-Claude; Williams, Thomas; Howden, William; Smith, Brian; Lewis, Carl M. (Editor)

    1989-01-01

    Every complex system built is validated in some manner. Computer validation begins with review of the system design. As systems became too complicated for one person to review, validation began to rely on the application of adhoc methods by many individuals. As the cost of the changes mounted and the expense of failure increased, more organized procedures became essential. Attempts at devising and carrying out those procedures showed that validation is indeed a difficult technical problem. The successful transformation of the validation process into a systematic series of formally sound, integrated steps is necessary if the liability inherent in the future digita-system-based avionic and space systems is to be minimized. A suggested framework and timetable for the transformtion are presented. Basic working definitions of two pivotal ideas (validation and system life-cyle) are provided and show how the two concepts interact. Many examples are given of past and present validation activities by NASA and others. A conceptual framework is presented for the validation process. Finally, important areas are listed for ongoing development of the validation process at NASA Langley Research Center.

  20. CANDU radiotoxicity inventories estimation: A calculated experiment cross-check for data verification and validation

    International Nuclear Information System (INIS)

    Pavelescu, Alexandru Octavian; Cepraga, Dan Gabriel

    2007-01-01

    This paper is related to the Clearance Potential Index, Ingestion and Inhalation Hazard Factors of the nuclear spent fuel and radioactive wastes. This study required a complex activity that consisted of various phases such us: the acquisition, setting up, validation and application of procedures, codes and libraries. The paper reflects the validation phase of this study. Its objective was to compare the measured inventories of selected actinide and fission products radionuclides in an element from a Pickering CANDU reactor with inventories predicted using a recent version of the ORIGEN-ARP from SCALE 5 coupled with the time dependent cross sections library, CANDU 28.lib, produced by the sequence SAS2H of SCALE 4.4a. In this way, the procedures, codes and libraries for the characterization of radioactive material in terms of radioactive inventories, clearance, and biological hazard factors are being qualified and validated, in support for the safety management of the radioactive wastes. (authors)