WorldWideScience

Sample records for test based models

  1. Model-based security testing

    OpenAIRE

    Schieferdecker, Ina; Großmann, Jürgen; Schneider, Martin

    2012-01-01

    Security testing aims at validating software system requirements related to security properties like confidentiality, integrity, authentication, authorization, availability, and non-repudiation. Although security testing techniques are available for many years, there has been little approaches that allow for specification of test cases at a higher level of abstraction, for enabling guidance on test identification and specification as well as for automated test generation. Model-based security...

  2. Model-Based Security Testing

    Directory of Open Access Journals (Sweden)

    Ina Schieferdecker

    2012-02-01

    Full Text Available Security testing aims at validating software system requirements related to security properties like confidentiality, integrity, authentication, authorization, availability, and non-repudiation. Although security testing techniques are available for many years, there has been little approaches that allow for specification of test cases at a higher level of abstraction, for enabling guidance on test identification and specification as well as for automated test generation. Model-based security testing (MBST is a relatively new field and especially dedicated to the systematic and efficient specification and documentation of security test objectives, security test cases and test suites, as well as to their automated or semi-automated generation. In particular, the combination of security modelling and test generation approaches is still a challenge in research and of high interest for industrial applications. MBST includes e.g. security functional testing, model-based fuzzing, risk- and threat-oriented testing, and the usage of security test patterns. This paper provides a survey on MBST techniques and the related models as well as samples of new methods and tools that are under development in the European ITEA2-project DIAMONDS.

  3. Model-based testing for software safety

    NARCIS (Netherlands)

    Gurbuz, Havva Gulay; Tekinerdogan, Bedir

    2017-01-01

    Testing safety-critical systems is crucial since a failure or malfunction may result in death or serious injuries to people, equipment, or environment. An important challenge in testing is the derivation of test cases that can identify the potential faults. Model-based testing adopts models of a

  4. Traceability in Model-Based Testing

    Directory of Open Access Journals (Sweden)

    Mathew George

    2012-11-01

    Full Text Available The growing complexities of software and the demand for shorter time to market are two important challenges that face today’s IT industry. These challenges demand the increase of both productivity and quality of software. Model-based testing is a promising technique for meeting these challenges. Traceability modeling is a key issue and challenge in model-based testing. Relationships between the different models will help to navigate from one model to another, and trace back to the respective requirements and the design model when the test fails. In this paper, we present an approach for bridging the gaps between the different models in model-based testing. We propose relation definition markup language (RDML for defining the relationships between models.

  5. Model-based testing for embedded systems

    CERN Document Server

    Zander, Justyna; Mosterman, Pieter J

    2011-01-01

    What the experts have to say about Model-Based Testing for Embedded Systems: "This book is exactly what is needed at the exact right time in this fast-growing area. From its beginnings over 10 years ago of deriving tests from UML statecharts, model-based testing has matured into a topic with both breadth and depth. Testing embedded systems is a natural application of MBT, and this book hits the nail exactly on the head. Numerous topics are presented clearly, thoroughly, and concisely in this cutting-edge book. The authors are world-class leading experts in this area and teach us well-used

  6. Observation-Based Modeling for Model-Based Testing

    NARCIS (Netherlands)

    Kanstrén, T.; Piel, E.; Gross, H.G.

    2009-01-01

    One of the single most important reasons that modeling and modelbased testing are not yet common practice in industry is the perceived difficulty of making the models up to the level of detail and quality required for their automated processing. Models unleash their full potential only through

  7. Divergence-based tests for model diagnostic

    Czech Academy of Sciences Publication Activity Database

    Hobza, Tomáš; Esteban, M. D.; Morales, D.; Marhuenda, Y.

    2008-01-01

    Roč. 78, č. 13 (2008), s. 1702-1710 ISSN 0167-7152 R&D Projects: GA MŠk 1M0572 Grant - others:Instituto Nacional de Estadistica (ES) MTM2006-05693 Institutional research plan: CEZ:AV0Z10750506 Keywords : goodness of fit * devergence statistics * GLM * model checking * bootstrap Subject RIV: BB - Applied Statistics, Operational Research Impact factor: 0.445, year: 2008 http://library.utia.cas.cz/separaty/2008/SI/hobza-divergence-based%20tests%20for%20model%20diagnostic.pdf

  8. Model-Based Software Testing for Object-Oriented Software

    Science.gov (United States)

    Biju, Soly Mathew

    2008-01-01

    Model-based testing is one of the best solutions for testing object-oriented software. It has a better test coverage than other testing styles. Model-based testing takes into consideration behavioural aspects of a class, which are usually unchecked in other testing methods. An increase in the complexity of software has forced the software industry…

  9. Model Based Analysis and Test Generation for Flight Software

    Science.gov (United States)

    Pasareanu, Corina S.; Schumann, Johann M.; Mehlitz, Peter C.; Lowry, Mike R.; Karsai, Gabor; Nine, Harmon; Neema, Sandeep

    2009-01-01

    We describe a framework for model-based analysis and test case generation in the context of a heterogeneous model-based development paradigm that uses and combines Math- Works and UML 2.0 models and the associated code generation tools. This paradigm poses novel challenges to analysis and test case generation that, to the best of our knowledge, have not been addressed before. The framework is based on a common intermediate representation for different modeling formalisms and leverages and extends model checking and symbolic execution tools for model analysis and test case generation, respectively. We discuss the application of our framework to software models for a NASA flight mission.

  10. Tree-Based Global Model Tests for Polytomous Rasch Models

    Science.gov (United States)

    Komboz, Basil; Strobl, Carolin; Zeileis, Achim

    2018-01-01

    Psychometric measurement models are only valid if measurement invariance holds between test takers of different groups. Global model tests, such as the well-established likelihood ratio (LR) test, are sensitive to violations of measurement invariance, such as differential item functioning and differential step functioning. However, these…

  11. Model-Based GUI Testing Using Uppaal at Novo Nordisk

    Science.gov (United States)

    Hjort, Ulrik H.; Illum, Jacob; Larsen, Kim G.; Petersen, Michael A.; Skou, Arne

    This paper details a collaboration between Aalborg University and Novo Nordiskin developing an automatic model-based test generation tool for system testing of the graphical user interface of a medical device on an embedded platform. The tool takes as input an UML Statemachine model and generates a test suite satisfying some testing criterion, such as edge or state coverage, and converts the individual test case into a scripting language that can be automatically executed against the target. The tool has significantly reduced the time required for test construction and generation, and reduced the number of test scripts while increasing the coverage.

  12. Model-Based GUI Testing Using Uppaal at Novo Nordisk

    DEFF Research Database (Denmark)

    H. Hjort, Ulrik; Rasmussen, Jacob Illum; Larsen, Kim Guldstrand

    2009-01-01

    This paper details a collaboration between Aalborg University and Novo Nordiskin developing an automatic model-based test generation tool for system testing of the graphical user interface of a medical device on an embedded platform. The tool takes as input an UML Statemachine model and generates...

  13. Towards model-based testing of electronic funds transfer systems

    OpenAIRE

    Asaadi, H.R.; Khosravi, R.; Mousavi, M.R.; Noroozi, N.

    2010-01-01

    We report on our first experience with applying model-based testing techniques to an operational Electronic Funds Transfer (EFT) switch. The goal is to test the conformance of the EFT switch to the standard flows described by the ISO 8583 standard. To this end, we first make a formalization of the transaction flows specified in the ISO 8583 standard in terms of a Labeled Transition System (LTS). This formalization paves the way for model-based testing based on the formal notion of Input-Outpu...

  14. Design Of Computer Based Test Using The Unified Modeling Language

    Science.gov (United States)

    Tedyyana, Agus; Danuri; Lidyawati

    2017-12-01

    The Admission selection of Politeknik Negeri Bengkalis through interest and talent search (PMDK), Joint Selection of admission test for state Polytechnics (SB-UMPN) and Independent (UM-Polbeng) were conducted by using paper-based Test (PBT). Paper Based Test model has some weaknesses. They are wasting too much paper, the leaking of the questios to the public, and data manipulation of the test result. This reasearch was Aimed to create a Computer-based Test (CBT) models by using Unified Modeling Language (UML) the which consists of Use Case diagrams, Activity diagram and sequence diagrams. During the designing process of the application, it is important to pay attention on the process of giving the password for the test questions before they were shown through encryption and description process. RSA cryptography algorithm was used in this process. Then, the questions shown in the questions banks were randomized by using the Fisher-Yates Shuffle method. The network architecture used in Computer Based test application was a client-server network models and Local Area Network (LAN). The result of the design was the Computer Based Test application for admission to the selection of Politeknik Negeri Bengkalis.

  15. Automated model-based testing of hybrid systems

    NARCIS (Netherlands)

    Osch, van M.P.W.J.

    2009-01-01

    In automated model-based input-output conformance testing, tests are automati- cally generated from a speci¯cation and automatically executed on an implemen- tation. Input is applied to the implementation and output is observed from the implementation. If the observed output is allowed according to

  16. Towards model-based testing of electronic funds transfer systems

    NARCIS (Netherlands)

    Asaadi, H.R.; Khosravi, R.; Mousavi, M.R.; Noroozi, N.; Arbab, F.; Sirjani, M.

    2012-01-01

    We report on our first experience with applying model-based testing techniques to an operational Electronic Funds Transfer (EFT) switch. The goal is to test the conformance of the EFT switch to the standard flows described by the ISO 8583 standard. To this end, we first make a formalization of the

  17. Towards model-based testing of electronic funds transfer systems

    NARCIS (Netherlands)

    Asaadi, H.R.; Khosravi, R.; Mousavi, M.R.; Noroozi, N.

    2010-01-01

    We report on our first experience with applying model-based testing techniques to an operational Electronic Funds Transfer (EFT) switch. The goal is to test the conformance of the EFT switch to the standard flows described by the ISO 8583 standard. To this end, we first make a formalization of the

  18. Testing R&D-Based Endogenous Growth Models

    DEFF Research Database (Denmark)

    Kruse-Andersen, Peter Kjær

    2017-01-01

    R&D-based growth models are tested using US data for the period 1953-2014. A general growth model is developed which nests the model varieties of interest. The model implies a cointegrating relationship between multifactor productivity, research intensity, and employment. This relationship...... is estimated using cointegrated VAR models. The results provide evidence against the widely used fully endogenous variety and in favor of the semi-endogenous variety. Forecasts based on the empirical estimates suggest that the slowdown in US productivity growth will continue. Particularly, the annual long...

  19. A model based security testing method for protocol implementation.

    Science.gov (United States)

    Fu, Yu Long; Xin, Xiao Long

    2014-01-01

    The security of protocol implementation is important and hard to be verified. Since the penetration testing is usually based on the experience of the security tester and the specific protocol specifications, a formal and automatic verification method is always required. In this paper, we propose an extended model of IOLTS to describe the legal roles and intruders of security protocol implementations, and then combine them together to generate the suitable test cases to verify the security of protocol implementation.

  20. Model-based automated testing of critical PLC programs.

    CERN Document Server

    Fernández Adiego, B; Tournier, J-C; González Suárez, V M; Bliudze, S

    2014-01-01

    Testing of critical PLC (Programmable Logic Controller) programs remains a challenging task for control system engineers as it can rarely be automated. This paper proposes a model based approach which uses the BIP (Behavior, Interactions and Priorities) framework to perform automated testing of PLC programs developed with the UNICOS (UNified Industrial COntrol System) framework. This paper defines the translation procedure and rules from UNICOS to BIP which can be fully automated in order to hide the complexity of the underlying model from the control engineers. The approach is illustrated and validated through the study of a water treatment process.

  1. Test-Driven, Model-Based Systems Engineering

    DEFF Research Database (Denmark)

    Munck, Allan

    Hearing systems have evolved over many years from simple mechanical devices (horns) to electronic units consisting of microphones, amplifiers, analog filters, loudspeakers, batteries, etc. Digital signal processors replaced analog filters to provide better performance end new features. Central....... This thesis concerns methods for identifying, selecting and implementing tools for various aspects of model-based systems engineering. A comprehensive method was proposed that include several novel steps such as techniques for analyzing the gap between requirements and tool capabilities. The method...... was verified with good results in two case studies for selection of a traceability tool (single-tool scenario) and a set of modeling tools (multi-tool scenarios). Models must be subjected to testing to allow engineers to predict functionality and performance of systems. Test-first strategies are known...

  2. Bayesian models based on test statistics for multiple hypothesis testing problems.

    Science.gov (United States)

    Ji, Yuan; Lu, Yiling; Mills, Gordon B

    2008-04-01

    We propose a Bayesian method for the problem of multiple hypothesis testing that is routinely encountered in bioinformatics research, such as the differential gene expression analysis. Our algorithm is based on modeling the distributions of test statistics under both null and alternative hypotheses. We substantially reduce the complexity of the process of defining posterior model probabilities by modeling the test statistics directly instead of modeling the full data. Computationally, we apply a Bayesian FDR approach to control the number of rejections of null hypotheses. To check if our model assumptions for the test statistics are valid for various bioinformatics experiments, we also propose a simple graphical model-assessment tool. Using extensive simulations, we demonstrate the performance of our models and the utility of the model-assessment tool. In the end, we apply the proposed methodology to an siRNA screening and a gene expression experiment.

  3. Strengthening Theoretical Testing in Criminology Using Agent-based Modeling.

    Science.gov (United States)

    Johnson, Shane D; Groff, Elizabeth R

    2014-07-01

    The Journal of Research in Crime and Delinquency ( JRCD ) has published important contributions to both criminological theory and associated empirical tests. In this article, we consider some of the challenges associated with traditional approaches to social science research, and discuss a complementary approach that is gaining popularity-agent-based computational modeling-that may offer new opportunities to strengthen theories of crime and develop insights into phenomena of interest. Two literature reviews are completed. The aim of the first is to identify those articles published in JRCD that have been the most influential and to classify the theoretical perspectives taken. The second is intended to identify those studies that have used an agent-based model (ABM) to examine criminological theories and to identify which theories have been explored. Ecological theories of crime pattern formation have received the most attention from researchers using ABMs, but many other criminological theories are amenable to testing using such methods. Traditional methods of theory development and testing suffer from a number of potential issues that a more systematic use of ABMs-not without its own issues-may help to overcome. ABMs should become another method in the criminologists toolbox to aid theory testing and falsification.

  4. Model Checking and Model-based Testing in the Railway Domain

    DEFF Research Database (Denmark)

    Haxthausen, Anne Elisabeth; Peleska, Jan

    2015-01-01

    This chapter describes some approaches and emerging trends for verification and model-based testing of railway control systems. We describe state-of-the-art methods and associated tools for verifying interlocking systems and their configuration data, using bounded model checking and k...... with good test strength are explained. Interlocking systems represent just one class of many others, where concrete system instances are created from generic representations, using configuration data for determining the behaviour of the instances. We explain how the systematic transition from generic...... to concrete instances in the development path is complemented by associated transitions in the verification and testing paths....

  5. Research on Generating Method of Embedded Software Test Document Based on Dynamic Model

    Science.gov (United States)

    Qu, MingCheng; Wu, XiangHu; Tao, YongChao; Liu, Ying

    2018-03-01

    This paper provides a dynamic model-based test document generation method for embedded software that provides automatic generation of two documents: test requirements specification documentation and configuration item test documentation. This method enables dynamic test requirements to be implemented in dynamic models, enabling dynamic test demand tracking to be easily generated; able to automatically generate standardized, standardized test requirements and test documentation, improved document-related content inconsistency and lack of integrity And other issues, improve the efficiency.

  6. Towards a Development Environment for Model Based Test Design

    OpenAIRE

    Jing, Han

    2008-01-01

    Within the UP IP I&V organization there is high focus on increasing the ability to predict product quality in a cost efficient way. Test automation has therefore been an important enabler for us. The IP test design environment is continuously evolving and the investigations will show which improvements that is most important to implement in short and long term. In Ericsson UP IP I&V, the test automation framework environments are severed to complete some process by automated method, f...

  7. Automatic Model-Based Generation of Parameterized Test Cases Using Data Abstraction

    NARCIS (Netherlands)

    Calamé, Jens R.; Ioustinova, Natalia; Romijn, J.M.T.; Smith, G.; van de Pol, Jan Cornelis

    2007-01-01

    Developing test suites is a costly and error-prone process. Model-based test generation tools facilitate this process by automatically generating test cases from system models. The applicability of these tools, however, depends on the size of the target systems. Here, we propose an approach to

  8. Model-based development and testing of advertising messages

    DEFF Research Database (Denmark)

    Bech-Larsen, Tino

    2000-01-01

    The implementation of valid and comprehensible guidelines for message development potentially enhances the effects of advertising messages and improves the possibility of measuring such effects. Moreover, such guidelines also have potential implications for the managerial communication processes...... (client-agency and intra-agency) involved in the development of advertising messages. The purpose of the study described in this paper is to compare the development and effects of two campaign proposals, with the common aim of increasing the consumption of apples among young Danes (18 to 35 years of age......-test with the target group (n=500), as well as interviews with the involved advertising agency and client2 staff....

  9. Model-Based Testing of a Reactive System with Coloured Petri Nets

    DEFF Research Database (Denmark)

    Tjell, Simon

    2006-01-01

    In this paper, a reactive and nondeterministic system is tested. This is doneby applying a generic model that has been specified as a configurable Coloured PetriNet. In this way, model-based testing is possible for a wide class of reactive system atthe level of discrete events. Concurrently...

  10. OOK power model based dynamic error testing for smart electricity meter

    International Nuclear Information System (INIS)

    Wang, Xuewei; Chen, Jingxia; Jia, Xiaolu; Zhu, Meng; Yuan, Ruiming; Jiang, Zhenyu

    2017-01-01

    This paper formulates the dynamic error testing problem for a smart meter, with consideration and investigation of both the testing signal and the dynamic error testing method. To solve the dynamic error testing problems, the paper establishes an on-off-keying (OOK) testing dynamic current model and an OOK testing dynamic load energy (TDLE) model. Then two types of TDLE sequences and three modes of OOK testing dynamic power are proposed. In addition, a novel algorithm, which helps to solve the problem of dynamic electric energy measurement’s traceability, is derived for dynamic errors. Based on the above researches, OOK TDLE sequence generation equipment is developed and a dynamic error testing system is constructed. Using the testing system, five kinds of meters were tested in the three dynamic power modes. The test results show that the dynamic error is closely related to dynamic power mode and the measurement uncertainty is 0.38%. (paper)

  11. OOK power model based dynamic error testing for smart electricity meter

    Science.gov (United States)

    Wang, Xuewei; Chen, Jingxia; Yuan, Ruiming; Jia, Xiaolu; Zhu, Meng; Jiang, Zhenyu

    2017-02-01

    This paper formulates the dynamic error testing problem for a smart meter, with consideration and investigation of both the testing signal and the dynamic error testing method. To solve the dynamic error testing problems, the paper establishes an on-off-keying (OOK) testing dynamic current model and an OOK testing dynamic load energy (TDLE) model. Then two types of TDLE sequences and three modes of OOK testing dynamic power are proposed. In addition, a novel algorithm, which helps to solve the problem of dynamic electric energy measurement’s traceability, is derived for dynamic errors. Based on the above researches, OOK TDLE sequence generation equipment is developed and a dynamic error testing system is constructed. Using the testing system, five kinds of meters were tested in the three dynamic power modes. The test results show that the dynamic error is closely related to dynamic power mode and the measurement uncertainty is 0.38%.

  12. Empirical Modeling of Lithium-ion Batteries Based on Electrochemical Impedance Spectroscopy Tests

    International Nuclear Information System (INIS)

    Samadani, Ehsan; Farhad, Siamak; Scott, William; Mastali, Mehrdad; Gimenez, Leonardo E.; Fowler, Michael; Fraser, Roydon A.

    2015-01-01

    Highlights: • Two commercial Lithium-ion batteries are studied through HPPC and EIS tests. • An equivalent circuit model is developed for a range of operating conditions. • This model improves the current battery empirical models for vehicle applications • This model is proved to be efficient in terms of predicting HPPC test resistances. - ABSTRACT: An empirical model for commercial lithium-ion batteries is developed based on electrochemical impedance spectroscopy (EIS) tests. An equivalent circuit is established according to EIS test observations at various battery states of charge and temperatures. A Laplace transfer time based model is developed based on the circuit which can predict the battery operating output potential difference in battery electric and plug-in hybrid vehicles at various operating conditions. This model demonstrates up to 6% improvement compared to simple resistance and Thevenin models and is suitable for modeling and on-board controller purposes. Results also show that this model can be used to predict the battery internal resistance obtained from hybrid pulse power characterization (HPPC) tests to within 20 percent, making it suitable for low to medium fidelity powertrain design purposes. In total, this simple battery model can be employed as a real-time model in electrified vehicle battery management systems

  13. Zero-inflated Poisson model based likelihood ratio test for drug safety signal detection.

    Science.gov (United States)

    Huang, Lan; Zheng, Dan; Zalkikar, Jyoti; Tiwari, Ram

    2017-02-01

    In recent decades, numerous methods have been developed for data mining of large drug safety databases, such as Food and Drug Administration's (FDA's) Adverse Event Reporting System, where data matrices are formed by drugs such as columns and adverse events as rows. Often, a large number of cells in these data matrices have zero cell counts and some of them are "true zeros" indicating that the drug-adverse event pairs cannot occur, and these zero counts are distinguished from the other zero counts that are modeled zero counts and simply indicate that the drug-adverse event pairs have not occurred yet or have not been reported yet. In this paper, a zero-inflated Poisson model based likelihood ratio test method is proposed to identify drug-adverse event pairs that have disproportionately high reporting rates, which are also called signals. The maximum likelihood estimates of the model parameters of zero-inflated Poisson model based likelihood ratio test are obtained using the expectation and maximization algorithm. The zero-inflated Poisson model based likelihood ratio test is also modified to handle the stratified analyses for binary and categorical covariates (e.g. gender and age) in the data. The proposed zero-inflated Poisson model based likelihood ratio test method is shown to asymptotically control the type I error and false discovery rate, and its finite sample performance for signal detection is evaluated through a simulation study. The simulation results show that the zero-inflated Poisson model based likelihood ratio test method performs similar to Poisson model based likelihood ratio test method when the estimated percentage of true zeros in the database is small. Both the zero-inflated Poisson model based likelihood ratio test and likelihood ratio test methods are applied to six selected drugs, from the 2006 to 2011 Adverse Event Reporting System database, with varying percentages of observed zero-count cells.

  14. Rank-based Tests of the Cointegrating Rank in Semiparametric Error Correction Models

    NARCIS (Netherlands)

    Hallin, M.; van den Akker, R.; Werker, B.J.M.

    2012-01-01

    Abstract: This paper introduces rank-based tests for the cointegrating rank in an Error Correction Model with i.i.d. elliptical innovations. The tests are asymptotically distribution-free, and their validity does not depend on the actual distribution of the innovations. This result holds despite the

  15. An Improved Test Selection Optimization Model Based on Fault Ambiguity Group Isolation and Chaotic Discrete PSO

    Directory of Open Access Journals (Sweden)

    Xiaofeng Lv

    2018-01-01

    Full Text Available Sensor data-based test selection optimization is the basis for designing a test work, which ensures that the system is tested under the constraint of the conventional indexes such as fault detection rate (FDR and fault isolation rate (FIR. From the perspective of equipment maintenance support, the ambiguity isolation has a significant effect on the result of test selection. In this paper, an improved test selection optimization model is proposed by considering the ambiguity degree of fault isolation. In the new model, the fault test dependency matrix is adopted to model the correlation between the system fault and the test group. The objective function of the proposed model is minimizing the test cost with the constraint of FDR and FIR. The improved chaotic discrete particle swarm optimization (PSO algorithm is adopted to solve the improved test selection optimization model. The new test selection optimization model is more consistent with real complicated engineering systems. The experimental result verifies the effectiveness of the proposed method.

  16. Model-based testing in powertrain development; Modellgestuetzte Erprobungsmethodik in der Antriebsstrangentwicklung

    Energy Technology Data Exchange (ETDEWEB)

    Albers, A.; Schyr, C. [Inst. fuer Produktentwicklung der Univ. Karlsruhe (T.H.) (Germany)

    2005-07-01

    The paper describes a new methodical approach for a model-based testing of powertrain components in vehicle development. The presented methodology is based on a holistic model environment which covers the major dynamic effects of the vehicle in an early development phase and refines the models during the testing phase on the different test bed configurations. This allows a realistic loading of the mechanical components and their electronic control units in parallel to a simulation based analysis of design and application variants in the mechanics and software and their influence onto the complete vehicle. In the first application example the development of a pre-adjustable transmission for passenger cars is presented. In the second example the testing concept for tracked vehicles with hydrostatic drivetrain is described. (orig.)

  17. A Rigorous Temperature-Dependent Stochastic Modelling and Testing for MEMS-Based Inertial Sensor Errors

    Directory of Open Access Journals (Sweden)

    Spiros Pagiatakis

    2009-10-01

    Full Text Available In this paper, we examine the effect of changing the temperature points on MEMS-based inertial sensor random error. We collect static data under different temperature points using a MEMS-based inertial sensor mounted inside a thermal chamber. Rigorous stochastic models, namely Autoregressive-based Gauss-Markov (AR-based GM models are developed to describe the random error behaviour. The proposed AR-based GM model is initially applied to short stationary inertial data to develop the stochastic model parameters (correlation times. It is shown that the stochastic model parameters of a MEMS-based inertial unit, namely the ADIS16364, are temperature dependent. In addition, field kinematic test data collected at about 17 °C are used to test the performance of the stochastic models at different temperature points in the filtering stage using Unscented Kalman Filter (UKF. It is shown that the stochastic model developed at 20 °C provides a more accurate inertial navigation solution than the ones obtained from the stochastic models developed at −40 °C, −20 °C, 0 °C, +40 °C, and +60 °C. The temperature dependence of the stochastic model is significant and should be considered at all times to obtain optimal navigation solution for MEMS-based INS/GPS integration.

  18. A Rigorous Temperature-Dependent Stochastic Modelling and Testing for MEMS-Based Inertial Sensor Errors.

    Science.gov (United States)

    El-Diasty, Mohammed; Pagiatakis, Spiros

    2009-01-01

    In this paper, we examine the effect of changing the temperature points on MEMS-based inertial sensor random error. We collect static data under different temperature points using a MEMS-based inertial sensor mounted inside a thermal chamber. Rigorous stochastic models, namely Autoregressive-based Gauss-Markov (AR-based GM) models are developed to describe the random error behaviour. The proposed AR-based GM model is initially applied to short stationary inertial data to develop the stochastic model parameters (correlation times). It is shown that the stochastic model parameters of a MEMS-based inertial unit, namely the ADIS16364, are temperature dependent. In addition, field kinematic test data collected at about 17 °C are used to test the performance of the stochastic models at different temperature points in the filtering stage using Unscented Kalman Filter (UKF). It is shown that the stochastic model developed at 20 °C provides a more accurate inertial navigation solution than the ones obtained from the stochastic models developed at -40 °C, -20 °C, 0 °C, +40 °C, and +60 °C. The temperature dependence of the stochastic model is significant and should be considered at all times to obtain optimal navigation solution for MEMS-based INS/GPS integration.

  19. A GMM-Based Test for Normal Disturbances of the Heckman Sample Selection Model

    Directory of Open Access Journals (Sweden)

    Michael Pfaffermayr

    2014-10-01

    Full Text Available The Heckman sample selection model relies on the assumption of normal and homoskedastic disturbances. However, before considering more general, alternative semiparametric models that do not need the normality assumption, it seems useful to test this assumption. Following Meijer and Wansbeek (2007, the present contribution derives a GMM-based pseudo-score LM test on whether the third and fourth moments of the disturbances of the outcome equation of the Heckman model conform to those implied by the truncated normal distribution. The test is easy to calculate and in Monte Carlo simulations it shows good performance for sample sizes of 1000 or larger.

  20. A Method for Modeling the Virtual Instrument Automatic Test System Based on the Petri Net

    Institute of Scientific and Technical Information of China (English)

    MA Min; CHEN Guang-ju

    2005-01-01

    Virtual instrument is playing the important role in automatic test system. This paper introduces a composition of a virtual instrument automatic test system and takes the VXIbus based a test software platform which is developed by CAT lab of the UESTC as an example. Then a method to model this system based on Petri net is proposed. Through this method, we can analyze the test task scheduling to prevent the deadlock or resources conflict. At last, this paper analyzes the feasibility of this method.

  1. Top-Down and Bottom-Up Approach for Model-Based Testing of Product Lines

    Directory of Open Access Journals (Sweden)

    Stephan Weißleder

    2013-03-01

    Full Text Available Systems tend to become more and more complex. This has a direct impact on system engineering processes. Two of the most important phases in these processes are requirements engineering and quality assurance. Two significant complexity drivers located in these phases are the growing number of product variants that have to be integrated into the requirements engineering and the ever growing effort for manual test design. There are modeling techniques to deal with both complexity drivers like, e.g., feature modeling and model-based test design. Their combination, however, has been seldom the focus of investigation. In this paper, we present two approaches to combine feature modeling and model-based testing as an efficient quality assurance technique for product lines. We present the corresponding difficulties and approaches to overcome them. All explanations are supported by an example of an online shop product line.

  2. Model-based testing with UML applied to a roaming algorithm for bluetooth devices.

    Science.gov (United States)

    Dai, Zhen Ru; Grabowski, Jens; Neukirchen, Helmut; Pals, Holger

    2004-11-01

    In late 2001, the Object Management Group issued a Request for Proposal to develop a testing profile for UML 2.0. In June 2003, the work on the UML 2.0 Testing Profile was finally adopted by the OMG. Since March 2004, it has become an official standard of the OMG. The UML 2.0 Testing Profile provides support for UML based model-driven testing. This paper introduces a methodology on how to use the testing profile in order to modify and extend an existing UML design model for test issues. The application of the methodology will be explained by applying it to an existing UML Model for a Bluetooth device.

  3. Using a data base management system for modelling SSME test history data

    Science.gov (United States)

    Abernethy, K.

    1985-01-01

    The usefulness of a data base management system (DBMS) for modelling historical test data for the complete series of static test firings for the Space Shuttle Main Engine (SSME) was assessed. From an analysis of user data base query requirements, it became clear that a relational DMBS which included a relationally complete query language would permit a model satisfying the query requirements. Representative models and sample queries are discussed. A list of environment-particular evaluation criteria for the desired DBMS was constructed; these criteria include requirements in the areas of user-interface complexity, program independence, flexibility, modifiability, and output capability. The evaluation process included the construction of several prototype data bases for user assessement. The systems studied, representing the three major DBMS conceptual models, were: MIRADS, a hierarchical system; DMS-1100, a CODASYL-based network system; ORACLE, a relational system; and DATATRIEVE, a relational-type system.

  4. Cyber-Physical Energy Systems Modeling, Test Specification, and Co-Simulation Based Testing

    DEFF Research Database (Denmark)

    van der Meer, A. A.; Palensky, P.; Heussen, Kai

    2017-01-01

    The gradual deployment of intelligent and coordinated devices in the electrical power system needs careful investigation of the interactions between the various domains involved. Especially due to the coupling between ICT and power systems a holistic approach for testing and validating is required....... Taking existing (quasi-) standardised smart grid system and test specification methods as a starting point, we are developing a holistic testing and validation approach that allows a very flexible way of assessing the system level aspects by various types of experiments (including virtual, real......, and mixed lab settings). This paper describes the formal holistic test case specification method and applies it to a particular co-simulation experimental setup. The various building blocks of such a simulation (i.e., FMI, mosaik, domain-specific simulation federates) are covered in more detail...

  5. Monte Carlo tests of the Rasch model based on scalability coefficients

    DEFF Research Database (Denmark)

    Christensen, Karl Bang; Kreiner, Svend

    2010-01-01

    that summarizes the number of Guttman errors in the data matrix. These coefficients are shown to yield efficient tests of the Rasch model using p-values computed using Markov chain Monte Carlo methods. The power of the tests of unequal item discrimination, and their ability to distinguish between local dependence......For item responses fitting the Rasch model, the assumptions underlying the Mokken model of double monotonicity are met. This makes non-parametric item response theory a natural starting-point for Rasch item analysis. This paper studies scalability coefficients based on Loevinger's H coefficient...

  6. Experimental Study of Dowel Bar Alternatives Based on Similarity Model Test

    Directory of Open Access Journals (Sweden)

    Chichun Hu

    2017-01-01

    Full Text Available In this study, a small-scaled accelerated loading test based on similarity theory and Accelerated Pavement Analyzer was developed to evaluate dowel bars with different materials and cross-sections. Jointed concrete specimen consisting of one dowel was designed as scaled model for the test, and each specimen was subjected to 864 thousand loading cycles. Deflections between jointed slabs were measured with dial indicators, and strains of the dowel bars were monitored with strain gauges. The load transfer efficiency, differential deflection, and dowel-concrete bearing stress for each case were calculated from these measurements. The test results indicated that the effect of the dowel modulus on load transfer efficiency can be characterized based on the similarity model test developed in the study. Moreover, round steel dowel was found to have similar performance to larger FRP dowel, and elliptical dowel can be preferentially considered in practice.

  7. Fuzzy delay model based fault simulator for crosstalk delay fault test ...

    Indian Academy of Sciences (India)

    In this paper, a fuzzy delay model based crosstalk delay fault simulator is proposed. As design .... To find the quality of non-robust tests, a fuzzy delay ..... Dubois D and Prade H 1989 Processing Fuzzy temporal knowledge. IEEE Transactions ...

  8. Modelling of XCO2 Surfaces Based on Flight Tests of TanSat Instruments

    Directory of Open Access Journals (Sweden)

    Li Li Zhang

    2016-11-01

    Full Text Available The TanSat carbon satellite is to be launched at the end of 2016. In order to verify the performance of its instruments, a flight test of TanSat instruments was conducted in Jilin Province in September, 2015. The flight test area covered a total area of about 11,000 km2 and the underlying surface cover included several lakes, forest land, grassland, wetland, farmland, a thermal power plant and numerous cities and villages. We modeled the column-average dry-air mole fraction of atmospheric carbon dioxide (XCO2 surface based on flight test data which measured the near- and short-wave infrared (NIR reflected solar radiation in the absorption bands at around 760 and 1610 nm. However, it is difficult to directly analyze the spatial distribution of XCO2 in the flight area using the limited flight test data and the approximate surface of XCO2, which was obtained by regression modeling, which is not very accurate either. We therefore used the high accuracy surface modeling (HASM platform to fill the gaps where there is no information on XCO2 in the flight test area, which takes the approximate surface of XCO2 as its driving field and the XCO2 observations retrieved from the flight test as its optimum control constraints. High accuracy surfaces of XCO2 were constructed with HASM based on the flight’s observations. The results showed that the mean XCO2 in the flight test area is about 400 ppm and that XCO2 over urban areas is much higher than in other places. Compared with OCO-2’s XCO2, the mean difference is 0.7 ppm and the standard deviation is 0.95 ppm. Therefore, the modelling of the XCO2 surface based on the flight test of the TanSat instruments fell within an expected and acceptable range.

  9. Class hierarchical test case generation algorithm based on expanded EMDPN model

    Institute of Scientific and Technical Information of China (English)

    LI Jun-yi; GONG Hong-fang; HU Ji-ping; ZOU Bei-ji; SUN Jia-guang

    2006-01-01

    A new model of event and message driven Petri network(EMDPN) based on the characteristic of class interaction for messages passing between two objects was extended. Using EMDPN interaction graph, a class hierarchical test-case generation algorithm with cooperated paths (copaths) was proposed, which can be used to solve the problems resulting from the class inheritance mechanism encountered in object-oriented software testing such as oracle, message transfer errors, and unreachable statement. Finally, the testing sufficiency was analyzed with the ordered sequence testing criterion(OSC). The results indicate that the test cases stemmed from newly proposed automatic algorithm of copaths generation satisfies synchronization message sequences testing criteria, therefore the proposed new algorithm of copaths generation has a good coverage rate.

  10. In silico model-based inference: a contemporary approach for hypothesis testing in network biology.

    Science.gov (United States)

    Klinke, David J

    2014-01-01

    Inductive inference plays a central role in the study of biological systems where one aims to increase their understanding of the system by reasoning backwards from uncertain observations to identify causal relationships among components of the system. These causal relationships are postulated from prior knowledge as a hypothesis or simply a model. Experiments are designed to test the model. Inferential statistics are used to establish a level of confidence in how well our postulated model explains the acquired data. This iterative process, commonly referred to as the scientific method, either improves our confidence in a model or suggests that we revisit our prior knowledge to develop a new model. Advances in technology impact how we use prior knowledge and data to formulate models of biological networks and how we observe cellular behavior. However, the approach for model-based inference has remained largely unchanged since Fisher, Neyman and Pearson developed the ideas in the early 1900s that gave rise to what is now known as classical statistical hypothesis (model) testing. Here, I will summarize conventional methods for model-based inference and suggest a contemporary approach to aid in our quest to discover how cells dynamically interpret and transmit information for therapeutic aims that integrates ideas drawn from high performance computing, Bayesian statistics, and chemical kinetics. © 2014 American Institute of Chemical Engineers.

  11. Creep Tests and Modeling Based on Continuum Damage Mechanics for T91 and T92 Steels

    Science.gov (United States)

    Pan, J. P.; Tu, S. H.; Zhu, X. W.; Tan, L. J.; Hu, B.; Wang, Q.

    2017-12-01

    9-11%Cr ferritic steels play an important role in high-temperature and high-pressure boilers of advanced power plants. In this paper, a continuum damage mechanics (CDM)-based creep model was proposed to study the creep behavior of T91 and T92 steels at high temperatures. Long-time creep tests were performed for both steels under different conditions. The creep rupture data and creep curves obtained from creep tests were captured well by theoretical calculation based on the CDM model over a long creep time. It is shown that the developed model is able to predict creep data for the two ferritic steels accurately up to tens of thousands of hours.

  12. Industrial-Strength Model-Based Testing - State of the Art and Current Challenges

    Directory of Open Access Journals (Sweden)

    Jan Peleska

    2013-03-01

    Full Text Available As of today, model-based testing (MBT is considered as leading-edge technology in industry. We sketch the different MBT variants that - according to our experience - are currently applied in practice, with special emphasis on the avionic, railway and automotive domains. The key factors for successful industrial-scale application of MBT are described, both from a scientific and a managerial point of view. With respect to the former view, we describe the techniques for automated test case, test data and test procedure generation for concurrent reactive real-time systems which are considered as the most important enablers for MBT in practice. With respect to the latter view, our experience with introducing MBT approaches in testing teams are sketched. Finally, the most challenging open scientific problems whose solutions are bound to improve the acceptance and effectiveness of MBT in industry are discussed.

  13. Field-based tests of geochemical modeling codes: New Zealand hydrothermal systems

    International Nuclear Information System (INIS)

    Bruton, C.J.; Glassley, W.E.; Bourcier, W.L.

    1993-12-01

    Hydrothermal systems in the Taupo Volcanic Zone, North Island, New Zealand are being used as field-based modeling exercises for the EQ3/6 geochemical modeling code package. Comparisons of the observed state and evolution of the hydrothermal systems with predictions of fluid-solid equilibria made using geochemical modeling codes will determine how the codes can be used to predict the chemical and mineralogical response of the environment to nuclear waste emplacement. Field-based exercises allow us to test the models on time scales unattainable in the laboratory. Preliminary predictions of mineral assemblages in equilibrium with fluids sampled from wells in the Wairakei and Kawerau geothermal field suggest that affinity-temperature diagrams must be used in conjunction with EQ6 to minimize the effect of uncertainties in thermodynamic and kinetic data on code predictions

  14. Field-based tests of geochemical modeling codes usign New Zealand hydrothermal systems

    International Nuclear Information System (INIS)

    Bruton, C.J.; Glassley, W.E.; Bourcier, W.L.

    1994-06-01

    Hydrothermal systems in the Taupo Volcanic Zone, North Island, New Zealand are being used as field-based modeling exercises for the EQ3/6 geochemical modeling code package. Comparisons of the observed state and evolution of the hydrothermal systems with predictions of fluid-solid equilibria made using geochemical modeling codes will determine how the codes can be used to predict the chemical and mineralogical response of the environment to nuclear waste emplacement. Field-based exercises allow us to test the models on time scales unattainable in the laboratory. Preliminary predictions of mineral assemblages in equilibrium with fluids sampled from wells in the Wairakei and Kawerau geothermal field suggest that affinity-temperature diagrams must be used in conjunction with EQ6 to minimize the effect of uncertainties in thermodynamic and kinetic data on code predictions

  15. Earthquake likelihood model testing

    Science.gov (United States)

    Schorlemmer, D.; Gerstenberger, M.C.; Wiemer, S.; Jackson, D.D.; Rhoades, D.A.

    2007-01-01

    INTRODUCTIONThe Regional Earthquake Likelihood Models (RELM) project aims to produce and evaluate alternate models of earthquake potential (probability per unit volume, magnitude, and time) for California. Based on differing assumptions, these models are produced to test the validity of their assumptions and to explore which models should be incorporated in seismic hazard and risk evaluation. Tests based on physical and geological criteria are useful but we focus on statistical methods using future earthquake catalog data only. We envision two evaluations: a test of consistency with observed data and a comparison of all pairs of models for relative consistency. Both tests are based on the likelihood method, and both are fully prospective (i.e., the models are not adjusted to fit the test data). To be tested, each model must assign a probability to any possible event within a specified region of space, time, and magnitude. For our tests the models must use a common format: earthquake rates in specified “bins” with location, magnitude, time, and focal mechanism limits.Seismology cannot yet deterministically predict individual earthquakes; however, it should seek the best possible models for forecasting earthquake occurrence. This paper describes the statistical rules of an experiment to examine and test earthquake forecasts. The primary purposes of the tests described below are to evaluate physical models for earthquakes, assure that source models used in seismic hazard and risk studies are consistent with earthquake data, and provide quantitative measures by which models can be assigned weights in a consensus model or be judged as suitable for particular regions.In this paper we develop a statistical method for testing earthquake likelihood models. A companion paper (Schorlemmer and Gerstenberger 2007, this issue) discusses the actual implementation of these tests in the framework of the RELM initiative.Statistical testing of hypotheses is a common task and a

  16. Testing a ground-based canopy model using the wind river canopy crane

    Science.gov (United States)

    Robert Van Pelt; Malcolm P. North

    1999-01-01

    A ground-based canopy model that estimates the volume of occupied space in forest canopies was tested using the Wind River Canopy Crane. A total of 126 trees in a 0.25 ha area were measured from the ground and directly from a gondola suspended from the crane. The trees were located in a low elevation, old-growth forest in the southern Washington Cascades. The ground-...

  17. Genetic screening and testing in an episode-based payment model: preserving patient autonomy.

    Science.gov (United States)

    Sutherland, Sharon; Farrell, Ruth M; Lockwood, Charles

    2014-11-01

    The State of Ohio is implementing an episode-based payment model for perinatal care. All costs of care will be tabulated for each live birth and assigned to the delivering provider, creating a three-tiered model for reimbursement for care. Providers will be reimbursed as usual for care that is average in cost and quality, while instituting rewards or penalties for those outside the expected range in either domain. There are few exclusions, and all methods of genetic screening and diagnostic testing are included in the episode cost calculation as proposed. Prenatal ultrasonography, genetic screening, and diagnostic testing are critical components of the delivery of high-quality, evidence-based prenatal care. These tests provide pregnant women with key information about the pregnancy, which, in turn, allows them to work closely with their health care provider to determine optimal prenatal care. The concepts of informed consent and decision-making, cornerstones of the ethical practice of medicine, are founded on the principles of autonomy and respect for persons. These principles recognize that patients' rights to make choices and take actions are based on their personal beliefs and values. Given the personal nature of such decisions, it is critical that patients have unbarred access to prenatal genetic tests if they elect to use them as part of their prenatal care. The proposed restructuring of reimbursement creates a clear conflict between patient autonomy and physician financial incentives.

  18. GOODNESS-OF-FIT TEST FOR THE ACCELERATED FAILURE TIME MODEL BASED ON MARTINGALE RESIDUALS

    Czech Academy of Sciences Publication Activity Database

    Novák, Petr

    2013-01-01

    Roč. 49, č. 1 (2013), s. 40-59 ISSN 0023-5954 R&D Projects: GA MŠk(CZ) 1M06047 Grant - others:GA MŠk(CZ) SVV 261315/2011 Keywords : accelerated failure time model * survival analysis * goodness-of-fit Subject RIV: BB - Applied Statistics, Operational Research Impact factor: 0.563, year: 2013 http://library.utia.cas.cz/separaty/2013/SI/novak-goodness-of-fit test for the aft model based on martingale residuals.pdf

  19. Model-Based Prediction of Pulsed Eddy Current Testing Signals from Stratified Conductive Structures

    International Nuclear Information System (INIS)

    Zhang, Jian Hai; Song, Sung Jin; Kim, Woong Ji; Kim, Hak Joon; Chung, Jong Duk

    2011-01-01

    Excitation and propagation of electromagnetic field of a cylindrical coil above an arbitrary number of conductive plates for pulsed eddy current testing(PECT) are very complex problems due to their complicated physical properties. In this paper, analytical modeling of PECT is established by Fourier series based on truncated region eigenfunction expansion(TREE) method for a single air-cored coil above stratified conductive structures(SCS) to investigate their integrity. From the presented expression of PECT, the coil impedance due to SCS is calculated based on analytical approach using the generalized reflection coefficient in series form. Then the multilayered structures manufactured by non-ferromagnetic (STS301L) and ferromagnetic materials (SS400) are investigated by the developed PECT model. Good prediction of analytical model of PECT not only contributes to the development of an efficient solver but also can be applied to optimize the conditions of experimental setup in PECT

  20. Comparing Science Virtual and Paper-Based Test to Measure Students’ Critical Thinking based on VAK Learning Style Model

    Science.gov (United States)

    Rosyidah, T. H.; Firman, H.; Rusyati, L.

    2017-02-01

    This research was comparing virtual and paper-based test to measure students’ critical thinking based on VAK (Visual-Auditory-Kynesthetic) learning style model. Quasi experiment method with one group post-test only design is applied in this research in order to analyze the data. There was 40 eight grade students at one of public junior high school in Bandung becoming the sample in this research. The quantitative data was obtained through 26 questions about living thing and environment sustainability which is constructed based on the eight elements of critical thinking and be provided in the form of virtual and paper-based test. Based on analysis of the result, it is shown that within visual, auditory, and kinesthetic were not significantly difference in virtual and paper-based test. Besides, all result was supported by quistionnaire about students’ respond on virtual test which shows 3.47 in the scale of 4. Means that student showed positive respond in all aspet measured, which are interest, impression, and expectation.

  1. Testing of technology readiness index model based on exploratory factor analysis approach

    Science.gov (United States)

    Ariani, AF; Napitupulu, D.; Jati, RK; Kadar, JA; Syafrullah, M.

    2018-04-01

    SMEs readiness in using ICT will determine the adoption of ICT in the future. This study aims to evaluate the model of technology readiness in order to apply the technology on SMEs. The model is tested to find if TRI model is relevant to measure ICT adoption, especially for SMEs in Indonesia. The research method used in this paper is survey to a group of SMEs in South Tangerang. The survey measures the readiness to adopt ICT based on four variables which is Optimism, Innovativeness, Discomfort, and Insecurity. Each variable contains several indicators to make sure the variable is measured thoroughly. The data collected through survey is analysed using factor analysis methodwith the help of SPSS software. The result of this study shows that TRI model gives more descendants on some indicators and variables. This result can be caused by SMEs owners’ knowledge is not homogeneous about either the technology that they are used, knowledge or the type of their business.

  2. Development and design of a late-model fitness test instrument based on LabView

    Science.gov (United States)

    Xie, Ying; Wu, Feiqing

    2010-12-01

    Undergraduates are pioneers of China's modernization program and undertake the historic mission of rejuvenating our nation in the 21st century, whose physical fitness is vital. A smart fitness test system can well help them understand their fitness and health conditions, thus they can choose more suitable approaches and make practical plans for exercising according to their own situation. following the future trends, a Late-model fitness test Instrument based on LabView has been designed to remedy defects of today's instruments. The system hardware consists of fives types of sensors with their peripheral circuits, an acquisition card of NI USB-6251 and a computer, while the system software, on the basis of LabView, includes modules of user register, data acquisition, data process and display, and data storage. The system, featured by modularization and an open structure, is able to be revised according to actual needs. Tests results have verified the system's stability and reliability.

  3. Error-Rate Estimation Based on Multi-Signal Flow Graph Model and Accelerated Radiation Tests.

    Directory of Open Access Journals (Sweden)

    Wei He

    Full Text Available A method of evaluating the single-event effect soft-error vulnerability of space instruments before launched has been an active research topic in recent years. In this paper, a multi-signal flow graph model is introduced to analyze the fault diagnosis and meantime to failure (MTTF for space instruments. A model for the system functional error rate (SFER is proposed. In addition, an experimental method and accelerated radiation testing system for a signal processing platform based on the field programmable gate array (FPGA is presented. Based on experimental results of different ions (O, Si, Cl, Ti under the HI-13 Tandem Accelerator, the SFER of the signal processing platform is approximately 10-3(error/particle/cm2, while the MTTF is approximately 110.7 h.

  4. Scopolamine provocation-based pharmacological MRI model for testing procognitive agents.

    Science.gov (United States)

    Hegedűs, Nikolett; Laszy, Judit; Gyertyán, István; Kocsis, Pál; Gajári, Dávid; Dávid, Szabolcs; Deli, Levente; Pozsgay, Zsófia; Tihanyi, Károly

    2015-04-01

    There is a huge unmet need to understand and treat pathological cognitive impairment. The development of disease modifying cognitive enhancers is hindered by the lack of correct pathomechanism and suitable animal models. Most animal models to study cognition and pathology do not fulfil either the predictive validity, face validity or construct validity criteria, and also outcome measures greatly differ from those of human trials. Fortunately, some pharmacological agents such as scopolamine evoke similar effects on cognition and cerebral circulation in rodents and humans and functional MRI enables us to compare cognitive agents directly in different species. In this paper we report the validation of a scopolamine based rodent pharmacological MRI provocation model. The effects of deemed procognitive agents (donepezil, vinpocetine, piracetam, alpha 7 selective cholinergic compounds EVP-6124, PNU-120596) were compared on the blood-oxygen-level dependent responses and also linked to rodent cognitive models. These drugs revealed significant effect on scopolamine induced blood-oxygen-level dependent change except for piracetam. In the water labyrinth test only PNU-120596 did not show a significant effect. This provocational model is suitable for testing procognitive compounds. These functional MR imaging experiments can be paralleled with human studies, which may help reduce the number of false cognitive clinical trials. © The Author(s) 2015.

  5. Testing of money multiplier model for Pakistan: does monetary base carry any information?

    Directory of Open Access Journals (Sweden)

    Muhammad Arshad Khan

    2010-02-01

    Full Text Available This paper tests the constancy and stationarity of mechanic version of the money multiplier model for Pakistan using monthly data over the period 1972M1-2009M2. We split the data into pre-liberalization (1972M1-1990M12 and post-liberalization (1991M1-2009M2 periods to examine the impact of financial sector reforms. We first examine the constancy and stationarity of the money multiplier and the results suggest the money multiplier remains non-stationary for the entire sample period and sub-periods. We then tested cointegration between money supply and monetary base and find the evidence of cointegration between two variables for the entire period and two sub-periods. The coefficient restrictions are satisfied only for the post-liberalization period. Two-way long-run causality between money supply and monetary base is found for the entire period and post-liberalization. For the post-liberalization period the evidence of short-run causality running from monetary base to money supply is also identified. On the whole, the results suggest that money multiplier model can serve as framework for conducting short-run monetary policy in Pakistan. However, the monetary authority may consider the co-movements between money supply and reserve money at the time of conducting monetary policy.

  6. A testing procedure for wind turbine generators based on the power grid statistical model

    DEFF Research Database (Denmark)

    Farajzadehbibalan, Saber; Ramezani, Mohammad Hossein; Nielsen, Peter

    2017-01-01

    In this study, a comprehensive test procedure is developed to test wind turbine generators with a hardware-in-loop setup. The procedure employs the statistical model of the power grid considering the restrictions of the test facility and system dynamics. Given the model in the latent space...

  7. Complete Model-Based Equivalence Class Testing for the ETCS Ceiling Speed Monitor

    DEFF Research Database (Denmark)

    Braunstein, Cécile; Haxthausen, Anne Elisabeth; Huang, Wen-ling

    2014-01-01

    In this paper we present a new test model written in SysML and an associated blackbox test suite for the Ceiling Speed Monitor (CSM) of the European Train Control System (ETCS). The model is publicly available and intended to serve as a novel benchmark for investigating new testing theories...

  8. Developing and testing transferability and feasibility of a model for educators using simulation-based learning - A European collaboration

    DEFF Research Database (Denmark)

    Bøje, Rikke Buus; Bland, Andrew; Sutton, Andrew

    2017-01-01

    of the study were to develop a model to educate the educators who deliver simulation-based learning and to test to which extent this model could be transferred to education providers in different national settings. METHODS: This model, its transferability and feasibility, was tested across three European...

  9. Environmental risk assessment of selected organic chemicals based on TOC test and QSAR estimation models.

    Science.gov (United States)

    Chi, Yulang; Zhang, Huanteng; Huang, Qiansheng; Lin, Yi; Ye, Guozhu; Zhu, Huimin; Dong, Sijun

    2018-02-01

    Environmental risks of organic chemicals have been greatly determined by their persistence, bioaccumulation, and toxicity (PBT) and physicochemical properties. Major regulations in different countries and regions identify chemicals according to their bioconcentration factor (BCF) and octanol-water partition coefficient (Kow), which frequently displays a substantial correlation with the sediment sorption coefficient (Koc). Half-life or degradability is crucial for the persistence evaluation of chemicals. Quantitative structure activity relationship (QSAR) estimation models are indispensable for predicting environmental fate and health effects in the absence of field- or laboratory-based data. In this study, 39 chemicals of high concern were chosen for half-life testing based on total organic carbon (TOC) degradation, and two widely accepted and highly used QSAR estimation models (i.e., EPI Suite and PBT Profiler) were adopted for environmental risk evaluation. The experimental results and estimated data, as well as the two model-based results were compared, based on the water solubility, Kow, Koc, BCF and half-life. Environmental risk assessment of the selected compounds was achieved by combining experimental data and estimation models. It was concluded that both EPI Suite and PBT Profiler were fairly accurate in measuring the physicochemical properties and degradation half-lives for water, soil, and sediment. However, the half-lives between the experimental and the estimated results were still not absolutely consistent. This suggests deficiencies of the prediction models in some ways, and the necessity to combine the experimental data and predicted results for the evaluation of environmental fate and risks of pollutants. Copyright © 2016. Published by Elsevier B.V.

  10. Models everywhere. How a fully integrated model-based test environment can enable progress in the future

    Energy Technology Data Exchange (ETDEWEB)

    Ben Gaid, Mongi; Lebas, Romain; Fremovici, Morgan; Font, Gregory; Le Solliec, Gunael [IFP Energies nouvelles, Rueil-Malmaison (France); Albrecht, Antoine [D2T Powertrain Engineering, Rueil-Malmaison (France)

    2011-07-01

    The aim of this paper is to demonstrate how advanced modelling approaches coupled with powerful tools allow to set up a complete and coherent test environment suite. Based on a real study focused on the development of a Euro 6 hybrid powertrain with a Euro 5 turbocharged diesel engine, the authors present how a diesel engine simulator including an in-cylinder phenomenological approach to predict the raw emissions can be coupled with a DOC and DPF after-treatment system and embedded in the complete hybrid powertrain to be used in various test environments: - coupled with the control software in a multi-model multi-core simulation platform with test automation features, allowing the simulation speed to be faster than the real-time; - exported in a real time hardware in the loop platform with the ECU and hardware actuators; embedded at the experimental engine test bed to perform driving cycles such as NEDC or FTP cycles with the hybrid powertrain management. Thanks to these complete and versatile test platform suite xMOD/Morphee, all the key issues of a full hybrid powertrain can be addressed efficiently and at low cost compared to the experimental powertrain prototypes: consumption minimisation, energy optimisation, thermal exhaust management. NOx/soots trade off, NO/NO2 ratios.. Having a good balance between versatility and compliancy of the model oriented test platforms such as presented in this paper is the best way to take the maximum benefit of the model developed at each stage of the powertrain development. (orig.)

  11. Agent-Based Modeling for Testing and Designing Novel Decentralized Command and Control System Paradigms

    National Research Council Canada - National Science Library

    Bonabeau, Eric; Hunt, Carl W; Gaudiano, Paolo

    2003-01-01

    Agent-based modeling (ABM) is a recent simulation modeling technique that consists of modeling a system from the bottom up, capturing the interactions taking place between the system's constituent units...

  12. Impact of Participatory Health Research: A Test of the Community-Based Participatory Research Conceptual Model

    Directory of Open Access Journals (Sweden)

    John G. Oetzel

    2018-01-01

    Full Text Available Objectives. A key challenge in evaluating the impact of community-based participatory research (CBPR is identifying what mechanisms and pathways are critical for health equity outcomes. Our purpose is to provide an empirical test of the CBPR conceptual model to address this challenge. Methods. A three-stage quantitative survey was completed: (1 294 US CBPR projects with US federal funding were identified; (2 200 principal investigators completed a questionnaire about project-level details; and (3 450 community or academic partners and principal investigators completed a questionnaire about perceived contextual, process, and outcome variables. Seven in-depth qualitative case studies were conducted to explore elements of the model not captured in the survey; one is presented due to space limitations. Results. We demonstrated support for multiple mechanisms illustrated by the conceptual model using a latent structural equation model. Significant pathways were identified, showing the positive association of context with partnership structures and dynamics. Partnership structures and dynamics showed similar associations with partnership synergy and community involvement in research; both of these had positive associations with intermediate community changes and distal health outcomes. The case study complemented and extended understandings of the mechanisms of how partnerships can improve community conditions. Conclusions. The CBPR conceptual model is well suited to explain key relational and structural pathways for impact on health equity outcomes.

  13. Modelling, Simulation and Testing of a Reconfigurable Cable-Based Parallel Manipulator as Motion Aiding System

    Directory of Open Access Journals (Sweden)

    Gianni Castelli

    2010-01-01

    Full Text Available This paper presents results on the modelling, simulation and experimental tests of a cable-based parallel manipulator to be used as an aiding or guiding system for people with motion disabilities. There is a high level of motivation for people with a motion disability or the elderly to perform basic daily-living activities independently. Therefore, it is of great interest to design and implement safe and reliable motion assisting and guiding devices that are able to help end-users. In general, a robot for a medical application should be able to interact with a patient in safety conditions, i.e. it must not damage people or surroundings; it must be designed to guarantee high accuracy and low acceleration during the operation. Furthermore, it should not be too bulky and it should exert limited wrenches after close interaction with people. It can be advisable to have a portable system which can be easily brought into and assembled in a hospital or a domestic environment. Cable-based robotic structures can fulfil those requirements because of their main characteristics that make them light and intrinsically safe. In this paper, a reconfigurable four-cable-based parallel manipulator has been proposed as a motion assisting and guiding device to help people to accomplish a number of tasks, such as an aiding or guiding system to move the upper and lower limbs or the whole body. Modelling and simulation are presented in the ADAMS environment. Moreover, experimental tests are reported as based on an available laboratory prototype.

  14. Large scale model testing

    International Nuclear Information System (INIS)

    Brumovsky, M.; Filip, R.; Polachova, H.; Stepanek, S.

    1989-01-01

    Fracture mechanics and fatigue calculations for WWER reactor pressure vessels were checked by large scale model testing performed using large testing machine ZZ 8000 (with a maximum load of 80 MN) at the SKODA WORKS. The results are described from testing the material resistance to fracture (non-ductile). The testing included the base materials and welded joints. The rated specimen thickness was 150 mm with defects of a depth between 15 and 100 mm. The results are also presented of nozzles of 850 mm inner diameter in a scale of 1:3; static, cyclic, and dynamic tests were performed without and with surface defects (15, 30 and 45 mm deep). During cyclic tests the crack growth rate in the elastic-plastic region was also determined. (author). 6 figs., 2 tabs., 5 refs

  15. Model-based framework for multi-axial real-time hybrid simulation testing

    Science.gov (United States)

    Fermandois, Gaston A.; Spencer, Billie F.

    2017-10-01

    Real-time hybrid simulation is an efficient and cost-effective dynamic testing technique for performance evaluation of structural systems subjected to earthquake loading with rate-dependent behavior. A loading assembly with multiple actuators is required to impose realistic boundary conditions on physical specimens. However, such a testing system is expected to exhibit significant dynamic coupling of the actuators and suffer from time lags that are associated with the dynamics of the servo-hydraulic system, as well as control-structure interaction (CSI). One approach to reducing experimental errors considers a multi-input, multi-output (MIMO) controller design, yielding accurate reference tracking and noise rejection. In this paper, a framework for multi-axial real-time hybrid simulation (maRTHS) testing is presented. The methodology employs a real-time feedback-feedforward controller for multiple actuators commanded in Cartesian coordinates. Kinematic transformations between actuator space and Cartesian space are derived for all six-degrees-offreedom of the moving platform. Then, a frequency domain identification technique is used to develop an accurate MIMO transfer function of the system. Further, a Cartesian-domain model-based feedforward-feedback controller is implemented for time lag compensation and to increase the robustness of the reference tracking for given model uncertainty. The framework is implemented using the 1/5th-scale Load and Boundary Condition Box (LBCB) located at the University of Illinois at Urbana- Champaign. To demonstrate the efficacy of the proposed methodology, a single-story frame subjected to earthquake loading is tested. One of the columns in the frame is represented physically in the laboratory as a cantilevered steel column. For realtime execution, the numerical substructure, kinematic transformations, and controllers are implemented on a digital signal processor. Results show excellent performance of the maRTHS framework when six

  16. Development and tests of a mouse voxel model dor MCNPX based on Digimouse images

    Energy Technology Data Exchange (ETDEWEB)

    Melo M, B.; Ferreira F, C. [Centro de Desenvolvimento da Tecnologia Nuclear / CNEN, Pte. Antonio Carlos No. 6627, Belo Horizonte 31270-901, Minas Gerais (Brazil); Garcia de A, I.; Machado T, B.; Passos Ribeiro de C, T., E-mail: bmm@cdtn.br [Universidade Federal de Minas Gerais, Departamento de Engenharia Nuclear, Pte. Antonio Carlos 6627, Belo Horizonte 31270-901, Minas Gerais (Brazil)

    2015-10-15

    Mice have been widely used in experimental protocols involving ionizing radiation. Biological effects (Be) induced by radiation can compromise studies results. Good estimates of mouse whole body and organs absorbed dose could provide valuable information to researchers. The aim of this study was to create and test a new voxel phantom for mice dosimetry from -Digimouse- project images. Micro CT images from Digimouse project were used in this work. Corel PHOTOPAINT software was utilized in segmentation process. The three-dimensional (3-D) model assembly and its voxel size manipulation were performed by Image J. SISCODES was used to adapt the model to run in MCNPX Monte Carlo code. The resulting model was called DM{sub B}RA. The volume and mass of segmented organs were compared with data available in literature. For the preliminary tests the heart was considered the source organ. Photons of diverse energies were simulated and Saf values obtained through F6:p and + F6 MCNPX tallies. The results were compared with reference data. 3-D picturing of absorbed doses patterns and relative errors distribution were generated by a C++ -in house- made program and visualized through Amide software. The organ masses of DM{sub B}RA correlated well with two models that were based on same set of images. However some organs, like eyes and adrenals, skeleton and brain showed large discrepancies. Segmentation of an identical image set by different persons and/or methods can result significant organ masses variations. We believe that the main causes of these differences were: i) operator dependent subjectivity in the definition of organ limits during the segmentation processes; and i i) distinct voxel dimensions between evaluated models. Lack of reference data for mice models construction and dosimetry was detected. Comparison with other models originated from different mice strains also demonstrated that the anatomical and size variability can be significant. Use of + F6 tally for mouse

  17. Development and tests of a mouse voxel model dor MCNPX based on Digimouse images

    International Nuclear Information System (INIS)

    Melo M, B.; Ferreira F, C.; Garcia de A, I.; Machado T, B.; Passos Ribeiro de C, T.

    2015-10-01

    Mice have been widely used in experimental protocols involving ionizing radiation. Biological effects (Be) induced by radiation can compromise studies results. Good estimates of mouse whole body and organs absorbed dose could provide valuable information to researchers. The aim of this study was to create and test a new voxel phantom for mice dosimetry from -Digimouse- project images. Micro CT images from Digimouse project were used in this work. Corel PHOTOPAINT software was utilized in segmentation process. The three-dimensional (3-D) model assembly and its voxel size manipulation were performed by Image J. SISCODES was used to adapt the model to run in MCNPX Monte Carlo code. The resulting model was called DM B RA. The volume and mass of segmented organs were compared with data available in literature. For the preliminary tests the heart was considered the source organ. Photons of diverse energies were simulated and Saf values obtained through F6:p and + F6 MCNPX tallies. The results were compared with reference data. 3-D picturing of absorbed doses patterns and relative errors distribution were generated by a C++ -in house- made program and visualized through Amide software. The organ masses of DM B RA correlated well with two models that were based on same set of images. However some organs, like eyes and adrenals, skeleton and brain showed large discrepancies. Segmentation of an identical image set by different persons and/or methods can result significant organ masses variations. We believe that the main causes of these differences were: i) operator dependent subjectivity in the definition of organ limits during the segmentation processes; and i i) distinct voxel dimensions between evaluated models. Lack of reference data for mice models construction and dosimetry was detected. Comparison with other models originated from different mice strains also demonstrated that the anatomical and size variability can be significant. Use of + F6 tally for mouse phantoms

  18. IRT-based test construction

    OpenAIRE

    van der Linden, Willem J.; Theunissen, T.J.J.M.; Boekkooi-Timminga, Ellen; Kelderman, Henk

    1987-01-01

    Four discussions of test construction based on item response theory (IRT) are presented. The first discussion, "Test Design as Model Building in Mathematical Programming" (T.J.J.M. Theunissen), presents test design as a decision process under certainty. A natural way of modeling this process leads to mathematical programming. General models of test construction are discussed, with information about algorithms and heuristics; ideas about the analysis and refinement of test constraints are also...

  19. Pedestrian simulation model based on principles of bounded rationality: results of validation tests

    NARCIS (Netherlands)

    Zhu, W.; Timmermans, H.J.P.; Lo, H.P.; Leung, Stephen C.H.; Tan, Susanna M.L.

    2009-01-01

    Over the years, different modelling approaches to simulating pedestrian movement have been suggested. The majority of pedestrian decision models are based on the concept of utility maximization. To explore alternatives, we developed the heterogeneous heuristic model (HHM), based on principles of

  20. Design of New Test Function Model Based on Multi-objective Optimization Method

    Directory of Open Access Journals (Sweden)

    Zhaoxia Shang

    2017-01-01

    Full Text Available Space partitioning method, as a new algorism, has been applied to planning and decision-making of investment portfolio more and more often. But currently there are so few testing function for this algorism, which has greatly restrained its further development and application. An innovative test function model is designed in this paper and is used to test the algorism. It is proved that for evaluation of space partitioning method in certain applications, this test function has fairly obvious advantage.

  1. A Review of Models for Computer-Based Testing. Research Report 2011-12

    Science.gov (United States)

    Luecht, Richard M.; Sireci, Stephen G.

    2011-01-01

    Over the past four decades, there has been incremental growth in computer-based testing (CBT) as a viable alternative to paper-and-pencil testing. However, the transition to CBT is neither easy nor inexpensive. As Drasgow, Luecht, and Bennett (2006) noted, many design engineering, test development, operations/logistics, and psychometric changes…

  2. T-UPPAAL: Online Model-based Testing of Real-Time Systems

    DEFF Research Database (Denmark)

    Mikucionis, Marius; Larsen, Kim Guldstrand; Nielsen, Brian

    2004-01-01

    The goal of testing is to gain confidence in a physical computer based system by means of executing it. More than one third of typical project resources is spent on testing embedded and real-time systems, but still it remains ad-hoc, based on heuristics, and error-prone. Therefore systematic...

  3. Testing the ontogenetic base for the transient model of inflorescence development.

    Science.gov (United States)

    Bull-Hereñu, Kester; Claßen-Bockhoff, Regine

    2013-11-01

    Current research in plant science has concentrated on revealing ontogenetic processes of key attributes in plant evolution. One recently discussed model is the 'transient model' successful in explaining some types of inflorescence architectures based on two main principles: the decline of the so called 'vegetativeness' (veg) factor and the transient nature of apical meristems in developing inflorescences. This study examines whether both principles find a concrete ontogenetic correlate in inflorescence development. To test the ontogenetic base of veg decline and the transient character of apical meristems the ontogeny of meristematic size in developing inflorescences was investigated under scanning electron microscopy. Early and late inflorescence meristems were measured and compared during inflorescence development in 13 eudicot species from 11 families. The initial size of the inflorescence meristem in closed inflorescences correlates with the number of nodes in the mature inflorescence. Conjunct compound inflorescences (panicles) show a constant decrease of meristematic size from early to late inflorescence meristems, while disjunct compound inflorescences present an enlargement by merging from early inflorescence meristems to late inflorescence meristems, implying a qualitative change of the apical meristems during ontogeny. Partial confirmation was found for the transient model for inflorescence architecture in the ontogeny: the initial size of the apical meristem in closed inflorescences is consistent with the postulated veg decline mechanism regulating the size of the inflorescence. However, the observed biphasic kinetics of the development of the apical meristem in compound racemes offers the primary explanation for their disjunct morphology, contrary to the putative exclusive transient mechanism in lateral axes as expected by the model.

  4. A test of the California competency-based differentiated role model.

    Science.gov (United States)

    Keating, Sarah B; Rutledge, Dana N; Sargent, Arlene; Walker, Polly

    2003-01-01

    To address the incongruence between the expectations of nursing service and education in California, the Education Industry Interface Task Force of the California Strategic Planning Committee for Nursing developed descriptions to assist employers and educators in clearly differentiating practice and educational competencies. The completion of the Competency-Based Role Differentiation Model resulted in the need to test the model for its utility in the service setting, in education, and for career planning for nurses. Three alpha demonstration sites were selected based on representative geographical regions of California. The sites were composed of tri-partnerships consisting of a medical center, an associate degree in nursing program, and a baccalaureate nursing program. Observers rated senior students and new graduates in medical-surgical units on their behaviors in teacher and leadership care provider and care coordinator roles. The alpha demonstration study results were as expected. That is, senior students practice predominantly at a novice level in teacher and management/leadership care provider functions and new graduates practice predominately at the competent level. New graduates are more likely to take on novice and competent care coordinator roles. The CBRDM may be useful for practice and education settings to evaluate student and nurse performance, to define role expectations, and to identify the preparation necessary for the roles. It is useful for all of nursing as it continues to define its levels of practice and their relationship to on-the-job performance, curriculum development, and carrier planning.

  5. Testing a model of codependency for college students in Taiwan based on Bowen's concept of differentiation.

    Science.gov (United States)

    Chang, Shih-Hua

    2018-04-01

    The purpose of this study was to test a model of codependency based on Bowen's concept of differentiation for college students in Taiwan. The relations between family-of-origin dysfunction, differentiation of self, codependency traits and related symptoms including low self-esteem, relationship distress and psychological adjustment problems were examined. Data were collected from 567 college students from 2 large, urban universities in northern Taiwan. Results indicated a significantly negative relationship between levels of codependency and self-differentiation and that self-differentiation partially mediated the relationship between family-of-origin dysfunction and codependency. The implications of these findings for counselling Taiwanese college students who experience codependency traits and related symptoms as well as suggestions for future research are discussed. © 2016 International Union of Psychological Science.

  6. Home-Based Risk of Falling Assessment Test Using a Closed-Loop Balance Model.

    Science.gov (United States)

    Ayena, Johannes C; Zaibi, Helmi; Otis, Martin J-D; Menelas, Bob-Antoine J

    2016-12-01

    The aim of this study is to improve and facilitate the methods used to assess risk of falling at home among older people through the computation of a risk of falling in real time in daily activities. In order to increase a real time computation of the risk of falling, a closed-loop balance model is proposed and compared with One-Leg Standing Test (OLST). This balance model allows studying the postural response of a person having an unpredictable perturbation. Twenty-nine volunteers participated in this study for evaluating the effectiveness of the proposed system which includes seventeen elder participants: ten healthy elderly ( 68.4 ±5.5 years), seven Parkinson's disease (PD) subjects ( 66.28 ±8.9 years), and twelve healthy young adults ( 28.27 ±3.74 years). Our work suggests that there is a relationship between OLST score and the risk of falling based on center of pressure measurement with four low cost force sensors located inside an instrumented insole, which could be predicted using our suggested closed-loop balance model. For long term monitoring at home, this system could be included in a medical electronic record and could be useful as a diagnostic aid tool.

  7. How "Does" the Comforting Process Work? An Empirical Test of an Appraisal-Based Model of Comforting

    Science.gov (United States)

    Jones, Susanne M.; Wirtz, John G.

    2006-01-01

    Burleson and Goldsmith's (1998) comforting model suggests an appraisal-based mechanism through which comforting messages can bring about a positive change in emotional states. This study is a first empirical test of three causal linkages implied by the appraisal-based comforting model. Participants (N=258) talked about an upsetting event with a…

  8. Fuzzy delay model based fault simulator for crosstalk delay fault test ...

    Indian Academy of Sciences (India)

    In this paper, a fuzzy delay model based crosstalk delay fault simulator is proposed. As design trends move towards nanometer technologies, more number of new parameters affects the delay of the component. Fuzzy delay models are ideal for modelling the uncertainty found in the design and manufacturing steps.

  9. An Outcrop-based Detailed Geological Model to Test Automated Interpretation of Seismic Inversion Results

    NARCIS (Netherlands)

    Feng, R.; Sharma, S.; Luthi, S.M.; Gisolf, A.

    2015-01-01

    Previously, Tetyukhina et al. (2014) developed a geological and petrophysical model based on the Book Cliffs outcrops that contained eight lithotypes. For reservoir modelling purposes, this model is judged to be too coarse because in the same lithotype it contains reservoir and non-reservoir

  10. A practical model-based statistical approach for generating functional test cases: application in the automotive industry

    OpenAIRE

    Awédikian , Roy; Yannou , Bernard

    2012-01-01

    International audience; With the growing complexity of industrial software applications, industrials are looking for efficient and practical methods to validate the software. This paper develops a model-based statistical testing approach that automatically generates online and offline test cases for embedded software. It discusses an integrated framework that combines solutions for three major software testing research questions: (i) how to select test inputs; (ii) how to predict the expected...

  11. Testing the Community-Based Learning Collaborative (CBLC) implementation model: a study protocol.

    Science.gov (United States)

    Hanson, Rochelle F; Schoenwald, Sonja; Saunders, Benjamin E; Chapman, Jason; Palinkas, Lawrence A; Moreland, Angela D; Dopp, Alex

    2016-01-01

    High rates of youth exposure to violence, either through direct victimization or witnessing, result in significant health/mental health consequences and high associated lifetime costs. Evidence-based treatments (EBTs), such as Trauma-Focused Cognitive Behavioral Therapy (TF-CBT), can prevent and/or reduce these negative effects, yet these treatments are not standard practice for therapists working with children identified by child welfare or mental health systems as needing services. While research indicates that collaboration among child welfare and mental health services sectors improves availability and sustainment of EBTs for children, few implementation strategies designed specifically to promote and sustain inter-professional collaboration (IC) and inter-organizational relationships (IOR) have undergone empirical investigation. A potential candidate for evaluation is the Community-Based Learning Collaborative (CBLC) implementation model, an adaptation of the Learning Collaborative which includes strategies designed to develop and strengthen inter-professional relationships between brokers and providers of mental health services to promote IC and IOR and achieve sustained implementation of EBTs for children within a community. This non-experimental, mixed methods study involves two phases: (1) analysis of existing prospective quantitative and qualitative quality improvement and project evaluation data collected pre and post, weekly, and monthly from 998 participants in one of seven CBLCs conducted as part of a statewide initiative; and (2) Phase 2 collection of new quantitative and qualitative (key informant interviews) data during the funded study period to evaluate changes in relations among IC, IOR, social networks and the penetration and sustainment of TF-CBT in targeted communities. Recruitment for Phase 2 is from the pool of 998 CBLC participants to achieve a targeted enrollment of n = 150. Study aims include: (1) Use existing quality improvement

  12. Opening the black box—Development, testing and documentation of a mechanistically rich agent-based model

    DEFF Research Database (Denmark)

    Topping, Chris J.; Høye, Toke; Olesen, Carsten Riis

    2010-01-01

    Although increasingly widely used in biology, complex adaptive simulation models such as agent-based models have been criticised for being difficult to communicate and test. This study demonstrates the application of pattern-oriented model testing, and a novel documentation procedure to present...... accessible description of the processes included in the model. Application of the model to a comprehensive historical data set supported the hypothesis that interference competition is the primary population regulating factor in the absence of mammal predators in the brown hare, and that the effect works...

  13. Feasibility and effectiveness of two community-based HIV testing models in rural Swaziland.

    Science.gov (United States)

    Parker, Lucy Anne; Jobanputra, Kiran; Rusike, Lorraine; Mazibuko, Sikhathele; Okello, Velephi; Kerschberger, Bernhard; Jouquet, Guillaume; Cyr, Joanne; Teck, Roger

    2015-07-01

    To evaluate the feasibility (population reached, costs) and effectiveness (positivity rates, linkage to care) of two strategies of community-based HIV testing and counselling (HTC) in rural Swaziland. Strategies used were mobile HTC (MHTC) and home-based HTC (HBHTC). Information on age, sex, previous testing and HIV results was obtained from routine HTC records. A consecutive series of individuals testing HIV-positive were followed up for 6 months from the test date to assess linkage to care. A total of 9 060 people were tested: 2 034 through MHTC and 7 026 through HBHTC. A higher proportion of children and adolescents (<20 years) were tested through HBHTC than MHTC (57% vs. 17%; P < 0.001). MHTC reached a higher proportion of adult men than HBHTC (42% vs. 39%; P = 0.015). Of 398 HIV-positive individuals, only 135 (34%) were enrolled in HIV care within 6 months. Of 42 individuals eligible for antiretroviral therapy, 22 (52%) started treatment within 6 months. Linkage to care was lowest among people who had tested previously and those aged 20-40 years. HBHTC was 50% cheaper (US$11 per person tested; $797 per individual enrolled in HIV care) than MHTC ($24 and $1698, respectively). In this high HIV prevalence setting, a community-based testing programme achieved high uptake of testing and appears to be an effective and affordable way to encourage large numbers of people to learn their HIV status (particularly underserved populations such as men and young people). However, for community HTC to impact mortality and incidence, strategies need to be implemented to ensure people testing HIV-positive in the community are linked to HIV care. © 2015 The Authors. Tropical Medicine & International Health Published by John Wiley & Sons Ltd.

  14. Fiber Bragg Grating-Based Performance Monitoring of Piles Fiber in a Geotechnical Centrifugal Model Test

    Directory of Open Access Journals (Sweden)

    Xiaolin Weng

    2014-01-01

    Full Text Available In centrifugal tests, conventional sensors can hardly capture the performance of reinforcement in small-scale models. However, recent advances in fiber optic sensing technologies enable the accurate and reliable monitoring of strain and temperature in laboratory geotechnical tests. This paper outlines a centrifugal model test, performed using a 60 g ton geocentrifuge, to investigate the performance of pipe piles used to reinforce the loess foundation below a widened embankment. Prior to the test, quasidistributed fiber Bragg grating (FBG strain sensors were attached to the surface of the pipe piles to measure the lateral friction resistance in real time. Via the centrifuge actuator, the driving of pipe piles was simulated. During testing, the variations of skin friction distribution along the pipe piles were measured automatically using an optical fiber interrogator. This paper represents the presentation and detailed analysis of monitoring results. Herein, we verify the reliability of the fiber optic sensors in monitoring the model piles without affecting the integrity of the centrifugal model. This paper, furthermore, shows that lateral friction resistance developed in stages with the pipe piles being pressed in and that this sometimes may become negative.

  15. Springback study in aluminum alloys based on the Demeri Benchmark Test : influence of material model

    International Nuclear Information System (INIS)

    Greze, R.; Laurent, H.; Manach, P. Y.

    2007-01-01

    Springback is a serious problem in sheet metal forming. Its origin lies in the elastic recovery of materials after a deep drawing operation. Springback modifies the final shape of the part when removed from the die after forming. This study deals with Springback in an Al5754-O aluminum alloy. An experimental test similar to the Demeri Benchmark Test has been developed. The experimentally measured Springback is compared to predicted Springback simulation using Abaqus software. Several material models are analyzed, all models using isotropic hardening of Voce type and plasticity criteria such as Von Mises and Hill48's yield criterion

  16. Performance Analysis of Spotify® for Android with Model-Based Testing

    Directory of Open Access Journals (Sweden)

    Ana Rosario Espada

    2017-01-01

    Full Text Available This paper presents the foundations and the real use of a tool to automatically detect anomalies in Internet traffic produced by mobile applications. In particular, our MVE tool is focused on analyzing the impact that user interactions have on the traffic produced and received by the smartphones. To make the analysis exhaustive with regard to the potential user behaviors, we follow a model-based approach to automatically generate test cases to be executed on the smartphones. In addition, we make use of a specification language to define traffic patterns to be compared with the actual traffic in the device. MVE also includes monitoring and verification support to detect executions that do not fit the patterns. In these cases, the developer will obtain detailed information on the user actions that produce the anomaly in order to improve the application. To validate the approach, the paper presents an experimental study with the well-known Spotify app for Android, in which we detected some interesting behaviors. For instance, some HTTP connections do not end successfully due to timeout errors from the remote Spotify service.

  17. Analytical Model of Coil Spring Damper Based on the Loading Test

    Energy Technology Data Exchange (ETDEWEB)

    Cho, Sung Gook; Park, Woong Ki [INNOSE TECH Co. LTD, Incheon (Korea, Republic of); Furuya, Osamu [Tokyo City University, Tokyo (Japan); Kurabayashi, Hiroshi [Vibro-System, Tokyo (Japan)

    2016-05-15

    The one way of solving such problems is to enhance and to develop an improved damping element used in base-isolation and response control system. A cost reduction of damper for a large scale structure is another important task to upgrade the total response control abilities in the near future. This study has examined a response control device using elastoplastic hysteresis damping of metal material. The proposed damper is designed to be coil spring element shape for a uniform stress of metal and for a reduction of low cyclic fatigue in large deformation to upgrade a repetitive strength during the earthquake motions. By using the metal material of SS400 general structural rolled steel, the corresponding cost issues of the damping element will be effectively reduced. The analytical of elasto-plastic coil spring damper (CSD) is introduced, and basic mechanical properties evaluated experimentally and analytically. This study has been examined the response control damper using elasto-plastic hysteresis characteristics of metal material. The paper described the design method of elasto-plastic coil spring damper, basic mechanical properties evaluated from loading test, and analytical model of damper are summarized. It was confirmed that the damping force and mechanical characteristics of elasto-plastic coil spring damper are almost satisfied the design specifications.

  18. Meaning-Based Scoring: A Systemic Functional Linguistics Model for Automated Test Tasks

    Science.gov (United States)

    Gleason, Jesse

    2014-01-01

    Communicative approaches to language teaching that emphasize the importance of speaking (e.g., task-based language teaching) require innovative and evidence-based means of assessing oral language. Nonetheless, research has yet to produce an adequate assessment model for oral language (Chun 2006; Downey et al. 2008). Limited by automatic speech…

  19. Model-based integration and testing : bridging the gap between academic theory and industrial practice

    NARCIS (Netherlands)

    Braspenning, N.C.W.M.

    2008-01-01

    For manufacturers of high-tech multi-disciplinary systems such as semiconductor equipment, the effort required for integration and system testing is ever increasing, while customers demand a shorter time-to-market.This book describes how executable models can replace unavailable component

  20. Loglinear Rasch model tests

    NARCIS (Netherlands)

    Kelderman, Hendrikus

    1984-01-01

    Existing statistical tests for the fit of the Rasch model have been criticized, because they are only sensitive to specific violations of its assumptions. Contingency table methods using loglinear models have been used to test various psychometric models. In this paper, the assumptions of the Rasch

  1. Testing a Dual Process Model of Gender-Based Violence: A Laboratory Examination.

    Science.gov (United States)

    Berke, Danielle S; Zeichner, Amos

    2016-01-01

    The dire impact of gender-based violence on society compels development of models comprehensive enough to capture the diversity of its forms. Research has established hostile sexism (HS) as a robust predictor of gender-based violence. However, to date, research has yet to link men's benevolent sexism (BS) to physical aggression toward women, despite correlations between BS and HS and between BS and victim blaming. One model, the opposing process model of benevolent sexism (Sibley & Perry, 2010), suggests that, for men, BS acts indirectly through HS to predict acceptance of hierarchy-enhancing social policy as an expression of a preference for in-group dominance (i. e., social dominance orientation [SDO]). The extent to which this model applies to gender-based violence remains untested. Therefore, in this study, 168 undergraduate men in a U. S. university participated in a competitive reaction time task, during which they had the option to shock an ostensible female opponent as a measure of gender-based violence. Results of multiple-mediation path analyses indicated dual pathways potentiating gender-based violence and highlight SDO as a particularly potent mechanism of this violence. Findings are discussed in terms of group dynamics and norm-based violence prevention.

  2. Measuring Japanese EFL Student Perceptions of Internet-Based Tests with the Technology Acceptance Model

    Science.gov (United States)

    Dizon, Gilbert

    2016-01-01

    The Internet has made it possible for teachers to administer online assessments with affordability and ease. However, little is known about Japanese English as a Foreign Language (EFL) students' attitudes of internet-based tests (IBTs). Therefore, this study aimed to measure the perceptions of IBTs among Japanese English language learners with the…

  3. Cohesive Zone Model Based Numerical Analysis of Steel-Concrete Composite Structure Push-Out Tests

    Directory of Open Access Journals (Sweden)

    J. P. Lin

    2014-01-01

    Full Text Available Push-out tests were widely used to determine the shear bearing capacity and shear stiffness of shear connectors in steel-concrete composite structures. The finite element method was one efficient alternative to push-out testing. This paper focused on a simulation analysis of the interface between concrete slabs and steel girder flanges as well as the interface of the shear connectors and the surrounding concrete. A cohesive zone model was used to simulate the tangential sliding and normal separation of the interfaces. Then, a zero-thickness cohesive element was implemented via the user-defined element subroutine UEL in the software ABAQUS, and a multiple broken line mode was used to define the constitutive relations of the cohesive zone. A three-dimensional numerical analysis model was established for push-out testing to analyze the load-displacement curves of the push-out test process, interface relative displacement, and interface stress distribution. This method was found to accurately calculate the shear capacity and shear stiffness of shear connectors. The numerical results showed that the multiple broken lines mode cohesive zone model could describe the nonlinear mechanical behavior of the interface between steel and concrete and that a discontinuous deformation numerical simulation could be implemented.

  4. Evaluation of the base/subgrade soil under repeated loading : phase I--laboratory testing and numerical modeling of geogrid reinforced bases in flexible pavement.

    Science.gov (United States)

    2009-10-01

    This report documents the results of a study that was conducted to characterize the behavior of geogrid reinforced base : course materials. The research was conducted through an experimental testing and numerical modeling programs. The : experimental...

  5. Performance Model for High-Power Lithium Titanate Oxide Batteries based on Extended Characterization Tests

    DEFF Research Database (Denmark)

    Stroe, Ana-Irina; Swierczynski, Maciej Jozef; Stroe, Daniel Ioan

    2015-01-01

    Lithium-ion (Li-ion) batteries are found nowadays not only in portable/consumer electronics but also in more power demanding applications, such as stationary renewable energy storage, automotive and back-up power supply, because of their superior characteristics in comparison to other energy...... storage technologies. Nevertheless, prior to be used in any of the aforementioned application, a Li-ion battery cell must be intensively characterized and its behavior needs to be understood. This can be realized by performing extended laboratory characterization tests and developing Li-ion battery...... performance models. Furthermore, accurate performance models are necessary in order to analyze the behavior of the battery cell under different mission profiles, by simulation; thus, avoiding time and cost demanding real life tests. This paper presents the development and the parametrization of a performance...

  6. Q-Matrix Optimization Based on the Linear Logistic Test Model.

    Science.gov (United States)

    Ma, Lin; Green, Kelly E

    This study explored optimization of item-attribute matrices with the linear logistic test model (Fischer, 1973), with optimal models explaining more variance in item difficulty due to identified item attributes. Data were 8th-grade mathematics test item responses of two TIMSS 2007 booklets. The study investigated three categories of attributes (content, cognitive process, and comprehensive cognitive process) at two grain levels (larger, smaller) and also compared results with random attribute matrices. The proposed attributes accounted for most of the variance in item difficulty for two assessment booklets (81% and 65%). The variance explained by the content attributes was very small (13% to 31%), less than variance explained by the comprehensive cognitive process attributes which explained much more variance than the content and cognitive process attributes. The variances explained by the grain level were similar to each other. However, the attributes did not predict the item difficulties of two assessment booklets equally.

  7. Radiation Belt Test Model

    Science.gov (United States)

    Freeman, John W.

    2000-10-01

    Rice University has developed a dynamic model of the Earth's radiation belts based on real-time data driven boundary conditions and full adiabaticity. The Radiation Belt Test Model (RBTM) successfully replicates the major features of storm-time behavior of energetic electrons: sudden commencement induced main phase dropout and recovery phase enhancement. It is the only known model to accomplish the latter. The RBTM shows the extent to which new energetic electrons introduced to the magnetosphere near the geostationary orbit drift inward due to relaxation of the magnetic field. It also shows the effects of substorm related rapid motion of magnetotail field lines for which the 3rd adiabatic invariant is violated. The radial extent of this violation is seen to be sharply delineated to a region outside of 5Re, although this distance is determined by the Hilmer-Voigt magnetic field model used by the RBTM. The RBTM appears to provide an excellent platform on which to build parameterized refinements to compensate for unknown acceleration processes inside 5Re where adiabaticity is seen to hold. Moreover, built within the framework of the MSFM, it offers the prospect of an operational forecast model for MeV electrons.

  8. Model-Based Structural Health Monitoring of Fatigue Damage Test-Bed Specimens

    Science.gov (United States)

    2011-11-15

    the hull welds or notches along component edges are good initial candidates for the hypothetical damage initiation areas. The branching process adds...to it off-center. The base plate and the stiffener plate are rigidly welded by a tungsten inert gas ( TIG ) weld . Three different crack paths...shown in Figure 9(a), an 18 in long stiffener plate has been welded to each of the tested plates with 0.625 in long discrete TIG welds at 5 locations

  9. Testing Models of Psychopathology in Preschool-aged Children Using a Structured Interview-based Assessment

    Science.gov (United States)

    Dougherty, Lea R.; Bufferd, Sara J.; Carlson, Gabrielle A.; Klein, Daniel N.

    2014-01-01

    A number of studies have found that broadband internalizing and externalizing factors provide a parsimonious framework for understanding the structure of psychopathology across childhood, adolescence, and adulthood. However, few of these studies have examined psychopathology in young children, and several recent studies have found support for alternative models, including a bi-factor model with common and specific factors. The present study used parents’ (typically mothers’) reports on a diagnostic interview in a community sample of 3-year old children (n=541; 53.9 % male) to compare the internalizing-externalizing latent factor model with a bi-factor model. The bi-factor model provided a better fit to the data. To test the concurrent validity of this solution, we examined associations between this model and paternal reports and laboratory observations of child temperament. The internalizing factor was associated with low levels of surgency and high levels of fear; the externalizing factor was associated with high levels of surgency and disinhibition and low levels of effortful control; and the common factor was associated with high levels of surgency and negative affect and low levels of effortful control. These results suggest that psychopathology in preschool-aged children may be explained by a single, common factor influencing nearly all disorders and unique internalizing and externalizing factors. These findings indicate that shared variance across internalizing and externalizing domains is substantial and are consistent with recent suggestions that emotion regulation difficulties may be a common vulnerability for a wide array of psychopathology. PMID:24652485

  10. Comparison of rate theory based modeling calculations with the surveillance test results of Korean light water reactors

    International Nuclear Information System (INIS)

    Lee, Gyeong Geun; Lee, Yong Bok; Kim, Min Chul; Kwon, Junh Yun

    2012-01-01

    Neutron irradiation to reactor pressure vessel (RPV) steels causes a decrease in fracture toughness and an increase in yield strength while in service. It is generally accepted that the growth of point defect cluster (PDC) and copper rich precipitate (CRP) affects radiation hardening of RPV steels. A number of models have been proposed to account for the embrittlement of RPV steels. The rate theory based modeling mathematically described the evolution of radiation induced microstructures of ferritic steels under neutron irradiation. In this work, we compared the rate theory based modeling calculation with the surveillance test results of Korean Light Water Reactors (LWRs)

  11. Improving Junior High Schools’ Critical Thinking Skills Based on Test Three Different Models of Learning

    Directory of Open Access Journals (Sweden)

    Nur Miftahul Fuad

    2017-01-01

    Full Text Available The aims of this study were (1 to find out the differences in critical thinking skills among students who were given three different learning models: differentiated science inquiry combined with mind map, differentiated science inquiry model, and conventional model, (2 to find out the differences of critical thinking skills among male and female students. This study is a quasi-experimental research with pretest-posttest nonequivalent control group design. The population in this research is the seventh grade students of junior high schools in Kediri, Indonesia. The sample of the research is in the number of 96 students distributed in three classes at different schools. The data of critical thinking skills are gained from test scores and then analyzed using descriptive and inferential statistics through ANCOVA. The results of research revealed that there are different skills in critical thinking in different models. The highest skills in critical thinking are reached by students who were given differentiated science inquiry model combined with mind map in their learning. There are also differences in critical thinking skills between male and female students.

  12. Trait-based representation of biological nitrification: Model development, testing, and predicted community composition

    Directory of Open Access Journals (Sweden)

    Nick eBouskill

    2012-10-01

    Full Text Available Trait-based microbial models show clear promise as tools to represent the diversity and activity of microorganisms across ecosystem gradients. These models parameterize specific traits that determine the relative fitness of an ‘organism’ in a given environment, and represent the complexity of biological systems across temporal and spatial scales. In this study we introduce a microbial community trait-based modeling framework (MicroTrait focused on nitrification (MicroTrait-N that represents the ammonia-oxidizing bacteria (AOB and ammonia-oxidizing archaea (AOA and nitrite oxidizing bacteria (NOB using traits related to enzyme kinetics and physiological properties. We used this model to predict nitrifier diversity, ammonia (NH3 oxidation rates and nitrous oxide (N2O production across pH, temperature and substrate gradients. Predicted nitrifier diversity was predominantly determined by temperature and substrate availability, the latter was strongly influenced by pH. The model predicted that transient N2O production rates are maximized by a decoupling of the AOB and NOB communities, resulting in an accumulation and detoxification of nitrite to N2O by AOB. However, cumulative N2O production (over six month simulations is maximized in a system where the relationship between AOB and NOB is maintained. When the reactions uncouple, the AOB become unstable and biomass declines rapidly, resulting in decreased NH3 oxidation and N2O production. We evaluated this model against site level chemical datasets from the interior of Alaska and accurately simulated NH3 oxidation rates and the relative ratio of AOA:AOB biomass. The predicted community structure and activity indicate (a parameterization of a small number of traits may be sufficient to broadly characterize nitrifying community structure and (b changing decadal trends in climate and edaphic conditions could impact nitrification rates in ways that are not captured by extant biogeochemical models.

  13. Dynamic analysis of ITER tokamak. Based on results of vibration test using scaled model

    International Nuclear Information System (INIS)

    Takeda, Nobukazu; Kakudate, Satoshi; Nakahira, Masataka

    2005-01-01

    The vibration experiments of the support structures with flexible plates for the ITER major components such as toroidal field coil (TF coil) and vacuum vessel (VV) were performed using small-sized flexible plates aiming to obtain its basic mechanical characteristics such as dependence of the stiffness on the loading angle. The experimental results were compared with the analytical ones in order to estimate an adequate analytical model for ITER support structure with flexible plates. As a result, the bolt connection of the flexible plates on the base plate strongly affected on the stiffness of the flexible plates. After studies of modeling the connection of the bolts, it is found that the analytical results modeling the bolts with finite stiffness only in the axial direction and infinite stiffness in the other directions agree well with the experimental ones. Based on this, numerical analysis regarding the actual support structure of the ITER VV and TF coil was performed. The support structure composed of flexible plates and connection bolts was modeled as a spring composed of only two spring elements simulating the in-plane and out-of-plane stiffness of the support structure with flexible plates including the effect of connection bolts. The stiffness of both spring models for VV and TF coil agree well with that of shell models, simulating actual structures such as flexible plates and connection bolts based on the experimental results. It is therefore found that the spring model with the only two values of stiffness enables to simplify the complicated support structure with flexible plates for the dynamic analysis of the VV and TF coil. Using the proposed spring model, the dynamic analysis of the VV and TF coil for the ITER were performed to estimate the integrity under the design earthquake. As a result, it is found that the maximum relative displacement of 8.6 mm between VV and TF coil is much less than 100 mm, so that the integrity of the VV and TF coil of the

  14. Physics-based modeling of live wildland fuel ignition experiments in the Forced Ignition and Flame Spread Test apparatus

    Science.gov (United States)

    C. Anand; B. Shotorban; S. Mahalingam; S. McAllister; D. R. Weise

    2017-01-01

    A computational study was performed to improve our understanding of the ignition of live fuel in the forced ignition and flame spread test apparatus, a setup where the impact of the heating mode is investigated by subjecting the fuel to forced convection and radiation. An improvement was first made in the physics-based model WFDS where the fuel is treated as fixed...

  15. Predictors of Willingness to Read in English: Testing a Model Based on Possible Selves and Self-Confidence

    Science.gov (United States)

    Khajavy, Gholam Hassan; Ghonsooly, Behzad

    2017-01-01

    The aim of the present study is twofold. First, it tests a model of willingness to read (WTR) based on L2 motivation and communication confidence (communication anxiety and perceived communicative competence). Second, it applies the recent theory of L2 motivation proposed by Dörnyei [2005. "The Psychology of Language Learner: Individual…

  16. Evidence used in model-based economic evaluations for evaluating pharmacogenetic and pharmacogenomic tests: a systematic review protocol.

    Science.gov (United States)

    Peters, Jaime L; Cooper, Chris; Buchanan, James

    2015-11-11

    Decision models can be used to conduct economic evaluations of new pharmacogenetic and pharmacogenomic tests to ensure they offer value for money to healthcare systems. These models require a great deal of evidence, yet research suggests the evidence used is diverse and of uncertain quality. By conducting a systematic review, we aim to investigate the test-related evidence used to inform decision models developed for the economic evaluation of genetic tests. We will search electronic databases including MEDLINE, EMBASE and NHS EEDs to identify model-based economic evaluations of pharmacogenetic and pharmacogenomic tests. The search will not be limited by language or date. Title and abstract screening will be conducted independently by 2 reviewers, with screening of full texts and data extraction conducted by 1 reviewer, and checked by another. Characteristics of the decision problem, the decision model and the test evidence used to inform the model will be extracted. Specifically, we will identify the reported evidence sources for the test-related evidence used, describe the study design and how the evidence was identified. A checklist developed specifically for decision analytic models will be used to critically appraise the models described in these studies. Variations in the test evidence used in the decision models will be explored across the included studies, and we will identify gaps in the evidence in terms of both quantity and quality. The findings of this work will be disseminated via a peer-reviewed journal publication and at national and international conferences. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/

  17. A Comparison of Computer-Based Classification Testing Approaches Using Mixed-Format Tests with the Generalized Partial Credit Model

    Science.gov (United States)

    Kim, Jiseon

    2010-01-01

    Classification testing has been widely used to make categorical decisions by determining whether an examinee has a certain degree of ability required by established standards. As computer technologies have developed, classification testing has become more computerized. Several approaches have been proposed and investigated in the context of…

  18. Evaluation of exercise tolerance patients in cardiac rehabilitation D model based on 6 Minute Walk Test

    Directory of Open Access Journals (Sweden)

    Bielawa Lukasz.

    2012-12-01

    Full Text Available Evaluation of the results of 6-minute walk test depending on gender, age, left ventricular ejection fraction, the primary disease and BMI. Patients underwent assessment of Cardiac Rehabilitation Department in Szymbark in 2012 (80 people. Duration of rehabilitation for all patients was 21 days. The test was performed at the beginning and end of the cycle. Following the 3-week cardiac rehabilitation in the model D in a group of 80 patients with a mean age of 72 years achieved a statistically significant improvement in exercise capacity, expressed in the increase in test 6MWT distance by an average of 52 meters. In the study, men received final results statistically superior to women. The largest increase in the distance gained to patients after aortic valve prosthesis. People who are obese with a body mass index BMI over 30 have an average trip distance underperform both at baseline, final, and in the resulting increase of the distance than those with a BMI under 30. Prevention of obesity, one of the modifiable risk factors for cardiovascular disease should be the goal of training during cardiac rehabilitation patient education.

  19. Verifying the functional ability of microstructured surfaces by model-based testing

    Science.gov (United States)

    Hartmann, Wito; Weckenmann, Albert

    2014-09-01

    Micro- and nanotechnology enables the use of new product features such as improved light absorption, self-cleaning or protection, which are based, on the one hand, on the size of functional nanostructures and the other hand, on material-specific properties. With the need to reliably measure progressively smaller geometric features, coordinate and surface-measuring instruments have been refined and now allow high-resolution topography and structure measurements down to the sub-nanometre range. Nevertheless, in many cases it is not possible to make a clear statement about the functional ability of the workpiece or its topography because conventional concepts of dimensioning and tolerancing are solely geometry oriented and standardized surface parameters are not sufficient to consider interaction with non-geometric parameters, which are dominant for functions such as sliding, wetting, sealing and optical reflection. To verify the functional ability of microstructured surfaces, a method was developed based on a parameterized mathematical-physical model of the function. From this model, function-related properties can be identified and geometric parameters can be derived, which may be different for the manufacturing and verification processes. With this method it is possible to optimize the definition of the shape of the workpiece regarding the intended function by applying theoretical and experimental knowledge, as well as modelling and simulation. Advantages of this approach will be discussed and demonstrated by the example of a microstructured inking roll.

  20. Agent based models for testing city evacuation strategies under a flood event as strategy to reduce flood risk

    Science.gov (United States)

    Medina, Neiler; Sanchez, Arlex; Nokolic, Igor; Vojinovic, Zoran

    2016-04-01

    This research explores the uses of Agent Based Models (ABM) and its potential to test large scale evacuation strategies in coastal cities at risk from flood events due to extreme hydro-meteorological events with the final purpose of disaster risk reduction by decreasing human's exposure to the hazard. The first part of the paper corresponds to the theory used to build the models such as: Complex adaptive systems (CAS) and the principles and uses of ABM in this field. The first section outlines the pros and cons of using AMB to test city evacuation strategies at medium and large scale. The second part of the paper focuses on the central theory used to build the ABM, specifically the psychological and behavioral model as well as the framework used in this research, specifically the PECS reference model is cover in this section. The last part of this section covers the main attributes or characteristics of human beings used to described the agents. The third part of the paper shows the methodology used to build and implement the ABM model using Repast-Symphony as an open source agent-based modelling and simulation platform. The preliminary results for the first implementation in a region of the island of Sint-Maarten a Dutch Caribbean island are presented and discussed in the fourth section of paper. The results obtained so far, are promising for a further development of the model and its implementation and testing in a full scale city

  1. Design and clinical pilot testing of the model-based dynamic insulin sensitivity and secretion test (DISST).

    Science.gov (United States)

    Lotz, Thomas F; Chase, J Geoffrey; McAuley, Kirsten A; Shaw, Geoffrey M; Docherty, Paul D; Berkeley, Juliet E; Williams, Sheila M; Hann, Christopher E; Mann, Jim I

    2010-11-01

    Insulin resistance is a significant risk factor in the pathogenesis of type 2 diabetes. This article presents pilot study results of the dynamic insulin sensitivity and secretion test (DISST), a high-resolution, low-intensity test to diagnose insulin sensitivity (IS) and characterize pancreatic insulin secretion in response to a (small) glucose challenge. This pilot study examines the effect of glucose and insulin dose on the DISST, and tests its repeatability. DISST tests were performed on 16 subjects randomly allocated to low (5 g glucose, 0.5 U insulin), medium (10 g glucose, 1 U insulin) and high dose (20 g glucose, 2 U insulin) protocols. Two or three tests were performed on each subject a few days apart. Average variability in IS between low and medium dose was 10.3% (p=.50) and between medium and high dose 6.0% (p=.87). Geometric mean variability between tests was 6.0% (multiplicative standard deviation (MSD) 4.9%). Geometric mean variability in first phase endogenous insulin response was 6.8% (MSD 2.2%). Results were most consistent in subjects with low IS. These findings suggest that DISST may be an easily performed dynamic test to quantify IS with high resolution, especially among those with reduced IS. © 2010 Diabetes Technology Society.

  2. Development, standardization and testing of a bacterial wound infection model based on ex vivo human skin.

    Directory of Open Access Journals (Sweden)

    Christoph Schaudinn

    Full Text Available Current research on wound infections is primarily conducted on animal models, which limits direct transferability of these studies to humans. Some of these limitations can be overcome by using-otherwise discarded-skin from cosmetic surgeries. Superficial wounds are induced in fresh ex vivo skin, followed by intradermal injection of Pseudomonas aeruginosa under the wound. Subsequently, the infected skin is incubated for 20 hours at 37°C and the CFU/wound are determined. Within 20 hours, the bacteria count increased from 107 to 109 bacteria per wound, while microscopy revealed a dense bacterial community in the collagen network of the upper wound layers as well as numerous bacteria scattered in the dermis. At the same time, IL-1alpha and IL-1beta amounts increased in all infected wounds, while-due to bacteria-induced cell lysis-the IL-6 and IL-8 concentrations rose only in the uninfected samples. High-dosage ciprofloxacin treatment resulted in a decisive decrease in bacteria, but consistently failed to eradicate all bacteria. The main benefits of the ex vivo wound model are the use of healthy human skin, a quantifiable bacterial infection, a measureable donor-dependent immune response and a good repeatability of the results. These properties turn the ex vivo wound model into a valuable tool to examine the mechanisms of host-pathogen interactions and to test antimicrobial agents.

  3. Modeling and Testing of Phase Transition-Based Deployable Systems for Small Body Sample Capture

    Science.gov (United States)

    Quadrelli, Marco; Backes, Paul; Wilkie, Keats; Giersch, Lou; Quijano, Ubaldo; Keim, Jason; Mukherjee, Rudranarayan

    2009-01-01

    This paper summarizes the modeling, simulation, and testing work related to the development of technology to investigate the potential that shape memory actuation has to provide mechanically simple and affordable solutions for delivering assets to a surface and for sample capture and return. We investigate the structural dynamics and controllability aspects of an adaptive beam carrying an end-effector which, by changing equilibrium phases is able to actively decouple the end-effector dynamics from the spacecraft dynamics during the surface contact phase. Asset delivery and sample capture and return are at the heart of several emerging potential missions to small bodies, such as asteroids and comets, and to the surface of large bodies, such as Titan.

  4. Neighborhood social capital and adult health: an empirical test of a Bourdieu-based model.

    Science.gov (United States)

    Carpiano, Richard M

    2007-09-01

    Drawing upon Bourdieu's [1986. The forms of capital. In: Richardson, J.G. (Ed.), Handbook of Theory and Research for the Sociology of Education. Greenwood, New York, pp. 241-258.] social capital theory, I test a conceptual model of neighborhood conditions and social capital - considering relationships between neighborhood social capital forms (social support, social leverage, informal social control, and neighborhood organization participation) and adult health behaviors (smoking, binge drinking) and perceived health, as well as interactions between neighborhood social capital and individuals' access to that social capital. Analyzing Los Angeles Family and Neighborhood Survey data linked with tract level census data, results suggest that specific social capital forms were directly associated with both positive and negative health outcomes. Additionally, residents' neighborhood attachment moderated relationships between various social capital forms and health. Future studies should consider social capital resources and the role of differential access to such resources for promoting or compromising health.

  5. Evaluation of liquefaction potential of soil based on standard penetration test using multi-gene genetic programming model

    Science.gov (United States)

    Muduli, Pradyut; Das, Sarat

    2014-06-01

    This paper discusses the evaluation of liquefaction potential of soil based on standard penetration test (SPT) dataset using evolutionary artificial intelligence technique, multi-gene genetic programming (MGGP). The liquefaction classification accuracy (94.19%) of the developed liquefaction index (LI) model is found to be better than that of available artificial neural network (ANN) model (88.37%) and at par with the available support vector machine (SVM) model (94.19%) on the basis of the testing data. Further, an empirical equation is presented using MGGP to approximate the unknown limit state function representing the cyclic resistance ratio (CRR) of soil based on developed LI model. Using an independent database of 227 cases, the overall rates of successful prediction of occurrence of liquefaction and non-liquefaction are found to be 87, 86, and 84% by the developed MGGP based model, available ANN and the statistical models, respectively, on the basis of calculated factor of safety (F s) against the liquefaction occurrence.

  6. Computer-Based Testing: Test Site Security.

    Science.gov (United States)

    Rosen, Gerald A.

    Computer-based testing places great burdens on all involved parties to ensure test security. A task analysis of test site security might identify the areas of protecting the test, protecting the data, and protecting the environment as essential issues in test security. Protecting the test involves transmission of the examinations, identifying the…

  7. SAR-based change detection using hypothesis testing and Markov random field modelling

    Science.gov (United States)

    Cao, W.; Martinis, S.

    2015-04-01

    The objective of this study is to automatically detect changed areas caused by natural disasters from bi-temporal co-registered and calibrated TerraSAR-X data. The technique in this paper consists of two steps: Firstly, an automatic coarse detection step is applied based on a statistical hypothesis test for initializing the classification. The original analytical formula as proposed in the constant false alarm rate (CFAR) edge detector is reviewed and rewritten in a compact form of the incomplete beta function, which is a builtin routine in commercial scientific software such as MATLAB and IDL. Secondly, a post-classification step is introduced to optimize the noisy classification result in the previous step. Generally, an optimization problem can be formulated as a Markov random field (MRF) on which the quality of a classification is measured by an energy function. The optimal classification based on the MRF is related to the lowest energy value. Previous studies provide methods for the optimization problem using MRFs, such as the iterated conditional modes (ICM) algorithm. Recently, a novel algorithm was presented based on graph-cut theory. This method transforms a MRF to an equivalent graph and solves the optimization problem by a max-flow/min-cut algorithm on the graph. In this study this graph-cut algorithm is applied iteratively to improve the coarse classification. At each iteration the parameters of the energy function for the current classification are set by the logarithmic probability density function (PDF). The relevant parameters are estimated by the method of logarithmic cumulants (MoLC). Experiments are performed using two flood events in Germany and Australia in 2011 and a forest fire on La Palma in 2009 using pre- and post-event TerraSAR-X data. The results show convincing coarse classifications and considerable improvement by the graph-cut post-classification step.

  8. A Multi-Marker Genetic Association Test Based on the Rasch Model Applied to Alzheimer's Disease.

    Directory of Open Access Journals (Sweden)

    Wenjia Wang

    Full Text Available Results from Genome-Wide Association Studies (GWAS have shown that the genetic basis of complex traits often include many genetic variants with small to moderate effects whose identification remains a challenging problem. In this context multi-marker analysis at the gene and pathway level can complement traditional point-wise approaches that treat the genetic markers individually. In this paper we propose a novel statistical approach for multi-marker analysis based on the Rasch model. The method summarizes the categorical genotypes of SNPs by a generalized logistic function into a genetic score that can be used for association analysis. Through different sets of simulations, the false-positive rate and power of the proposed approach are compared to a set of existing methods, and shows good performances. The application of the Rasch model on Alzheimer's Disease (AD ADNI GWAS dataset also allows a coherent interpretation of the results. Our analysis supports the idea that APOE is a major susceptibility gene for AD. In the top genes selected by proposed method, several could be functionally linked to AD. In particular, a pathway analysis of these genes also highlights the metabolism of cholesterol, that is known to play a key role in AD pathogenesis. Interestingly, many of these top genes can be integrated in a hypothetic signalling network.

  9. A spheroid-based 3-D culture model for pancreatic cancer drug testing, using the acid phosphatase assay

    International Nuclear Information System (INIS)

    Wen, Z.; Liao, Q.; Hu, Y.; You, L.; Zhou, L.; Zhao, Y.

    2013-01-01

    Current therapy for pancreatic cancer is multimodal, involving surgery and chemotherapy. However, development of pancreatic cancer therapies requires a thorough evaluation of drug efficacy in vitro before animal testing and subsequent clinical trials. Compared to two-dimensional culture of cell monolayer, three-dimensional (3-D) models more closely mimic native tissues, since the tumor microenvironment established in 3-D models often plays a significant role in cancer progression and cellular responses to the drugs. Accumulating evidence has highlighted the benefits of 3-D in vitro models of various cancers. In the present study, we have developed a spheroid-based, 3-D culture of pancreatic cancer cell lines MIAPaCa-2 and PANC-1 for pancreatic drug testing, using the acid phosphatase assay. Drug efficacy testing showed that spheroids had much higher drug resistance than monolayers. This model, which is characteristically reproducible and easy and offers rapid handling, is the preferred choice for filling the gap between monolayer cell cultures and in vivo models in the process of drug development and testing for pancreatic cancer

  10. A spheroid-based 3-D culture model for pancreatic cancer drug testing, using the acid phosphatase assay

    Directory of Open Access Journals (Sweden)

    Z. Wen

    2013-08-01

    Full Text Available Current therapy for pancreatic cancer is multimodal, involving surgery and chemotherapy. However, development of pancreatic cancer therapies requires a thorough evaluation of drug efficacy in vitro before animal testing and subsequent clinical trials. Compared to two-dimensional culture of cell monolayer, three-dimensional (3-D models more closely mimic native tissues, since the tumor microenvironment established in 3-D models often plays a significant role in cancer progression and cellular responses to the drugs. Accumulating evidence has highlighted the benefits of 3-D in vitro models of various cancers. In the present study, we have developed a spheroid-based, 3-D culture of pancreatic cancer cell lines MIAPaCa-2 and PANC-1 for pancreatic drug testing, using the acid phosphatase assay. Drug efficacy testing showed that spheroids had much higher drug resistance than monolayers. This model, which is characteristically reproducible and easy and offers rapid handling, is the preferred choice for filling the gap between monolayer cell cultures and in vivo models in the process of drug development and testing for pancreatic cancer.

  11. Fault Modeling and Testing for Analog Circuits in Complex Space Based on Supply Current and Output Voltage

    Directory of Open Access Journals (Sweden)

    Hongzhi Hu

    2015-01-01

    Full Text Available This paper deals with the modeling of fault for analog circuits. A two-dimensional (2D fault model is first proposed based on collaborative analysis of supply current and output voltage. This model is a family of circle loci on the complex plane, and it simplifies greatly the algorithms for test point selection and potential fault simulations, which are primary difficulties in fault diagnosis of analog circuits. Furthermore, in order to reduce the difficulty of fault location, an improved fault model in three-dimensional (3D complex space is proposed, which achieves a far better fault detection ratio (FDR against measurement error and parametric tolerance. To address the problem of fault masking in both 2D and 3D fault models, this paper proposes an effective design for testability (DFT method. By adding redundant bypassing-components in the circuit under test (CUT, this method achieves excellent fault isolation ratio (FIR in ambiguity group isolation. The efficacy of the proposed model and testing method is validated through experimental results provided in this paper.

  12. Numerical model of the nanoindentation test based on the digital material representation of the Ti/TiN multilayers

    Directory of Open Access Journals (Sweden)

    Perzyński Konrad

    2015-06-01

    Full Text Available The developed numerical model of a local nanoindentation test, based on the digital material representation (DMR concept, has been presented within the paper. First, an efficient algorithm describing the pulsed laser deposition (PLD process was proposed to realistically recreate the specific morphology of a nanolayered material in an explicit manner. The nanolayered Ti/TiN composite was selected for the investigation. Details of the developed cellular automata model of the PLD process were presented and discussed. Then, the Ti/TiN DMR was incorporated into the finite element software and numerical model of the nanoindentation test was established. Finally, examples of obtained results presenting capabilities of the proposed approach were highlighted.

  13. Direct-to-consumer advertising of predictive genetic tests: a health belief model based examination of consumer response.

    Science.gov (United States)

    Rollins, Brent L; Ramakrishnan, Shravanan; Perri, Matthew

    2014-01-01

    Direct-to-consumer (DTC) advertising of predictive genetic tests (PGTs) has added a new dimension to health advertising. This study used an online survey based on the health belief model framework to examine and more fully understand consumers' responses and behavioral intentions in response to a PGT DTC advertisement. Overall, consumers reported moderate intentions to talk with their doctor and seek more information about PGTs after advertisement exposure, though consumers did not seem ready to take the advertised test or engage in active information search. Those who perceived greater threat from the disease, however, had significantly greater behavioral intentions and information search behavior.

  14. Base Station Performance Model

    OpenAIRE

    Walsh, Barbara; Farrell, Ronan

    2005-01-01

    At present the testing of power amplifiers within base station transmitters is limited to testing at component level as opposed to testing at the system level. While the detection of catastrophic failure is possible, that of performance degradation is not. This paper proposes a base station model with respect to transmitter output power with the aim of introducing system level monitoring of the power amplifier behaviour within the base station. Our model reflects the expe...

  15. Modeling and test of a kinaesthetic actuator based on MR fluid for haptic applications.

    Science.gov (United States)

    Yang, Tae-Heon; Koo, Jeong-Hoi; Kim, Sang-Youn; Kwon, Dong-Soo

    2017-03-01

    Haptic display units have been widely used for conveying button sensations to users, primarily employing vibrotactile actuators. However, the human feeling for pressing buttons mainly relies on kinaesthetic sensations (rather than vibrotactile sensations), and little studies exist on small-scale kinaesthetic haptic units. Thus, the primary goals of this paper are to design a miniature kinaesthetic actuator based on Magneto-Rheological (MR) fluid that can convey various button-clicking sensations and to experimentally evaluate its haptic performance. The design focuses of the proposed actuator were to produce sufficiently large actuation forces (resistive forces) for human users in a given size constraint and to offer a wide range of actuation forces for conveying vivid haptic sensations to users. To this end, this study first performed a series of parametric studies using mathematical force models for multiple operating modes of MR fluid in conjunction with finite element electromagnetism analysis. After selecting design parameters based on parametric studies, a prototype actuator was constructed, and its performance was evaluated using a dynamic mechanical analyzer. It measured the actuator's resistive force with a varying stroke (pressed depth) up to 1 mm and a varying input current from 0 A to 200 mA. The results show that the proposed actuator creates a wide range of resistive forces from around 2 N (off-state) to over 9.5 N at 200 mA. In order to assess the prototype's performance in the terms of the haptic application prospective, a maximum force rate was calculated to determine just noticeable difference in force changes for the 1 mm stoke of the actuator. The results show that the force rate is sufficient to mimic various levels of button sensations, indicating that the proposed kinaesthetic actuator can offer a wide range of resistive force changes that can be conveyed to human operators.

  16. TestDose: A nuclear medicine software based on Monte Carlo modeling for generating gamma camera acquisitions and dosimetry

    Energy Technology Data Exchange (ETDEWEB)

    Garcia, Marie-Paule, E-mail: marie-paule.garcia@univ-brest.fr; Villoing, Daphnée [UMR 1037 INSERM/UPS, CRCT, 133 Route de Narbonne, 31062 Toulouse (France); McKay, Erin [St George Hospital, Gray Street, Kogarah, New South Wales 2217 (Australia); Ferrer, Ludovic [ICO René Gauducheau, Boulevard Jacques Monod, St Herblain 44805 (France); Cremonesi, Marta; Botta, Francesca; Ferrari, Mahila [European Institute of Oncology, Via Ripamonti 435, Milano 20141 (Italy); Bardiès, Manuel [UMR 1037 INSERM/UPS, CRCT, 133 Route de Narbonne, Toulouse 31062 (France)

    2015-12-15

    Purpose: The TestDose platform was developed to generate scintigraphic imaging protocols and associated dosimetry by Monte Carlo modeling. TestDose is part of a broader project (www.dositest.com) whose aim is to identify the biases induced by different clinical dosimetry protocols. Methods: The TestDose software allows handling the whole pipeline from virtual patient generation to resulting planar and SPECT images and dosimetry calculations. The originality of their approach relies on the implementation of functional segmentation for the anthropomorphic model representing a virtual patient. Two anthropomorphic models are currently available: 4D XCAT and ICRP 110. A pharmacokinetic model describes the biodistribution of a given radiopharmaceutical in each defined compartment at various time-points. The Monte Carlo simulation toolkit GATE offers the possibility to accurately simulate scintigraphic images and absorbed doses in volumes of interest. The TestDose platform relies on GATE to reproduce precisely any imaging protocol and to provide reference dosimetry. For image generation, TestDose stores user’s imaging requirements and generates automatically command files used as input for GATE. Each compartment is simulated only once and the resulting output is weighted using pharmacokinetic data. Resulting compartment projections are aggregated to obtain the final image. For dosimetry computation, emission data are stored in the platform database and relevant GATE input files are generated for the virtual patient model and associated pharmacokinetics. Results: Two samples of software runs are given to demonstrate the potential of TestDose. A clinical imaging protocol for the Octreoscan™ therapeutical treatment was implemented using the 4D XCAT model. Whole-body “step and shoot” acquisitions at different times postinjection and one SPECT acquisition were generated within reasonable computation times. Based on the same Octreoscan™ kinetics, a dosimetry

  17. Testing the standard model

    International Nuclear Information System (INIS)

    Gordon, H.; Marciano, W.; Williams, H.H.

    1982-01-01

    We summarize here the results of the standard model group which has studied the ways in which different facilities may be used to test in detail what we now call the standard model, that is SU/sub c/(3) x SU(2) x U(1). The topics considered are: W +- , Z 0 mass, width; sin 2 theta/sub W/ and neutral current couplings; W + W - , Wγ; Higgs; QCD; toponium and naked quarks; glueballs; mixing angles; and heavy ions

  18. Reactor noise diagnostics based on multivariate autoregressive modeling: Application to LOFT [Loss-of-Fluid-Test] reactor process noise

    International Nuclear Information System (INIS)

    Gloeckler, O.; Upadhyaya, B.R.

    1987-01-01

    Multivariate noise analysis of power reactor operating signals is useful for plant diagnostics, for isolating process and sensor anomalies, and for automated plant monitoring. In order to develop a reliable procedure, the previously established techniques for empirical modeling of fluctuation signals in power reactors have been improved. Application of the complete algorithm to operational data from the Loss-of-Fluid-Test (LOFT) Reactor showed that earlier conjectures (based on physical modeling) regarding the perturbation sources in a Pressurized Water Reactor (PWR) affecting coolant temperature and neutron power fluctuations can be systematically explained. This advanced methodology has important implication regarding plant diagnostics, and system or sensor anomaly isolation. 6 refs., 24 figs

  19. Wave Reflection Model Tests

    DEFF Research Database (Denmark)

    Burcharth, H. F.; Larsen, Brian Juul

    The investigation concerns the design of a new internal breakwater in the main port of Ibiza. The objective of the model tests was in the first hand to optimize the cross section to make the wave reflection low enough to ensure that unacceptable wave agitation will not occur in the port. Secondly...

  20. Work stress and work-based learning in secondary education : testing the Karasek model

    NARCIS (Netherlands)

    Kwakman, Kitty

    2001-01-01

    In this study the Job Demand-Control model was used to study the quality of working life of Dutch secondary teachers. The Job Demand-Control model of Karasek is a theoretical model in which stress and learning are both considered as dependent variables which are influenced by three different task

  1. Work stress and work based learning in secondary education: Testing the Karasek model

    NARCIS (Netherlands)

    Kwakman, Kitty

    1999-01-01

    In this study the Job Demand-Control model was used to study the quality of working life of Dutch secondary teachers. The Job Demand-Control model of Karasek is a theoretical model in which stress and learning are both considered as dependent variables which are influenced by three different task

  2. Improving Junior High Schools' Critical Thinking Skills Based on Test Three Different Models of Learning

    Science.gov (United States)

    Fuad, Nur Miftahul; Zubaidah, Siti; Mahanal, Susriyati; Suarsini, Endang

    2017-01-01

    The aims of this study were (1) to find out the differences in critical thinking skills among students who were given three different learning models: differentiated science inquiry combined with mind map, differentiated science inquiry model, and conventional model, (2) to find out the differences of critical thinking skills among male and female…

  3. Testing the Standard Model

    CERN Document Server

    Riles, K

    1998-01-01

    The Large Electron Project (LEP) accelerator near Geneva, more than any other instrument, has rigorously tested the predictions of the Standard Model of elementary particles. LEP measurements have probed the theory from many different directions and, so far, the Standard Model has prevailed. The rigour of these tests has allowed LEP physicists to determine unequivocally the number of fundamental 'generations' of elementary particles. These tests also allowed physicists to ascertain the mass of the top quark in advance of its discovery. Recent increases in the accelerator's energy allow new measurements to be undertaken, measurements that may uncover directly or indirectly the long-sought Higgs particle, believed to impart mass to all other particles.

  4. A Range-Based Test for the Parametric Form of the Volatility in Diffusion Models

    DEFF Research Database (Denmark)

    Podolskij, Mark; Ziggel, Daniel

    statistic. Under rather weak assumptions on the drift and volatility we prove weak convergence of the test statistic to a centered mixed Gaussian distribution. As a consequence we obtain a test, which is consistent for any fixed alternative. Moreover, we present a parametric bootstrap procedure which...

  5. Tuning and Test of Fragmentation Models Based on Identified Particles and Precision Event Shape Data

    CERN Document Server

    Abreu, P; Adye, T; Ajinenko, I; Alekseev, G D; Alemany, R; Allport, P P; Almehed, S; Amaldi, Ugo; Amato, S; Andreazza, A; Andrieux, M L; Antilogus, P; Apel, W D; Åsman, B; Augustin, J E; Augustinus, A; Baillon, Paul; Bambade, P; Barão, F; Barate, R; Barbi, M S; Bardin, Dimitri Yuri; Baroncelli, A; Bärring, O; Barrio, J A; Bartl, Walter; Bates, M J; Battaglia, Marco; Baubillier, M; Baudot, J; Becks, K H; Begalli, M; Beillière, P; Belokopytov, Yu A; Belous, K S; Benvenuti, Alberto C; Berggren, M; Bertini, D; Bertrand, D; Besançon, M; Bianchi, F; Bigi, M; Bilenky, S M; Billoir, P; Bloch, D; Blume, M; Bolognese, T; Bonesini, M; Bonivento, W; Booth, P S L; Bosio, C; Botner, O; Boudinov, E; Bouquet, B; Bourdarios, C; Bowcock, T J V; Bozzo, M; Branchini, P; Brand, K D; Brenke, T; Brenner, R A; Bricman, C; Brown, R C A; Brückman, P; Brunet, J M; Bugge, L; Buran, T; Burgsmüller, T; Buschmann, P; Buys, A; Cabrera, S; Caccia, M; Calvi, M; Camacho-Rozas, A J; Camporesi, T; Canale, V; Canepa, M; Cankocak, K; Cao, F; Carena, F; Carroll, L; Caso, Carlo; Castillo-Gimenez, M V; Cattai, A; Cavallo, F R; Chabaud, V; Charpentier, P; Chaussard, L; Checchia, P; Chelkov, G A; Chen, M; Chierici, R; Chliapnikov, P V; Chochula, P; Chorowicz, V; Chudoba, J; Cindro, V; Collins, P; Contreras, J L; Contri, R; Cortina, E; Cosme, G; Cossutti, F; Cowell, J H; Crawley, H B; Crennell, D J; Crosetti, G; Cuevas-Maestro, J; Czellar, S; Dahl-Jensen, Erik; Dahm, J; D'Almagne, B; Dam, M; Damgaard, G; Dauncey, P D; Davenport, Martyn; Da Silva, W; Defoix, C; Deghorain, A; Della Ricca, G; Delpierre, P A; Demaria, N; De Angelis, A; de Boer, Wim; De Brabandere, S; De Clercq, C; La Vaissière, C de; De Lotto, B; De Min, A; De Paula, L S; De Saint-Jean, C; Dijkstra, H; Di Ciaccio, Lucia; Di Diodato, A; Djama, F; Dolbeau, J; Dönszelmann, M; Doroba, K; Dracos, M; Drees, J; Drees, K A; Dris, M; Durand, J D; Edsall, D M; Ehret, R; Eigen, G; Ekelöf, T J C; Ekspong, Gösta; Elsing, M; Engel, J P; Erzen, B; Espirito-Santo, M C; Falk, E; Fassouliotis, D; Feindt, Michael; Ferrer, A; Fichet, S; Filippas-Tassos, A; Firestone, A; Fischer, P A; Föth, H; Fokitis, E; Fontanelli, F; Formenti, F; Franek, B J; Frenkiel, P; Fries, D E C; Frodesen, A G; Frühwirth, R; Fulda-Quenzer, F; Fuster, J A; Galloni, A; Gamba, D; Gandelman, M; García, C; García, J; Gaspar, C; Gasparini, U; Gavillet, P; Gazis, E N; Gelé, D; Gerber, J P; Gokieli, R; Golob, B; Gopal, Gian P; Gorn, L; Górski, M; Guz, Yu; Gracco, Valerio; Graziani, E; Green, C; Grefrath, A; Gris, P; Grosdidier, G; Grzelak, K; Gumenyuk, S A; Gunnarsson, P; Günther, M; Guy, J; Hahn, F; Hahn, S; Hajduk, Z; Hallgren, A; Hamacher, K; Harris, F J; Hedberg, V; Henriques, R P; Hernández, J J; Herquet, P; Herr, H; Hessing, T L; Higón, E; Hilke, Hans Jürgen; Hill, T S; Holmgren, S O; Holt, P J; Holthuizen, D J; Hoorelbeke, S; Houlden, M A; Hrubec, Josef; Huet, K; Hultqvist, K; Jackson, J N; Jacobsson, R; Jalocha, P; Janik, R; Jarlskog, C; Jarlskog, G; Jarry, P; Jean-Marie, B; Johansson, E K; Jönsson, L B; Jönsson, P E; Joram, Christian; Juillot, P; Kaiser, M; Kapusta, F; Karafasoulis, K; Karlsson, M; Karvelas, E; Katsanevas, S; Katsoufis, E C; Keränen, R; Khokhlov, Yu A; Khomenko, B A; Khovanskii, N N; King, B J; Kjaer, N J; Klapp, O; Klein, H; Klovning, A; Kluit, P M; Köne, B; Kokkinias, P; Koratzinos, M; Korcyl, K; Kostyukhin, V; Kourkoumelis, C; Kuznetsov, O; Kreuter, C; Kronkvist, I J; Krumshtein, Z; Krupinski, W; Kubinec, P; Kucewicz, W; Kurvinen, K L; Lacasta, C; Laktineh, I; Lamsa, J; Lanceri, L; Lane, D W; Langefeld, P; Lapin, V; Laugier, J P; Lauhakangas, R; Leder, Gerhard; Ledroit, F; Lefébure, V; Legan, C K; Leitner, R; Lemonne, J; Lenzen, Georg; Lepeltier, V; Lesiak, T; Libby, J; Liko, D; Lindner, R; Lipniacka, A; Lippi, I; Lörstad, B; Loken, J G; López, J M; Loukas, D; Lutz, P; Lyons, L; Naughton, J M; Maehlum, G; Mahon, J R; Maio, A; Malmgren, T G M; Malychev, V; Mandl, F; Marco, J; Marco, R P; Maréchal, B; Margoni, M; Marin, J C; Mariotti, C; Markou, A; Martínez-Rivero, C; Martínez-Vidal, F; Martí i García, S; Masik, J; Matorras, F; Matteuzzi, C; Matthiae, Giorgio; Mazzucato, M; McCubbin, M L; McKay, R; McNulty, R; Medbo, J; Merk, M; Meroni, C; Meyer, S; Meyer, W T; Myagkov, A; Michelotto, M; Migliore, E; Mirabito, L; Mitaroff, Winfried A; Mjörnmark, U; Moa, T; Møller, R; Mönig, K; Monge, M R; Morettini, P; Müller, H; Mulders, M; Mundim, L M; Murray, W J; Muryn, B; Myatt, Gerald; Naraghi, F; Navarria, Francesco Luigi; Navas, S; Nawrocki, K; Negri, P; Neumann, W; Neumeister, N; Nicolaidou, R; Nielsen, B S; Nieuwenhuizen, M; Nikolaenko, V; Niss, P; Nomerotski, A; Normand, Ainsley; Oberschulte-Beckmann, W; Obraztsov, V F; Olshevskii, A G; Onofre, A; Orava, Risto; Österberg, K; Ouraou, A; Paganini, P; Paganoni, M; Pagès, P; Pain, R; Palka, H; Papadopoulou, T D; Papageorgiou, K; Pape, L; Parkes, C; Parodi, F; Passeri, A; Pegoraro, M; Peralta, L; Pernegger, H; Pernicka, Manfred; Perrotta, A; Petridou, C; Petrolini, A; Petrovykh, M; Phillips, H T; Piana, G; Pierre, F; Plaszczynski, S; Podobrin, O; Pol, M E; Polok, G; Poropat, P; Pozdnyakov, V; Privitera, P; Pukhaeva, N; Pullia, Antonio; Radojicic, D; Ragazzi, S; Rahmani, H; Rames, J; Ratoff, P N; Read, A L; Reale, M; Rebecchi, P; Redaelli, N G; Regler, Meinhard; Reid, D; Renton, P B; Resvanis, L K; Richard, F; Richardson, J; Rídky, J; Rinaudo, G; Ripp, I; Romero, A; Roncagliolo, I; Ronchese, P; Roos, L; Rosenberg, E I; Rosso, E; Roudeau, Patrick; Rovelli, T; Rückstuhl, W; Ruhlmann-Kleider, V; Ruiz, A; Rybicki, K; Saarikko, H; Sacquin, Yu; Sadovskii, A; Sahr, O; Sajot, G; Salt, J; Sánchez, J; Sannino, M; Schimmelpfennig, M; Schneider, H; Schwickerath, U; Schyns, M A E; Sciolla, G; Scuri, F; Seager, P; Sedykh, Yu; Segar, A M; Seitz, A; Sekulin, R L; Serbelloni, L; Shellard, R C; Siegrist, P; Silvestre, R; Simonetti, S; Simonetto, F; Sissakian, A N; Sitár, B; Skaali, T B; Smadja, G; Smirnov, N; Smirnova, O G; Smith, G R; Sokolov, A; Sosnowski, R; Souza-Santos, D; Spassoff, Tz; Spiriti, E; Sponholz, P; Squarcia, S; Stanescu, C; Stapnes, Steinar; Stavitski, I; Stevenson, K; Stichelbaut, F; Stocchi, A; Strauss, J; Strub, R; Stugu, B; Szczekowski, M; Szeptycka, M; Tabarelli de Fatis, T; Tavernet, J P; Chikilev, O G; Thomas, J; Tilquin, A; Timmermans, J; Tkatchev, L G; Todorov, T; Todorova, S; Toet, D Z; Tomaradze, A G; Tomé, B; Tonazzo, A; Tortora, L; Tranströmer, G; Treille, D; Trischuk, W; Tristram, G; Trombini, A; Troncon, C; Tsirou, A L; Turluer, M L; Tyapkin, I A; Tyndel, M; Tzamarias, S; Überschär, B; Ullaland, O; Uvarov, V; Valenti, G; Vallazza, E; van Apeldoorn, G W; van Dam, P; Van Eldik, J; Vassilopoulos, N; Vegni, G; Ventura, L; Venus, W A; Verbeure, F; Verlato, M; Vertogradov, L S; Vilanova, D; Vincent, P; Vitale, L; Vlasov, E; Vodopyanov, A S; Vrba, V; Wahlen, H; Walck, C; Waldner, F; Weierstall, M; Weilhammer, Peter; Weiser, C; Wetherell, Alan M; Wicke, D; Wickens, J H; Wielers, M; Wilkinson, G R; Williams, W S C; Winter, M; Witek, M; Woschnagg, K; Yip, K; Yushchenko, O P; Zach, F; Zaitsev, A; Zalewska-Bak, A; Zalewski, Piotr; Zavrtanik, D; Zevgolatakos, E; Zimin, N I; Zito, M; Zontar, D; Zucchelli, G C; Zumerle, G

    1996-01-01

    Event shape and charged particle inclusive distributions are measured using 750000 decays of the $Z$ to hadrons from the DELPHI detector at LEP. These precise data allow a decisive confrontation with models of the hadronization process. Improved tunings of the JETSET ARIADNE and HERWIG parton shower models and the JETSET matrix element model are obtained by fitting the models to these DELPHI data as well as to identified particle distributions from all LEP experiments. The description of the data distributions by the models is critically reviewed with special importance attributed to identified particles.

  6. Mechanistic-empirical subgrade design model based on heavy vehicle simulator test results

    CSIR Research Space (South Africa)

    Theyse, HL

    2006-06-01

    Full Text Available Although Accelerated Pavement Testing (APT) is often done with specific objectives, valuable pavement performance data is generated over the long-term that may be used to investigate pavement behaviour in general and calibrate mechanistic...

  7. Development of an evaluation method for fracture mechanical tests on small samples based on a cohesive zone model

    International Nuclear Information System (INIS)

    Mahler, Michael

    2016-01-01

    The safety and reliability of nuclear power plants of the fourth generation is an important issue. It is based on a reliable interpretation of the components for which, among other fracture mechanical material properties are required. The existing irradiation in the power plants significantly affects the material properties which therefore need to be determined on irradiated material. Often only small amounts of irradiated material are available for characterization. In that case it is not possible to manufacture sufficiently large specimens, which are necessary for fracture mechanical testing in agreement with the standard. Small specimens must be used. From this follows the idea of this study, in which the fracture toughness can be predicted with the developed method based on tests of small specimens. For this purpose, the fracture process including the crack growth is described with a continuum mechanical approach using the finite element method and the cohesive zone model. The experiments on small specimens are used for parameter identification of the cohesive zone model. The two parameters of the cohesive zone model are determined by tensile tests on notched specimens (cohesive stress) and by parameter fitting to the fracture behavior of smalls specimens (cohesive energy). To account the different triaxialities of the specimens, the cohesive stress is used depending on the triaxiality. After parameter identification a large specimen can be simulated with the cohesive zone parameters derived from small specimens. The predicted fracture toughness of this big specimen fulfills the size requirements in the standard (ASTM E1820 or ASTM E399) in contrast to the small specimen. This method can be used for ductile and brittle material behavior and was validated in this work. In summary, this method offers the possibility to determine the fracture toughness indirectly based on small specimen testing. Main advantage is the low required specimen volume. Thereby massively

  8. Residuals and the Residual-Based Statistic for Testing Goodness of Fit of Structural Equation Models

    Science.gov (United States)

    Foldnes, Njal; Foss, Tron; Olsson, Ulf Henning

    2012-01-01

    The residuals obtained from fitting a structural equation model are crucial ingredients in obtaining chi-square goodness-of-fit statistics for the model. The authors present a didactic discussion of the residuals, obtaining a geometrical interpretation by recognizing the residuals as the result of oblique projections. This sheds light on the…

  9. Memory-Based Simple Heuristics as Attribute Substitution: Competitive Tests of Binary Choice Inference Models

    Science.gov (United States)

    Honda, Hidehito; Matsuka, Toshihiko; Ueda, Kazuhiro

    2017-01-01

    Some researchers on binary choice inference have argued that people make inferences based on simple heuristics, such as recognition, fluency, or familiarity. Others have argued that people make inferences based on available knowledge. To examine the boundary between heuristic and knowledge usage, we examine binary choice inference processes in…

  10. Wood-based Tri-Axial Sandwich Composite Materials: Design, Fabrication, Testing, Modeling and Application

    Science.gov (United States)

    Jinghao Li; John F. Hunt; Shaoqin Gong; Zhiyong Cai

    2014-01-01

    As the demand for sustainable materials increases, there are unique challenges and opportunities to develop light-weight green composites materials for a wide range of applications. Thus wood-based composite materials from renewable forests may provide options for some niche applications while helping to protect our environment. In this paper, the wood-based tri-axial...

  11. Species delineation using Bayesian model-based assignment tests: a case study using Chinese toad-headed agamas (genus Phrynocephalus

    Directory of Open Access Journals (Sweden)

    Fu Jinzhong

    2010-06-01

    Full Text Available Abstract Background Species are fundamental units in biology, yet much debate exists surrounding how we should delineate species in nature. Species discovery now requires the use of separate, corroborating datasets to quantify independently evolving lineages and test species criteria. However, the complexity of the speciation process has ushered in a need to infuse studies with new tools capable of aiding in species delineation. We suggest that model-based assignment tests are one such tool. This method circumvents constraints with traditional population genetic analyses and provides a novel means of describing cryptic and complex diversity in natural systems. Using toad-headed agamas of the Phrynocephalus vlangalii complex as a case study, we apply model-based assignment tests to microsatellite DNA data to test whether P. putjatia, a controversial species that closely resembles P. vlangalii morphologically, represents a valid species. Mitochondrial DNA and geographic data are also included to corroborate the assignment test results. Results Assignment tests revealed two distinct nuclear DNA clusters with 95% (230/243 of the individuals being assigned to one of the clusters with > 90% probability. The nuclear genomes of the two clusters remained distinct in sympatry, particularly at three syntopic sites, suggesting the existence of reproductive isolation between the identified clusters. In addition, a mitochondrial ND2 gene tree revealed two deeply diverged clades, which were largely congruent with the two nuclear DNA clusters, with a few exceptions. Historical mitochondrial introgression events between the two groups might explain the disagreement between the mitochondrial and nuclear DNA data. The nuclear DNA clusters and mitochondrial clades corresponded nicely to the hypothesized distributions of P. vlangalii and P. putjatia. Conclusions These results demonstrate that assignment tests based on microsatellite DNA data can be powerful tools

  12. Empirical component model to predict the overall performance of heating coils: Calibrations and tests based on manufacturer catalogue data

    International Nuclear Information System (INIS)

    Ruivo, Celestino R.; Angrisani, Giovanni

    2015-01-01

    Highlights: • An empirical model for predicting the performance of heating coils is presented. • Low and high heating capacity cases are used for calibration. • Versions based on several effectiveness correlations are tested. • Catalogue data are considered in approach testing. • The approach is a suitable component model to be used in dynamic simulation tools. - Abstract: A simplified methodology for predicting the overall behaviour of heating coils is presented in this paper. The coil performance is predicted by the ε-NTU method. Usually manufacturers do not provide information about the overall thermal resistance or the geometric details that are required either for the device selection or to apply known empirical correlations for the estimation of the involved thermal resistances. In the present work, heating capacity tables from the manufacturer catalogue are used to calibrate simplified approaches based on the classical theory of heat exchangers, namely the effectiveness method. Only two reference operating cases are required to calibrate each approach. The validity of the simplified approaches is investigated for a relatively high number of operating cases, listed in the technical catalogue of a manufacturer. Four types of coils of three sizes of air handling units are considered. A comparison is conducted between the heating coil capacities provided by the methodology and the values given by the manufacturer catalogue. The results show that several of the proposed approaches are suitable component models to be integrated in dynamic simulation tools of air conditioning systems such as TRNSYS or EnergyPlus

  13. Modified calibration protocol evaluated in a model-based testing of SBR flexibility

    DEFF Research Database (Denmark)

    Corominas, Lluís; Sin, Gürkan; Puig, Sebastià

    2011-01-01

    The purpose of this paper is to refine the BIOMATH calibration protocol for SBR systems, in particular to develop a pragmatic calibration protocol that takes advantage of SBR information-rich data, defines a simulation strategy to obtain proper initial conditions for model calibration and provide...

  14. Model based monitoring of urban traffic noise : Field test results for road side and shielded sides

    NARCIS (Netherlands)

    Eerden, F.J.M. van der; Lutgendorf, D.; Wessels, P.W.; Basten, T.G.H.

    2012-01-01

    Urban traffic noise can be a major issue for people and (local) governments. On a local scale the use of measurements is increasing, especially when measures or changes to the local infrastructure are proposed. However, measuring (only) urban traffic noise is a challenging task. By using a model

  15. Shelf-Life Prediction of Extra Virgin Olive Oils Using an Empirical Model Based on Standard Quality Tests

    Directory of Open Access Journals (Sweden)

    Claudia Guillaume

    2016-01-01

    Full Text Available Extra virgin olive oil shelf-life could be defined as the length of time under normal storage conditions within which no off-flavours or defects are developed and quality parameters such as peroxide value and specific absorbance are retained within accepted limits for this commercial category. Prediction of shelf-life is a desirable goal in the food industry. Even when extra virgin olive oil shelf-life should be one of the most important quality markers for extra virgin olive oil, it is not recognised as a legal parameter in most regulations and standards around the world. The proposed empirical formula to be evaluated in the present study is based on common quality tests with known and predictable result changes over time and influenced by different aspects of extra virgin olive oil with a meaningful influence over its shelf-life. The basic quality tests considered in the formula are Rancimat® or induction time (IND; 1,2-diacylglycerols (DAGs; pyropheophytin a (PPP; and free fatty acids (FFA. This paper reports research into the actual shelf-life of commercially packaged extra virgin olive oils versus the predicted shelf-life of those oils determined by analysing the expected deterioration curves for the three basic quality tests detailed above. Based on the proposed model, shelf-life is predicted by choosing the lowest predicted shelf-life of any of those three tests.

  16. Identification of a Cessna Citation II Model Based on Flight Test Data

    NARCIS (Netherlands)

    de Visser, C.C.; Pool, D.M.

    2017-01-01

    As a result of new aviation legislation, from 2019 on all air-carrier pilots are obliged to go through flight simulator-based stall recovery training. For this reason the Control and Simulation division at Delft University of Technology has set up a task force to develop a new methodology for

  17. Profile control simulations and experiments on TCV: a controller test environment and results using a model-based predictive controller

    Science.gov (United States)

    Maljaars, E.; Felici, F.; Blanken, T. C.; Galperti, C.; Sauter, O.; de Baar, M. R.; Carpanese, F.; Goodman, T. P.; Kim, D.; Kim, S. H.; Kong, M.; Mavkov, B.; Merle, A.; Moret, J. M.; Nouailletas, R.; Scheffer, M.; Teplukhina, A. A.; Vu, N. M. T.; The EUROfusion MST1-team; The TCV-team

    2017-12-01

    The successful performance of a model predictive profile controller is demonstrated in simulations and experiments on the TCV tokamak, employing a profile controller test environment. Stable high-performance tokamak operation in hybrid and advanced plasma scenarios requires control over the safety factor profile (q-profile) and kinetic plasma parameters such as the plasma beta. This demands to establish reliable profile control routines in presently operational tokamaks. We present a model predictive profile controller that controls the q-profile and plasma beta using power requests to two clusters of gyrotrons and the plasma current request. The performance of the controller is analyzed in both simulation and TCV L-mode discharges where successful tracking of the estimated inverse q-profile as well as plasma beta is demonstrated under uncertain plasma conditions and the presence of disturbances. The controller exploits the knowledge of the time-varying actuator limits in the actuator input calculation itself such that fast transitions between targets are achieved without overshoot. A software environment is employed to prepare and test this and three other profile controllers in parallel in simulations and experiments on TCV. This set of tools includes the rapid plasma transport simulator RAPTOR and various algorithms to reconstruct the plasma equilibrium and plasma profiles by merging the available measurements with model-based predictions. In this work the estimated q-profile is merely based on RAPTOR model predictions due to the absence of internal current density measurements in TCV. These results encourage to further exploit model predictive profile control in experiments on TCV and other (future) tokamaks.

  18. Testing the Community-Based Learning Collaborative (CBLC) implementation model: a study protocol

    OpenAIRE

    Hanson, Rochelle F.; Schoenwald, Sonja; Saunders, Benjamin E.; Chapman, Jason; Palinkas, Lawrence A.; Moreland, Angela D.; Dopp, Alex

    2016-01-01

    Background High rates of youth exposure to violence, either through direct victimization or witnessing, result in significant health/mental health consequences and high associated lifetime costs. Evidence-based treatments (EBTs), such as Trauma-Focused Cognitive Behavioral Therapy (TF-CBT), can prevent and/or reduce these negative effects, yet these treatments are not standard practice for therapists working with children identified by child welfare or mental health systems as needing service...

  19. Model Based approach to Predict Boundary Conditions of a Single  Cylinder Test Engine

    OpenAIRE

    Reddy, Goutham; Khambaty, Murtaza

    2016-01-01

    Huvudämnet i denna avhandling är användningen av prediktiva modeller för att styra randvillkor i en encylindrig motor. Encylindriga motorer används i utvecklingen av nya motorer för att studera förbränningskoncept. De utgör en modulär plattform för utveckling av bland annat nya ventilkoncept, förbränningsmetoder, bränsleinsprutningsmetoder och portkonstruktioner. I en produktionsmotor representeras turboaggregatet och motorn av ett kopplat dynamiskt system där motorns driftspunkt bestämmer av...

  20. Probabilistic model of bridge vehicle loads in port area based on in-situ load testing

    Science.gov (United States)

    Deng, Ming; Wang, Lei; Zhang, Jianren; Wang, Rei; Yan, Yanhong

    2017-11-01

    Vehicle load is an important factor affecting the safety and usability of bridges. An statistical analysis is carried out in this paper to investigate the vehicle load data of Tianjin Haibin highway in Tianjin port of China, which are collected by the Weigh-in- Motion (WIM) system. Following this, the effect of the vehicle load on test bridge is calculated, and then compared with the calculation result according to HL-93(AASHTO LRFD). Results show that the overall vehicle load follows a distribution with a weighted sum of four normal distributions. The maximum vehicle load during the design reference period follows a type I extremum distribution. The vehicle load effect also follows a weighted sum of four normal distributions, and the standard value of the vehicle load is recommended as 1.8 times that of the calculated value according to HL-93.

  1. Research on oral test modeling based on multi-feature fusion

    Science.gov (United States)

    Shi, Yuliang; Tao, Yiyue; Lei, Jun

    2018-04-01

    In this paper, the spectrum of speech signal is taken as an input of feature extraction. The advantage of PCNN in image segmentation and other processing is used to process the speech spectrum and extract features. And a new method combining speech signal processing and image processing is explored. At the same time of using the features of the speech map, adding the MFCC to establish the spectral features and integrating them with the features of the spectrogram to further improve the accuracy of the spoken language recognition. Considering that the input features are more complicated and distinguishable, we use Support Vector Machine (SVM) to construct the classifier, and then compare the extracted test voice features with the standard voice features to achieve the spoken standard detection. Experiments show that the method of extracting features from spectrograms using PCNN is feasible, and the fusion of image features and spectral features can improve the detection accuracy.

  2. Validation through model testing

    International Nuclear Information System (INIS)

    1995-01-01

    Geoval-94 is the third Geoval symposium arranged jointly by the OECD/NEA and the Swedish Nuclear Power Inspectorate. Earlier symposia in this series took place in 1987 and 1990. In many countries, the ongoing programmes to site and construct deep geological repositories for high and intermediate level nuclear waste are close to realization. A number of studies demonstrates the potential barrier function of the geosphere, but also that there are many unresolved issues. A key to these problems are the possibilities to gain knowledge by model testing with experiments and to increase confidence in models used for prediction. The sessions cover conclusions from the INTRAVAL-project, experiences from integrated experimental programs and underground research laboratories as well as the integration between performance assessment and site characterisation. Technical issues ranging from waste and buffer interactions with the rock to radionuclide migration in different geological media is addressed. (J.S.)

  3. Transoral endoscopic esophageal myotomy based on esophageal function testing in a survival porcine model.

    Science.gov (United States)

    Perretta, Silvana; Dallemagne, Bernard; Donatelli, Gianfranco; Diemunsch, Pierre; Marescaux, Jacques

    2011-01-01

    The most effective treatment of achalasia is Heller myotomy. To explore a submucosal endoscopic myotomy technique tailored on esophageal physiology testing and to compare it with the open technique. Prospective acute and survival comparative study in pigs (n = 12; 35 kg). University animal research center. Eight acute-4 open and 4 endoscopic-myotomies followed by 4 survival endoscopic procedures. Preoperative and postoperative manometry; esophagogastric junction (EGJ) distensibility before and after selective division of muscular fibers at the EGJ and after the myotomy was prolonged to a standard length by using the EndoFLIP Functional Lumen Imaging Probe (Crospon, Galway, Ireland). All procedures were successful, with no intraoperative and postoperative complications. In the survival group, the animals recovered promptly from surgery. Postoperative manometry demonstrated a 50% drop in mean lower esophageal sphincter pressure (LESp) in the endoscopic group (mean preoperative LESp, 22.2 ± 3.3 mm Hg; mean postoperative LESp, 11.34 ± 2.7 mm Hg; P open procedure group (mean preoperative LESp, 24.2 ± 3.2 mm Hg; mean postoperative LESp, 7.4 ± 4 mm Hg; P myotomy is feasible and safe. The lack of a significant difference in EGJ distensibility between the open and endoscopic procedure is very appealing. Were it to be perfected in a human population, this endoscopic approach could suggest a new strategy in the treatment of selected achalasia patients. Copyright © 2011 American Society for Gastrointestinal Endoscopy. Published by Mosby, Inc. All rights reserved.

  4. Using Evidence Based Practice in LIS Education: Results of a Test of a Communities of Practice Model

    Directory of Open Access Journals (Sweden)

    Joyce Yukawa

    2010-03-01

    Full Text Available Objective ‐ This study investigated the use of a communities of practice (CoP model for blended learning in library and information science (LIS graduate courses. The purposes were to: (1 test the model’s efficacy in supporting student growth related to core LIS concepts, practices, professional identity, and leadership skills, and (2 develop methods for formative and summative assessment using the model.Methods ‐ Using design‐based research principles to guide the formative and summative assessments, pre‐, mid‐, and post‐course questionnaires were constructed to test the model and administered to students in three LIS courses taught by the author. Participation was voluntary and anonymous. A total of 34 students completed the three courses; response rate for the questionnaires ranged from 47% to 95%. The pre‐course questionnaire addressed attitudes toward technology and the use of technology for learning. The mid‐course questionnaire addressed strengths and weaknesses of the course and suggestions for improvement. The post‐course questionnaire addressed what students valued about their learning and any changes in attitude toward technology for learning. Data were analyzed on three levels. Micro‐level analysis addressed technological factors related to usability and participant skills and attitudes. Meso‐level analysis addressed social and pedagogical factors influencing community learning. Macro‐level analysis addressed CoP learning outcomes, namely, knowledge of core concepts and practices, and the development of professional identity and leadership skills.Results ‐ The students can be characterized as adult learners who were neither early nor late adopters of technology. At the micro‐level, responses indicate that the online tools met high standards of usability and effectively supported online communication and learning. Moreover, the increase in positive attitudes toward the use of technology for learning at

  5. NET model coil test possibilities

    International Nuclear Information System (INIS)

    Erb, J.; Gruenhagen, A.; Herz, W.; Jentzsch, K.; Komarek, P.; Lotz, E.; Malang, S.; Maurer, W.; Noether, G.; Ulbricht, A.; Vogt, A.; Zahn, G.; Horvath, I.; Kwasnitza, K.; Marinucci, C.; Pasztor, G.; Sborchia, C.; Weymuth, P.; Peters, A.; Roeterdink, A.

    1987-11-01

    A single full size coil for NET/INTOR represents an investment of the order of 40 MUC (Million Unit Costs). Before such an amount of money or even more for the 16 TF coils is invested as much risks as possible must be eliminated by a comprehensive development programme. In the course of such a programme a coil technology verification test should finally prove the feasibility of NET/INTOR TF coils. This study report is almost exclusively dealing with such a verification test by model coil testing. These coils will be built out of two Nb 3 Sn-conductors based on two concepts already under development and investigation. Two possible coil arrangements are discussed: A cluster facility, where two model coils out of the two Nb 3 TF-conductors are used, and the already tested LCT-coils producing a background field. A solenoid arrangement, where in addition to the two TF model coils another model coil out of a PF-conductor for the central PF-coils of NET/INTOR is used instead of LCT background coils. Technical advantages and disadvantages are worked out in order to compare and judge both facilities. Costs estimates and the time schedules broaden the base for a decision about the realisation of such a facility. (orig.) [de

  6. Model-Based Systems Engineering for Capturing Mission Architecture System Processes with an Application Case Study - Orion Flight Test 1

    Science.gov (United States)

    Bonanne, Kevin H.

    2011-01-01

    Model-based Systems Engineering (MBSE) is an emerging methodology that can be leveraged to enhance many system development processes. MBSE allows for the centralization of an architecture description that would otherwise be stored in various locations and formats, thus simplifying communication among the project stakeholders, inducing commonality in representation, and expediting report generation. This paper outlines the MBSE approach taken to capture the processes of two different, but related, architectures by employing the Systems Modeling Language (SysML) as a standard for architecture description and the modeling tool MagicDraw. The overarching goal of this study was to demonstrate the effectiveness of MBSE as a means of capturing and designing a mission systems architecture. The first portion of the project focused on capturing the necessary system engineering activities that occur when designing, developing, and deploying a mission systems architecture for a space mission. The second part applies activities from the first to an application problem - the system engineering of the Orion Flight Test 1 (OFT-1) End-to-End Information System (EEIS). By modeling the activities required to create a space mission architecture and then implementing those activities in an application problem, the utility of MBSE as an approach to systems engineering can be demonstrated.

  7. Quantitative Assessment of Optical Coherence Tomography Imaging Performance with Phantom-Based Test Methods And Computational Modeling

    Science.gov (United States)

    Agrawal, Anant

    Optical coherence tomography (OCT) is a powerful medical imaging modality that uniquely produces high-resolution cross-sectional images of tissue using low energy light. Its clinical applications and technological capabilities have grown substantially since its invention about twenty years ago, but efforts have been limited to develop tools to assess performance of OCT devices with respect to the quality and content of acquired images. Such tools are important to ensure information derived from OCT signals and images is accurate and consistent, in order to support further technology development, promote standardization, and benefit public health. The research in this dissertation investigates new physical and computational models which can provide unique insights into specific performance characteristics of OCT devices. Physical models, known as phantoms, are fabricated and evaluated in the interest of establishing standardized test methods to measure several important quantities relevant to image quality. (1) Spatial resolution is measured with a nanoparticle-embedded phantom and model eye which together yield the point spread function under conditions where OCT is commonly used. (2) A multi-layered phantom is constructed to measure the contrast transfer function along the axis of light propagation, relevant for cross-sectional imaging capabilities. (3) Existing and new methods to determine device sensitivity are examined and compared, to better understand the detection limits of OCT. A novel computational model based on the finite-difference time-domain (FDTD) method, which simulates the physics of light behavior at the sub-microscopic level within complex, heterogeneous media, is developed to probe device and tissue characteristics influencing the information content of an OCT image. This model is first tested in simple geometric configurations to understand its accuracy and limitations, then a highly realistic representation of a biological cell, the retinal

  8. Mechanism-based population modelling for assessment of L-cell function based on total GLP-1 response following an oral glucose tolerance test

    DEFF Research Database (Denmark)

    Møller, Jonas B.; Jusko, William J.; Gao, Wei

    2011-01-01

    was to build a mechanism-based population model that describes the time course of total GLP-1 and provides indices for capability of secretion in each subject. The goal was thus to model the secretion of GLP-1, and not its effect on insulin production. Single 75 g doses of glucose were administered orally......GLP-1 is an insulinotropic hormone that synergistically with glucose gives rise to an increased insulin response. Its secretion is increased following a meal and it is thus of interest to describe the secretion of this hormone following an oral glucose tolerance test (OGTT). The aim of this study....... The individual estimates of absorption rate constants were used in the model for GLP-1 secretion. Estimation of parameters was performed using the FOCE method with interaction implemented in NONMEM VI. The final transit/indirect-response model obtained for GLP-1 production following an OGTT included two...

  9. Branched Adaptive Testing with a Rasch-Model-Calibrated Test: Analysing Item Presentation's Sequence Effects Using the Rasch-Model-Based LLTM

    Science.gov (United States)

    Kubinger, Klaus D.; Reif, Manuel; Yanagida, Takuya

    2011-01-01

    Item position effects provoke serious problems within adaptive testing. This is because different testees are necessarily presented with the same item at different presentation positions, as a consequence of which comparing their ability parameter estimations in the case of such effects would not at all be fair. In this article, a specific…

  10. Interface-based software testing

    Directory of Open Access Journals (Sweden)

    Aziz Ahmad Rais

    2016-10-01

    Full Text Available Software quality is determined by assessing the characteristics that specify how it should work, which are verified through testing. If it were possible to touch, see, or measure software, it would be easier to analyze and prove its quality. Unfortunately, software is an intangible asset, which makes testing complex. This is especially true when software quality is not a question of particular functions that can be tested through a graphical user interface. The primary objective of software architecture is to design quality of software through modeling and visualization. There are many methods and standards that define how to control and manage quality. However, many IT software development projects still fail due to the difficulties involved in measuring, controlling, and managing software quality. Software quality failure factors are numerous. Examples include beginning to test software too late in the development process, or failing properly to understand, or design, the software architecture and the software component structure. The goal of this article is to provide an interface-based software testing technique that better measures software quality, automates software quality testing, encourages early testing, and increases the software’s overall testability

  11. Testing a bioenergetics-based habitat choice model: bluegill (Lepomis macrochirus) responses to food availability and temperature

    Science.gov (United States)

    2011-01-01

    Using an automated shuttlebox system, we conducted patch choice experiments with 32, 8–12 g bluegill sunfish (Lepomis macrochirus) to test a behavioral energetics hypothesis of habitat choice. When patch temperature and food levels were held constant within patches but different between patches, we expected bluegill to choose patches that maximized growth based on the bioenergetic integration of food and temperature as predicted by a bioenergetics model. Alternative hypotheses were that bluegill may choose patches based only on food (optimal foraging) or temperature (behavioral thermoregulation). The behavioral energetics hypothesis was not a good predictor of short-term (from minutes to weeks) patch choice by bluegill; the behavioral thermoregulation hypothesis was the best predictor. In the short-term, food and temperature appeared to affect patch choice hierarchically; temperature was more important, although food can alter temperature preference during feeding periods. Over a 19-d experiment, mean temperatures occupied by fish offered low rations did decline as predicted by the behavioral energetics hypothesis, but the decline was less than 1.0 °C as opposed to a possible 5 °C decline. A short-term, bioenergetic response to food and temperature may be precluded by physiological costs of acclimation not considered explicitly in the behavioral energetics hypothesis.

  12. Properties of hypothesis testing techniques and (Bayesian) model selection for exploration-based and theory-based (order-restricted) hypotheses.

    Science.gov (United States)

    Kuiper, Rebecca M; Nederhoff, Tim; Klugkist, Irene

    2015-05-01

    In this paper, the performance of six types of techniques for comparisons of means is examined. These six emerge from the distinction between the method employed (hypothesis testing, model selection using information criteria, or Bayesian model selection) and the set of hypotheses that is investigated (a classical, exploration-based set of hypotheses containing equality constraints on the means, or a theory-based limited set of hypotheses with equality and/or order restrictions). A simulation study is conducted to examine the performance of these techniques. We demonstrate that, if one has specific, a priori specified hypotheses, confirmation (i.e., investigating theory-based hypotheses) has advantages over exploration (i.e., examining all possible equality-constrained hypotheses). Furthermore, examining reasonable order-restricted hypotheses has more power to detect the true effect/non-null hypothesis than evaluating only equality restrictions. Additionally, when investigating more than one theory-based hypothesis, model selection is preferred over hypothesis testing. Because of the first two results, we further examine the techniques that are able to evaluate order restrictions in a confirmatory fashion by examining their performance when the homogeneity of variance assumption is violated. Results show that the techniques are robust to heterogeneity when the sample sizes are equal. When the sample sizes are unequal, the performance is affected by heterogeneity. The size and direction of the deviations from the baseline, where there is no heterogeneity, depend on the effect size (of the means) and on the trend in the group variances with respect to the ordering of the group sizes. Importantly, the deviations are less pronounced when the group variances and sizes exhibit the same trend (e.g., are both increasing with group number). © 2014 The British Psychological Society.

  13. Modeling the Test-Taking Motivation Construct through Investigation of Psychometric Properties of an Expectancy-Value-Based Questionnaire

    Science.gov (United States)

    Knekta, Eva; Eklöf, Hanna

    2015-01-01

    The aim of this study was to evaluate the psychometric properties of an expectancy-value-based questionnaire measuring five aspects of test-taking motivation (effort, expectancies, importance, interest, and test anxiety). The questionnaire was distributed to a sample of Swedish Grade 9 students taking a low-stakes (n = 1,047) or a high-stakes (n =…

  14. Testing homogeneity in Weibull-regression models.

    Science.gov (United States)

    Bolfarine, Heleno; Valença, Dione M

    2005-10-01

    In survival studies with families or geographical units it may be of interest testing whether such groups are homogeneous for given explanatory variables. In this paper we consider score type tests for group homogeneity based on a mixing model in which the group effect is modelled as a random variable. As opposed to hazard-based frailty models, this model presents survival times that conditioned on the random effect, has an accelerated failure time representation. The test statistics requires only estimation of the conventional regression model without the random effect and does not require specifying the distribution of the random effect. The tests are derived for a Weibull regression model and in the uncensored situation, a closed form is obtained for the test statistic. A simulation study is used for comparing the power of the tests. The proposed tests are applied to real data sets with censored data.

  15. Strategy for a Rock Mechanics Site Descriptive Model. A test case based on data from the Aespoe HRL

    International Nuclear Information System (INIS)

    Hudson, John A

    2002-06-01

    In anticipation of the SKB Site Investigations for radioactive waste disposal, an approach has been developed for the Rock Mechanics Site Descriptive Model. This approach was tested by predicting the rock mechanics properties of a 600 m x 180 m x 120 m rock volume at the Aespoe Hard Rock Laboratory (HRL) using limited borehole data of the type typically obtained during a site investigation. These predicted properties were then compared with 'best estimate' properties obtained from a study of the test rock volume using additional information, mainly tunnel data. The exercise was known as the Test Case, and is the subject of this Report. Three modelling techniques were used to predict the rock properties: the 'empirical approach' - the rock properties were estimated using rock mass classification schemes and empirical correlation formulae; the 'theoretical approach' - the rock properties were estimated using numerical modelling techniques; and the 'stress approach' - the rock stress state was estimated using primary data and numerical modelling. These approaches are described separately and respectively. Following an explanation of the context for the Test Case within the strategy for developing the Rock Mechanics Site Descriptive Model, conditions at the Aespoe HRL are described in Chapter 2. The Test Case organization and the suite of nine Protocols used to ensure that the work was appropriately guided and co-ordinated are described in Chapter 3. The methods for predicting the rock properties and the rock stress, and comparisons with the 'best estimate' properties of the actual conditions, are presented in Chapters 4 and 5. Finally, the conclusions from this Test Case exercise are given in Chapter 6. General recommendations for the management of this type of Test Case are also included

  16. Strategy for a Rock Mechanics Site Descriptive Model. A test case based on data from the Aespoe HRL

    Energy Technology Data Exchange (ETDEWEB)

    Hudson, John A (ed.) [Rock Engineering Consultants, Welwyn Garden City (United Kingdom)

    2002-06-01

    In anticipation of the SKB Site Investigations for radioactive waste disposal, an approach has been developed for the Rock Mechanics Site Descriptive Model. This approach was tested by predicting the rock mechanics properties of a 600 m x 180 m x 120 m rock volume at the Aespoe Hard Rock Laboratory (HRL) using limited borehole data of the type typically obtained during a site investigation. These predicted properties were then compared with 'best estimate' properties obtained from a study of the test rock volume using additional information, mainly tunnel data. The exercise was known as the Test Case, and is the subject of this Report. Three modelling techniques were used to predict the rock properties: the 'empirical approach' - the rock properties were estimated using rock mass classification schemes and empirical correlation formulae; the 'theoretical approach' - the rock properties were estimated using numerical modelling techniques; and the 'stress approach' - the rock stress state was estimated using primary data and numerical modelling. These approaches are described separately and respectively. Following an explanation of the context for the Test Case within the strategy for developing the Rock Mechanics Site Descriptive Model, conditions at the Aespoe HRL are described in Chapter 2. The Test Case organization and the suite of nine Protocols used to ensure that the work was appropriately guided and co-ordinated are described in Chapter 3. The methods for predicting the rock properties and the rock stress, and comparisons with the 'best estimate' properties of the actual conditions, are presented in Chapters 4 and 5. Finally, the conclusions from this Test Case exercise are given in Chapter 6. General recommendations for the management of this type of Test Case are also included.

  17. Development of a lifetime prediction model for lithium-ion batteries based on extended accelerated aging test data

    Science.gov (United States)

    Ecker, Madeleine; Gerschler, Jochen B.; Vogel, Jan; Käbitz, Stefan; Hust, Friedrich; Dechent, Philipp; Sauer, Dirk Uwe

    2012-10-01

    Battery lifetime prognosis is a key requirement for successful market introduction of electric and hybrid vehicles. This work aims at the development of a lifetime prediction approach based on an aging model for lithium-ion batteries. A multivariable analysis of a detailed series of accelerated lifetime experiments representing typical operating conditions in hybrid electric vehicle is presented. The impact of temperature and state of charge on impedance rise and capacity loss is quantified. The investigations are based on a high-power NMC/graphite lithium-ion battery with good cycle lifetime. The resulting mathematical functions are physically motivated by the occurring aging effects and are used for the parameterization of a semi-empirical aging model. An impedance-based electric-thermal model is coupled to the aging model to simulate the dynamic interaction between aging of the battery and the thermal as well as electric behavior. Based on these models different drive cycles and management strategies can be analyzed with regard to their impact on lifetime. It is an important tool for vehicle designers and for the implementation of business models. A key contribution of the paper is the parameterization of the aging model by experimental data, while aging simulation in the literature usually lacks a robust empirical foundation.

  18. Experiences with formal engineering: model-based specification, implementation and testing of a software bus at Neopost

    NARCIS (Netherlands)

    Sijtema, M.; Salaün, G.; Schätz, B.; Belinfante, Axel; Stoelinga, Mariëlle Ida Antoinette; Marinelli, L.

    2014-01-01

    We report on the actual industrial use of formal methods during the development of a software bus. During an internship at Neopost Inc., of 14 weeks, we developed the server component of a software bus, called the XBus, using formal methods during the design, validation and testing phase: we modeled

  19. Experiences with Formal Engineering : Model-Based Specification, Implementation and Testing of a Software Bus at Neopost

    NARCIS (Netherlands)

    Sijtema, Marten; Stoelinga, Mariëlle Ida Antoinette; Belinfante, Axel; Marinelli, Lawrence; Salaün, Gwen; Schätz, Bernhard

    We report on the actual industrial use of formal methods during the development of a software bus. At Neopost Inc., we developed the server component of a software bus, called the XBus, using formal methods during the design, validation and testing phase: We modeled our design of the XBus in the

  20. Profile control simulations and experiments on TCV : A controller test environment and results using a model-based predictive controller

    NARCIS (Netherlands)

    Maljaars, E.; Felici, F.; Blanken, T.C.; Galperti, C.; Sauter, O.; de Baar, M.R.; Carpanese, F.; Goodman, T.P.; Kim, D.; Kim, S.H.; Kong, M.G.; Mavkov, B.; Merle, A.; Moret, J.M.; Nouailletas, R.; Scheffer, M.; Teplukhina, A.A.; Vu, N.M.T.

    2017-01-01

    The successful performance of a model predictive profile controller is demonstrated in simulations and experiments on the TCV tokamak, employing a profile controller test environment. Stable high-performance tokamak operation in hybrid and advanced plasma scenarios requires control over the safety

  1. Profile control simulations and experiments on TCV: a controller test environment and results using a model-based predictive controller

    NARCIS (Netherlands)

    Maljaars, B.; Felici, F.; Blanken, T. C.; Galperti, C.; Sauter, O.; de Baar, M. R.; Carpanese, F.; Goodman, T. P.; Kim, D.; Kim, S. H.; Kong, M.; Mavkov, B.; Merle, A.; Moret, J.; Nouailletas, R.; Scheffer, M.; Teplukhina, A.; Vu, T.

    2017-01-01

    The successful performance of a model predictive profile controller is demonstrated in simulations and experiments on the TCV tokamak, employing a profile controller test environment. Stable high-performance tokamak operation in hybrid and advanced plasma scenarios requires control over the safety

  2. Testing a Web-Based, Trained-Peer Model to Build Capacity for Evidence-Based Practices in Community Mental Health Systems.

    Science.gov (United States)

    German, Ramaris E; Adler, Abby; Frankel, Sarah A; Stirman, Shannon Wiltsey; Pinedo, Paola; Evans, Arthur C; Beck, Aaron T; Creed, Torrey A

    2018-03-01

    Use of expert-led workshops plus consultation has been established as an effective strategy for training community mental health (CMH) clinicians in evidence-based practices (EBPs). Because of high rates of staff turnover, this strategy inadequately addresses the need to maintain capacity to deliver EBPs. This study examined knowledge, competency, and retention outcomes of a two-phase model developed to build capacity for an EBP in CMH programs. In the first phase, an initial training cohort in each CMH program participated in in-person workshops followed by expert-led consultation (in-person, expert-led [IPEL] phase) (N=214 clinicians). After this cohort completed training, new staff members participated in Web-based training (in place of in-person workshops), followed by peer-led consultation with the initial cohort (Web-based, trained-peer [WBTP] phase) (N=148). Tests of noninferiority assessed whether WBTP was not inferior to IPEL at increasing clinician cognitive-behavioral therapy (CBT) competency, as measured by the Cognitive Therapy Rating Scale. WBTP was not inferior to IPEL at developing clinician competency. Hierarchical linear models showed no significant differences in CBT knowledge acquisition between the two phases. Survival analyses indicated that WBTP trainees were less likely than IPEL trainees to complete training. In terms of time required from experts, WBTP required 8% of the resources of IPEL. After an initial investment to build in-house CBT expertise, CMH programs were able to use a WBTP model to broaden their own capacity for high-fidelity CBT. IPEL followed by WBTP offers an effective alternative to build EBP capacity in CMH programs, rather than reliance on external experts.

  3. Evidence synthesis to inform model-based cost-effectiveness evaluations of diagnostic tests: a methodological review of health technology assessments

    Directory of Open Access Journals (Sweden)

    Bethany Shinkins

    2017-04-01

    Full Text Available Abstract Background Evaluations of diagnostic tests are challenging because of the indirect nature of their impact on patient outcomes. Model-based health economic evaluations of tests allow different types of evidence from various sources to be incorporated and enable cost-effectiveness estimates to be made beyond the duration of available study data. To parameterize a health-economic model fully, all the ways a test impacts on patient health must be quantified, including but not limited to diagnostic test accuracy. Methods We assessed all UK NIHR HTA reports published May 2009-July 2015. Reports were included if they evaluated a diagnostic test, included a model-based health economic evaluation and included a systematic review and meta-analysis of test accuracy. From each eligible report we extracted information on the following topics: 1 what evidence aside from test accuracy was searched for and synthesised, 2 which methods were used to synthesise test accuracy evidence and how did the results inform the economic model, 3 how/whether threshold effects were explored, 4 how the potential dependency between multiple tests in a pathway was accounted for, and 5 for evaluations of tests targeted at the primary care setting, how evidence from differing healthcare settings was incorporated. Results The bivariate or HSROC model was implemented in 20/22 reports that met all inclusion criteria. Test accuracy data for health economic modelling was obtained from meta-analyses completely in four reports, partially in fourteen reports and not at all in four reports. Only 2/7 reports that used a quantitative test gave clear threshold recommendations. All 22 reports explored the effect of uncertainty in accuracy parameters but most of those that used multiple tests did not allow for dependence between test results. 7/22 tests were potentially suitable for primary care but the majority found limited evidence on test accuracy in primary care settings

  4. Do Test Design and Uses Influence Test Preparation? Testing a Model of Washback with Structural Equation Modeling

    Science.gov (United States)

    Xie, Qin; Andrews, Stephen

    2013-01-01

    This study introduces Expectancy-value motivation theory to explain the paths of influences from perceptions of test design and uses to test preparation as a special case of washback on learning. Based on this theory, two conceptual models were proposed and tested via Structural Equation Modeling. Data collection involved over 870 test takers of…

  5. TESTING GARCH-X TYPE MODELS

    DEFF Research Database (Denmark)

    Pedersen, Rasmus Søndergaard; Rahbek, Anders

    2017-01-01

    We present novel theory for testing for reduction of GARCH-X type models with an exogenous (X) covariate to standard GARCH type models. To deal with the problems of potential nuisance parameters on the boundary of the parameter space as well as lack of identification under the null, we exploit...... a noticeable property of specific zero-entries in the inverse information of the GARCH-X type models. Specifically, we consider sequential testing based on two likelihood ratio tests and as demonstrated the structure of the inverse information implies that the proposed test neither depends on whether...... the nuisance parameters lie on the boundary of the parameter space, nor on lack of identification. Our general results on GARCH-X type models are applied to Gaussian based GARCH-X models, GARCH-X models with Student's t-distributed innovations as well as the integer-valued GARCH-X (PAR-X) models....

  6. Testing Social-driven Forces on the Evolution of Sahelian Rural Systems: A Combined Agent-based Modeling and Anthropological Approach

    OpenAIRE

    Saqalli , Mehdi; Gérard , B.; Bielders , C.; Defourny , Pierre

    2010-01-01

    International audience; This article presents the results of a methodology combining an extensive fieldwork, a formalization of field-based individual rules and norms into an agent-based model and the implementation of scenarios analyzing the effects of social and agro-ecological constraints on rural farmers through the study of three different sites in Nigerien Sahel. Two family transition processes are here tested, following field observations and literature-based hypotheses: family organiz...

  7. Model-based control of a fuel cell cooling circuit with automatic software testing; Modellbasierte Steuerung des Kuehlkreislaufes einer Brennstoffzelle mit automatisiertem Test der Software

    Energy Technology Data Exchange (ETDEWEB)

    Schaefer, Sascha

    2012-07-01

    For reconstruction and control of the volume flow in a cooling circuit a fuel cell system is analyzed, and physical models of the fluid temperature and pump volume flow are derived. On this basis, functional models for derivation of software algorithms are presented which enable model-based calculation of the volume flow in a fluid circuit and coolant pump control on the basis of detailed system modelling. The available functions enable complete control and diagnosis of the volume flow. The functional models in Matlab/Simulink will be implemented in a control unit; they are discussed in the context of a software development process. It is stressed that correctness of the functions must be assured, and a trial automation system for functional models of this type is presented. (orig./AKB) [German] Zur Rekonstruktion und Steuerung des Volumenstroms in einem Kuehlkreislauf, wird ein Brennstoffzellensystem analysiert und physikalische Modelle fuer die Fluidtemperatur und den Pumpenvolumenstrom werden hergeleitet. Basierend auf diesen Zusammenhaengen werden Funktionsmodelle zur Ableitung von Softwarealgorithmen vorgestellt. Diese ermoeglichen es den Volumenstrom in einem Fluidkreislauf modellbasiert zu bestimmen bzw. die Kuehlmittelpumpe basierend auf einer detaillierten Systemmodellierung zu steuern. Die zur Verfuegung stehenden Funktionen erlauben eine komplette Regelung, Steuerung und Diagnose des Volumenstroms. Die Funktionsmodelle, in Matlab/Simulink realisiert und zur Ausfuehrung auf einem Steuergeraet vorgesehen, werden im Kontext eines Softwareentwicklungsprozesses diskutiert. Es wird die Notwendigkeit der Sicherstellung der Korrektheit der Funktionen herausgearbeitet und ein Testautomatisierungssystem fuer solche Funktionsmodelle vorgestellt.

  8. Simulation-based Testing of Control Software

    Energy Technology Data Exchange (ETDEWEB)

    Ozmen, Ozgur [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Nutaro, James J. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Sanyal, Jibonananda [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Olama, Mohammed M. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)

    2017-02-10

    It is impossible to adequately test complex software by examining its operation in a physical prototype of the system monitored. Adequate test coverage can require millions of test cases, and the cost of equipment prototypes combined with the real-time constraints of testing with them makes it infeasible to sample more than a small number of these tests. Model based testing seeks to avoid this problem by allowing for large numbers of relatively inexpensive virtual prototypes that operate in simulation time at a speed limited only by the available computing resources. In this report, we describe how a computer system emulator can be used as part of a model based testing environment; specifically, we show that a complete software stack including operating system and application software - can be deployed within a simulated environment, and that these simulations can proceed as fast as possible. To illustrate this approach to model based testing, we describe how it is being used to test several building control systems that act to coordinate air conditioning loads for the purpose of reducing peak demand. These tests involve the use of ADEVS (A Discrete Event System Simulator) and QEMU (Quick Emulator) to host the operational software within the simulation, and a building model developed with the MODELICA programming language using Buildings Library and packaged as an FMU (Functional Mock-up Unit) that serves as the virtual test environment.

  9. Properties of hypothesis testing techniques and (Bayesian) model selection for exploration-based and theory-based (order-restricted) hypotheses

    NARCIS (Netherlands)

    Kuiper, Rebecca M.; Nederhoff, Tim; Klugkist, Irene

    2015-01-01

    In this paper, the performance of six types of techniques for comparisons of means is examined. These six emerge from the distinction between the method employed (hypothesis testing, model selection using information criteria, or Bayesian model selection) and the set of hypotheses that is

  10. A Gaussian mixture model based adaptive classifier for fNIRS brain-computer interfaces and its testing via simulation

    Science.gov (United States)

    Li, Zheng; Jiang, Yi-han; Duan, Lian; Zhu, Chao-zhe

    2017-08-01

    Objective. Functional near infra-red spectroscopy (fNIRS) is a promising brain imaging technology for brain-computer interfaces (BCI). Future clinical uses of fNIRS will likely require operation over long time spans, during which neural activation patterns may change. However, current decoders for fNIRS signals are not designed to handle changing activation patterns. The objective of this study is to test via simulations a new adaptive decoder for fNIRS signals, the Gaussian mixture model adaptive classifier (GMMAC). Approach. GMMAC can simultaneously classify and track activation pattern changes without the need for ground-truth labels. This adaptive classifier uses computationally efficient variational Bayesian inference to label new data points and update mixture model parameters, using the previous model parameters as priors. We test GMMAC in simulations in which neural activation patterns change over time and compare to static decoders and unsupervised adaptive linear discriminant analysis classifiers. Main results. Our simulation experiments show GMMAC can accurately decode under time-varying activation patterns: shifts of activation region, expansions of activation region, and combined contractions and shifts of activation region. Furthermore, the experiments show the proposed method can track the changing shape of the activation region. Compared to prior work, GMMAC performed significantly better than the other unsupervised adaptive classifiers on a difficult activation pattern change simulation: 99% versus  brain-computer interfaces, including neurofeedback training systems, where operation over long time spans is required.

  11. Model-based development and testing of advertising messages: A comparative study of two campaign proposals based on the MECCAS model and a conventional approach

    DEFF Research Database (Denmark)

    Bech-Larsen, Tino

    2001-01-01

    Traditionally, the development of advertising messages has been based on "creative independence", sometimes catalysed by inductively generated empirical data. Due to the recent intensified focus on advertising effectiveness, this state of affairs is beginning to change. The purpose of the study....... The comparison involved the efficiency of the managerial communication taking place in the message development process as well as target group communication effects. The managerial communication was studied by interviews with the involved advertising agency (Midtmarketing, Ikast, Denmark) and client staff...... described in this article is to compare the development and effects of two campaign proposals, with the sommon aim of increasing the consumption of apples among young Danes (18 to 35 years of age). One of the proposals is the result of an inductive-creative process, while the other is base on the MECCAS...

  12. Model-based development and testing of advertising messages: A comparative study of two campaign proposals based on the MECCAS model and a conventional approach

    DEFF Research Database (Denmark)

    Bech-Larsen, Tino

    2001-01-01

    Traditionally, the development of advertising messages has been based on "creative independence", sometimes catalysed by inductively generated empirical data. Due to the recent intensified focus on advertising effectiveness, this state of affairs is beginning to change. The purpose of the study....... The comparison involved the efficiency of the managerial communication taking place in the message development process as well as target group communication effects. The managerial communication was studied by interviews with the involved advertising agency (Midtmarketing, Ikast, Denmark) and client staff...

  13. Testing seasonal and long-term controls of streamwater DOC using empirical and process-based models.

    Science.gov (United States)

    Futter, Martyn N; de Wit, Heleen A

    2008-12-15

    Concentrations of dissolved organic carbon (DOC) in surface waters are increasing across Europe and parts of North America. Several mechanisms have been proposed to explain these increases including reductions in acid deposition, change in frequency of winter storms and changes in temperature and precipitation patterns. We used two modelling approaches to identify the mechanisms responsible for changing surface water DOC concentrations. Empirical regression analysis and INCA-C, a process-based model of stream-water DOC, were used to simulate long-term (1986--2003) patterns in stream water DOC concentrations in a small boreal stream. Both modelling approaches successfully simulated seasonal and inter-annual patterns in DOC concentration. In both models, seasonal patterns of DOC concentration were controlled by hydrology and inter-annual patterns were explained by climatic variation. There was a non-linear relationship between warmer summer temperatures and INCA-C predicted DOC. Only the empirical model was able to satisfactorily simulate the observed long-term increase in DOC. The observed long-term trends in DOC are likely to be driven by in-soil processes controlled by SO4(2-) and Cl(-) deposition, and to a lesser extent by temperature-controlled processes. Given the projected changes in climate and deposition, future modelling and experimental research should focus on the possible effects of soil temperature and moisture on organic carbon production, sorption and desorption rates, and chemical controls on organic matter solubility.

  14. Hand interception of occluded motion in humans: a test of model-based vs. on-line control.

    Science.gov (United States)

    La Scaleia, Barbara; Zago, Myrka; Lacquaniti, Francesco

    2015-09-01

    Two control schemes have been hypothesized for the manual interception of fast visual targets. In the model-free on-line control, extrapolation of target motion is based on continuous visual information, without resorting to physical models. In the model-based control, instead, a prior model of target motion predicts the future spatiotemporal trajectory. To distinguish between the two hypotheses in the case of projectile motion, we asked participants to hit a ball that rolled down an incline at 0.2 g and then fell in air at 1 g along a parabola. By varying starting position, ball velocity and trajectory differed between trials. Motion on the incline was always visible, whereas parabolic motion was either visible or occluded. We found that participants were equally successful at hitting the falling ball in both visible and occluded conditions. Moreover, in different trials the intersection points were distributed along the parabolic trajectories of the ball, indicating that subjects were able to extrapolate an extended segment of the target trajectory. Remarkably, this trend was observed even at the very first repetition of movements. These results are consistent with the hypothesis of model-based control, but not with on-line control. Indeed, ball path and speed during the occlusion could not be extrapolated solely from the kinematic information obtained during the preceding visible phase. The only way to extrapolate ball motion correctly during the occlusion was to assume that the ball would fall under gravity and air drag when hidden from view. Such an assumption had to be derived from prior experience. Copyright © 2015 the American Physiological Society.

  15. Modelling the pile load test

    Directory of Open Access Journals (Sweden)

    Prekop Ľubomír

    2017-01-01

    Full Text Available This paper deals with the modelling of the load test of horizontal resistance of reinforced concrete piles. The pile belongs to group of piles with reinforced concrete heads. The head is pressed with steel arches of a bridge on motorway D1 Jablonov - Studenec. Pile model was created in ANSYS with several models of foundation having properties found out from geotechnical survey. Finally some crucial results obtained from computer models are presented and compared with these obtained from experiment.

  16. Modelling the pile load test

    OpenAIRE

    Prekop Ľubomír

    2017-01-01

    This paper deals with the modelling of the load test of horizontal resistance of reinforced concrete piles. The pile belongs to group of piles with reinforced concrete heads. The head is pressed with steel arches of a bridge on motorway D1 Jablonov - Studenec. Pile model was created in ANSYS with several models of foundation having properties found out from geotechnical survey. Finally some crucial results obtained from computer models are presented and compared with these obtained from exper...

  17. Modeling and Testing of Growth Status for Chinese Cabbage and White Radish with UAV-Based RGB Imagery

    Directory of Open Access Journals (Sweden)

    Dong-Wook Kim

    2018-04-01

    Full Text Available Conventional crop-monitoring methods are time-consuming and labor-intensive, necessitating new techniques to provide faster measurements and higher sampling intensity. This study reports on mathematical modeling and testing of growth status for Chinese cabbage and white radish using unmanned aerial vehicle-red, green and blue (UAV-RGB imagery for measurement of their biophysical properties. Chinese cabbage seedlings and white radish seeds were planted at 7–10-day intervals to provide a wide range of growth rates. Remotely sensed digital imagery data were collected for test fields at approximately one-week intervals using a UAV platform equipped with an RGB digital camera flying at 2 m/s at 20 m above ground. Radiometric calibrations for the RGB band sensors were performed on every UAV flight using standard calibration panels to minimize the effect of ever-changing light conditions on the RGB images. Vegetation fractions (VFs of crops in each region of interest from the mosaicked ortho-images were calculated as the ratio of pixels classified as crops segmented using the Otsu threshold method and a vegetation index of excess green (ExG. Plant heights (PHs were estimated using the structure from motion (SfM algorithm to create 3D surface models from crop canopy data. Multiple linear regression equations consisting of three predictor variables (VF, PH, and VF × PH and four different response variables (fresh weight, leaf length, leaf width, and leaf count provided good fits with coefficients of determination (R2 ranging from 0.66 to 0.90. The validation results using a dataset of crop growth obtained in a different year also showed strong linear relationships (R2 > 0.76 between the developed regression models and standard methods, confirming that the models make it possible to use UAV-RGB images for quantifying spatial and temporal variability in biophysical properties of Chinese cabbage and white radish over the growing season.

  18. Modelling skylarks (Alauda arvensis) to predict impacts of changes in land management and policy: development and testing of an agent-based model

    DEFF Research Database (Denmark)

    Topping, Christopher John; Odderskær, Peter; Kahlert, Johnny Abildgaard

    2013-01-01

    of distribution and density, reproductive performance and seasonal changes in territory numbers. Data to support this was collected over a 13-year period and comprised detailed field observations of breeding birds and intensive surveys. The model was able to recreate the real world data patterns accurately......; it was also able to simultaneously fit a number of other secondary system properties which were not formally a part of the testing procedure. The correspondence of model output to real world data and sensitivity analysis are presented and discussed, and the model’s description is provided in ODdox format (a...... formal description inter-linked to the program code). Detailed and stringent tests for model performance were carried out, and standardised model description and open access to the source code were provided to open development of the skylark model to others. Over and above documenting the utility...

  19. Model-based development and testing of advertising messages: A comparative study of two campaign proposals based on the MECCAS model and a conventional approach

    DEFF Research Database (Denmark)

    Bech-Larsen, Tino

    theoretically valid and comprehensible guidelines for message development potentially enhances the effects of advertising messages and improves the possibility of measuring such effects. Moreover, such guidelines also have potential implications for the managerial communication processes (client......1. Traditionally the development of advertising messages has been based on "creative independence", sometimes catalysed by inductively generated empirical data. Due to the recent intensified focus on advertising effectiveness, this state of affair is now beginning to change. 2. Implementing......-agency and intra-agency) involved in the development of advertising messages. 3. The purpose of the study described in this paper is to compare the development and effects of two campaign proposals, with the common aim of increasing the consumption of apples among young Danes (18 to 35 years of age). One...

  20. Nonlinear modeling, strength-based design, and testing of flexible piezoelectric energy harvesters under large dynamic loads for rotorcraft applications

    Science.gov (United States)

    Leadenham, Stephen; Erturk, Alper

    2014-04-01

    There has been growing interest in enabling wireless health and usage monitoring for rotorcraft applications, such as helicopter rotor systems. Large dynamic loads and acceleration fluctuations available in these environments make the implementation of vibration-based piezoelectric energy harvesters a very promising choice. However, such extreme loads transmitted to the harvester can also be detrimental to piezoelectric laminates and overall system reliability. Particularly flexible resonant cantilever configurations tuned to match the dominant excitation frequency can be subject to very large deformations and failure of brittle piezoelectric laminates due to excessive bending stresses at the root of the harvester. Design of resonant piezoelectric energy harvesters for use in these environments require nonlinear electroelastic dynamic modeling and strength-based analysis to maximize the power output while ensuring that the harvester is still functional. This paper presents a mathematical framework to design and analyze the dynamics of nonlinear flexible piezoelectric energy harvesters under large base acceleration levels. A strength-based limit is imposed to design the piezoelectric energy harvester with a proof mass while accounting for material, geometric, and dissipative nonlinearities, with a focus on two demonstrative case studies having the same linear fundamental resonance frequency but different overhang length and proof mass values. Experiments are conducted at different excitation levels for validation of the nonlinear design approach proposed in this work. The case studies in this work reveal that harvesters exhibiting similar behavior and power generation performance at low excitation levels (e.g. less than 0.1g) can have totally different strength-imposed performance limitations under high excitations (e.g. above 1g). Nonlinear modeling and strength-based design is necessary for such excitation levels especially when using resonant cantilevers with no

  1. Test-driven modeling of embedded systems

    DEFF Research Database (Denmark)

    Munck, Allan; Madsen, Jan

    2015-01-01

    To benefit maximally from model-based systems engineering (MBSE) trustworthy high quality models are required. From the software disciplines it is known that test-driven development (TDD) can significantly increase the quality of the products. Using a test-driven approach with MBSE may have...... a similar positive effect on the quality of the system models and the resulting products and may therefore be desirable. To define a test-driven model-based systems engineering (TD-MBSE) approach, we must define this approach for numerous sub disciplines such as modeling of requirements, use cases...... suggest that our method provides a sound foundation for rapid development of high quality system models....

  2. Sample Size Determination for Rasch Model Tests

    Science.gov (United States)

    Draxler, Clemens

    2010-01-01

    This paper is concerned with supplementing statistical tests for the Rasch model so that additionally to the probability of the error of the first kind (Type I probability) the probability of the error of the second kind (Type II probability) can be controlled at a predetermined level by basing the test on the appropriate number of observations.…

  3. IN VITRO MODELS TO EVALUATE DRUG-INDUCED HYPERSENSITIVITY: POTENTIAL TEST BASED ON ACTIVATION OF DENDRITIC CELLS

    Directory of Open Access Journals (Sweden)

    Valentina Galbiati

    2016-07-01

    Full Text Available Hypersensitivity drug reactions (HDRs are the adverse effect of pharmaceuticals that clinically resemble allergy. HDRs account for approximately 1/6 of drug-induced adverse effects, and include immune-mediated ('allergic' and non immune-mediated ('pseudo allergic' reactions. In recent years, the severe and unpredicted drug adverse events clearly indicate that the immune system can be a critical target of drugs. Enhanced prediction in preclinical safety evaluation is, therefore, crucial. Nowadays, there are no validated in vitro or in vivo methods to screen the sensitizing potential of drugs in the pre-clinical phase. The problem of non-predictability of immunologically-based hypersensitivity reactions is related to the lack of appropriate experimental models rather than to the lack of -understanding of the adverse phenomenon.We recently established experimental conditions and markers to correctly identify drug associated with in vivo hypersensitivity reactions using THP-1 cells and IL-8 production, CD86 and CD54 expression. The proposed in vitro method benefits from a rationalistic approach with the idea that allergenic drugs share with chemical allergens common mechanisms of cell activation. This assay can be easily incorporated into drug development for hazard identification of drugs, which may have the potential to cause in vivo hypersensitivity reactions. The purpose of this review is to assess the state of the art of in vitro models to assess the allergenic potential of drugs based on the activation of dendritic cells.

  4. A priori tests of combustion models based on a CH{sub 4}/H{sub 2} Triple Flame

    Energy Technology Data Exchange (ETDEWEB)

    Dombard, J.; Naud, B.; Jimenez Sanchez, C.

    2008-07-01

    This document reproduces the final project of Jerome Dombard, presented on June 25, 2008, for the obtention of the Master degree MIMSE (Master Ingenierie Mathematique, Statistique et Economique) of Bordeaux University (Universite Bordeaux 1). We make an a priori study of FPI/FGM-type turbulent combustion models using a 2D DNS of a triple flame. A reduced chemical scheme of 16 species and 12 reactions is used (ARM1, proposed by J.-Y. Chen at Berkeley University). The fuel (CH4/H2 mixture) and oxidizer (air) correspond to the inlet composition of the Sydney bluff-body stabilised flame experiments (flames HM1-3). First, we compute 1D laminar premixed flames. The purpose of those calculations is twofold: 1. check the differences between different computer programs and different treatments of molecular diffusion, and 2. calibrate the 2D-DNS of the laminar triple flame (mainly decide on the grid resolution). Then, the solution of the 2D laminar triple flame is used to test a priori FPI/FGM tables. Finally, preliminary considerations on sub-grid scale modelling in the context of Large Eddy Simulation are made. (Author) 14 refs.

  5. Modelling and Testing of Friction in Forging

    DEFF Research Database (Denmark)

    Bay, Niels

    2007-01-01

    Knowledge about friction is still limited in forging. The theoretical models applied presently for process analysis are not satisfactory compared to the advanced and detailed studies possible to carry out by plastic FEM analyses and more refined models have to be based on experimental testing...

  6. Methods for testing transport models

    International Nuclear Information System (INIS)

    Singer, C.; Cox, D.

    1991-01-01

    Substantial progress has been made over the past year on six aspects of the work supported by this grant. As a result, we have in hand for the first time a fairly complete set of transport models and improved statistical methods for testing them against large databases. We also have initial results of such tests. These results indicate that careful application of presently available transport theories can reasonably well produce a remarkably wide variety of tokamak data

  7. Test of Antifibrotic Drugs in a Cellular Model of Fibrosis Based on Muscle-Derived Fibroblasts from Duchenne Muscular Dystrophy Patients.

    Science.gov (United States)

    Zanotti, Simona; Mora, Marina

    2018-01-01

    An in vitro model of muscle fibrosis, based on the use of primary human fibroblasts isolated from muscle biopsies of patients affected by Duchenne muscular dystrophies (DMD) and cultivated in monolayer and 3D conditions, is used to test the potential antifibrotic activity of pirfenidone (PFD). This in vitro model may be usefully also to evaluate the toxicity and efficacy of other candidate molecules for the treatment of fibrosis. The drug toxicity is evaluated using a colorimetric assay based on the conversion of tetrazolium salt (MTT) to insoluble formazan, while the effect of the drug on cell proliferation is measured with the bromodeoxyuridine incorporation assay. The efficacy of the drug is evaluated in fibroblast monolayers by quantitating synthesis and deposition of intracellular collagen with a spectrophotometric picrosirius red-based assay, and by quantitating cell migration using a "scratch" assay. The efficacy of PFD as antifibrotic drug is also evaluated in a 3D fibroblast model by measuring diameters and number of nodules.

  8. Linking community, parenting, and depressive symptom trajectories: testing resilience models of adolescent agency based on race/ethnicity and gender.

    Science.gov (United States)

    Williams, Amanda L; Merten, Michael J

    2014-09-01

    Family stress models illustrate how communities affect youth outcomes through effects on parents and studies consistently show the enduring effects of early community context. The present study takes a different approach identifying human agency during adolescence as a potentially significant promotive factor mediating the relationship between community, parenting, and mental health. While agency is an important part of resilience, its longitudinal effects are unknown, particularly based on gender and race/ethnicity. The purpose of this research was to model the long-term effects of community structural adversity and social resources as predictors of adolescent depressive symptom trajectories via indirect effects of parental happiness, parent-child relationships, and human agency. Latent growth analyses were conducted with 1,796 participants (53% female; 56% White) across four waves of the National Longitudinal Study of Adolescent Health spanning adolescence (Wave 1) through adulthood (Wave 4). The results identified agency as an important promotive factor during adolescence with long-term mental health benefits, but only for White and male participants. For these individuals, community social resources and the quality of the parent-child relationship were related to higher levels of agency and more positive mental health trajectories. Although community social resources similarly benefitted parenting and agency among females and non-White participants, there were no significant links between agency and depressive symptoms for these youth. The results suggest that agency remains an important, but poorly understood concept and additional work is necessary to continue unpacking its meaning for diverse groups of youth.

  9. Risk Based Optimal Fatigue Testing

    DEFF Research Database (Denmark)

    Sørensen, John Dalsgaard; Faber, M.H.; Kroon, I.B.

    1992-01-01

    Optimal fatigue life testing of materials is considered. Based on minimization of the total expected costs of a mechanical component a strategy is suggested to determine the optimal stress range levels for which additional experiments are to be performed together with an optimal value...

  10. Testing a social cognitive theory-based model of indoor tanning: implications for skin cancer prevention messages.

    Science.gov (United States)

    Noar, Seth M; Myrick, Jessica Gall; Zeitany, Alexandra; Kelley, Dannielle; Morales-Pico, Brenda; Thomas, Nancy E

    2015-01-01

    The lack of a theory-based understanding of indoor tanning is a major impediment to the development of effective messages to prevent or reduce this behavior. This study applied the Comprehensive Indoor Tanning Expectations (CITE) scale in an analysis of indoor tanning behavior among sorority women (total N = 775). Confirmatory factor analyses indicated that CITE positive and negative expectations were robust, multidimensional factors and that a hierarchical structure fit the data well. Social cognitive theory-based structural equation models demonstrated that appearance-oriented variables were significantly associated with outcome expectations. Outcome expectations were, in turn, significantly associated with temptations to tan, intention to tan indoors, and indoor tanning behavior. The implications of these findings for the development of messages to prevent and reduce indoor tanning behavior are discussed in two domains: (a) messages that attempt to change broader societal perceptions about tan skin, and (b) messages that focus more narrowly on indoor tanning-challenging positive expectations, enhancing negative expectations, and encouraging substitution of sunless tanning products.

  11. Testing a hydraulic trait based model of stomatal control: results from a controlled drought experiment on aspen (Populus tremuloides, Michx.) and ponderosa pine (Pinus ponderosa, Douglas)

    Science.gov (United States)

    Love, D. M.; Venturas, M.; Sperry, J.; Wang, Y.; Anderegg, W.

    2017-12-01

    Modeling approaches for tree stomatal control often rely on empirical fitting to provide accurate estimates of whole tree transpiration (E) and assimilation (A), which are limited in their predictive power by the data envelope used to calibrate model parameters. Optimization based models hold promise as a means to predict stomatal behavior under novel climate conditions. We designed an experiment to test a hydraulic trait based optimization model, which predicts stomatal conductance from a gain/risk approach. Optimal stomatal conductance is expected to maximize the potential carbon gain by photosynthesis, and minimize the risk to hydraulic transport imposed by cavitation. The modeled risk to the hydraulic network is assessed from cavitation vulnerability curves, a commonly measured physiological trait in woody plant species. Over a growing season garden grown plots of aspen (Populus tremuloides, Michx.) and ponderosa pine (Pinus ponderosa, Douglas) were subjected to three distinct drought treatments (moderate, severe, severe with rehydration) relative to a control plot to test model predictions. Model outputs of predicted E, A, and xylem pressure can be directly compared to both continuous data (whole tree sapflux, soil moisture) and point measurements (leaf level E, A, xylem pressure). The model also predicts levels of whole tree hydraulic impairment expected to increase mortality risk. This threshold is used to estimate survivorship in the drought treatment plots. The model can be run at two scales, either entirely from climate (meteorological inputs, irrigation) or using the physiological measurements as a starting point. These data will be used to study model performance and utility, and aid in developing the model for larger scale applications.

  12. The Physical Clogging of the Landfill Leachate Collection System in China: Based on Filtration Test and Numerical Modelling

    Directory of Open Access Journals (Sweden)

    Yili Liu

    2018-02-01

    Full Text Available Clogging of the leachate collection system (LCS has been a common operation problem in municipal solid waste (MSW landfills in China, which can result in high water levels that threaten the safety of landfill operations. To determine the cause of failure in an LCS, raw leachate from a municipal solid waste transfer station was collected and the high content of particulate matter was characterized. Based on the parameters obtained in a filtration test, a numerical simulation was performed to estimate the influence of particle deposition on drainage system clogging. The results showed that LCSs were confronted with the risk of clogging due to the deposition of particulate matter resulting from the higher concentration of total suspended solids (TSS level > 2200 mg L−1 and larger particle size (>30% TSS particles > 15 μm in the leachate. On one hand, the non-woven geotextile, as the upper layer of the LCS, retained most particulate matter of large diameters, reducing its hydraulic conductivity to approximately 10−8 to 10−9 m s−1 after 1–2 years of operation and perching significant leachate above it (0.6–0.7 m. On the other hand, the geotextile prevented the gravel layer from physically clogging and minimized the leachate head above the bottom liner. Therefore, the role of geotextile should be balanced to optimize the LCS in MSW landfills in China.

  13. Test and application of a general process-based dynamic coastal mass-balance model for contaminants using data for radionuclides in the Dnieper-Bug estuary

    International Nuclear Information System (INIS)

    Hakanson, Lars; Lindgren, Dan

    2009-01-01

    In this work a general, process-based mass-balance model for water contaminants for coastal areas at the ecosystem scale (CoastMab) is presented and for the first time tested for radionuclides. The model is dynamic, based on ordinary differential equations and gives monthly predictions. Connected to the core model there is also a sub-model for contaminant concentrations in fish. CoastMab calculates sedimentation, resuspension, diffusion, mixing, burial and retention of the given contaminant. The model contains both general algorithms, which apply to all contaminants, and substance-specific parts (such as algorithms for the particulate fraction, diffusion, biouptake and biological half-life). CoastMab and the sub-model for fish are simple to apply in practice since all driving variables may be readily accessed from maps or regular monitoring programs. The separation between the surface-water layer and the deep-water layer is not done as in most traditional models from water temperature data but from sedimentological criteria. Previous versions of the models for phosphorus and suspended particulate matter (in the Baltic Sea) have been validated and shown to predict well. This work presents modifications of the model and tests using two tracers, radiocesium and radiostrontium (from the Chernobyl fallout) in the Dnieper-Bug estuary (the Black Sea). Good correlations are shown between modeled and empirical data, except for the month directly after the fallout. We have, e.g., shown that: 1. The conditions in the sea outside the bay are important for the concentrations of the substances in water, sediments and fish within the bay, 2. We have demonstrated 'biological,' 'chemical' and 'water' dilution, 3. That the water chemical conditions in the bay influence biouptake and concentrations in fish of the radionuclides and 4. That the feeding behaviour of the coastal fish is very important for the biouptake of the radionuclides

  14. The design and testing of a caring teaching model based on the theoretical framework of caring in the Chinese Context: a mixed-method study.

    Science.gov (United States)

    Guo, Yujie; Shen, Jie; Ye, Xuchun; Chen, Huali; Jiang, Anli

    2013-08-01

    This paper aims to report the design and test the effectiveness of an innovative caring teaching model based on the theoretical framework of caring in the Chinese context. Since the 1970's, caring has been a core value in nursing education. In a previous study, a theoretical framework of caring in the Chinese context is explored employing a grounded theory study, considered beneficial for caring education. A caring teaching model was designed theoretically and a one group pre- and post-test quasi-experimental study was administered to test its effectiveness. From Oct, 2009 to Jul, 2010, a cohort of grade-2 undergraduate nursing students (n=64) in a Chinese medical school was recruited to participate in the study. Data were gathered through quantitative and qualitative methods to evaluate the effectiveness of the caring teaching model. The caring teaching model created an esthetic situation and experiential learning style for teaching caring that was integrated within the curricula. Quantitative data from the quasi-experimental study showed that the post-test scores of each item were higher than those on the pre-test (p<0.01). Thematic analysis of 1220 narratives from students' caring journals and reports of participant class observation revealed two main thematic categories, which reflected, from the students' points of view, the development of student caring character and the impact that the caring teaching model had on this regard. The model could be used as an integrated approach to teach caring in nursing curricula. It would also be beneficial for nursing administrators in cultivating caring nurse practitioners. Copyright © 2012 Elsevier Ltd. All rights reserved.

  15. Collaborative testing of turbulence models

    Science.gov (United States)

    Bradshaw, P.

    1992-12-01

    This project, funded by AFOSR, ARO, NASA, and ONR, was run by the writer with Profs. Brian E. Launder, University of Manchester, England, and John L. Lumley, Cornell University. Statistical data on turbulent flows, from lab. experiments and simulations, were circulated to modelers throughout the world. This is the first large-scale project of its kind to use simulation data. The modelers returned their predictions to Stanford, for distribution to all modelers and to additional participants ('experimenters')--over 100 in all. The object was to obtain a consensus on the capabilities of present-day turbulence models and identify which types most deserve future support. This was not completely achieved, mainly because not enough modelers could produce results for enough test cases within the duration of the project. However, a clear picture of the capabilities of various modeling groups has appeared, and the interaction has been helpful to the modelers. The results support the view that Reynolds-stress transport models are the most accurate.

  16. Developing Pairwise Preference-Based Personality Test and Experimental Investigation of Its Resistance to Faking Effect by Item Response Model

    Science.gov (United States)

    Usami, Satoshi; Sakamoto, Asami; Naito, Jun; Abe, Yu

    2016-01-01

    Recent years have shown increased awareness of the importance of personality tests in educational, clinical, and occupational settings, and developing faking-resistant personality tests is a very pragmatic issue for achieving more precise measurement. Inspired by Stark (2002) and Stark, Chernyshenko, and Drasgow (2005), we develop a pairwise…

  17. Choosing wisely: a model-based analysis evaluating the trade-offs in cancer benefit and diagnostic referrals among alternative HPV testing strategies in Norway.

    Science.gov (United States)

    Burger, Emily A; Pedersen, Kine; Sy, Stephen; Kristiansen, Ivar Sønbø; Kim, Jane J

    2017-09-05

    Forthcoming cervical cancer screening strategies involving human papillomavirus (HPV) testing for women not vaccinated against HPV infections may increase colposcopy referral rates. We quantified health and resource trade-offs associated with alternative HPV-based algorithms to inform decision-makers when choosing between candidate algorithms. We used a mathematical simulation model of HPV-induced cervical carcinogenesis in Norway. We compared the current cytology-based strategy to alternative strategies that varied by the switching age to primary HPV testing (ages 25-34 years), the routine screening frequency (every 3-10 years), and management of HPV-positive, cytology-negative women. Model outcomes included reductions in lifetime cervical cancer risk, relative colposcopy rates, and colposcopy rates per cervical cancer prevented. The age of switching to primary HPV testing and the screening frequency had the largest impacts on cancer risk reductions, which ranged from 90.9% to 96.3% compared to no screening. In contrast, increasing the follow-up intensity of HPV-positive, cytology-negative women provided only minor improvements in cancer benefits, but generally required considerably higher rates of colposcopy referrals compared to current levels, resulting in less efficient cervical cancer prevention. We found that in order to maximise cancer benefits HPV-based screening among unvaccinated women should not be delayed: rather, policy makers should utilise the triage mechanism to control colposcopy referrals.

  18. Point-of-Care Testing for Anemia, Diabetes, and Hypertension: A Pharmacy-Based Model in Lima, Peru.

    Science.gov (United States)

    Saldarriaga, Enrique M; Vodicka, Elisabeth; La Rosa, Sayda; Valderrama, Maria; Garcia, Patricia J

    Prevention and control of chronic diseases is a high priority for many low- and middle-income countries. This study evaluated the feasibility and acceptability of training pharmacy workers to provide point-of-care testing for 3 chronic diseases-hypertension, diabetes, and anemia-to improve disease detection and awareness through private pharmacies. We developed a multiphase training curriculum for pharmacists and pharmacy technicians to build capacity for identification of risk factors, patient education, point-of-care testing, and referral for abnormal results. We conducted a pre-post evaluation with participants and evaluated results using Student t test for proportions. We conducted point-of-care testing with pharmacy clients and evaluated acceptability by patient characteristics (age, gender, and type of patient) using multiple logistic regression. In total, 72 pharmacy workers (66%) completed the full training curriculum. Pretest scores indicated that pharmacists had more knowledge and skills in chronic disease risk factors, patient education, and testing than pharmacy technicians. All participants improved their knowledge and skills after the training, and post-test scores indicated that pharmacy technicians achieved the same level of competency as pharmacists (P anemia than a traditional health center. Fast service was very important: 41% ranked faster results and 30% ranked faster attention as the most important factor for receiving diagnostic testing in the pharmacy. We found that it is both feasible for pharmacies and acceptable to clients to train pharmacy workers to provide point-of-care testing for anemia, diabetes, and hypertension. This innovative approach holds potential to increase early detection of risk factors and bolster disease prevention and management efforts in Peru and other low- and middle-income settings. Copyright © 2017. Published by Elsevier Inc.

  19. Testlet-Based Multidimensional Adaptive Testing.

    Science.gov (United States)

    Frey, Andreas; Seitz, Nicki-Nils; Brandt, Steffen

    2016-01-01

    Multidimensional adaptive testing (MAT) is a highly efficient method for the simultaneous measurement of several latent traits. Currently, no psychometrically sound approach is available for the use of MAT in testlet-based tests. Testlets are sets of items sharing a common stimulus such as a graph or a text. They are frequently used in large operational testing programs like TOEFL, PISA, PIRLS, or NAEP. To make MAT accessible for such testing programs, we present a novel combination of MAT with a multidimensional generalization of the random effects testlet model (MAT-MTIRT). MAT-MTIRT compared to non-adaptive testing is examined for several combinations of testlet effect variances (0.0, 0.5, 1.0, and 1.5) and testlet sizes (3, 6, and 9 items) with a simulation study considering three ability dimensions with simple loading structure. MAT-MTIRT outperformed non-adaptive testing regarding the measurement precision of the ability estimates. Further, the measurement precision decreased when testlet effect variances and testlet sizes increased. The suggested combination of the MTIRT model therefore provides a solution to the substantial problems of testlet-based tests while keeping the length of the test within an acceptable range.

  20. Testlet-based Multidimensional Adaptive Testing

    Directory of Open Access Journals (Sweden)

    Andreas Frey

    2016-11-01

    Full Text Available Multidimensional adaptive testing (MAT is a highly efficient method for the simultaneous measurement of several latent traits. Currently, no psychometrically sound approach is available for the use of MAT in testlet-based tests. Testlets are sets of items sharing a common stimulus such as a graph or a text. They are frequently used in large operational testing programs like TOEFL, PISA, PIRLS, or NAEP. To make MAT accessible for such testing programs, we present a novel combination of MAT with a multidimensional generalization of the random effects testlet model (MAT-MTIRT. MAT-MTIRT compared to non-adaptive testing is examined for several combinations of testlet effect variances (0.0, 0.5, 1.0, 1.5 and testlet sizes (3 items, 6 items, 9 items with a simulation study considering three ability dimensions with simple loading structure. MAT-MTIRT outperformed non-adaptive testing regarding the measurement precision of the ability estimates. Further, the measurement precision decreased when testlet effect variances and testlet sizes increased. The suggested combination of the MTIRT model therefore provides a solution to the substantial problems of testlet-based tests while keeping the length of the test within an acceptable range.

  1. Developing confidence in a coupled TH model based on the results of experiment by using engineering scale test facility, 'COUPLE'

    International Nuclear Information System (INIS)

    Fujisaki, Kiyoshi; Suzuki, Hideaki; Fujita, Tomoo

    2008-03-01

    It is necessary to understand quantitative changes of near-field conditions and processes over time and space for modeling the near-field evolution after emplacement of engineered barriers. However, the coupled phenomena in near-field are complicated because thermo-, hydro-, mechanical, chemical processes will interact each other. The question is, therefore, whether the applied model will represent the coupled behavior adequately or not. In order to develop confidence in the modeling, it is necessary to compare with results of coupled behavior experiments in laboratory or in site. In this report, we evaluated the applicability of a coupled T-H model under the conditions of simulated near-field for the results of coupled T-H experiment in laboratory. As a result, it has been shown that the fitting by the modeling with the measured data is reasonable under this condition. (author)

  2. Methods for testing transport models

    International Nuclear Information System (INIS)

    Singer, C.; Cox, D.

    1993-01-01

    This report documents progress to date under a three-year contract for developing ''Methods for Testing Transport Models.'' The work described includes (1) choice of best methods for producing ''code emulators'' for analysis of very large global energy confinement databases, (2) recent applications of stratified regressions for treating individual measurement errors as well as calibration/modeling errors randomly distributed across various tokamaks, (3) Bayesian methods for utilizing prior information due to previous empirical and/or theoretical analyses, (4) extension of code emulator methodology to profile data, (5) application of nonlinear least squares estimators to simulation of profile data, (6) development of more sophisticated statistical methods for handling profile data, (7) acquisition of a much larger experimental database, and (8) extensive exploratory simulation work on a large variety of discharges using recently improved models for transport theories and boundary conditions. From all of this work, it has been possible to define a complete methodology for testing new sets of reference transport models against much larger multi-institutional databases

  3. Modelling inorganic and organic biocide leaching from CBA-amine (Copper–Boron–Azole) treated wood based on characterisation leaching tests

    Energy Technology Data Exchange (ETDEWEB)

    Lupsea, Maria [University of Toulouse, INSA, UPS, INP, LISBP, 135 Avenue de Rangueil, F–31077 Toulouse (France); INRA, UMR 792, F-31400 Toulouse (France); CNRS, UMR 5504, F-31400 Toulouse (France); Paris-Est University, CSTB — Scientific and Technical Centre for the Building Industry, DEE/Environment and Life Cycle Engineering Team, 24 Rue Joseph Fourier, F-38400 Saint Martin d' Hères (France); Tiruta-Barna, Ligia, E-mail: ligia.barna@insa-toulouse.fr [University of Toulouse, INSA, UPS, INP, LISBP, 135 Avenue de Rangueil, F–31077 Toulouse (France); INRA, UMR 792, F-31400 Toulouse (France); CNRS, UMR 5504, F-31400 Toulouse (France); Schiopu, Nicoleta [Paris-Est University, CSTB — Scientific and Technical Centre for the Building Industry, DEE/Environment and Life Cycle Engineering Team, 24 Rue Joseph Fourier, F-38400 Saint Martin d' Hères (France); Schoknecht, Ute [BAM — Federal Institute for Materials Research and Testing, Division 4.1, Unter den Eichen 87, 12205 Berlin (Germany)

    2013-09-01

    Numerical simulation of the leaching behaviour of treated wood is the most pertinent and less expensive method for the prediction of biocides' release in water. Few studies based on mechanistic leaching models have been carried out so far. In this work, a coupled chemistry-mass transport model is developed for simulating the leaching behaviour of inorganic (Cu, B) and organic (Tebuconazole) biocides from CBA-amine treated wood. The model is based on experimental investigations (lab-scale leaching tests coupled with chemical and structural analysis). It considers biocides' interactions with wood solid components and with extractives (literature confirmed reactions), as well as transport mechanisms (diffusion, convection) in different compartments. Simulation results helped at identifying the main fixation mechanisms, like (i) direct complexation of Cu by wood-phenolic and -carboxylic sites (and not via monoethanolamine; complex) on lignin and hemicellulose and strong dependence on extractives' nature, (ii) pH dependent binding of tebuconazole on polarized -OH moieties on wood. The role of monoethanolamine is to provide a pore-solution pH of about 7.5, when copper solubility is found to be weakest. The capability of the developed model to simulate the chemical and transport behaviour is the main result of this study. Moreover, it proved that characterization leaching tests (pH dependency and dynamic tests), combined with appropriate analytical methods are useful experimental tools. Due to its flexibility for representing and simulating various leaching conditions, chemical-transport model developed could be used to further simulate the leaching behaviour of CBA treated wood at larger scales. - Highlights: • Biocide and extractives leaching from ammonia-CBA treated wood were modelled. • The chemical-transport model identifies the main fixation/solubilisation mechanisms. • The model describes well the results of equilibrium and dynamic leaching

  4. Results and Lessons Learned from a Coupled Social and Physical Hydrology Model: Testing Alternative Water Management Policies and Institutional Structures Using Agent-Based Modeling and Regional Hydrology

    Science.gov (United States)

    Murphy, J.; Lammers, R. B.; Prousevitch, A.; Ozik, J.; Altaweel, M.; Collier, N. T.; Kliskey, A. D.; Alessa, L.

    2015-12-01

    Water Management in the U.S. Southwest is under increasing scrutiny as many areas endure persistent drought. The impact of these prolonged dry conditions is a product of regional climate and hydrological conditions, but also of a highly engineered water management infrastructure and a complex web of social arrangements whereby water is allocated, shared, exchanged, used, re-used, and finally consumed. We coupled an agent-based model with a regional hydrological model to understand the dynamics in one richly studied and highly populous area: southern Arizona, U.S.A., including metropolitan Phoenix and Tucson. There, multiple management entities representing an array of municipalities and other water providers and customers, including private companies and Native American tribes are enmeshed in a complex legal and economic context in which water is bought, leased, banked, and exchanged in a variety of ways and on multiple temporal and physical scales. A recurrent question in the literature of adaptive management is the impact of management structure on overall system performance. To explore this, we constructed an agent-based model to capture this social complexity, and coupled this with a physical hydrological model that we used to drive the system under a variety of water stress scenarios and to assess the regional impact of the social system's performance. We report the outcomes of ensembles of runs in which varieties of alternative policy constraints and management strategies are considered. We hope to contribute to policy discussions in this area and connected and legislatively similar areas (such as California) as current conditions change and existing legal and policy structures are revised. Additionally, we comment on the challenges of integrating models that ostensibly are in different domains (physical and social) but that independently represent a system in which physical processes and human actions are closely intertwined and difficult to disentangle.

  5. A simulation-based goodness-of-fit test for random effects in generalized linear mixed models

    DEFF Research Database (Denmark)

    Waagepetersen, Rasmus

    2006-01-01

    The goodness-of-fit of the distribution of random effects in a generalized linear mixed model is assessed using a conditional simulation of the random effects conditional on the observations. Provided that the specified joint model for random effects and observations is correct, the marginal...... distribution of the simulated random effects coincides with the assumed random effects distribution. In practice, the specified model depends on some unknown parameter which is replaced by an estimate. We obtain a correction for this by deriving the asymptotic distribution of the empirical distribution...

  6. A simulation-based goodness-of-fit test for random effects in generalized linear mixed models

    DEFF Research Database (Denmark)

    Waagepetersen, Rasmus Plenge

    The goodness-of-fit of the distribution of random effects in a generalized linear mixed model is assessed using a conditional simulation of the random effects conditional on the observations. Provided that the specified joint model for random effects and observations is correct, the marginal...... distribution of the simulated random effects coincides with the assumed random effects distribution. In practice the specified model depends on some unknown parameter which is replaced by an estimate. We obtain a correction for this by deriving the asymptotic distribution of the empirical distribution function...

  7. A person fit test for IRT models for polytomous items

    NARCIS (Netherlands)

    Glas, Cornelis A.W.; Dagohoy, A.V.

    2007-01-01

    A person fit test based on the Lagrange multiplier test is presented for three item response theory models for polytomous items: the generalized partial credit model, the sequential model, and the graded response model. The test can also be used in the framework of multidimensional ability

  8. High-Fidelity Solar Power Income Modeling for Solar-Electric UAVs: Development and Flight Test Based Verification

    OpenAIRE

    Oettershagen, Philipp

    2017-01-01

    Solar power models are a crucial element of solar-powered UAV design and performance analysis. During the conceptual design phase, their accuracy directly relates to the accuracy of the predicted performance metrics and thus the final design characteristics of the solar-powered UAV. Likewise, during the operations phase of a solar-powered UAV accurate solar power income models are required to predict and assess the solar power system performance. However, the existing literature on solar-powe...

  9. catcher: A Software Program to Detect Answer Copying in Multiple-Choice Tests Based on Nominal Response Model

    Science.gov (United States)

    Kalender, Ilker

    2012-01-01

    catcher is a software program designed to compute the [omega] index, a common statistical index for the identification of collusions (cheating) among examinees taking an educational or psychological test. It requires (a) responses and (b) ability estimations of individuals, and (c) item parameters to make computations and outputs the results of…

  10. Simulation-based evaluation of the performance of the F test in a linear multilevel model setting with sparseness at the level of the primary unit.

    Science.gov (United States)

    Bruyndonckx, Robin; Aerts, Marc; Hens, Niel

    2016-09-01

    In a linear multilevel model, significance of all fixed effects can be determined using F tests under maximum likelihood (ML) or restricted maximum likelihood (REML). In this paper, we demonstrate that in the presence of primary unit sparseness, the performance of the F test under both REML and ML is rather poor. Using simulations based on the structure of a data example on ceftriaxone consumption in hospitalized children, we studied variability, type I error rate and power in scenarios with a varying number of secondary units within the primary units. In general, the variability in the estimates for the effect of the primary unit decreased as the number of secondary units increased. In the presence of singletons (i.e., only one secondary unit within a primary unit), REML consistently outperformed ML, although even under REML the performance of the F test was found inadequate. When modeling the primary unit as a random effect, the power was lower while the type I error rate was unstable. The options of dropping, regrouping, or splitting the singletons could solve either the problem of a high type I error rate or a low power, while worsening the other. The permutation test appeared to be a valid alternative as it outperformed the F test, especially under REML. We conclude that in the presence of singletons, one should be careful in using the F test to determine the significance of the fixed effects, and propose the permutation test (under REML) as an alternative. © 2016 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  11. Building and verifying a severity prediction model of acute pancreatitis (AP) based on BISAP, MEWS and routine test indexes.

    Science.gov (United States)

    Ye, Jiang-Feng; Zhao, Yu-Xin; Ju, Jian; Wang, Wei

    2017-10-01

    To discuss the value of the Bedside Index for Severity in Acute Pancreatitis (BISAP), Modified Early Warning Score (MEWS), serum Ca2+, similarly hereinafter, and red cell distribution width (RDW) for predicting the severity grade of acute pancreatitis and to develop and verify a more accurate scoring system to predict the severity of AP. In 302 patients with AP, we calculated BISAP and MEWS scores and conducted regression analyses on the relationships of BISAP scoring, RDW, MEWS, and serum Ca2+ with the severity of AP using single-factor logistics. The variables with statistical significance in the single-factor logistic regression were used in a multi-factor logistic regression model; forward stepwise regression was used to screen variables and build a multi-factor prediction model. A receiver operating characteristic curve (ROC curve) was constructed, and the significance of multi- and single-factor prediction models in predicting the severity of AP using the area under the ROC curve (AUC) was evaluated. The internal validity of the model was verified through bootstrapping. Among 302 patients with AP, 209 had mild acute pancreatitis (MAP) and 93 had severe acute pancreatitis (SAP). According to single-factor logistic regression analysis, we found that BISAP, MEWS and serum Ca2+ are prediction indexes of the severity of AP (P-value0.05). The multi-factor logistic regression analysis showed that BISAP and serum Ca2+ are independent prediction indexes of AP severity (P-value0.05); BISAP is negatively related to serum Ca2+ (r=-0.330, P-valuemodel is as follows: ln()=7.306+1.151*BISAP-4.516*serum Ca2+. The predictive ability of each model for SAP follows the order of the combined BISAP and serum Ca2+ prediction model>Ca2+>BISAP. There is no statistical significance for the predictive ability of BISAP and serum Ca2+ (P-value>0.05); however, there is remarkable statistical significance for the predictive ability using the newly built prediction model as well as BISAP

  12. Progress Toward Quality Assurance Standards for Advanced Hydrocarbon Fuels Based on Thermal Performance Testing and Chemometric Modeling

    Science.gov (United States)

    2015-12-15

    4AP a < 2 < 2 < 2 < 2 < 2 - Maximum ΔP, mmHg 0.1 280.1 0.1 0 0.1 0 0 0 a “A” denotes abnormal deposit; “P” denotes peacock deposit. 6...threshold. Furthermore, the ability of the CRAFTI metrology to reproduce carbon deposit behavior after a full year of testing with a variety of special

  13. Which randomizations detect convergence and divergence in trait-based community assembly? A test of commonly used null models

    Czech Academy of Sciences Publication Activity Database

    Götzenberger, Lars; Botta-Dukát, Z.; Lepš, Jan; Pärtel, M.; Zobel, M.; de Bello, Francesco

    2016-01-01

    Roč. 27, č. 6 (2016), s. 1275-1287 ISSN 1100-9233 R&D Projects: GA ČR GAP505/12/1296 Institutional support: RVO:67985939 ; RVO:60077344 Keywords : Community ecology * Competition * Null model Subject RIV: EH - Ecology, Behaviour Impact factor: 2.924, year: 2016

  14. Indirect determination of material model parameters for single trabecula based on nanoindentation and three point bending test

    Czech Academy of Sciences Publication Activity Database

    Zlámal, P.; Jiroušek, Ondřej; Kytýř, Daniel; Doktor, Tomáš

    2013-01-01

    Roč. 58, č. 2 (2013), s. 157-171 ISSN 0001-7043 R&D Projects: GA ČR(CZ) GAP105/10/2305 Institutional support: RVO:68378297 Keywords : elasto-visco-plastic-damage model * FEM * nanoindentation * trabecular bone Subject RIV: FI - Traumatology, Orthopedics http://journal.it.cas.cz/index.php?stranka= contents

  15. Test model of WWER core

    International Nuclear Information System (INIS)

    Tikhomirov, A. V.; Gorokhov, A. K.

    2007-01-01

    The objective of this paper is creation of precision test model for WWER RP neutron-physics calculations. The model is considered as a tool for verification of deterministic computer codes that enables to reduce conservatism of design calculations and enhance WWER RP competitiveness. Precision calculations were performed using code MCNP5/1/ (Monte Carlo method). Engineering computer package Sapfir 9 5andRC V VER/2/ is used in comparative analysis of the results, it was certified for design calculations of WWER RU neutron-physics characteristic. The object of simulation is the first fuel loading of Volgodon NPP RP. Peculiarities of transition in calculation using MCNP5 from 2D geometry to 3D geometry are shown on the full-scale model. All core components as well as radial and face reflectors, automatic regulation in control and protection system control rod are represented in detail description according to the design. The first stage of application of the model is assessment of accuracy of calculation of the core power. At the second stage control and protection system control rod worth was assessed. Full scale RP representation in calculation using code MCNP5 is time consuming that calls for parallelization of computational problem on multiprocessing computer (Authors)

  16. Development of a Porcine Delayed Wound-Healing Model and Its Use in Testing a Novel Cell-Based Therapy

    International Nuclear Information System (INIS)

    Hadad, Ivan; Johnstone, Brian H.; Brabham, Jeffrey G.; Blanton, Matthew W.; Rogers, Pamela I.; Fellers, Cory; Solomon, James L.; Merfeld-Clauss, Stephanie; DesRosiers, Colleen M.; Dynlacht, Joseph R.; Coleman, John J.; March, Keith L.

    2010-01-01

    Purpose: A delayed full-thickness wound-healing model was developed and used for examining the capacity of adipose-derived stem cells (ASCs), either alone or in platelet-rich fibrin gels, to promote healing. Methods and Materials: Four pigs received electron beam radiation to the dorsal skin surface. Five weeks after radiation, subcutaneous fat was harvested from nonirradiated areas and processed to yield ASCs. Two weeks later, 28 to 30 full-thickness 1.5-cm 2 wounds were made in irradiated and nonirradiated skin. Wounds were treated with either saline solution, ASCs in saline solution, platelet-rich plasma (PRP) fibrin gel, ASCs in PRP, or non-autologous green fluorescence protein-labeled ASCs. Results: The single radiation dose produced a significant loss of dermal microvasculature density (75%) by 7 weeks. There was a significant difference in the rate of healing between irradiated and nonirradiated skin treated with saline solution. The ASCs in PRP-treated wounds exhibited a significant 11.2% improvement in wound healing compared with saline solution. Enhancement was dependent on the combination of ASCs and PRP, because neither ASCs nor PRP alone had an effect. Conclusions: We have created a model that simulates the clinically relevant late radiation effects of delayed wound healing. Using this model, we showed that a combination of ASCs and PRP improves the healing rates of perfusion-depleted tissues, possibly through enhancing local levels of growth factors.

  17. Comparison between traditional laboratory tests, permeability measurements and CT-based fluid flow modelling for cultural heritage applications

    Energy Technology Data Exchange (ETDEWEB)

    De Boever, Wesley, E-mail: Wesley.deboever@ugent.be [UGCT/PProGRess, Dept. of Geology, Ghent University, Krijgslaan 281, 9000 Ghent (Belgium); Bultreys, Tom; Derluyn, Hannelore [UGCT/PProGRess, Dept. of Geology, Ghent University, Krijgslaan 281, 9000 Ghent (Belgium); Van Hoorebeke, Luc [UGCT/Radiation Physics, Dept. of Physics & Astronomy, Ghent University, Proeftuinstraat 86, 9000 Ghent (Belgium); Cnudde, Veerle [UGCT/PProGRess, Dept. of Geology, Ghent University, Krijgslaan 281, 9000 Ghent (Belgium)

    2016-06-01

    In this paper, we examine the possibility to use on-site permeability measurements for cultural heritage applications as an alternative for traditional laboratory tests such as determination of the capillary absorption coefficient. These on-site measurements, performed with a portable air permeameter, were correlated with the pore network properties of eight sandstones and one granular limestone that are discussed in this paper. The network properties of the 9 materials tested in this study were obtained from micro-computed tomography (μCT) and compared to measurements and calculations of permeability and the capillary absorption rate of the stones under investigation, in order to find the correlation between pore network characteristics and fluid management characteristics of these sandstones. Results show a good correlation between capillary absorption, permeability and network properties, opening the possibility of using on-site permeability measurements as a standard method in cultural heritage applications. - Highlights: • Measurements of capillary absorption are compared to in-situ permeability. • We obtain pore size distribution and connectivity by using micro-CT. • These properties explain correlation between permeability and capillarity. • Correlation between both methods is good to excellent. • Permeability measurements could be a good alternative to capillarity measurement.

  18. Regression-Based Norms for a Bi-factor Model for Scoring the Brief Test of Adult Cognition by Telephone (BTACT).

    Science.gov (United States)

    Gurnani, Ashita S; John, Samantha E; Gavett, Brandon E

    2015-05-01

    The current study developed regression-based normative adjustments for a bi-factor model of the The Brief Test of Adult Cognition by Telephone (BTACT). Archival data from the Midlife Development in the United States-II Cognitive Project were used to develop eight separate linear regression models that predicted bi-factor BTACT scores, accounting for age, education, gender, and occupation-alone and in various combinations. All regression models provided statistically significant fit to the data. A three-predictor regression model fit best and accounted for 32.8% of the variance in the global bi-factor BTACT score. The fit of the regression models was not improved by gender. Eight different regression models are presented to allow the user flexibility in applying demographic corrections to the bi-factor BTACT scores. Occupation corrections, while not widely used, may provide useful demographic adjustments for adult populations or for those individuals who have attained an occupational status not commensurate with expected educational attainment. © The Author 2015. Published by Oxford University Press. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.

  19. Base pressure and heat transfer tests of the 0.0225-scale space shuttle plume simulation model (19-OTS) in yawed flight conditions in the NASA-Lewis 10x10-foot supersonic wind tunnel (test IH83)

    Science.gov (United States)

    Foust, J. W.

    1979-01-01

    Wind tunnel tests were performed to determine pressures, heat transfer rates, and gas recovery temperatures in the base region of a rocket firing model of the space shuttle integrated vehicle during simulated yawed flight conditions. First and second stage flight of the space shuttle were simulated by firing the main engines in conjunction with the SRB rocket motors or only the SSME's into the continuous tunnel airstream. For the correct rocket plume environment, the simulated altitude pressures were halved to maintain the rocket chamber/altitude pressure ratio. Tunnel freestream Mach numbers from 2.2 to 3.5 were simulated over an altitude range of 60 to 130 thousand feet with varying angle of attack, yaw angle, nozzle gimbal angle and SRB chamber pressure. Gas recovery temperature data derived from nine gas temperature probe runs are presented. The model configuration, instrumentation, test procedures, and data reduction are described.

  20. Understanding native Russian listeners' errors on an English word recognition test: model-based analysis of phoneme confusion.

    Science.gov (United States)

    Shi, Lu-Feng; Morozova, Natalia

    2012-08-01

    Word recognition is a basic component in a comprehensive hearing evaluation, but data are lacking for listeners speaking two languages. This study obtained such data for Russian natives in the US and analysed the data using the perceptual assimilation model (PAM) and speech learning model (SLM). Listeners were randomly presented 200 NU-6 words in quiet. Listeners responded verbally and in writing. Performance was scored on words and phonemes (word-initial consonants, vowels, and word-final consonants). Seven normal-hearing, adult monolingual English natives (NM), 16 English-dominant (ED), and 15 Russian-dominant (RD) Russian natives participated. ED and RD listeners differed significantly in their language background. Consistent with the SLM, NM outperformed ED listeners and ED outperformed RD listeners, whether responses were scored on words or phonemes. NM and ED listeners shared similar phoneme error patterns, whereas RD listeners' errors had unique patterns that could be largely understood via the PAM. RD listeners had particular difficulty differentiating vowel contrasts /i-I/, /æ-ε/, and /ɑ-Λ/, word-initial consonant contrasts /p-h/ and /b-f/, and word-final contrasts /f-v/. Both first-language phonology and second-language learning history affect word and phoneme recognition. Current findings may help clinicians differentiate word recognition errors due to language background from hearing pathologies.

  1. GENERATING TEST CASES FOR PLATFORM INDEPENDENT MODEL BY USING USE CASE MODEL

    OpenAIRE

    Hesham A. Hassan,; Zahraa. E. Yousif

    2010-01-01

    Model-based testing refers to testing and test case generation based on a model that describes the behavior of the system. Extensive use of models throughout all the phases of software development starting from the requirement engineering phase has led to increased importance of Model Based Testing. The OMG initiative MDA has revolutionized the way models would be used for software development. Ensuring that all user requirements are addressed in system design and the design is getting suffic...

  2. Estimation of Design Wave Loads on the SSG WEC Pilot Plant based on 3-D Model Tests

    DEFF Research Database (Denmark)

    Kofoed, Jens Peter; Vicinanza, Diego; Osaland, Espen

    2006-01-01

    This paper discuss wave loadings acting on the Wave Energy Converter (WEC) Seawave Slot-Cone Generator (SSG). The SSG is a new type of structure for wave energy conversion based on storing the incoming waves in several reservoirs placed one above the other. The wave forces on the main structure c...... to be of direct use to engineers analyzing design and stability of the pilot plant under construction at Kvitsøy island, partly funded by the EU 6th framework program (WAVESSG)....

  3. Model-based testing for space-time interaction using point processes: An application to psychiatric hospital admissions in an urban area.

    Science.gov (United States)

    Meyer, Sebastian; Warnke, Ingeborg; Rössler, Wulf; Held, Leonhard

    2016-05-01

    Spatio-temporal interaction is inherent to cases of infectious diseases and occurrences of earthquakes, whereas the spread of other events, such as cancer or crime, is less evident. Statistical significance tests of space-time clustering usually assess the correlation between the spatial and temporal (transformed) distances of the events. Although appealing through simplicity, these classical tests do not adjust for the underlying population nor can they account for a distance decay of interaction. We propose to use the framework of an endemic-epidemic point process model to jointly estimate a background event rate explained by seasonal and areal characteristics, as well as a superposed epidemic component representing the hypothesis of interest. We illustrate this new model-based test for space-time interaction by analysing psychiatric inpatient admissions in Zurich, Switzerland (2007-2012). Several socio-economic factors were found to be associated with the admission rate, but there was no evidence of general clustering of the cases. Copyright © 2016 Elsevier Ltd. All rights reserved.

  4. Modeling and Testing Legacy Data Consistency Requirements

    DEFF Research Database (Denmark)

    Nytun, J. P.; Jensen, Christian Søndergaard

    2003-01-01

    An increasing number of data sources are available on the Internet, many of which offer semantically overlapping data, but based on different schemas, or models. While it is often of interest to integrate such data sources, the lack of consistency among them makes this integration difficult....... This paper addresses the need for new techniques that enable the modeling and consistency checking for legacy data sources. Specifically, the paper contributes to the development of a framework that enables consistency testing of data coming from different types of data sources. The vehicle is UML and its...... accompanying XMI. The paper presents techniques for modeling consistency requirements using OCL and other UML modeling elements: it studies how models that describe the required consistencies among instances of legacy models can be designed in standard UML tools that support XMI. The paper also considers...

  5. Field testing of bioenergetic models

    International Nuclear Information System (INIS)

    Nagy, K.A.

    1985-01-01

    Doubly labeled water provides a direct measure of the rate of carbon dioxide production by free-living animals. With appropriate conversion factors, based on chemical composition of the diet and assimilation efficiency, field metabolic rate (FMR), in units of energy expenditure, and field feeding rate can be estimated. Validation studies indicate that doubly labeled water measurements of energy metabolism are accurate to within 7% in reptiles, birds, and mammals. This paper discusses the use of doubly labeled water to generate empirical models for FMR and food requirements for a variety of animals

  6. A 'Turing' Test for Landscape Evolution Models

    Science.gov (United States)

    Parsons, A. J.; Wise, S. M.; Wainwright, J.; Swift, D. A.

    2008-12-01

    Resolving the interactions among tectonics, climate and surface processes at long timescales has benefited from the development of computer models of landscape evolution. However, testing these Landscape Evolution Models (LEMs) has been piecemeal and partial. We argue that a more systematic approach is required. What is needed is a test that will establish how 'realistic' an LEM is and thus the extent to which its predictions may be trusted. We propose a test based upon the Turing Test of artificial intelligence as a way forward. In 1950 Alan Turing posed the question of whether a machine could think. Rather than attempt to address the question directly he proposed a test in which an interrogator asked questions of a person and a machine, with no means of telling which was which. If the machine's answer could not be distinguished from those of the human, the machine could be said to demonstrate artificial intelligence. By analogy, if an LEM cannot be distinguished from a real landscape it can be deemed to be realistic. The Turing test of intelligence is a test of the way in which a computer behaves. The analogy in the case of an LEM is that it should show realistic behaviour in terms of form and process, both at a given moment in time (punctual) and in the way both form and process evolve over time (dynamic). For some of these behaviours, tests already exist. For example there are numerous morphometric tests of punctual form and measurements of punctual process. The test discussed in this paper provides new ways of assessing dynamic behaviour of an LEM over realistically long timescales. However challenges remain in developing an appropriate suite of challenging tests, in applying these tests to current LEMs and in developing LEMs that pass them.

  7. Model test of boson mappings

    International Nuclear Information System (INIS)

    Navratil, P.; Dobes, J.

    1992-01-01

    Methods of boson mapping are tested in calculations for a simple model system of four protons and four neutrons in single-j distinguishable orbits. Two-body terms in the boson images of the fermion operators are considered. Effects of the seniority v=4 states are thus included. The treatment of unphysical states and the influence of boson space truncation are particularly studied. Both the Dyson boson mapping and the seniority boson mapping as dictated by the similarity transformed Dyson mapping do not seem to be simply amenable to truncation. This situation improves when the one-body form of the seniority image of the quadrupole operator is employed. Truncation of the boson space is addressed by using the effective operator theory with a notable improvement of results

  8. MATLAB-implemented estimation procedure for model-based assessment of hepatic insulin degradation from standard intravenous glucose tolerance test data.

    Science.gov (United States)

    Di Nardo, Francesco; Mengoni, Michele; Morettini, Micaela

    2013-05-01

    Present study provides a novel MATLAB-based parameter estimation procedure for individual assessment of hepatic insulin degradation (HID) process from standard frequently-sampled intravenous glucose tolerance test (FSIGTT) data. Direct access to the source code, offered by MATLAB, enabled us to design an optimization procedure based on the alternating use of Gauss-Newton's and Levenberg-Marquardt's algorithms, which assures the full convergence of the process and the containment of computational time. Reliability was tested by direct comparison with the application, in eighteen non-diabetic subjects, of well-known kinetic analysis software package SAAM II, and by application on different data. Agreement between MATLAB and SAAM II was warranted by intraclass correlation coefficients ≥0.73; no significant differences between corresponding mean parameter estimates and prediction of HID rate; and consistent residual analysis. Moreover, MATLAB optimization procedure resulted in a significant 51% reduction of CV% for the worst-estimated parameter by SAAM II and in maintaining all model-parameter CV% MATLAB-based procedure was suggested as a suitable tool for the individual assessment of HID process. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.

  9. BRCA mutation carrier detection. A model-based cost-effectiveness analysis comparing the traditional family history approach and the testing of all patients with breast cancer

    Science.gov (United States)

    Norum, Jan; Grindedal, Eli Marie; Heramb, Cecilie; Karsrud, Inga; Ariansen, Sarah Louise; Undlien, Dag Erik; Schlichting, Ellen; Mæhle, Lovise

    2018-01-01

    Background Identification of BRCA mutation carriers among patients with breast cancer (BC) involves costs and gains. Testing has been performed according to international guidelines, focusing on family history (FH) of breast and/or ovarian cancer. An alternative is testing all patients with BC employing sequencing of the BRCA genes and Multiplex Ligation Probe Amplification (MLPA). Patients and methods A model-based cost-effectiveness analysis, employing data from Oslo University Hospital, Ullevål (OUH-U) and a decision tree, was done. The societal and the healthcare perspectives were focused and a lifetime perspective employed. The comparators were the traditional FH approach used as standard of care at OUH-U in 2013 and the intervention (testing all patients with BC) performed in 2014 and 2015 at the same hospital. During the latter period, 535 patients with BC were offered BRCA testing with sequencing and MLPA. National 2014 data on mortality rates and costs were implemented, a 3% discount rate used and the costing year was 2015. The incremental cost-effectiveness ratio was calculated in euros (€) per life-year gained (LYG). Results The net healthcare cost (healthcare perspective) was €40 503/LYG. Including all resource use (societal perspective), the cost was €5669/LYG. The univariate sensitivity analysis documented the unit cost of the BRCA test and the number of LYGs the prominent parameters affecting the result. Diagnostic BRCA testing of all patients with BC was superior to the FH approach and cost-effective within the frequently used thresholds (healthcare perspective) in Norway (€60 000–€80 000/LYG). PMID:29682331

  10. BRCA mutation carrier detection. A model-based cost-effectiveness analysis comparing the traditional family history approach and the testing of all patients with breast cancer.

    Science.gov (United States)

    Norum, Jan; Grindedal, Eli Marie; Heramb, Cecilie; Karsrud, Inga; Ariansen, Sarah Louise; Undlien, Dag Erik; Schlichting, Ellen; Mæhle, Lovise

    2018-01-01

    Identification of BRCA mutation carriers among patients with breast cancer (BC) involves costs and gains. Testing has been performed according to international guidelines, focusing on family history (FH) of breast and/or ovarian cancer. An alternative is testing all patients with BC employing sequencing of the BRCA genes and Multiplex Ligation Probe Amplification (MLPA). A model-based cost-effectiveness analysis, employing data from Oslo University Hospital, Ullevål (OUH-U) and a decision tree, was done. The societal and the healthcare perspectives were focused and a lifetime perspective employed. The comparators were the traditional FH approach used as standard of care at OUH-U in 2013 and the intervention (testing all patients with BC) performed in 2014 and 2015 at the same hospital. During the latter period, 535 patients with BC were offered BRCA testing with sequencing and MLPA. National 2014 data on mortality rates and costs were implemented, a 3% discount rate used and the costing year was 2015. The incremental cost-effectiveness ratio was calculated in euros (€) per life-year gained (LYG). The net healthcare cost (healthcare perspective) was €40 503/LYG. Including all resource use (societal perspective), the cost was €5669/LYG. The univariate sensitivity analysis documented the unit cost of the BRCA test and the number of LYGs the prominent parameters affecting the result.Diagnostic BRCA testing of all patients with BC was superior to the FH approach and cost-effective within the frequently used thresholds (healthcare perspective) in Norway (€60 000-€80 000/LYG).

  11. Pile Model Tests Using Strain Gauge Technology

    Science.gov (United States)

    Krasiński, Adam; Kusio, Tomasz

    2015-09-01

    Ordinary pile bearing capacity tests are usually carried out to determine the relationship between load and displacement of pile head. The measurement system required in such tests consists of force transducer and three or four displacement gauges. The whole system is installed at the pile head above the ground level. This approach, however, does not give us complete information about the pile-soil interaction. We can only determine the total bearing capacity of the pile, without the knowledge of its distribution into the shaft and base resistances. Much more information can be obtained by carrying out a test of instrumented pile equipped with a system for measuring the distribution of axial force along its core. In the case of pile model tests the use of such measurement is difficult due to small scale of the model. To find a suitable solution for axial force measurement, which could be applied to small scale model piles, we had to take into account the following requirements: - a linear and stable relationship between measured and physical values, - the force measurement accuracy of about 0.1 kN, - the range of measured forces up to 30 kN, - resistance of measuring gauges against aggressive counteraction of concrete mortar and against moisture, - insensitivity to pile bending, - economical factor. These requirements can be fulfilled by strain gauge sensors if an appropriate methodology is used for test preparation (Hoffmann [1]). In this paper, we focus on some aspects of the application of strain gauge sensors for model pile tests. The efficiency of the method is proved on the examples of static load tests carried out on SDP model piles acting as single piles and in a group.

  12. Low-order model of the Loss-of-Fluid Test (LOFT) reactor plant for use in Kalman filter-based optimal estimators

    International Nuclear Information System (INIS)

    Tylee, J.L.

    1980-01-01

    A low-order, nonlinear model of the Loss-of-Fluid Test (LOFT) reactor plant, for use in Kalman filter estimators, is developed, described, and evaluated. This model consists of 31 differential equations and represents all major subsystems of both the primary and secondary sides of the LOFT plant. Comparisons between model calculations and available LOFT power range testing transients demonstrate the accuracy of the low-order model. The nonlinear model is numerically linearized for future implementation in Kalman filter and optimal control algorithms. The linearized model is shown to be an adequate representation of the nonlinear plant dynamics

  13. Model-based analysis of costs and outcomes of non-invasive prenatal testing for Down's syndrome using cell free fetal DNA in the UK National Health Service.

    Directory of Open Access Journals (Sweden)

    Stephen Morris

    Full Text Available Non-invasive prenatal testing (NIPT for Down's syndrome (DS using cell free fetal DNA in maternal blood has the potential to dramatically alter the way prenatal screening and diagnosis is delivered. Before NIPT can be implemented into routine practice, information is required on its costs and benefits. We investigated the costs and outcomes of NIPT for DS as contingent testing and as first-line testing compared with the current DS screening programme in the UK National Health Service.We used a pre-existing model to evaluate the costs and outcomes associated with NIPT compared with the current DS screening programme. The analysis was based on a hypothetical screening population of 10,000 pregnant women. Model inputs were taken from published sources. The main outcome measures were number of DS cases detected, number of procedure-related miscarriages and total cost.At a screening risk cut-off of 1∶150 NIPT as contingent testing detects slightly fewer DS cases, has fewer procedure-related miscarriages, and costs the same as current DS screening (around UK£280,000 at a cost of £500 per NIPT. As first-line testing NIPT detects more DS cases, has fewer procedure-related miscarriages, and is more expensive than current screening at a cost of £50 per NIPT. When NIPT uptake increases, NIPT detects more DS cases with a small increase in procedure-related miscarriages and costs.NIPT is currently available in the private sector in the UK at a price of £400-£900. If the NHS cost was at the lower end of this range then at a screening risk cut-off of 1∶150 NIPT as contingent testing would be cost neutral or cost saving compared with current DS screening. As first-line testing NIPT is likely to produce more favourable outcomes but at greater cost. Further research is needed to evaluate NIPT under real world conditions.

  14. Physical modelling and testing in environmental geotechnics

    International Nuclear Information System (INIS)

    Garnier, J.; Thorel, L.; Haza, E.

    2000-01-01

    The preservation of natural environment has become a major concern, which affects nowadays a wide range of professionals from local communities administrators to natural resources managers (water, wildlife, flora, etc) and, in the end, to the consumers that we all are. Although totally ignored some fifty years ago, environmental geotechnics has become an emergent area of study and research which borders on the traditional domains, with which the geo-technicians are confronted (soil and rock mechanics, engineering geology, natural and anthropogenic risk management). Dedicated to experimental approaches (in-situ investigations and tests, laboratory tests, small-scale model testing), the Symposium fits in with the geotechnical domains of environment and transport of soil pollutants. These proceedings report some progress of developments in measurement techniques and studies of transport of pollutants in saturated and unsaturated soils in order to improve our understanding of such phenomena within multiphase environments. Experimental investigations on decontamination and isolation methods for polluted soils are discussed. The intention is to assess the impact of in-situ and laboratory tests, as well as small-scale model testing, on engineering practice. One paper is analysed in INIS data base for its specific interest in nuclear industry. The other ones, concerning the energy, are analyzed in ETDE data base

  15. Behavior of sulfur species in steam generator conditions of PWRs - towards an update of the secondary side corrosion cracking model based on laboratory tests in sulfate environments

    International Nuclear Information System (INIS)

    Mansour, C.; Legras, L.; Catalette, H.; Lefevre, G.; Fedoroff, M.; Pavageau, E.-M.

    2007-01-01

    Secondary side corrosion cracking affects Mill Annealed Alloy 600 steam generator (SG) tubes of PWRs in flow restricted areas where pollutants, such as sulfate, can concentrate and form various aggressive local environments. The 'sulfate model', based on laboratory tests in sulfate environments, was developed to predict the degradation of SG tubes. Such prediction is aimed to be done after having evaluated the chemistry in the flow restricted areas where the degradation occurs. For such purpose, a better knowledge of the behavior of sulfur species in SG conditions is needed. After a brief description of the sulfate model, this paper focuses on the latest experimental results that have been obtained : sorption of sulfur species over magnetite in SG temperature conditions, thermodynamical calculations as well as transmission electron microscopy and X-ray photoelectron spectroscopy observations of magnetite after sorption and C-Ring specimens tested in sulfate environments. Then, all these results are discussed in order to contribute to a better understanding of the secondary side corrosion cracking of SG tubes. (author)

  16. Algorithms to analyze the quality test parameter values of seafood in the proposed ontology based seafood quality analyzer and miner (ONTO SQAM model

    Directory of Open Access Journals (Sweden)

    Vinu Sherimon

    2017-07-01

    Full Text Available Ensuring the quality of food, particularly seafood has increasingly become an important issue nowadays. Quality Management Systems empower any organization to identify, measure, control and improve the quality of the products manufactured that will eventually lead to improved business performance. With the advent of new technologies, now intelligent systems are being developed. To ensure the quality of seafood, an ontology based seafood quality analyzer and miner (ONTO SQAM model is proposed. The knowledge is represented using ontology. The domain concepts are defined using ontology. This paper presents the initial part of the proposed model – the analysis of quality test parameter values. Two algorithms are proposed to do the analysis – Comparison Algorithm and Data Store Updater algorithm. The algorithms ensure that the values of various quality tests are in the acceptable range. The real data sets taken from different seafood companies in Kerala, India, and validated by the Marine Product Export Development Authority of India (MPEDA are used for the experiments. The performance of the algorithms is evaluated using standard performance metrics such as precision, recall, and accuracy. The results obtained show that all the three measures achieved good results.

  17. Interface-based software testing

    OpenAIRE

    Aziz Ahmad Rais

    2016-01-01

    Software quality is determined by assessing the characteristics that specify how it should work, which are verified through testing. If it were possible to touch, see, or measure software, it would be easier to analyze and prove its quality. Unfortunately, software is an intangible asset, which makes testing complex. This is especially true when software quality is not a question of particular functions that can be tested through a graphical user interface. The primary objective of softwar...

  18. Accuracy tests of the tessellated SLBM model

    International Nuclear Information System (INIS)

    Ramirez, A L; Myers, S C

    2007-01-01

    We have compared the Seismic Location Base Model (SLBM) tessellated model (version 2.0 Beta, posted July 3, 2007) with the GNEMRE Unified Model. The comparison is done on a layer/depth-by-layer/depth and layer/velocity-by-layer/velocity comparison. The SLBM earth model is defined on a tessellation that spans the globe at a constant resolution of about 1 degree (Ballard, 2007). For the tests, we used the earth model in file ''unified( ) iasp.grid''. This model contains the top 8 layers of the Unified Model (UM) embedded in a global IASP91 grid. Our test queried the same set of nodes included in the UM model file. To query the model stored in memory, we used some of the functionality built into the SLBMInterface object. We used the method get InterpolatedPoint() to return desired values for each layer at user-specified points. The values returned include: depth to the top of each layer, layer velocity, layer thickness and (for the upper-mantle layer) velocity gradient. The SLBM earth model has an extra middle crust layer whose values are used when Pg/Lg phases are being calculated. This extra layer was not accessed by our tests. Figures 1 to 8 compare the layer depths, P velocities and P gradients in the UM and SLBM models. The figures show results for the three sediment layers, three crustal layers and the upper mantle layer defined in the UM model. Each layer in the models (sediment1, sediment2, sediment3, upper crust, middle crust, lower crust and upper mantle) is shown on a separate figure. The upper mantle P velocity and gradient distribution are shown on Figures 7 and 8. The left and center images in the top row of each figure is the rendering of depth to the top of the specified layer for the UM and SLBM models. When a layer has zero thickness, its depth is the same as that of the layer above. The right image in the top row is the difference between in layer depth for the UM and SLBM renderings. The left and center images in the bottom row of the figures are

  19. Computerized Classification Testing with the Rasch Model

    Science.gov (United States)

    Eggen, Theo J. H. M.

    2011-01-01

    If classification in a limited number of categories is the purpose of testing, computerized adaptive tests (CATs) with algorithms based on sequential statistical testing perform better than estimation-based CATs (e.g., Eggen & Straetmans, 2000). In these computerized classification tests (CCTs), the Sequential Probability Ratio Test (SPRT) (Wald,…

  20. Processes of behavior change and weight loss in a theory-based weight loss intervention program: a test of the process model for lifestyle behavior change.

    Science.gov (United States)

    Gillison, Fiona; Stathi, Afroditi; Reddy, Prasuna; Perry, Rachel; Taylor, Gordon; Bennett, Paul; Dunbar, James; Greaves, Colin

    2015-01-16

    Process evaluation is important for improving theories of behavior change and behavioral intervention methods. The present study reports on the process outcomes of a pilot test of the theoretical model (the Process Model for Lifestyle Behavior Change; PMLBC) underpinning an evidence-informed, theory-driven, group-based intervention designed to promote healthy eating and physical activity for people with high cardiovascular risk. 108 people at high risk of diabetes or heart disease were randomized to a group-based weight management intervention targeting diet and physical activity plus usual care, or to usual care. The intervention comprised nine group based sessions designed to promote motivation, social support, self-regulation and understanding of the behavior change process. Weight loss, diet, physical activity and theoretically defined mediators of change were measured pre-intervention, and after four and 12 months. The intervention resulted in significant improvements in fiber intake (M between-group difference = 5.7 g/day, p behavior change, and the predicted mechanisms of change specified in the PMBLC were largely supported. Improvements in self-efficacy and understanding of the behavior change process were associated with engagement in coping planning and self-monitoring activities, and successful dietary change at four and 12 months. While participants reported improvements in motivational and social support variables, there was no effect of these, or of the intervention overall, on physical activity. The data broadly support the theoretical model for supporting some dietary changes, but not for physical activity. Systematic intervention design allowed us to identify where improvements to the intervention may be implemented to promote change in all proposed mediators. More work is needed to explore effective mechanisms within interventions to promote physical activity behavior.

  1. Effect of Item Response Theory (IRT) Model Selection on Testlet-Based Test Equating. Research Report. ETS RR-14-19

    Science.gov (United States)

    Cao, Yi; Lu, Ru; Tao, Wei

    2014-01-01

    The local item independence assumption underlying traditional item response theory (IRT) models is often not met for tests composed of testlets. There are 3 major approaches to addressing this issue: (a) ignore the violation and use a dichotomous IRT model (e.g., the 2-parameter logistic [2PL] model), (b) combine the interdependent items to form a…

  2. Enhancing SAT-Based Test Pattern Generation

    Institute of Scientific and Technical Information of China (English)

    LIU Xin; XIONG You-lun

    2005-01-01

    This paper presents modeling tools based on Boolean satisfiability (SAT) to solve problems of test generation for combinational circuits. It exploits an added layer to maintain circuit-related information and value justification relations to a generic SAT algorithm. It dovetails binary decision graphs (BDD) and SAT techniques to improve the efficiency of automatic test pattern generation (ATPG). More specifically, it first exploits inexpensive reconvergent fanout analysis of circuit to gather information on the local signal correlation by using BDD learning, then uses the above learned information to restrict and focus the overall search space of SAT-based ATPG. Its learning technique is effective and lightweight. The experimental results demonstrate the effectiveness of the approach.

  3. 46 CFR 154.431 - Model test.

    Science.gov (United States)

    2010-10-01

    ... 46 Shipping 5 2010-10-01 2010-10-01 false Model test. 154.431 Section 154.431 Shipping COAST GUARD... Model test. (a) The primary and secondary barrier of a membrane tank, including the corners and joints...(c). (b) Analyzed data of a model test for the primary and secondary barrier of the membrane tank...

  4. 46 CFR 154.449 - Model test.

    Science.gov (United States)

    2010-10-01

    ... 46 Shipping 5 2010-10-01 2010-10-01 false Model test. 154.449 Section 154.449 Shipping COAST GUARD, DEPARTMENT OF HOMELAND SECURITY (CONTINUED) CERTAIN BULK DANGEROUS CARGOES SAFETY STANDARDS FOR SELF... § 154.449 Model test. The following analyzed data of a model test of structural elements for independent...

  5. Prospective Tests on Biological Models of Acupuncture

    Directory of Open Access Journals (Sweden)

    Charles Shang

    2009-01-01

    Full Text Available The biological effects of acupuncture include the regulation of a variety of neurohumoral factors and growth control factors. In science, models or hypotheses with confirmed predictions are considered more convincing than models solely based on retrospective explanations. Literature review showed that two biological models of acupuncture have been prospectively tested with independently confirmed predictions: The neurophysiology model on the long-term effects of acupuncture emphasizes the trophic and anti-inflammatory effects of acupuncture. Its prediction on the peripheral effect of endorphin in acupuncture has been confirmed. The growth control model encompasses the neurophysiology model and suggests that a macroscopic growth control system originates from a network of organizers in embryogenesis. The activity of the growth control system is important in the formation, maintenance and regulation of all the physiological systems. Several phenomena of acupuncture such as the distribution of auricular acupuncture points, the long-term effects of acupuncture and the effect of multimodal non-specific stimulation at acupuncture points are consistent with the growth control model. The following predictions of the growth control model have been independently confirmed by research results in both acupuncture and conventional biomedical sciences: (i Acupuncture has extensive growth control effects. (ii Singular point and separatrix exist in morphogenesis. (iii Organizers have high electric conductance, high current density and high density of gap junctions. (iv A high density of gap junctions is distributed as separatrices or boundaries at body surface after early embryogenesis. (v Many acupuncture points are located at transition points or boundaries between different body domains or muscles, coinciding with the connective tissue planes. (vi Some morphogens and organizers continue to function after embryogenesis. Current acupuncture research suggests a

  6. Physical modelling and testing in environmental geotechnics

    Energy Technology Data Exchange (ETDEWEB)

    Garnier, J.; Thorel, L.; Haza, E. [Laboratoire Central des Ponts et Chaussees a Nantes, 44 - Nantes (France)

    2000-07-01

    The preservation of natural environment has become a major concern, which affects nowadays a wide range of professionals from local communities administrators to natural resources managers (water, wildlife, flora, etc) and, in the end, to the consumers that we all are. Although totally ignored some fifty years ago, environmental geotechnics has become an emergent area of study and research which borders on the traditional domains, with which the geo-technicians are confronted (soil and rock mechanics, engineering geology, natural and anthropogenic risk management). Dedicated to experimental approaches (in-situ investigations and tests, laboratory tests, small-scale model testing), the Symposium fits in with the geotechnical domains of environment and transport of soil pollutants. These proceedings report some progress of developments in measurement techniques and studies of transport of pollutants in saturated and unsaturated soils in order to improve our understanding of such phenomena within multiphase environments. Experimental investigations on decontamination and isolation methods for polluted soils are discussed. The intention is to assess the impact of in-situ and laboratory tests, as well as small-scale model testing, on engineering practice. One paper has been analyzed in INIS data base for its specific interest in nuclear industry.

  7. Semantics-based Automated Web Testing

    Directory of Open Access Journals (Sweden)

    Hai-Feng Guo

    2015-08-01

    Full Text Available We present TAO, a software testing tool performing automated test and oracle generation based on a semantic approach. TAO entangles grammar-based test generation with automated semantics evaluation using a denotational semantics framework. We show how TAO can be incorporated with the Selenium automation tool for automated web testing, and how TAO can be further extended to support automated delta debugging, where a failing web test script can be systematically reduced based on grammar-directed strategies. A real-life parking website is adopted throughout the paper to demonstrate the effectivity of our semantics-based web testing approach.

  8. Vehicle rollover sensor test modeling

    NARCIS (Netherlands)

    McCoy, R.W.; Chou, C.C.; Velde, R. van de; Twisk, D.; Schie, C. van

    2007-01-01

    A computational model of a mid-size sport utility vehicle was developed using MADYMO. The model includes a detailed description of the suspension system and tire characteristics that incorporated the Delft-Tyre magic formula description. The model was correlated by simulating a vehicle suspension

  9. Web Based VRML Modelling

    NARCIS (Netherlands)

    Kiss, S.; Sarfraz, M.

    2004-01-01

    Presents a method to connect VRML (Virtual Reality Modeling Language) and Java components in a Web page using EAI (External Authoring Interface), which makes it possible to interactively generate and edit VRML meshes. The meshes used are based on regular grids, to provide an interaction and modeling

  10. Testing the compounding structure of the CP-INARCH model

    OpenAIRE

    Weiß, Christian H.; Gonçalves, Esmeralda; Lopes, Nazaré Mendes

    2017-01-01

    A statistical test to distinguish between a Poisson INARCH model and a Compound Poisson INARCH model is proposed, based on the form of the probability generating function of the compounding distribution of the conditional law of the model. For first-order autoregression, the normality of the test statistics’ asymptotic distribution is established, either in the case where the model parameters are specified, or when such parameters are consistently estimated. As the test statistics’ law involv...

  11. Black hole based tests of general relativity

    International Nuclear Information System (INIS)

    Yagi, Kent; Stein, Leo C

    2016-01-01

    General relativity has passed all solar system experiments and neutron star based tests, such as binary pulsar observations, with flying colors. A more exotic arena for testing general relativity is in systems that contain one or more black holes. Black holes are the most compact objects in the Universe, providing probes of the strongest-possible gravitational fields. We are motivated to study strong-field gravity since many theories give large deviations from general relativity only at large field strengths, while recovering the weak-field behavior. In this article, we review how one can probe general relativity and various alternative theories of gravity by using electromagnetic waves from a black hole with an accretion disk, and gravitational waves from black hole binaries. We first review model-independent ways of testing gravity with electromagnetic/gravitational waves from a black hole system. We then focus on selected examples of theories that extend general relativity in rather simple ways. Some important characteristics of general relativity include (but are not limited to) (i) only tensor gravitational degrees of freedom, (ii) the graviton is massless, (iii) no quadratic or higher curvatures in the action, and (iv) the theory is four-dimensional. Altering a characteristic leads to a different extension of general relativity: (i) scalar–tensor theories, (ii) massive gravity theories, (iii) quadratic gravity, and (iv) theories with large extra dimensions. Within each theory, we describe black hole solutions, their properties, and current and projected constraints on each theory using black hole based tests of gravity. We close this review by listing some of the open problems in model-independent tests and within each specific theory. (paper)

  12. Engineering model cryocooler test results

    International Nuclear Information System (INIS)

    Skimko, M.A.; Stacy, W.D.; McCormick, J.A.

    1992-01-01

    This paper reports that recent testing of diaphragm-defined, Stirling-cycle machines and components has demonstrated cooling performance potential, validated the design code, and confirmed several critical operating characteristics. A breadboard cryocooler was rebuilt and tested from cryogenic to near-ambient cold end temperatures. There was a significant increase in capacity at cryogenic temperatures and the performance results compared will with code predictions at all temperatures. Further testing on a breadboard diaphragm compressor validated the calculated requirement for a minimum axial clearance between diaphragms and mating heads

  13. Blast Testing and Modelling of Composite Structures

    DEFF Research Database (Denmark)

    Giversen, Søren

    The motivation for this work is based on a desire for finding light weight alternatives to high strength steel as the material to use for armouring in military vehicles. With the use of high strength steel, an increase in the level of armouring has a significant impact on the vehicle weight......, affecting for example the manoeuvrability and top speed negatively, which ultimately affects the safety of the personal in the vehicle. Strong and light materials, such as fibre reinforced composites, could therefore act as substitutes for the high strength steel, and minimize the impact on the vehicle...... work this set-up should be improved such that the modelled pressure can be validated. For tests performed with a 250g charge load comparisons with model data showed poor agreement. This was found to be due to improper design of the modelled laminate panels, where the layer interface delamination...

  14. BIOMOVS test scenario model comparison using BIOPATH

    International Nuclear Information System (INIS)

    Grogan, H.A.; Van Dorp, F.

    1986-07-01

    This report presents the results of the irrigation test scenario, presented in the BIOMOVS intercomparison study, calculated by the computer code BIOPATH. This scenario defines a constant release of Tc-99 and Np-237 into groundwater that is used for irrigation. The system of compartments used to model the biosphere is based upon an area in northern Switzerland and is essentially the same as that used in Projekt Gewaehr to assess the radiological impact of a high level waste repository. Two separate irrigation methods are considered, namely ditch and overhead irrigation. Their influence on the resultant activities calculated in the groundwater, soil and different foodproducts, as a function of time, is evaluated. The sensitivity of the model to parameter variations is analysed which allows a deeper understanding of the model chain. These results are assessed subjectively in a first effort to realistically quantify the uncertainty associated with each calculated activity. (author)

  15. Inverse hydrochemical models of aqueous extracts tests

    Energy Technology Data Exchange (ETDEWEB)

    Zheng, L.; Samper, J.; Montenegro, L.

    2008-10-10

    Aqueous extract test is a laboratory technique commonly used to measure the amount of soluble salts of a soil sample after adding a known mass of distilled water. Measured aqueous extract data have to be re-interpreted in order to infer porewater chemical composition of the sample because porewater chemistry changes significantly due to dilution and chemical reactions which take place during extraction. Here we present an inverse hydrochemical model to estimate porewater chemical composition from measured water content, aqueous extract, and mineralogical data. The model accounts for acid-base, redox, aqueous complexation, mineral dissolution/precipitation, gas dissolution/ex-solution, cation exchange and surface complexation reactions, of which are assumed to take place at local equilibrium. It has been solved with INVERSE-CORE{sup 2D} and been tested with bentonite samples taken from FEBEX (Full-scale Engineered Barrier EXperiment) in situ test. The inverse model reproduces most of the measured aqueous data except bicarbonate and provides an effective, flexible and comprehensive method to estimate porewater chemical composition of clays. Main uncertainties are related to kinetic calcite dissolution and variations in CO2(g) pressure.

  16. From GenBank to GBIF: Phylogeny-Based Predictive Niche Modeling Tests Accuracy of Taxonomic Identifications in Large Occurrence Data Repositories.

    Science.gov (United States)

    Smith, B Eugene; Johnston, Mark K; Lücking, Robert

    2016-01-01

    Accuracy of taxonomic identifications is crucial to data quality in online repositories of species occurrence data, such as the Global Biodiversity Information Facility (GBIF), which have accumulated several hundred million records over the past 15 years. These data serve as basis for large scale analyses of macroecological and biogeographic patterns and to document environmental changes over time. However, taxonomic identifications are often unreliable, especially for non-vascular plants and fungi including lichens, which may lack critical revisions of voucher specimens. Due to the scale of the problem, restudy of millions of collections is unrealistic and other strategies are needed. Here we propose to use verified, georeferenced occurrence data of a given species to apply predictive niche modeling that can then be used to evaluate unverified occurrences of that species. Selecting the charismatic lichen fungus, Usnea longissima, as a case study, we used georeferenced occurrence records based on sequenced specimens to model its predicted niche. Our results suggest that the target species is largely restricted to a narrow range of boreal and temperate forest in the Northern Hemisphere and that occurrence records in GBIF from tropical regions and the Southern Hemisphere do not represent this taxon, a prediction tested by comparison with taxonomic revisions of Usnea for these regions. As a novel approach, we employed Principal Component Analysis on the environmental grid data used for predictive modeling to visualize potential ecogeographical barriers for the target species; we found that tropical regions conform a strong barrier, explaining why potential niches in the Southern Hemisphere were not colonized by Usnea longissima and instead by morphologically similar species. This approach is an example of how data from two of the most important biodiversity repositories, GenBank and GBIF, can be effectively combined to remotely address the problem of inaccuracy of

  17. Team-Based Testing Improves Individual Learning

    Science.gov (United States)

    Vogler, Jane S.; Robinson, Daniel H.

    2016-01-01

    In two experiments, 90 undergraduates took six tests as part of an educational psychology course. Using a crossover design, students took three tests individually without feedback and then took the same test again, following the process of team-based testing (TBT), in teams in which the members reached consensus for each question and answered…

  18. The Model Identification Test: A Limited Verbal Science Test

    Science.gov (United States)

    McIntyre, P. J.

    1972-01-01

    Describes the production of a test with a low verbal load for use with elementary school science students. Animated films were used to present appropriate and inappropriate models of the behavior of particles of matter. (AL)

  19. Theoretical Models, Assessment Frameworks and Test Construction.

    Science.gov (United States)

    Chalhoub-Deville, Micheline

    1997-01-01

    Reviews the usefulness of proficiency models influencing second language testing. Findings indicate that several factors contribute to the lack of congruence between models and test construction and make a case for distinguishing between theoretical models. Underscores the significance of an empirical, contextualized and structured approach to the…

  20. Experimental Evaluation for the Microvibration Performance of a Segmented PC Method Based High Technology Industrial Facility Using 1/2 Scale Test Models

    Directory of Open Access Journals (Sweden)

    Sijun Kim

    2017-01-01

    Full Text Available The precast concrete (PC method used in the construction process of high technology industrial facilities is limited when applied to those with greater span lengths, due to the transport length restriction (maximum length of 15~16 m in Korea set by traffic laws. In order to resolve this, this study introduces a structural system with a segmented PC system, and a 1/2 scale model with a width of 9000 mm (hereafter Segmented Model is manufactured to evaluate vibration performance. Since a real vibrational environment cannot be reproduced for vibration testing using a scale model, a comparative analysis of their relative performances is conducted in this study. For this purpose, a 1/2 scale model with a width of 7200 mm (hereafter Nonsegmented Model of a high technology industrial facility is additionally prepared using the conventional PC method. By applying the same experiment method for both scale models and comparing the results, the relative vibration performance of the Segmented Model is observed. Through impact testing, the natural frequencies of the two scale models are compared. Also, in order to analyze the estimated response induced by the equipment, the vibration responses due to the exciter are compared. The experimental results show that the Segmented Model exhibits similar or superior performances when compared to the Nonsegmented Model.

  1. Geochemical Testing And Model Development - Residual Tank Waste Test Plan

    International Nuclear Information System (INIS)

    Cantrell, K.J.; Connelly, M.P.

    2010-01-01

    This Test Plan describes the testing and chemical analyses release rate studies on tank residual samples collected following the retrieval of waste from the tank. This work will provide the data required to develop a contaminant release model for the tank residuals from both sludge and salt cake single-shell tanks. The data are intended for use in the long-term performance assessment and conceptual model development.

  2. Hydraulic Model Tests on Modified Wave Dragon

    DEFF Research Database (Denmark)

    Hald, Tue; Lynggaard, Jakob

    A floating model of the Wave Dragon (WD) was built in autumn 1998 by the Danish Maritime Institute in scale 1:50, see Sørensen and Friis-Madsen (1999) for reference. This model was subjected to a series of model tests and subsequent modifications at Aalborg University and in the following...... are found in Hald and Lynggaard (2001). Model tests and reconstruction are carried out during the phase 3 project: ”Wave Dragon. Reconstruction of an existing model in scale 1:50 and sequentiel tests of changes to the model geometry and mass distribution parameters” sponsored by the Danish Energy Agency...

  3. A model for optimal constrained adaptive testing

    NARCIS (Netherlands)

    van der Linden, Willem J.; Reese, Lynda M.

    2001-01-01

    A model for constrained computerized adaptive testing is proposed in which the information on the test at the ability estimate is maximized subject to a large variety of possible constraints on the contents of the test. At each item-selection step, a full test is first assembled to have maximum

  4. A model for optimal constrained adaptive testing

    NARCIS (Netherlands)

    van der Linden, Willem J.; Reese, Lynda M.

    1997-01-01

    A model for constrained computerized adaptive testing is proposed in which the information in the test at the ability estimate is maximized subject to a large variety of possible constraints on the contents of the test. At each item-selection step, a full test is first assembled to have maximum

  5. Model Based Temporal Reasoning

    Science.gov (United States)

    Rabin, Marla J.; Spinrad, Paul R.; Fall, Thomas C.

    1988-03-01

    Systems that assess the real world must cope with evidence that is uncertain, ambiguous, and spread over time. Typically, the most important function of an assessment system is to identify when activities are occurring that are unusual or unanticipated. Model based temporal reasoning addresses both of these requirements. The differences among temporal reasoning schemes lies in the methods used to avoid computational intractability. If we had n pieces of data and we wanted to examine how they were related, the worst case would be where we had to examine every subset of these points to see if that subset satisfied the relations. This would be 2n, which is intractable. Models compress this; if several data points are all compatible with a model, then that model represents all those data points. Data points are then considered related if they lie within the same model or if they lie in models that are related. Models thus address the intractability problem. They also address the problem of determining unusual activities if the data do not agree with models that are indicated by earlier data then something out of the norm is taking place. The models can summarize what we know up to that time, so when they are not predicting correctly, either something unusual is happening or we need to revise our models. The model based reasoner developed at Advanced Decision Systems is thus both intuitive and powerful. It is currently being used on one operational system and several prototype systems. It has enough power to be used in domains spanning the spectrum from manufacturing engineering and project management to low-intensity conflict and strategic assessment.

  6. Cluster Based Text Classification Model

    DEFF Research Database (Denmark)

    Nizamani, Sarwat; Memon, Nasrullah; Wiil, Uffe Kock

    2011-01-01

    We propose a cluster based classification model for suspicious email detection and other text classification tasks. The text classification tasks comprise many training examples that require a complex classification model. Using clusters for classification makes the model simpler and increases...... the accuracy at the same time. The test example is classified using simpler and smaller model. The training examples in a particular cluster share the common vocabulary. At the time of clustering, we do not take into account the labels of the training examples. After the clusters have been created......, the classifier is trained on each cluster having reduced dimensionality and less number of examples. The experimental results show that the proposed model outperforms the existing classification models for the task of suspicious email detection and topic categorization on the Reuters-21578 and 20 Newsgroups...

  7. Seepage Calibration Model and Seepage Testing Data

    International Nuclear Information System (INIS)

    Dixon, P.

    2004-01-01

    The purpose of this Model Report is to document the Seepage Calibration Model (SCM). The SCM is developed (1) to establish the conceptual basis for the Seepage Model for Performance Assessment (SMPA), and (2) to derive seepage-relevant, model-related parameters and their distributions for use in the SMPA and seepage abstraction in support of the Total System Performance Assessment for License Application (TSPA-LA). The SCM is intended to be used only within this Model Report for the estimation of seepage-relevant parameters through calibration of the model against seepage-rate data from liquid-release tests performed in several niches along the Exploratory Studies Facility (ESF) Main Drift and in the Cross Drift. The SCM does not predict seepage into waste emplacement drifts under thermal or ambient conditions. Seepage predictions for waste emplacement drifts under ambient conditions will be performed with the SMPA (see upcoming REV 02 of CRWMS M and O 2000 [153314]), which inherits the conceptual basis and model-related parameters from the SCM. Seepage during the thermal period is examined separately in the Thermal Hydrologic (TH) Seepage Model (see BSC 2003 [161530]). The scope of this work is (1) to evaluate seepage rates measured during liquid-release experiments performed in several niches in the Exploratory Studies Facility (ESF) and in the Cross Drift, which was excavated for enhanced characterization of the repository block (ECRB); (2) to evaluate air-permeability data measured in boreholes above the niches and the Cross Drift to obtain the permeability structure for the seepage model; (3) to use inverse modeling to calibrate the SCM and to estimate seepage-relevant, model-related parameters on the drift scale; (4) to estimate the epistemic uncertainty of the derived parameters, based on the goodness-of-fit to the observed data and the sensitivity of calculated seepage with respect to the parameters of interest; (5) to characterize the aleatory uncertainty

  8. Strain measurement based battery testing

    Science.gov (United States)

    Xu, Jeff Qiang; Steiber, Joe; Wall, Craig M.; Smith, Robert; Ng, Cheuk

    2017-05-23

    A method and system for strain-based estimation of the state of health of a battery, from an initial state to an aged state, is provided. A strain gauge is applied to the battery. A first strain measurement is performed on the battery, using the strain gauge, at a selected charge capacity of the battery and at the initial state of the battery. A second strain measurement is performed on the battery, using the strain gauge, at the selected charge capacity of the battery and at the aged state of the battery. The capacity degradation of the battery is estimated as the difference between the first and second strain measurements divided by the first strain measurement.

  9. Experimental-based Modelling and Simulation of Water Hydraulic Mechatronics Test Facilities for Motion Control and Operation in Environmental Sensitive Applications` Areas

    DEFF Research Database (Denmark)

    Conrad, Finn; Pobedza, J.; Sobczyk, A.

    2003-01-01

    The paper presents experimental-based modelling, simulation, analysis and design of water hydraulic actuators for motion control of machines, lifts, cranes and robots. The contributions includes results from on-going research projects on fluid power and mechatronics based on tap water hydraulic...

  10. Assessment of the eye irritation potential of chemicals: A comparison study between two test methods based on human 3D hemi-cornea models.

    Science.gov (United States)

    Tandon, R; Bartok, M; Zorn-Kruppa, M; Brandner, J M; Gabel, D; Engelke, M

    2015-12-25

    We have recently developed two hemi-cornea models (Bartok et al., Toxicol in Vitro 29, 72, 2015; Zorn-Kruppa et al. PLoS One 9, e114181, 2014), which allow the correct prediction of eye irritation potential of chemicals according to the United Nations globally harmonized system of classification and labeling of chemicals (UN GHS). Both models comprise a multilayered epithelium and a stroma with embedded keratocytes in a collagenous matrix. These two models were compared, using a set of fourteen test chemicals. Their effects after 10 and 60 minutes (min) exposure were assessed from the quantification of cell viability using the MTT reduction assay. The first approach separately quantifies the damage inflicted to the epithelium and the stroma. The second approach quantifies the depth of injury by recording cell death as a function of depth. The classification obtained by the two models was compared to the Draize rabbit eye test and an ex vivo model using rabbit cornea (Jester et al. Toxicol in Vitro. 24, 597-604, 2010). With a 60 min exposure, both of our models are able to clearly differentiate UN GHS Category 1 and UN GHS Category 2 test chemicals. Copyright © 2015 Elsevier Ltd. All rights reserved.

  11. Test facility TIMO for testing the ITER model cryopump

    International Nuclear Information System (INIS)

    Haas, H.; Day, C.; Mack, A.; Methe, S.; Boissin, J.C.; Schummer, P.; Murdoch, D.K.

    2001-01-01

    Within the framework of the European Fusion Technology Programme, FZK is involved in the research and development process for a vacuum pump system of a future fusion reactor. As a result of these activities, the concept and the necessary requirements for the primary vacuum system of the ITER fusion reactor were defined. Continuing that development process, FZK has been preparing the test facility TIMO (Test facility for ITER Model pump) since 1996. This test facility provides for testing a cryopump all needed infrastructure as for example a process gas supply including a metering system, a test vessel, the cryogenic supply for the different temperature levels and a gas analysing system. For manufacturing the ITER model pump an order was given to the company L' Air Liquide in the form of a NET contract. (author)

  12. Test facility TIMO for testing the ITER model cryopump

    International Nuclear Information System (INIS)

    Haas, H.; Day, C.; Mack, A.; Methe, S.; Boissin, J.C.; Schummer, P.; Murdoch, D.K.

    1999-01-01

    Within the framework of the European Fusion Technology Programme, FZK is involved in the research and development process for a vacuum pump system of a future fusion reactor. As a result of these activities, the concept and the necessary requirements for the primary vacuum system of the ITER fusion reactor were defined. Continuing that development process, FZK has been preparing the test facility TIMO (Test facility for ITER Model pump) since 1996. This test facility provides for testing a cryopump all needed infrastructure as for example a process gas supply including a metering system, a test vessel, the cryogenic supply for the different temperature levels and a gas analysing system. For manufacturing the ITER model pump an order was given to the company L'Air Liquide in the form of a NET contract. (author)

  13. Adversarial life testing: A Bayesian negotiation model

    International Nuclear Information System (INIS)

    Rufo, M.J.; Martín, J.; Pérez, C.J.

    2014-01-01

    Life testing is a procedure intended for facilitating the process of making decisions in the context of industrial reliability. On the other hand, negotiation is a process of making joint decisions that has one of its main foundations in decision theory. A Bayesian sequential model of negotiation in the context of adversarial life testing is proposed. This model considers a general setting for which a manufacturer offers a product batch to a consumer. It is assumed that the reliability of the product is measured in terms of its lifetime. Furthermore, both the manufacturer and the consumer have to use their own information with respect to the quality of the product. Under these assumptions, two situations can be analyzed. For both of them, the main aim is to accept or reject the product batch based on the product reliability. This topic is related to a reliability demonstration problem. The procedure is applied to a class of distributions that belong to the exponential family. Thus, a unified framework addressing the main topics in the considered Bayesian model is presented. An illustrative example shows that the proposed technique can be easily applied in practice

  14. Statistical Tests for Mixed Linear Models

    CERN Document Server

    Khuri, André I; Sinha, Bimal K

    2011-01-01

    An advanced discussion of linear models with mixed or random effects. In recent years a breakthrough has occurred in our ability to draw inferences from exact and optimum tests of variance component models, generating much research activity that relies on linear models with mixed and random effects. This volume covers the most important research of the past decade as well as the latest developments in hypothesis testing. It compiles all currently available results in the area of exact and optimum tests for variance component models and offers the only comprehensive treatment for these models a

  15. Results of steel containment vessel model test

    International Nuclear Information System (INIS)

    Luk, V.K.; Ludwigsen, J.S.; Hessheimer, M.F.; Komine, Kuniaki; Matsumoto, Tomoyuki; Costello, J.F.

    1998-05-01

    A series of static overpressurization tests of scale models of nuclear containment structures is being conducted by Sandia National Laboratories for the Nuclear Power Engineering Corporation of Japan and the US Nuclear Regulatory Commission. Two tests are being conducted: (1) a test of a model of a steel containment vessel (SCV) and (2) a test of a model of a prestressed concrete containment vessel (PCCV). This paper summarizes the conduct of the high pressure pneumatic test of the SCV model and the results of that test. Results of this test are summarized and are compared with pretest predictions performed by the sponsoring organizations and others who participated in a blind pretest prediction effort. Questions raised by this comparison are identified and plans for posttest analysis are discussed

  16. lmerTest Package: Tests in Linear Mixed Effects Models

    DEFF Research Database (Denmark)

    Kuznetsova, Alexandra; Brockhoff, Per B.; Christensen, Rune Haubo Bojesen

    2017-01-01

    One of the frequent questions by users of the mixed model function lmer of the lme4 package has been: How can I get p values for the F and t tests for objects returned by lmer? The lmerTest package extends the 'lmerMod' class of the lme4 package, by overloading the anova and summary functions...... by providing p values for tests for fixed effects. We have implemented the Satterthwaite's method for approximating degrees of freedom for the t and F tests. We have also implemented the construction of Type I - III ANOVA tables. Furthermore, one may also obtain the summary as well as the anova table using...

  17. Movable scour protection. Model test report

    Energy Technology Data Exchange (ETDEWEB)

    Lorenz, R.

    2002-07-01

    This report presents the results of a series of model tests with scour protection of marine structures. The objective of the model tests is to investigate the integrity of the scour protection during a general lowering of the surrounding seabed, for instance in connection with movement of a sand bank or with general subsidence. The scour protection in the tests is made out of stone material. Two different fractions have been used: 4 mm and 40 mm. Tests with current, with waves and with combined current and waves were carried out. The scour protection material was placed after an initial scour hole has evolved in the seabed around the structure. This design philosophy has been selected because the situation often is that the scour hole starts to generate immediately after the structure has been placed. It is therefore difficult to establish a scour protection at the undisturbed seabed if the scour material is placed after the main structure. Further, placing the scour material in the scour hole increases the stability of the material. Two types of structure have been used for the test, a Monopile and a Tripod foundation. Test with protection mats around the Monopile model was also carried out. The following main conclusions have emerged form the model tests with flat bed (i.e. no general seabed lowering): 1. The maximum scour depth found in steady current on sand bed was 1.6 times the cylinder diameter, 2. The minimum horizontal extension of the scour hole (upstream direction) was 2.8 times the cylinder diameter, corresponding to a slope of 30 degrees, 3. Concrete protection mats do not meet the criteria for a strongly erodible seabed. In the present test virtually no reduction in the scour depth was obtained. The main problem is the interface to the cylinder. If there is a void between the mats and the cylinder, scour will develop. Even with the protection mats that are tightly connected to the cylinder, scour is expected to develop as long as the mats allow for

  18. Automation for a base station stability testing

    OpenAIRE

    Punnek, Elvis

    2016-01-01

    This Batchelor’s thesis was commissioned by Oy LM Ericsson Ab Oulu. The aim of it was to help to investigate and create a test automation solution for the stability testing of the LTE base station. The main objective was to create a test automation for a predefined test set. This test automation solution had to be created for specific environments and equipment. This work included creating the automation for the test cases and putting them to daily test automation jobs. The key factor...

  19. Linear Logistic Test Modeling with R

    Science.gov (United States)

    Baghaei, Purya; Kubinger, Klaus D.

    2015-01-01

    The present paper gives a general introduction to the linear logistic test model (Fischer, 1973), an extension of the Rasch model with linear constraints on item parameters, along with eRm (an R package to estimate different types of Rasch models; Mair, Hatzinger, & Mair, 2014) functions to estimate the model and interpret its parameters. The…

  20. Contribution to the improved ultrasonic testing of glass fiber-reinforced polymers based on analytic modeling; Beitrag zur Verbesserung der Ultraschallpruefung glasfaserverstaerkter Polymere auf der Grundlage analytischer Modellierung

    Energy Technology Data Exchange (ETDEWEB)

    Gripp, S.

    2001-04-01

    The non-destructive testing of acoustic anisotropic materials like fiber composites has been evaluated. Modelling enabled predictions about wave deflection, direction of wave propagation, and refractive angle of ultrasonic waves in these media. Thus, measurements could be carried out using unidirectional glass-fiber composites.

  1. A Bootstrap Cointegration Rank Test for Panels of VAR Models

    DEFF Research Database (Denmark)

    Callot, Laurent

    functions of the individual Cointegrated VARs (CVAR) models. A bootstrap based procedure is used to compute empirical distributions of the trace test statistics for these individual models. From these empirical distributions two panel trace test statistics are constructed. The satisfying small sample...

  2. MATT: Multi Agents Testing Tool Based Nets within Nets

    Directory of Open Access Journals (Sweden)

    Sara Kerraoui

    2016-12-01

    As part of this effort, we propose a model based testing approach for multi agent systems based on such a model called Reference net, where a tool, which aims to providing a uniform and automated approach is developed. The feasibility and the advantage of the proposed approach are shown through a short case study.

  3. Skull base tumor model.

    Science.gov (United States)

    Gragnaniello, Cristian; Nader, Remi; van Doormaal, Tristan; Kamel, Mahmoud; Voormolen, Eduard H J; Lasio, Giovanni; Aboud, Emad; Regli, Luca; Tulleken, Cornelius A F; Al-Mefty, Ossama

    2010-11-01

    Resident duty-hours restrictions have now been instituted in many countries worldwide. Shortened training times and increased public scrutiny of surgical competency have led to a move away from the traditional apprenticeship model of training. The development of educational models for brain anatomy is a fascinating innovation allowing neurosurgeons to train without the need to practice on real patients and it may be a solution to achieve competency within a shortened training period. The authors describe the use of Stratathane resin ST-504 polymer (SRSP), which is inserted at different intracranial locations to closely mimic meningiomas and other pathological entities of the skull base, in a cadaveric model, for use in neurosurgical training. Silicone-injected and pressurized cadaveric heads were used for studying the SRSP model. The SRSP presents unique intrinsic metamorphic characteristics: liquid at first, it expands and foams when injected into the desired area of the brain, forming a solid tumorlike structure. The authors injected SRSP via different passages that did not influence routes used for the surgical approach for resection of the simulated lesion. For example, SRSP injection routes included endonasal transsphenoidal or transoral approaches if lesions were to be removed through standard skull base approach, or, alternatively, SRSP was injected via a cranial approach if the removal was planned to be via the transsphenoidal or transoral route. The model was set in place in 3 countries (US, Italy, and The Netherlands), and a pool of 13 physicians from 4 different institutions (all surgeons and surgeons in training) participated in evaluating it and provided feedback. All 13 evaluating physicians had overall positive impressions of the model. The overall score on 9 components evaluated--including comparison between the tumor model and real tumor cases, perioperative requirements, general impression, and applicability--was 88% (100% being the best possible

  4. The Couplex test cases: models and lessons

    International Nuclear Information System (INIS)

    Bourgeat, A.; Kern, M.; Schumacher, S.; Talandier, J.

    2003-01-01

    The Couplex test cases are a set of numerical test models for nuclear waste deep geological disposal simulation. They are centered around the numerical issues arising in the near and far field transport simulation. They were used in an international contest, and are now becoming a reference in the field. We present the models used in these test cases, and show sample results from the award winning teams. (authors)

  5. Seepage Calibration Model and Seepage Testing Data

    Energy Technology Data Exchange (ETDEWEB)

    S. Finsterle

    2004-09-02

    The purpose of this Model Report is to document the Seepage Calibration Model (SCM). The SCM was developed (1) to establish the conceptual basis for the Seepage Model for Performance Assessment (SMPA), and (2) to derive seepage-relevant, model-related parameters and their distributions for use in the SMPA and seepage abstraction in support of the Total System Performance Assessment for License Application (TSPA-LA). This Model Report has been revised in response to a comprehensive, regulatory-focused evaluation performed by the Regulatory Integration Team [''Technical Work Plan for: Regulatory Integration Evaluation of Analysis and Model Reports Supporting the TSPA-LA'' (BSC 2004 [DIRS 169653])]. The SCM is intended to be used only within this Model Report for the estimation of seepage-relevant parameters through calibration of the model against seepage-rate data from liquid-release tests performed in several niches along the Exploratory Studies Facility (ESF) Main Drift and in the Cross-Drift. The SCM does not predict seepage into waste emplacement drifts under thermal or ambient conditions. Seepage predictions for waste emplacement drifts under ambient conditions will be performed with the SMPA [''Seepage Model for PA Including Drift Collapse'' (BSC 2004 [DIRS 167652])], which inherits the conceptual basis and model-related parameters from the SCM. Seepage during the thermal period is examined separately in the Thermal Hydrologic (TH) Seepage Model [see ''Drift-Scale Coupled Processes (DST and TH Seepage) Models'' (BSC 2004 [DIRS 170338])]. The scope of this work is (1) to evaluate seepage rates measured during liquid-release experiments performed in several niches in the Exploratory Studies Facility (ESF) and in the Cross-Drift, which was excavated for enhanced characterization of the repository block (ECRB); (2) to evaluate air-permeability data measured in boreholes above the niches and the Cross

  6. Seepage Calibration Model and Seepage Testing Data

    International Nuclear Information System (INIS)

    Finsterle, S.

    2004-01-01

    The purpose of this Model Report is to document the Seepage Calibration Model (SCM). The SCM was developed (1) to establish the conceptual basis for the Seepage Model for Performance Assessment (SMPA), and (2) to derive seepage-relevant, model-related parameters and their distributions for use in the SMPA and seepage abstraction in support of the Total System Performance Assessment for License Application (TSPA-LA). This Model Report has been revised in response to a comprehensive, regulatory-focused evaluation performed by the Regulatory Integration Team [''Technical Work Plan for: Regulatory Integration Evaluation of Analysis and Model Reports Supporting the TSPA-LA'' (BSC 2004 [DIRS 169653])]. The SCM is intended to be used only within this Model Report for the estimation of seepage-relevant parameters through calibration of the model against seepage-rate data from liquid-release tests performed in several niches along the Exploratory Studies Facility (ESF) Main Drift and in the Cross-Drift. The SCM does not predict seepage into waste emplacement drifts under thermal or ambient conditions. Seepage predictions for waste emplacement drifts under ambient conditions will be performed with the SMPA [''Seepage Model for PA Including Drift Collapse'' (BSC 2004 [DIRS 167652])], which inherits the conceptual basis and model-related parameters from the SCM. Seepage during the thermal period is examined separately in the Thermal Hydrologic (TH) Seepage Model [see ''Drift-Scale Coupled Processes (DST and TH Seepage) Models'' (BSC 2004 [DIRS 170338])]. The scope of this work is (1) to evaluate seepage rates measured during liquid-release experiments performed in several niches in the Exploratory Studies Facility (ESF) and in the Cross-Drift, which was excavated for enhanced characterization of the repository block (ECRB); (2) to evaluate air-permeability data measured in boreholes above the niches and the Cross-Drift to obtain the permeability structure for the seepage model

  7. Large scale injection test (LASGIT) modelling

    International Nuclear Information System (INIS)

    Arnedo, D.; Olivella, S.; Alonso, E.E.

    2010-01-01

    . In addition the horizontal interfaces between the bentonite blocks and the vertical interfaces corresponding to the host rock and the canister walls contacts are considered different materials but the properties are similar to the ones in the bentonite. Actually, this is done because interfaces are believed to be a potential preferential path for gas migration through the buffer. The host rock and the canister are not included in the model due to its high stiffness with respect to bentonite. A constitutive model that considers non-linear elasticity and visco-plasticity based on BBM model is adopted for the bentonite and the interfaces. An embedded fracture permeability model in which permeability and retention curve depend on strains through a fracture aperture is considered in the hydraulic problem. The following stages of the experiment are simulated: - Construction of the isolation barrier inside the deposition hole, - Hydration stage 1 in which the liquid pressure is increased at the Filter Mats and de canister Injection Filters up to 1.5 MPa in an initial stage and up to 2.35 MPa in a second stage in order to saturate the buffer, - Hydraulic test 1, - Gas injection test 1. The simulation results will be compared to the experimental record of different variables: total stresses and liquid pressure at the rock wall, at the canister wall, and at some points within the bentonite buffer. It will be interesting to try to explain if preferential paths develop and where. This is controlled by the swelling capacity of the buffer and its ability to seal the interfaces initially not closed between the clay and the wall and between the blocks. (authors)

  8. Model tests for prestressed concrete pressure vessels

    International Nuclear Information System (INIS)

    Stoever, R.

    1975-01-01

    Investigations with models of reactor pressure vessels are used to check results of three dimensional calculation methods and to predict the behaviour of the prototype. Model tests with 1:50 elastic pressure vessel models and with a 1:5 prestressed concrete pressure vessel are described and experimental results are presented. (orig.) [de

  9. A direct test of the Weinberg-Salam model based on the investigation of p-odd asymmetries in the processes l-++N → l-++X

    International Nuclear Information System (INIS)

    Bilen'ky, S.M.; Motz, G.

    1978-01-01

    In the framework of the Weinberg-Salam model the p-odd asymmetries Asub(+-) of the processes l -+ +N → l -+ +X with polarized leptons are shown to be related to the inclusive cross sections νsub(μ)(anti vsub(μ))+N → μ - (μ + )+X on an isoscalar target. It is shown that measurements of both Asub(-) and Asub(+) would permit one to test the Weinberg-Salam theory without dynamical assumptions

  10. A Dutch test with the NewProd-model

    NARCIS (Netherlands)

    Bronnenberg, J.J.A.M.; van Engelen, M.L.

    1988-01-01

    The paper contains a report of a test of Cooper's NewProd model for predicting success and failure of product development projects. Based on Canadian data, the model has been shown to make predictions which are 84% correct. Having reservations on the reliability and validity of the model on

  11. Pion interferometric tests of transport models

    Energy Technology Data Exchange (ETDEWEB)

    Padula, S.S.; Gyulassy, M.; Gavin, S. (Lawrence Berkeley Lab., CA (USA). Nuclear Science Div.)

    1990-01-08

    In hadronic reactions, the usual space-time interpretation of pion interferometry often breaks down due to strong correlations between spatial and momentum coordinates. We derive a general interferometry formula based on the Wigner density formalism that allows for arbitrary phase space and multiparticle correlations. Correction terms due to intermediate state pion cascading are derived using semiclassical hadronic transport theory. Finite wave packets are used to reveal the sensitivity of pion interference effects on the details of the production dynamics. The covariant generalization of the formula is shown to be equivalent to the formula derived via an alternate current ensemble formalism for minimal wave packets and reduces in the nonrelativistic limit to a formula derived by Pratt. The final expression is ideally suited for pion interferometric tests of Monte Carlo transport models. Examples involving gaussian and inside-outside phase space distributions are considered. (orig.).

  12. Pion interferometric tests of transport models

    International Nuclear Information System (INIS)

    Padula, S.S.; Gyulassy, M.; Gavin, S.

    1990-01-01

    In hadronic reactions, the usual space-time interpretation of pion interferometry often breaks down due to strong correlations between spatial and momentum coordinates. We derive a general interferometry formula based on the Wigner density formalism that allows for arbitrary phase space and multiparticle correlations. Correction terms due to intermediate state pion cascading are derived using semiclassical hadronic transport theory. Finite wave packets are used to reveal the sensitivity of pion interference effects on the details of the production dynamics. The covariant generalization of the formula is shown to be equivalent to the formula derived via an alternate current ensemble formalism for minimal wave packets and reduces in the nonrelativistic limit to a formula derived by Pratt. The final expression is ideally suited for pion interferometric tests of Monte Carlo transport models. Examples involving gaussian and inside-outside phase space distributions are considered. (orig.)

  13. A Monte Carlo-based method to estimate radiation dose from spiral CT: from phantom testing to patient-specific models

    International Nuclear Information System (INIS)

    Jarry, G; De Marco, J J; Beifuss, U; Cagnon, C H; McNitt-Gray, M F

    2003-01-01

    The purpose of this work is to develop and test a method to estimate the relative and absolute absorbed radiation dose from axial and spiral CT scans using a Monte Carlo approach. Initial testing was done in phantoms and preliminary results were obtained from a standard mathematical anthropomorphic model (MIRD V) and voxelized patient data. To accomplish this we have modified a general purpose Monte Carlo transport code (MCNP4B) to simulate the CT x-ray source and movement, and then to calculate absorbed radiation dose in desired objects. The movement of the source in either axial or spiral modes was modelled explicitly while the CT system components were modelled using published information about x-ray spectra as well as information provided by the manufacturer. Simulations were performed for single axial scans using the head and body computed tomography dose index (CTDI) polymethylmethacrylate phantoms at both central and peripheral positions for all available beam energies and slice thicknesses. For comparison, corresponding physical measurements of CTDI in phantom were made with an ion chamber. To obtain absolute dose values, simulations and measurements were performed in air at the scanner isocentre for each beam energy. To extend the verification, the CT scanner model was applied to the MIRD V model and compared with published results using similar technical factors. After verification of the model, the generalized source was simulated and applied to voxelized models of patient anatomy. The simulated and measured absolute dose data in phantom agreed to within 2% for the head phantom and within 4% for the body phantom at 120 and 140 kVp; this extends to 8% for the head and 9% for the body phantom across all available beam energies and positions. For the head phantom, the simulated and measured absolute dose data agree to within 2% across all slice thicknesses at 120 kVp. Our results in the MIRD phantom agree within 11% of all the different organ dose values

  14. Experiments towards model-based testing using Plan 9: Labelled transition file systems, stacking file systems, on-the-fly coverage measuring

    NARCIS (Netherlands)

    Belinfante, Axel; Guardiola, G.; Soriano, E.; Ballesteros, F.J.

    2006-01-01

    We report on experiments that we did on Plan 9/Inferno to gain more experience with the file-system-as-tool-interface approach. We reimplemented functionality that we earlier worked on in Unix, trying to use Plan 9 file system interfaces. The application domain for those experiments was model-based

  15. Simulating political stability and change in the Netherlands (1998-2002). An agent-based model of party competition with media effects empirically tested.

    NARCIS (Netherlands)

    Muis, J.C.

    2010-01-01

    Agent-based models of political party competition in a multidimensional policy space have been developed in order to reflect adaptive learning by party leaders with very limited information feedback. The key assumption is that two categories of actors continually make decisions: voters choose which

  16. Mixed Portmanteau Test for Diagnostic Checking of Time Series Models

    Directory of Open Access Journals (Sweden)

    Sohail Chand

    2014-01-01

    Full Text Available Model criticism is an important stage of model building and thus goodness of fit tests provides a set of tools for diagnostic checking of the fitted model. Several tests are suggested in literature for diagnostic checking. These tests use autocorrelation or partial autocorrelation in the residuals to criticize the adequacy of fitted model. The main idea underlying these portmanteau tests is to identify if there is any dependence structure which is yet unexplained by the fitted model. In this paper, we suggest mixed portmanteau tests based on autocorrelation and partial autocorrelation functions of the residuals. We derived the asymptotic distribution of the mixture test and studied its size and power using Monte Carlo simulations.

  17. Conformance test development with the Java modeling language

    DEFF Research Database (Denmark)

    Søndergaard, Hans; Korsholm, Stephan E.; Ravn, Anders P.

    2017-01-01

    In order to claim conformance with a Java Specification Request, a Java implementation has to pass all tests in an associated Technology Compatibility Kit (TCK). This paper presents a model-based development of a TCK test suite and a test execution tool for the draft Safety-Critical Java (SCJ......) profile specification. The Java Modeling Language (JML) is used to model conformance constraints for the profile. JML annotations define contracts for classes and interfaces. The annotations are translated by a tool into runtime assertion checks.Hereby the design and elaboration of the concrete test cases...

  18. Kernel-based tests for joint independence

    DEFF Research Database (Denmark)

    Pfister, Niklas; Bühlmann, Peter; Schölkopf, Bernhard

    2018-01-01

    if the $d$ variables are jointly independent, as long as the kernel is characteristic. Based on an empirical estimate of dHSIC, we define three different non-parametric hypothesis tests: a permutation test, a bootstrap test and a test based on a Gamma approximation. We prove that the permutation test......We investigate the problem of testing whether $d$ random variables, which may or may not be continuous, are jointly (or mutually) independent. Our method builds on ideas of the two variable Hilbert-Schmidt independence criterion (HSIC) but allows for an arbitrary number of variables. We embed...... the $d$-dimensional joint distribution and the product of the marginals into a reproducing kernel Hilbert space and define the $d$-variable Hilbert-Schmidt independence criterion (dHSIC) as the squared distance between the embeddings. In the population case, the value of dHSIC is zero if and only...

  19. Validity evidence based on test content.

    Science.gov (United States)

    Sireci, Stephen; Faulkner-Bond, Molly

    2014-01-01

    Validity evidence based on test content is one of the five forms of validity evidence stipulated in the Standards for Educational and Psychological Testing developed by the American Educational Research Association, American Psychological Association, and National Council on Measurement in Education. In this paper, we describe the logic and theory underlying such evidence and describe traditional and modern methods for gathering and analyzing content validity data. A comprehensive review of the literature and of the aforementioned Standards is presented. For educational tests and other assessments targeting knowledge and skill possessed by examinees, validity evidence based on test content is necessary for building a validity argument to support the use of a test for a particular purpose. By following the methods described in this article, practitioners have a wide arsenal of tools available for determining how well the content of an assessment is congruent with and appropriate for the specific testing purposes.

  20. 1/3-scale model testing program

    International Nuclear Information System (INIS)

    Yoshimura, H.R.; Attaway, S.W.; Bronowski, D.R.; Uncapher, W.L.; Huerta, M.; Abbott, D.G.

    1989-01-01

    This paper describes the drop testing of a one-third scale model transport cask system. Two casks were supplied by Transnuclear, Inc. (TN) to demonstrate dual purpose shipping/storage casks. These casks will be used to ship spent fuel from DOEs West Valley demonstration project in New York to the Idaho National Engineering Laboratory (INEL) for long term spent fuel dry storage demonstration. As part of the certification process, one-third scale model tests were performed to obtain experimental data. Two 9-m (30-ft) drop tests were conducted on a mass model of the cask body and scaled balsa and redwood filled impact limiters. In the first test, the cask system was tested in an end-on configuration. In the second test, the system was tested in a slap-down configuration where the axis of the cask was oriented at a 10 degree angle with the horizontal. Slap-down occurs for shallow angle drops where the primary impact at one end of the cask is followed by a secondary impact at the other end. The objectives of the testing program were to (1) obtain deceleration and displacement information for the cask and impact limiter system, (2) obtain dynamic force-displacement data for the impact limiters, (3) verify the integrity of the impact limiter retention system, and (4) examine the crush behavior of the limiters. This paper describes both test results in terms of measured deceleration, post test deformation measurements, and the general structural response of the system

  1. Superconducting solenoid model magnet test results

    Energy Technology Data Exchange (ETDEWEB)

    Carcagno, R.; Dimarco, J.; Feher, S.; Ginsburg, C.M.; Hess, C.; Kashikhin, V.V.; Orris, D.F.; Pischalnikov, Y.; Sylvester, C.; Tartaglia, M.A.; Terechkine, I.; /Fermilab

    2006-08-01

    Superconducting solenoid magnets suitable for the room temperature front end of the Fermilab High Intensity Neutrino Source (formerly known as Proton Driver), an 8 GeV superconducting H- linac, have been designed and fabricated at Fermilab, and tested in the Fermilab Magnet Test Facility. We report here results of studies on the first model magnets in this program, including the mechanical properties during fabrication and testing in liquid helium at 4.2 K, quench performance, and magnetic field measurements. We also describe new test facility systems and instrumentation that have been developed to accomplish these tests.

  2. Superconducting solenoid model magnet test results

    International Nuclear Information System (INIS)

    Carcagno, R.; Dimarco, J.; Feher, S.; Ginsburg, C.M.; Hess, C.; Kashikhin, V.V.; Orris, D.F.; Pischalnikov, Y.; Sylvester, C.; Tartaglia, M.A.; Terechkine, I.; Tompkins, J.C.; Wokas, T.; Fermilab

    2006-01-01

    Superconducting solenoid magnets suitable for the room temperature front end of the Fermilab High Intensity Neutrino Source (formerly known as Proton Driver), an 8 GeV superconducting H- linac, have been designed and fabricated at Fermilab, and tested in the Fermilab Magnet Test Facility. We report here results of studies on the first model magnets in this program, including the mechanical properties during fabrication and testing in liquid helium at 4.2 K, quench performance, and magnetic field measurements. We also describe new test facility systems and instrumentation that have been developed to accomplish these tests

  3. Methodology for testing and validating knowledge bases

    Science.gov (United States)

    Krishnamurthy, C.; Padalkar, S.; Sztipanovits, J.; Purves, B. R.

    1987-01-01

    A test and validation toolset developed for artificial intelligence programs is described. The basic premises of this method are: (1) knowledge bases have a strongly declarative character and represent mostly structural information about different domains, (2) the conditions for integrity, consistency, and correctness can be transformed into structural properties of knowledge bases, and (3) structural information and structural properties can be uniformly represented by graphs and checked by graph algorithms. The interactive test and validation environment have been implemented on a SUN workstation.

  4. Space Launch System Scale Model Acoustic Test Ignition Overpressure Testing

    Science.gov (United States)

    Nance, Donald; Liever, Peter; Nielsen, Tanner

    2015-01-01

    The overpressure phenomenon is a transient fluid dynamic event occurring during rocket propulsion system ignition. This phenomenon results from fluid compression of the accelerating plume gas, subsequent rarefaction, and subsequent propagation from the exhaust trench and duct holes. The high-amplitude unsteady fluid-dynamic perturbations can adversely affect the vehicle and surrounding structure. Commonly known as ignition overpressure (IOP), this is an important design-to environment for the Space Launch System (SLS) that NASA is currently developing. Subscale testing is useful in validating and verifying the IOP environment. This was one of the objectives of the Scale Model Acoustic Test, conducted at Marshall Space Flight Center. The test data quantifies the effectiveness of the SLS IOP suppression system and improves the analytical models used to predict the SLS IOP environments. The reduction and analysis of the data gathered during the SMAT IOP test series requires identification and characterization of multiple dynamic events and scaling of the event waveforms to provide the most accurate comparisons to determine the effectiveness of the IOP suppression systems. The identification and characterization of the overpressure events, the waveform scaling, the computation of the IOP suppression system knockdown factors, and preliminary comparisons to the analytical models are discussed.

  5. Space Launch System Scale Model Acoustic Test Ignition Overpressure Testing

    Science.gov (United States)

    Nance, Donald K.; Liever, Peter A.

    2015-01-01

    The overpressure phenomenon is a transient fluid dynamic event occurring during rocket propulsion system ignition. This phenomenon results from fluid compression of the accelerating plume gas, subsequent rarefaction, and subsequent propagation from the exhaust trench and duct holes. The high-amplitude unsteady fluid-dynamic perturbations can adversely affect the vehicle and surrounding structure. Commonly known as ignition overpressure (IOP), this is an important design-to environment for the Space Launch System (SLS) that NASA is currently developing. Subscale testing is useful in validating and verifying the IOP environment. This was one of the objectives of the Scale Model Acoustic Test (SMAT), conducted at Marshall Space Flight Center (MSFC). The test data quantifies the effectiveness of the SLS IOP suppression system and improves the analytical models used to predict the SLS IOP environments. The reduction and analysis of the data gathered during the SMAT IOP test series requires identification and characterization of multiple dynamic events and scaling of the event waveforms to provide the most accurate comparisons to determine the effectiveness of the IOP suppression systems. The identification and characterization of the overpressure events, the waveform scaling, the computation of the IOP suppression system knockdown factors, and preliminary comparisons to the analytical models are discussed.

  6. Validation of a CFD model simulating charge and discharge of a small heat storage test module based on a sodium acetate water mixture

    DEFF Research Database (Denmark)

    Dannemand, Mark; Fan, Jianhua; Furbo, Simon

    2014-01-01

    Experimental and theoretical investigations are carried out to study the heating of a 302 x 302 x 55 mm test box of steel containing a sodium acetate water mixture. A thermostatic bath has been set up to control the charging and discharging of the steel box. The charging and discharging has been...... for a Computational Fluid Dynamics (CFD) model. The CFD calculated temperatures are compared to measured temperatures internally in the box to validate the CFD model. Four cases are investigated; heating the test module with the sodium acetate water mixture in solid phase from ambient temperature to 52˚C; heating...... the module starting with the salt water mixture in liquid phase from 72˚C to 95˚C; heating up the module from ambient temperature with the salt water mixture in solid phase, going through melting, ending in liquid phase at 78˚C/82˚C; and discharging the test module from liquid phase at 82˚C, going through...

  7. 基于SPSS的试卷质量分析与建模%Analysis and Modeling of Test Paper Quality Base on SPSS

    Institute of Scientific and Technical Information of China (English)

    敖希琴; 蒋静; 曹莉

    2016-01-01

    试卷质量分析是考察试卷能否客观准确的检测出学生对知识的掌握程度。为此借助SPSS对试卷考核结果进行各类基本统计描述分析,将难度、区分度、信度以及效度作为试卷综合质量评价的指标体系,并通过调研确定各个指标间重要程度系数,然后采用古林法确定各个指标的权重,最终建立一个计算试卷综合质量系数的通用数学模型,从而为科学编制试题提供依据。%The analysis of test paper quality is investigated for detection of the test paper, which is able to detect students' mastery of knowledge objectively and accurately. For this purpose, we use SPSS to analyze the results of the test papers. On the basis of summarizing the existing research, the difficulty, discrimination, reliability and validity are used as the index system of the comprehensive quality evaluation of the test paper, and the important degree coefficient of each index is determined through the investigation. Then the weight of each index is determined by the A.I.Klee method. Finally, a general mathematical model is established to calculate the comprehensive quality coefficient of the test paper, the purpose is to provide the basis for the scientific establishment of the test questions.

  8. Test cell data-based predictive modelling to determine HVAC energy consumption for three façade solutions in Madrid

    Directory of Open Access Journals (Sweden)

    J. Guerrero-Rubio

    2018-01-01

    Full Text Available This study aims to narrow the gap between predicted and actual energy performance in buildings. Predictive models were established that relate the electric consumption by HVAC systems to maintain certain indoor environmental conditions in variable weather to the type of façade. The models were developed using data gathered from test cells with adiabatic envelopes on all but the façade to be tested. Three façade types were studied. The first, the standard solution, consisted in a double wythe brick wall with an intermediate air space, the configuration most commonly deployed in multi-family dwellings built in Spain between 1940 and 1980 (prior to the enactment of the first building codes that limited overall energy demand in buildings. The other two were retrofits frequently found in such buildings: ventilated façades and ETICS (external thermal insulation composite systems. Two predictive models were designed for each type of façade, one for summer and the other for winter. The linear regression equations and the main statistical parameters are reported.

  9. Is the standard model really tested?

    International Nuclear Information System (INIS)

    Takasugi, E.

    1989-01-01

    It is discussed how the standard model is really tested. Among various tests, I concentrate on the CP violation phenomena in K and B meson system. Especially, the resent hope to overcome the theoretical uncertainty in the evaluation on the CP violation of K meson system is discussed. (author)

  10. Geometrical error calibration in reflective surface testing based on reverse Hartmann test

    Science.gov (United States)

    Gong, Zhidong; Wang, Daodang; Xu, Ping; Wang, Chao; Liang, Rongguang; Kong, Ming; Zhao, Jun; Mo, Linhai; Mo, Shuhui

    2017-08-01

    In the fringe-illumination deflectometry based on reverse-Hartmann-test configuration, ray tracing of the modeled testing system is performed to reconstruct the test surface error. Careful calibration of system geometry is required to achieve high testing accuracy. To realize the high-precision surface testing with reverse Hartmann test, a computer-aided geometrical error calibration method is proposed. The aberrations corresponding to various geometrical errors are studied. With the aberration weights for various geometrical errors, the computer-aided optimization of system geometry with iterative ray tracing is carried out to calibration the geometrical error, and the accuracy in the order of subnanometer is achieved.

  11. PV panel model based on datasheet values

    DEFF Research Database (Denmark)

    Sera, Dezso; Teodorescu, Remus; Rodriguez, Pedro

    2007-01-01

    This work presents the construction of a model for a PV panel using the single-diode five-parameters model, based exclusively on data-sheet parameters. The model takes into account the series and parallel (shunt) resistance of the panel. The equivalent circuit and the basic equations of the PV cell....... Based on these equations, a PV panel model, which is able to predict the panel behavior in different temperature and irradiance conditions, is built and tested....

  12. Remote sensing and GIS-based landslide hazard analysis and cross-validation using multivariate logistic regression model on three test areas in Malaysia

    Science.gov (United States)

    Pradhan, Biswajeet

    2010-05-01

    This paper presents the results of the cross-validation of a multivariate logistic regression model using remote sensing data and GIS for landslide hazard analysis on the Penang, Cameron, and Selangor areas in Malaysia. Landslide locations in the study areas were identified by interpreting aerial photographs and satellite images, supported by field surveys. SPOT 5 and Landsat TM satellite imagery were used to map landcover and vegetation index, respectively. Maps of topography, soil type, lineaments and land cover were constructed from the spatial datasets. Ten factors which influence landslide occurrence, i.e., slope, aspect, curvature, distance from drainage, lithology, distance from lineaments, soil type, landcover, rainfall precipitation, and normalized difference vegetation index (ndvi), were extracted from the spatial database and the logistic regression coefficient of each factor was computed. Then the landslide hazard was analysed using the multivariate logistic regression coefficients derived not only from the data for the respective area but also using the logistic regression coefficients calculated from each of the other two areas (nine hazard maps in all) as a cross-validation of the model. For verification of the model, the results of the analyses were then compared with the field-verified landslide locations. Among the three cases of the application of logistic regression coefficient in the same study area, the case of Selangor based on the Selangor logistic regression coefficients showed the highest accuracy (94%), where as Penang based on the Penang coefficients showed the lowest accuracy (86%). Similarly, among the six cases from the cross application of logistic regression coefficient in other two areas, the case of Selangor based on logistic coefficient of Cameron showed highest (90%) prediction accuracy where as the case of Penang based on the Selangor logistic regression coefficients showed the lowest accuracy (79%). Qualitatively, the cross

  13. On selection of optimal stochastic model for accelerated life testing

    International Nuclear Information System (INIS)

    Volf, P.; Timková, J.

    2014-01-01

    This paper deals with the problem of proper lifetime model selection in the context of statistical reliability analysis. Namely, we consider regression models describing the dependence of failure intensities on a covariate, for instance, a stressor. Testing the model fit is standardly based on the so-called martingale residuals. Their analysis has already been studied by many authors. Nevertheless, the Bayes approach to the problem, in spite of its advantages, is just developing. We shall present the Bayes procedure of estimation in several semi-parametric regression models of failure intensity. Then, our main concern is the Bayes construction of residual processes and goodness-of-fit tests based on them. The method is illustrated with both artificial and real-data examples. - Highlights: • Statistical survival and reliability analysis and Bayes approach. • Bayes semi-parametric regression modeling in Cox's and AFT models. • Bayes version of martingale residuals and goodness-of-fit test

  14. Canine distemper virus (CDV) infection of ferrets as a model for testing Morbillivirus vaccine strategies: NYVAC- and ALVAC-based CDV recombinants protect against symptomatic infection.

    Science.gov (United States)

    Stephensen, C B; Welter, J; Thaker, S R; Taylor, J; Tartaglia, J; Paoletti, E

    1997-02-01

    Canine distemper virus (CDV) infection of ferrets causes an acute systemic disease involving multiple organ systems, including the respiratory tract, lymphoid system, and central nervous system (CNS). We have tested candidate CDV vaccines incorporating the fusion (F) and hemagglutinin (HA) proteins in the highly attenuated NYVAC strain of vaccinia virus and in the ALVAC strain of canarypox virus, which does not productively replicate in mammalian hosts. Juvenile ferrets were vaccinated twice with these constructs, or with an attenuated live-virus vaccine, while controls received saline or the NYVAC and ALVAC vectors expressing rabies virus glycoprotein. Control animals did not develop neutralizing antibody and succumbed to distemper after developing fever, weight loss, leukocytopenia, decreased activity, conjunctivitis, an erythematous rash typical of distemper, CNS signs, and viremia in peripheral blood mononuclear cells (as measured by reverse transcription-PCR). All three CDV vaccines elicited neutralizing titers of at least 1:96. All vaccinated ferrets survived, and none developed viremia. Both recombinant vaccines also protected against the development of symptomatic distemper. However, ferrets receiving the live-virus vaccine lost weight, became lymphocytopenic, and developed the erythematous rash typical of CDV. These data show that ferrets are an excellent model for evaluating the ability of CDV vaccines to protect against symptomatic infection. Because the pathogenesis and clinical course of CDV infection of ferrets is quite similar to that of other Morbillivirus infections, including measles, this model will be useful in testing new candidate Morbillivirus vaccines.

  15. Development of dynamic Bayesian models for web application test management

    Science.gov (United States)

    Azarnova, T. V.; Polukhin, P. V.; Bondarenko, Yu V.; Kashirina, I. L.

    2018-03-01

    The mathematical apparatus of dynamic Bayesian networks is an effective and technically proven tool that can be used to model complex stochastic dynamic processes. According to the results of the research, mathematical models and methods of dynamic Bayesian networks provide a high coverage of stochastic tasks associated with error testing in multiuser software products operated in a dynamically changing environment. Formalized representation of the discrete test process as a dynamic Bayesian model allows us to organize the logical connection between individual test assets for multiple time slices. This approach gives an opportunity to present testing as a discrete process with set structural components responsible for the generation of test assets. Dynamic Bayesian network-based models allow us to combine in one management area individual units and testing components with different functionalities and a direct influence on each other in the process of comprehensive testing of various groups of computer bugs. The application of the proposed models provides an opportunity to use a consistent approach to formalize test principles and procedures, methods used to treat situational error signs, and methods used to produce analytical conclusions based on test results.

  16. Model based design introduction: modeling game controllers to microprocessor architectures

    Science.gov (United States)

    Jungwirth, Patrick; Badawy, Abdel-Hameed

    2017-04-01

    We present an introduction to model based design. Model based design is a visual representation, generally a block diagram, to model and incrementally develop a complex system. Model based design is a commonly used design methodology for digital signal processing, control systems, and embedded systems. Model based design's philosophy is: to solve a problem - a step at a time. The approach can be compared to a series of steps to converge to a solution. A block diagram simulation tool allows a design to be simulated with real world measurement data. For example, if an analog control system is being upgraded to a digital control system, the analog sensor input signals can be recorded. The digital control algorithm can be simulated with the real world sensor data. The output from the simulated digital control system can then be compared to the old analog based control system. Model based design can compared to Agile software develop. The Agile software development goal is to develop working software in incremental steps. Progress is measured in completed and tested code units. Progress is measured in model based design by completed and tested blocks. We present a concept for a video game controller and then use model based design to iterate the design towards a working system. We will also describe a model based design effort to develop an OS Friendly Microprocessor Architecture based on the RISC-V.

  17. Kinematic tests of exotic flat cosmological models

    International Nuclear Information System (INIS)

    Charlton, J.C.; Turner, M.S.; NASA/Fermilab Astrophysics Center, Batavia, IL)

    1987-01-01

    Theoretical prejudice and inflationary models of the very early universe strongly favor the flat, Einstein-de Sitter model of the universe. At present the observational data conflict with this prejudice. This conflict can be resolved by considering flat models of the universe which posses a smooth component of energy density. The kinematics of such models, where the smooth component is relativistic particles, a cosmological term, a network of light strings, or fast-moving, light strings is studied in detail. The observational tests which can be used to discriminate between these models are also discussed. These tests include the magnitude-redshift, lookback time-redshift, angular size-redshift, and comoving volume-redshift diagrams and the growth of density fluctuations. 58 references

  18. Kinematic tests of exotic flat cosmological models

    International Nuclear Information System (INIS)

    Charlton, J.C.; Turner, M.S.

    1986-05-01

    Theoretical prejudice and inflationary models of the very early Universe strongly favor the flat, Einstein-deSitter model of the Universe. At present the observational data conflict with this prejudice. This conflict can be resolved by considering flat models of the Universe which possess a smooth component by energy density. We study in detail the kinematics of such models, where the smooth component is relativistic particles, a cosmological term, a network of light strings, or fast-moving, light strings. We also discuss the observational tests which can be used to discriminate between these models. These tests include the magnitude-redshift, lookback time-redshift, angular size-redshift, and comoving volume-redshift diagrams and the growth of density fluctuations

  19. Kinematic tests of exotic flat cosmological models

    Energy Technology Data Exchange (ETDEWEB)

    Charlton, J.C.; Turner, M.S.

    1986-05-01

    Theoretical prejudice and inflationary models of the very early Universe strongly favor the flat, Einstein-deSitter model of the Universe. At present the observational data conflict with this prejudice. This conflict can be resolved by considering flat models of the Universe which possess a smooth component by energy density. We study in detail the kinematics of such models, where the smooth component is relativistic particles, a cosmological term, a network of light strings, or fast-moving, light strings. We also discuss the observational tests which can be used to discriminate between these models. These tests include the magnitude-redshift, lookback time-redshift, angular size-redshift, and comoving volume-redshift diagrams and the growth of density fluctuations.

  20. An Empirical Test of a Model of Resistance to Persuasion.

    Science.gov (United States)

    And Others; Burgoon, Michael

    1978-01-01

    Tests a model of resistance to persuasion based upon variables not considered by earlier congruity and inoculation models. Supports the prediction that the kind of critical response set induced and the target of the criticism are mediators of resistance to persuasion. (JMF)

  1. Engineering Abstractions in Model Checking and Testing

    DEFF Research Database (Denmark)

    Achenbach, Michael; Ostermann, Klaus

    2009-01-01

    Abstractions are used in model checking to tackle problems like state space explosion or modeling of IO. The application of these abstractions in real software development processes, however, lacks engineering support. This is one reason why model checking is not widely used in practice yet...... and testing is still state of the art in falsification. We show how user-defined abstractions can be integrated into a Java PathFinder setting with tools like AspectJ or Javassist and discuss implications of remaining weaknesses of these tools. We believe that a principled engineering approach to designing...... and implementing abstractions will improve the applicability of model checking in practice....

  2. Diagnostic tests based on human basophils

    DEFF Research Database (Denmark)

    Kleine-Tebbe, Jörg; Erdmann, Stephan; Knol, Edward F

    2006-01-01

    -maximal responses, termed 'intrinsic sensitivity'. These variables give rise to shifts in the dose-response curves which, in a diagnostic setting where only a single antigen concentration is employed, may produce false-negative data. Thus, in order to meaningfully utilize the current basophil activation tests....... Diagnostic studies using CD63 or CD203c in hymenoptera, food and drug allergy are critically discussed. Basophil-based tests are indicated for allergy testing in selected cases but should only be performed by experienced laboratories....

  3. TREAT (TREe-based Association Test)

    Science.gov (United States)

    TREAT is an R package for detecting complex joint effects in case-control studies. The test statistic is derived from a tree-structure model by recursive partitioning the data. Ultra-fast algorithm is designed to evaluate the significance of association between candidate gene and disease outcome

  4. Conditional Monte Carlo randomization tests for regression models.

    Science.gov (United States)

    Parhat, Parwen; Rosenberger, William F; Diao, Guoqing

    2014-08-15

    We discuss the computation of randomization tests for clinical trials of two treatments when the primary outcome is based on a regression model. We begin by revisiting the seminal paper of Gail, Tan, and Piantadosi (1988), and then describe a method based on Monte Carlo generation of randomization sequences. The tests based on this Monte Carlo procedure are design based, in that they incorporate the particular randomization procedure used. We discuss permuted block designs, complete randomization, and biased coin designs. We also use a new technique by Plamadeala and Rosenberger (2012) for simple computation of conditional randomization tests. Like Gail, Tan, and Piantadosi, we focus on residuals from generalized linear models and martingale residuals from survival models. Such techniques do not apply to longitudinal data analysis, and we introduce a method for computation of randomization tests based on the predicted rate of change from a generalized linear mixed model when outcomes are longitudinal. We show, by simulation, that these randomization tests preserve the size and power well under model misspecification. Copyright © 2014 John Wiley & Sons, Ltd.

  5. Unit testing, model validation, and biological simulation.

    Science.gov (United States)

    Sarma, Gopal P; Jacobs, Travis W; Watts, Mark D; Ghayoomie, S Vahid; Larson, Stephen D; Gerkin, Richard C

    2016-01-01

    The growth of the software industry has gone hand in hand with the development of tools and cultural practices for ensuring the reliability of complex pieces of software. These tools and practices are now acknowledged to be essential to the management of modern software. As computational models and methods have become increasingly common in the biological sciences, it is important to examine how these practices can accelerate biological software development and improve research quality. In this article, we give a focused case study of our experience with the practices of unit testing and test-driven development in OpenWorm, an open-science project aimed at modeling Caenorhabditis elegans. We identify and discuss the challenges of incorporating test-driven development into a heterogeneous, data-driven project, as well as the role of model validation tests, a category of tests unique to software which expresses scientific models.

  6. Variable amplitude fatigue, modelling and testing

    International Nuclear Information System (INIS)

    Svensson, Thomas.

    1993-01-01

    Problems related to metal fatigue modelling and testing are here treated in four different papers. In the first paper different views of the subject are summarised in a literature survey. In the second paper a new model for fatigue life is investigated. Experimental results are established which are promising for further development of the mode. In the third paper a method is presented that generates a stochastic process, suitable to fatigue testing. The process is designed in order to resemble certain fatigue related features in service life processes. In the fourth paper fatigue problems in transport vibrations are treated

  7. Design, modeling and testing of data converters

    CERN Document Server

    Kiaei, Sayfe; Xu, Fang

    2014-01-01

    This book presents the a scientific discussion of the state-of-the-art techniques and designs for modeling, testing and for the performance analysis of data converters. The focus is put on sustainable data conversion. Sustainability has become a public issue that industries and users can not ignore. Devising environmentally friendly solutions for data conversion designing, modeling and testing is nowadays a requirement that researchers and practitioners must consider in their activities. This book presents the outcome of the IWADC workshop 2011, held in Orvieto, Italy.

  8. Allele-sharing models: LOD scores and accurate linkage tests.

    Science.gov (United States)

    Kong, A; Cox, N J

    1997-11-01

    Starting with a test statistic for linkage analysis based on allele sharing, we propose an associated one-parameter model. Under general missing-data patterns, this model allows exact calculation of likelihood ratios and LOD scores and has been implemented by a simple modification of existing software. Most important, accurate linkage tests can be performed. Using an example, we show that some previously suggested approaches to handling less than perfectly informative data can be unacceptably conservative. Situations in which this model may not perform well are discussed, and an alternative model that requires additional computations is suggested.

  9. Flight Test Maneuvers for Efficient Aerodynamic Modeling

    Science.gov (United States)

    Morelli, Eugene A.

    2011-01-01

    Novel flight test maneuvers for efficient aerodynamic modeling were developed and demonstrated in flight. Orthogonal optimized multi-sine inputs were applied to aircraft control surfaces to excite aircraft dynamic response in all six degrees of freedom simultaneously while keeping the aircraft close to chosen reference flight conditions. Each maneuver was designed for a specific modeling task that cannot be adequately or efficiently accomplished using conventional flight test maneuvers. All of the new maneuvers were first described and explained, then demonstrated on a subscale jet transport aircraft in flight. Real-time and post-flight modeling results obtained using equation-error parameter estimation in the frequency domain were used to show the effectiveness and efficiency of the new maneuvers, as well as the quality of the aerodynamic models that can be identified from the resultant flight data.

  10. Using Virtual ATE Model to Migrate Test Programs

    Institute of Scientific and Technical Information of China (English)

    王晓明; 杨乔林

    1995-01-01

    Bacause of high development costs of IC (Integrated Circuit)test programs,recycling existing test programs from one kind of ATE (Automatic Test Equipment) to another or generating directly from CAD simulation modules to ATE is more and more valuable.In this paper,a new approach to migrating test programs is presented.A virtual ATE model based on object-oriented paradigm is developed;it runs Test C++ (an intermediate test control language) programs and TeIF(Test Inftermediate Format-an intermediate pattern),migrates test programs among three kinds of ATE (Ando DIC8032,Schlumberger S15 and GenRad 1732) and generates test patterns from two kinds of CAD 9Daisy and Panda) automatically.

  11. Xenobiotic metabolism capacities of human skin in comparison with a 3D-epidermis model and keratinocyte-based cell culture as in vitro alternatives for chemical testing: phase II enzymes.

    Science.gov (United States)

    Götz, Christine; Pfeiffer, Roland; Tigges, Julia; Ruwiedel, Karsten; Hübenthal, Ulrike; Merk, Hans F; Krutmann, Jean; Edwards, Robert J; Abel, Josef; Pease, Camilla; Goebel, Carsten; Hewitt, Nicola; Fritsche, Ellen

    2012-05-01

    The 7th Amendment to the EU Cosmetics Directive prohibits the use of animals in cosmetic testing for certain endpoints, such as genotoxicity. Therefore, skin in vitro models have to replace chemical testing in vivo. However, the metabolic competence neither of human skin nor of alternative in vitro models has so far been fully characterized, although skin is the first-pass organ for accidentally or purposely (cosmetics and pharmaceuticals) applied chemicals. Thus, there is an urgent need to understand the xenobiotic-metabolizing capacities of human skin and to compare these activities to models developed to replace animal testing. We have measured the activity of the phase II enzymes glutathione S-transferase, UDP-glucuronosyltransferase and N-acetyltransferase in ex vivo human skin, the 3D epidermal model EpiDerm 200 (EPI-200), immortalized keratinocyte-based cell lines (HaCaT and NCTC 2544) and primary normal human epidermal keratinocytes. We show that all three phase II enzymes are present and highly active in skin as compared to phase I. Human skin, therefore, represents a more detoxifying than activating organ. This work systematically compares the activities of three important phase II enzymes in four different in vitro models directly to human skin. We conclude from our studies that 3D epidermal models, like the EPI-200 employed here, are superior over monolayer cultures in mimicking human skin xenobiotic metabolism and thus better suited for dermatotoxicity testing. © 2012 John Wiley & Sons A/S.

  12. Safety Testing of Ammonium Nitrate Based Mixtures

    Science.gov (United States)

    Phillips, Jason; Lappo, Karmen; Phelan, James; Peterson, Nathan; Gilbert, Don

    2013-06-01

    Ammonium nitrate (AN)/ammonium nitrate based explosives have a lengthy documented history of use by adversaries in acts of terror. While historical research has been conducted on AN-based explosive mixtures, it has primarily focused on detonation performance while varying the oxygen balance between the oxidizer and fuel components. Similarly, historical safety data on these materials is often lacking in pertinent details such as specific fuel type, particle size parameters, oxidizer form, etc. A variety of AN-based fuel-oxidizer mixtures were tested for small-scale sensitivity in preparation for large-scale testing. Current efforts focus on maintaining a zero oxygen-balance (a stoichiometric ratio for active chemical participants) while varying factors such as charge geometry, oxidizer form, particle size, and inert diluent ratios. Small-scale safety testing was conducted on various mixtures and fuels. It was found that ESD sensitivity is significantly affected by particle size, while this is less so for impact and friction. Thermal testing is in progress to evaluate hazards that may be experienced during large-scale testing.

  13. Simple sorting algorithm test based on CUDA

    OpenAIRE

    Meng, Hongyu; Guo, Fangjin

    2015-01-01

    With the development of computing technology, CUDA has become a very important tool. In computer programming, sorting algorithm is widely used. There are many simple sorting algorithms such as enumeration sort, bubble sort and merge sort. In this paper, we test some simple sorting algorithm based on CUDA and draw some useful conclusions.

  14. Forum: Is Test-Based Accountability Dead?

    Science.gov (United States)

    Polikoff, Morgan S.; Greene, Jay P.; Huffman, Kevin

    2017-01-01

    Since the 2001 passage of the No Child Left Behind Act (NCLB), test-based accountability has been an organizing principle--perhaps "the" organizing principle--of efforts to improve American schools. But lately, accountability has been under fire from many critics, including Common Core opponents and those calling for more multifaceted…

  15. Numerical modelling of concentrated leak erosion during Hole Erosion Tests

    OpenAIRE

    Mercier, F.; Bonelli, S.; Golay, F.; Anselmet, F.; Philippe, P.; Borghi, R.

    2015-01-01

    This study focuses on the numerical modelling of concentrated leak erosion of a cohesive soil by a turbulent flow in axisymmetrical geometry, with application to the Hole Erosion Test (HET). The numerical model is based on adaptive remeshing of the water/soil interface to ensure accurate description of the mechanical phenomena occurring near the soil/water interface. The erosion law governing the interface motion is based on two erosion parameters: the critical shear stress and the erosion co...

  16. Realistic evaluation of tester exposure based on Florida testing experience

    International Nuclear Information System (INIS)

    Schreiber, R.A.

    1990-01-01

    This paper reports on a radon decay product exposure model for Florida Certified Radon Measurement Technicians that has been formulated based on the guidance of 10CFR20. This model was used to estimate the exposure of 44 Florida measurement technicians from January through November of 1989. Comparing estimated testing and home exposure shows that 100% of the technicians observed received more exposure in the home than during testing activities. Exposure during normal office hours also exceed testing exposure in 86% of the technicians observed. Health and safety exposure data for radon measurement technicians does not follow the standard concepts of occupational radiation exposure normally accepted in 10CFR20

  17. Testing and Modeling of Mechanical Characteristics of Resistance Welding Machines

    DEFF Research Database (Denmark)

    Wu, Pei; Zhang, Wenqi; Bay, Niels

    2003-01-01

    for both upper and lower electrode systems. This has laid a foundation for modeling the welding process and selecting the welding parameters considering the machine factors. The method is straightforward and easy to be applied in industry since the whole procedure is based on tests with no requirements......The dynamic mechanical response of resistance welding machine is very important to the weld quality in resistance welding especially in projection welding when collapse or deformation of work piece occurs. It is mainly governed by the mechanical parameters of machine. In this paper, a mathematical...... model for characterizing the dynamic mechanical responses of machine and a special test set-up called breaking test set-up are developed. Based on the model and the test results, the mechanical parameters of machine are determined, including the equivalent mass, damping coefficient, and stiffness...

  18. Testing of a steel containment vessel model

    International Nuclear Information System (INIS)

    Luk, V.K.; Hessheimer, M.F.; Matsumoto, T.; Komine, K.; Costello, J.F.

    1997-01-01

    A mixed-scale containment vessel model, with 1:10 in containment geometry and 1:4 in shell thickness, was fabricated to represent an improved, boiling water reactor (BWR) Mark II containment vessel. A contact structure, installed over the model and separated at a nominally uniform distance from it, provided a simplified representation of a reactor shield building in the actual plant. This paper describes the pretest preparations and the conduct of the high pressure test of the model performed on December 11-12, 1996. 4 refs., 2 figs

  19. Precision tests of the Standard Model

    International Nuclear Information System (INIS)

    Ol'shevskij, A.G.

    1996-01-01

    The present status of the precision measurements of electroweak observables is discussed with the special emphasis on the results obtained recently. All together these measurements provide the basis for the stringent test of the Standard Model and determination of the SM parameters. 22 refs., 23 figs., 11 tabs

  20. Binomial test models and item difficulty

    NARCIS (Netherlands)

    van der Linden, Willem J.

    1979-01-01

    In choosing a binomial test model, it is important to know exactly what conditions are imposed on item difficulty. In this paper these conditions are examined for both a deterministic and a stochastic conception of item responses. It appears that they are more restrictive than is generally

  1. Shallow foundation model tests in Europe

    Czech Academy of Sciences Publication Activity Database

    Feda, Jaroslav; Simonini, P.; Arslan, U.; Georgiodis, M.; Laue, J.; Pinto, I.

    1999-01-01

    Roč. 2, č. 4 (1999), s. 447-475 ISSN 1436-6517. [Int. Conf. on Soil - Structure Interaction in Urban Civ. Engineering. Darmstadt, 08.10.1999-09.10.1999] R&D Projects: GA MŠk OC C7.10 Keywords : shallow foundations * model tests * sandy subsoil * bearing capacity * settlement Subject RIV: JM - Building Engineering

  2. Testing for Statistical Discrimination based on Gender

    OpenAIRE

    Lesner, Rune Vammen

    2016-01-01

    This paper develops a model which incorporates the two most commonly cited strands of the literature on statistical discrimination, namely screening discrimination and stereotyping. The model is used to provide empirical evidence of statistical discrimination based on gender in the labour market. It is shown that the implications of both screening discrimination and stereotyping are consistent with observable wage dynamics. In addition, it is found that the gender wage gap decreases in tenure...

  3. Animal models of toxicology testing: the role of pigs.

    Science.gov (United States)

    Helke, Kristi L; Swindle, Marvin Michael

    2013-02-01

    In regulatory toxicological testing, both a rodent and non-rodent species are required. Historically, dogs and non-human primates (NHP) have been the species of choice of the non-rodent portion of testing. The pig is an appropriate option for these tests based on metabolic pathways utilized in xenobiotic biotransformation. This review focuses on the Phase I and Phase II biotransformation pathways in humans and pigs and highlights the similarities and differences of these models. This is a growing field and references are sparse. Numerous breeds of pigs are discussed along with specific breed differences in these enzymes that are known. While much available data are presented, it is grossly incomplete and sometimes contradictory based on methods used. There is no ideal species to use in toxicology. The use of dogs and NHP in xenobiotic testing continues to be the norm. Pigs present a viable and perhaps more reliable model of non-rodent testing.

  4. Testing mechanistic models of growth in insects.

    Science.gov (United States)

    Maino, James L; Kearney, Michael R

    2015-11-22

    Insects are typified by their small size, large numbers, impressive reproductive output and rapid growth. However, insect growth is not simply rapid; rather, insects follow a qualitatively distinct trajectory to many other animals. Here we present a mechanistic growth model for insects and show that increasing specific assimilation during the growth phase can explain the near-exponential growth trajectory of insects. The presented model is tested against growth data on 50 insects, and compared against other mechanistic growth models. Unlike the other mechanistic models, our growth model predicts energy reserves per biomass to increase with age, which implies a higher production efficiency and energy density of biomass in later instars. These predictions are tested against data compiled from the literature whereby it is confirmed that insects increase their production efficiency (by 24 percentage points) and energy density (by 4 J mg(-1)) between hatching and the attainment of full size. The model suggests that insects achieve greater production efficiencies and enhanced growth rates by increasing specific assimilation and increasing energy reserves per biomass, which are less costly to maintain than structural biomass. Our findings illustrate how the explanatory and predictive power of mechanistic growth models comes from their grounding in underlying biological processes. © 2015 The Author(s).

  5. Testing proton spin models with polarized beams

    International Nuclear Information System (INIS)

    Ramsey, G.P.

    1991-01-01

    We review models for spin-weighted parton distributions in a proton. Sum rules involving the nonsinglet components of the structure function xg 1 p help narrow the range of parameters in these models. The contribution of the γ 5 anomaly term depends on the size of the integrated polarized gluon distribution and experimental predictions depend on its size. We have proposed three models for the polarized gluon distributions, whose range is considerable. These model distributions give an overall range is considerable. These model distributions give an overall range of parameters that can be tested with polarized beam experiments. These are discussed with regard to specific predictions for polarized beam experiments at energies typical of UNK

  6. Testing and Modeling of Machine Properties in Resistance Welding

    DEFF Research Database (Denmark)

    Wu, Pei

    The objective of this work has been to test and model the machine properties including the mechanical properties and the electrical properties in resistance welding. The results are used to simulate the welding process more accurately. The state of the art in testing and modeling machine properties...... as real projection welding tests, is easy to realize in industry, since tests may be performed in situ. In part II, an approach of characterizing the electrical properties of AC resistance welding machines is presented, involving testing and mathematical modelling of the weld current, the firing angle...... in resistance welding has been described based on a comprehensive literature study. The present thesis has been subdivided into two parts: Part I: Mechanical properties of resistance welding machines. Part II: Electrical properties of resistance welding machines. In part I, the electrode force in the squeeze...

  7. Fault tolerant system based on IDDQ testing

    Science.gov (United States)

    Guibane, Badi; Hamdi, Belgacem; Mtibaa, Abdellatif; Bensalem, Brahim

    2018-06-01

    Offline test is essential to ensure good manufacturing quality. However, for permanent or transient faults that occur during the use of the integrated circuit in an application, an online integrated test is needed as well. This procedure should ensure the detection and possibly the correction or the masking of these faults. This requirement of self-correction is sometimes necessary, especially in critical applications that require high security such as automotive, space or biomedical applications. We propose a fault-tolerant design for analogue and mixed-signal design complementary metal oxide (CMOS) circuits based on the quiescent current supply (IDDQ) testing. A defect can cause an increase in current consumption. IDDQ testing technique is based on the measurement of power supply current to distinguish between functional and failed circuits. The technique has been an effective testing method for detecting physical defects such as gate-oxide shorts, floating gates (open) and bridging defects in CMOS integrated circuits. An architecture called BICS (Built In Current Sensor) is used for monitoring the supply current (IDDQ) of the connected integrated circuit. If the measured current is not within the normal range, a defect is signalled and the system switches connection from the defective to a functional integrated circuit. The fault-tolerant technique is composed essentially by a double mirror built-in current sensor, allowing the detection of abnormal current consumption and blocks allowing the connection to redundant circuits, if a defect occurs. Spices simulations are performed to valid the proposed design.

  8. Using a micro computer based test bank

    International Nuclear Information System (INIS)

    Hamel, R.T.

    1987-01-01

    Utilizing a micro computer based test bank offers a training department many advantages and can have a positive impact upon training procedures and examination standards. Prior to data entry, Training Department management must pre-review the examination questions and answers to ensure compliance with examination standards and to verify the validity of all questions. Management must adhere to the TSD format since all questions require an enabling objective numbering scheme. Each question is entered under the enabling objective upon which it is based. Then the question is selected via the enabling objective. This eliminates any instructor bias because a random number generator chooses the test question. However, the instructor may load specific questions to create an emphasis theme for any test. The examination, answer and cover sheets are produced and printed within minutes. The test bank eliminates the large amount of time that is normally required for an instructor to formulate an examination. The need for clerical support is reduced by the elimination of typing examinations and also by the software's ability to maintain and generate student/course lists, attendance sheets, and grades. Software security measures limit access to the test bank, and the impromptu method used to generate and print an examination enhance its security

  9. Testing Parametric versus Semiparametric Modelling in Generalized Linear Models

    NARCIS (Netherlands)

    Härdle, W.K.; Mammen, E.; Müller, M.D.

    1996-01-01

    We consider a generalized partially linear model E(Y|X,T) = G{X'b + m(T)} where G is a known function, b is an unknown parameter vector, and m is an unknown function.The paper introduces a test statistic which allows to decide between a parametric and a semiparametric model: (i) m is linear, i.e.

  10. OTEC riser cable model and prototype testing

    Science.gov (United States)

    Kurt, J. P.; Schultz, J. A.; Roblee, L. H. S.

    1981-12-01

    Two different OTEC riser cables have been developed to span the distance between a floating OTEC power plant and the ocean floor. The major design concerns for a riser cable in the dynamic OTEC environment are fatigue, corrosion, and electrical/mechanical aging of the cable components. The basic properties of the cable materials were studied through tests on model cables and on samples of cable materials. Full-scale prototype cables were manufactured and were tested to measure their electrical and mechanical properties and performance. The full-scale testing was culminated by the electrical/mechanical fatigue test, which exposes full-scale cables to simultaneous tension, bending and electrical loads, all in a natural seawater environment.

  11. Model-Based Reasoning

    Science.gov (United States)

    Ifenthaler, Dirk; Seel, Norbert M.

    2013-01-01

    In this paper, there will be a particular focus on mental models and their application to inductive reasoning within the realm of instruction. A basic assumption of this study is the observation that the construction of mental models and related reasoning is a slowly developing capability of cognitive systems that emerges effectively with proper…

  12. Universal Verification Methodology Based Register Test Automation Flow.

    Science.gov (United States)

    Woo, Jae Hun; Cho, Yong Kwan; Park, Sun Kyu

    2016-05-01

    In today's SoC design, the number of registers has been increased along with complexity of hardware blocks. Register validation is a time-consuming and error-pron task. Therefore, we need an efficient way to perform verification with less effort in shorter time. In this work, we suggest register test automation flow based UVM (Universal Verification Methodology). UVM provides a standard methodology, called a register model, to facilitate stimulus generation and functional checking of registers. However, it is not easy for designers to create register models for their functional blocks or integrate models in test-bench environment because it requires knowledge of SystemVerilog and UVM libraries. For the creation of register models, many commercial tools support a register model generation from register specification described in IP-XACT, but it is time-consuming to describe register specification in IP-XACT format. For easy creation of register model, we propose spreadsheet-based register template which is translated to IP-XACT description, from which register models can be easily generated using commercial tools. On the other hand, we also automate all the steps involved integrating test-bench and generating test-cases, so that designers may use register model without detailed knowledge of UVM or SystemVerilog. This automation flow involves generating and connecting test-bench components (e.g., driver, checker, bus adaptor, etc.) and writing test sequence for each type of register test-case. With the proposed flow, designers can save considerable amount of time to verify functionality of registers.

  13. Agent-based modelling of cholera diffusion

    NARCIS (Netherlands)

    Augustijn-Beckers, Petronella; Doldersum, Tom; Useya, Juliana; Augustijn, Dionysius C.M.

    2016-01-01

    This paper introduces a spatially explicit agent-based simulation model for micro-scale cholera diffusion. The model simulates both an environmental reservoir of naturally occurring V.cholerae bacteria and hyperinfectious V. cholerae. Objective of the research is to test if runoff from open refuse

  14. Hypervapotron flow testing with rapid prototype models

    International Nuclear Information System (INIS)

    Driemeyer, D.; Hellwig, T.; Kubik, D.; Langenderfer, E.; Mantz, H.; McSmith, M.; Jones, B.; Butler, J.

    1995-01-01

    A flow test model of the inlet section of a three channel hypervapotron plate that has been proposed as a heat sink in the ITER divertor was prepared using a rapid prototyping stereolithography process that is widely used for component development in US industry. An existing water flow loop at the University of Illinois is being used for isothermal flow tests to collect pressure drop data for comparison with proposed vapotron friction factor correlations. Differential pressure measurements are taken, across the test section inlet manifold, the vapotron channel (about a seven inch length), the outlet manifold and the inlet-to-outlet. The differential pressures are currently measured with manometers. Tests were conducted at flow velocities from 1--10 m/s to cover the full range of ITER interest. A tap was also added for a small hypodermic needle to inject dye into the flow channel at several positions to examine the nature of the developing flow field at the entrance to the vapotron section. Follow-on flow tests are planned using a model with adjustable flow channel dimensions to permit more extensive pressure drop data to be collected. This information will be used to update vapotron design correlations for ITER

  15. EPR-based material modelling of soils

    Science.gov (United States)

    Faramarzi, Asaad; Alani, Amir M.

    2013-04-01

    In the past few decades, as a result of the rapid developments in computational software and hardware, alternative computer aided pattern recognition approaches have been introduced to modelling many engineering problems, including constitutive modelling of materials. The main idea behind pattern recognition systems is that they learn adaptively from experience and extract various discriminants, each appropriate for its purpose. In this work an approach is presented for developing material models for soils based on evolutionary polynomial regression (EPR). EPR is a recently developed hybrid data mining technique that searches for structured mathematical equations (representing the behaviour of a system) using genetic algorithm and the least squares method. Stress-strain data from triaxial tests are used to train and develop EPR-based material models for soil. The developed models are compared with some of the well-known conventional material models and it is shown that EPR-based models can provide a better prediction for the behaviour of soils. The main benefits of using EPR-based material models are that it provides a unified approach to constitutive modelling of all materials (i.e., all aspects of material behaviour can be implemented within a unified environment of an EPR model); it does not require any arbitrary choice of constitutive (mathematical) models. In EPR-based material models there are no material parameters to be identified. As the model is trained directly from experimental data therefore, EPR-based material models are the shortest route from experimental research (data) to numerical modelling. Another advantage of EPR-based constitutive model is that as more experimental data become available, the quality of the EPR prediction can be improved by learning from the additional data, and therefore, the EPR model can become more effective and robust. The developed EPR-based material models can be incorporated in finite element (FE) analysis.

  16. Horns Rev II, 2-D Model Tests

    DEFF Research Database (Denmark)

    Andersen, Thomas Lykke; Frigaard, Peter

    This report present the results of 2D physical model tests carried out in the shallow wave flume at Dept. of Civil Engineering, Aalborg University (AAU), on behalf of Energy E2 A/S part of DONG Energy A/S, Denmark. The objective of the tests was: to investigate the combined influence of the pile...... diameter to water depth ratio and the wave hight to water depth ratio on wave run-up of piles. The measurements should be used to design access platforms on piles....

  17. A Comparison of Item Selection Procedures Using Different Ability Estimation Methods in Computerized Adaptive Testing Based on the Generalized Partial Credit Model

    Science.gov (United States)

    Ho, Tsung-Han

    2010-01-01

    Computerized adaptive testing (CAT) provides a highly efficient alternative to the paper-and-pencil test. By selecting items that match examinees' ability levels, CAT not only can shorten test length and administration time but it can also increase measurement precision and reduce measurement error. In CAT, maximum information (MI) is the most…

  18. Optimization models for flight test scheduling

    Science.gov (United States)

    Holian, Derreck

    with restriction removal is based on heuristic approaches to support the reality of flight test in both solution space and computational time. Exact methods for yielding an optimized solution will be discussed however they are not directly applicable to the flight test problem and therefore have not been included in the system.

  19. Temperature Buffer Test. Final THM modelling

    International Nuclear Information System (INIS)

    Aakesson, Mattias; Malmberg, Daniel; Boergesson, Lennart; Hernelind, Jan; Ledesma, Alberto; Jacinto, Abel

    2012-01-01

    The Temperature Buffer Test (TBT) is a joint project between SKB/ANDRA and supported by ENRESA (modelling) and DBE (instrumentation), which aims at improving the understanding and to model the thermo-hydro-mechanical behavior of buffers made of swelling clay submitted to high temperatures (over 100 deg C) during the water saturation process. The test has been carried out in a KBS-3 deposition hole at Aespoe HRL. It was installed during the spring of 2003. Two heaters (3 m long, 0.6 m diameter) and two buffer arrangements have been investigated: the lower heater was surrounded by bentonite only, whereas the upper heater was surrounded by a composite barrier, with a sand shield between the heater and the bentonite. The test was dismantled and sampled during the winter of 2009/2010. This report presents the final THM modelling which was resumed subsequent to the dismantling operation. The main part of this work has been numerical modelling of the field test. Three different modelling teams have presented several model cases for different geometries and different degree of process complexity. Two different numerical codes, Code B right and Abaqus, have been used. The modelling performed by UPC-Cimne using Code B right, has been divided in three subtasks: i) analysis of the response observed in the lower part of the test, by inclusion of a number of considerations: (a) the use of the Barcelona Expansive Model for MX-80 bentonite; (b) updated parameters in the vapour diffusive flow term; (c) the use of a non-conventional water retention curve for MX-80 at high temperature; ii) assessment of a possible relation between the cracks observed in the bentonite blocks in the upper part of TBT, and the cycles of suction and stresses registered in that zone at the start of the experiment; and iii) analysis of the performance, observations and interpretation of the entire test. It was however not possible to carry out a full THM analysis until the end of the test due to

  20. Temperature Buffer Test. Final THM modelling

    Energy Technology Data Exchange (ETDEWEB)

    Aakesson, Mattias; Malmberg, Daniel; Boergesson, Lennart; Hernelind, Jan [Clay Technology AB, Lund (Sweden); Ledesma, Alberto; Jacinto, Abel [UPC, Universitat Politecnica de Catalunya, Barcelona (Spain)

    2012-01-15

    The Temperature Buffer Test (TBT) is a joint project between SKB/ANDRA and supported by ENRESA (modelling) and DBE (instrumentation), which aims at improving the understanding and to model the thermo-hydro-mechanical behavior of buffers made of swelling clay submitted to high temperatures (over 100 deg C) during the water saturation process. The test has been carried out in a KBS-3 deposition hole at Aespoe HRL. It was installed during the spring of 2003. Two heaters (3 m long, 0.6 m diameter) and two buffer arrangements have been investigated: the lower heater was surrounded by bentonite only, whereas the upper heater was surrounded by a composite barrier, with a sand shield between the heater and the bentonite. The test was dismantled and sampled during the winter of 2009/2010. This report presents the final THM modelling which was resumed subsequent to the dismantling operation. The main part of this work has been numerical modelling of the field test. Three different modelling teams have presented several model cases for different geometries and different degree of process complexity. Two different numerical codes, Code{sub B}right and Abaqus, have been used. The modelling performed by UPC-Cimne using Code{sub B}right, has been divided in three subtasks: i) analysis of the response observed in the lower part of the test, by inclusion of a number of considerations: (a) the use of the Barcelona Expansive Model for MX-80 bentonite; (b) updated parameters in the vapour diffusive flow term; (c) the use of a non-conventional water retention curve for MX-80 at high temperature; ii) assessment of a possible relation between the cracks observed in the bentonite blocks in the upper part of TBT, and the cycles of suction and stresses registered in that zone at the start of the experiment; and iii) analysis of the performance, observations and interpretation of the entire test. It was however not possible to carry out a full THM analysis until the end of the test due to

  1. Observational tests of FRW world models

    International Nuclear Information System (INIS)

    Lahav, Ofer

    2002-01-01

    Observational tests for the cosmological principle are reviewed. Assuming the FRW metric we then summarize estimates of cosmological parameters from various datasets, in particular the cosmic microwave background and the 2dF galaxy redshift survey. These and other analyses suggest a best-fit Λ-cold dark matter model with Ω m = 1 - Ω l ∼ 0.3 and H 0 ∼ 70 km s -1 Mpc -1 . It is remarkable that different measurements converge to this 'concordance model', although it remains to be seen if the two main components of this model, the dark matter and the dark energy, are real entities or just 'epicycles'. We point out some open questions related to this fashionable model

  2. Model-based Software Engineering

    DEFF Research Database (Denmark)

    Kindler, Ekkart

    2010-01-01

    The vision of model-based software engineering is to make models the main focus of software development and to automatically generate software from these models. Part of that idea works already today. But, there are still difficulties when it comes to behaviour. Actually, there is no lack in models...

  3. Preliminary Test for Constitutive Models of CAP

    Energy Technology Data Exchange (ETDEWEB)

    Choo, Yeon Joon; Hong, Soon Joon; Hwang, Su Hyun; Lee, Keo Hyung; Kim, Min Ki; Lee, Byung Chul [FNC Tech., Seoul (Korea, Republic of); Ha, Sang Jun; Choi, Hoon [Korea Electric Power Research Institute, Daejeon (Korea, Republic of)

    2010-05-15

    The development project for the domestic design code was launched to be used for the safety and performance analysis of pressurized light water reactors. As a part of this project, CAP (Containment Analysis Package) code has been developing for the containment safety and performance analysis side by side with SPACE. The CAP code treats three fields (vapor, continuous liquid and dispersed drop) for the assessment of containment specific phenomena, and is featured by assessment capabilities in multi-dimensional and lumped parameter thermal hydraulic cell. Thermal hydraulics solver was developed and has a significant progress now. Implementation of the well proven constitutive models and correlations are essential in other for a containment code to be used with the generalized or optimized purposes. Generally, constitutive equations are composed of interfacial and wall transport models and correlations. These equations are included in the source terms of the governing field equations. In order to develop the best model and correlation package of the CAP code, various models currently used in major containment analysis codes, such as GOTHIC, CONTAIN2.0 and CONTEMPT-LT are reviewed. Several models and correlations were incorporated for the preliminary test of CAP's performance and test results and future plans to improve the level of execution besides will be discussed in this paper

  4. Preliminary Test for Constitutive Models of CAP

    International Nuclear Information System (INIS)

    Choo, Yeon Joon; Hong, Soon Joon; Hwang, Su Hyun; Lee, Keo Hyung; Kim, Min Ki; Lee, Byung Chul; Ha, Sang Jun; Choi, Hoon

    2010-01-01

    The development project for the domestic design code was launched to be used for the safety and performance analysis of pressurized light water reactors. As a part of this project, CAP (Containment Analysis Package) code has been developing for the containment safety and performance analysis side by side with SPACE. The CAP code treats three fields (vapor, continuous liquid and dispersed drop) for the assessment of containment specific phenomena, and is featured by assessment capabilities in multi-dimensional and lumped parameter thermal hydraulic cell. Thermal hydraulics solver was developed and has a significant progress now. Implementation of the well proven constitutive models and correlations are essential in other for a containment code to be used with the generalized or optimized purposes. Generally, constitutive equations are composed of interfacial and wall transport models and correlations. These equations are included in the source terms of the governing field equations. In order to develop the best model and correlation package of the CAP code, various models currently used in major containment analysis codes, such as GOTHIC, CONTAIN2.0 and CONTEMPT-LT are reviewed. Several models and correlations were incorporated for the preliminary test of CAP's performance and test results and future plans to improve the level of execution besides will be discussed in this paper

  5. Business model stress testing : A practical approach to test the robustness of a business model

    NARCIS (Netherlands)

    Haaker, T.I.; Bouwman, W.A.G.A.; Janssen, W; de Reuver, G.A.

    Business models and business model innovation are increasingly gaining attention in practice as well as in academic literature. However, the robustness of business models (BM) is seldom tested vis-à-vis the fast and unpredictable changes in digital technologies, regulation and markets. The

  6. Principles of models based engineering

    Energy Technology Data Exchange (ETDEWEB)

    Dolin, R.M.; Hefele, J.

    1996-11-01

    This report describes a Models Based Engineering (MBE) philosophy and implementation strategy that has been developed at Los Alamos National Laboratory`s Center for Advanced Engineering Technology. A major theme in this discussion is that models based engineering is an information management technology enabling the development of information driven engineering. Unlike other information management technologies, models based engineering encompasses the breadth of engineering information, from design intent through product definition to consumer application.

  7. A valuation-Based Test of Market Timing

    NARCIS (Netherlands)

    Koeter-Kant, J.; Elliott, W.B.; Warr, R.S.

    2007-01-01

    We implement an earnings-based fundamental valuation model to test the impact of market timing on the firm's method of funding the financing deficit. We argue that our valuation metric provides a superior measure of equity misvaluation because it avoids multiple interpretation problems faced by the

  8. Risk based modelling

    International Nuclear Information System (INIS)

    Chapman, O.J.V.; Baker, A.E.

    1993-01-01

    Risk based analysis is a tool becoming available to both engineers and managers to aid decision making concerning plant matters such as In-Service Inspection (ISI). In order to develop a risk based method, some form of Structural Reliability Risk Assessment (SRRA) needs to be performed to provide a probability of failure ranking for all sites around the plant. A Probabilistic Risk Assessment (PRA) can then be carried out to combine these possible events with the capability of plant safety systems and procedures, to establish the consequences of failure for the sites. In this way the probability of failures are converted into a risk based ranking which can be used to assist the process of deciding which sites should be included in an ISI programme. This paper reviews the technique and typical results of a risk based ranking assessment carried out for nuclear power plant pipework. (author)

  9. Overload prevention in model supports for wind tunnel model testing

    Directory of Open Access Journals (Sweden)

    Anton IVANOVICI

    2015-09-01

    Full Text Available Preventing overloads in wind tunnel model supports is crucial to the integrity of the tested system. Results can only be interpreted as valid if the model support, conventionally called a sting remains sufficiently rigid during testing. Modeling and preliminary calculation can only give an estimate of the sting’s behavior under known forces and moments but sometimes unpredictable, aerodynamically caused model behavior can cause large transient overloads that cannot be taken into account at the sting design phase. To ensure model integrity and data validity an analog fast protection circuit was designed and tested. A post-factum analysis was carried out to optimize the overload detection and a short discussion on aeroelastic phenomena is included to show why such a detector has to be very fast. The last refinement of the concept consists in a fast detector coupled with a slightly slower one to differentiate between transient overloads that decay in time and those that are the result of aeroelastic unwanted phenomena. The decision to stop or continue the test is therefore conservatively taken preserving data and model integrity while allowing normal startup loads and transients to manifest.

  10. Designing healthy communities: Testing the walkability model

    OpenAIRE

    Zuniga-Teran, Adriana; Orr, Barron; Gimblett, Randy; Chalfoun, Nader; Marsh, Stuart; Guertin, David; Going, Scott

    2017-01-01

    Research from multiple domains has provided insights into how neighborhood design can be improved to have a more favorable effect on physical activity, a concept known as walkability. The relevant research findings/hypotheses have been integrated into a Walkability Framework, which organizes the design elements into nine walkability categories. The purpose of this study was to test whether this conceptual framework can be used as a model to measure the interactions between the built environme...

  11. 2-D Model Test of Dolosse Breakwater

    DEFF Research Database (Denmark)

    Burcharth, Hans F.; Liu, Zhou

    1994-01-01

    ). To extend the design diagram to cover Dolos breakwaters with superstructure, 2-D model tests of Dolos breakwater with wave wall is included in the project Rubble Mound Breakwater Failure Modes sponsored by the Directorate General XII of the Commission of the European Communities under Contract MAS-CT92......The rational design diagram for Dolos armour should incorporate both the hydraulic stability and the structural integrity. The previous tests performed by Aalborg University (AU) made available such design diagram for the trunk of Dolos breakwater without superstructures (Burcharth et al. 1992...... was on the Dolos breakwater with a high superstructure, where there was almost no overtopping. This case is believed to be the most dangerous one. The test of the Dolos breakwater with a low superstructure was also performed. The objective of the last part of the experiment is to investigate the influence...

  12. Optimisation of test and maintenance based on probabilistic methods

    International Nuclear Information System (INIS)

    Cepin, M.

    2001-01-01

    This paper presents a method, which based on models and results of probabilistic safety assessment, minimises the nuclear power plant risk by optimisation of arrangement of safety equipment outages. The test and maintenance activities of the safety equipment are timely arranged, so the classical static fault tree models are extended with the time requirements to be capable to model real plant states. A house event matrix is used, which enables modelling of the equipment arrangements through the discrete points of time. The result of the method is determination of such configuration of equipment outages, which result in the minimal risk. Minimal risk is represented by system unavailability. (authors)

  13. Testing for Statistical Discrimination based on Gender

    DEFF Research Database (Denmark)

    Lesner, Rune Vammen

    . It is shown that the implications of both screening discrimination and stereotyping are consistent with observable wage dynamics. In addition, it is found that the gender wage gap decreases in tenure but increases in job transitions and that the fraction of women in high-ranking positions within a firm does......This paper develops a model which incorporates the two most commonly cited strands of the literature on statistical discrimination, namely screening discrimination and stereotyping. The model is used to provide empirical evidence of statistical discrimination based on gender in the labour market...... not affect the level of statistical discrimination by gender....

  14. A general diagnostic model applied to language testing data.

    Science.gov (United States)

    von Davier, Matthias

    2008-11-01

    Probabilistic models with one or more latent variables are designed to report on a corresponding number of skills or cognitive attributes. Multidimensional skill profiles offer additional information beyond what a single test score can provide, if the reported skills can be identified and distinguished reliably. Many recent approaches to skill profile models are limited to dichotomous data and have made use of computationally intensive estimation methods such as Markov chain Monte Carlo, since standard maximum likelihood (ML) estimation techniques were deemed infeasible. This paper presents a general diagnostic model (GDM) that can be estimated with standard ML techniques and applies to polytomous response variables as well as to skills with two or more proficiency levels. The paper uses one member of a larger class of diagnostic models, a compensatory diagnostic model for dichotomous and partial credit data. Many well-known models, such as univariate and multivariate versions of the Rasch model and the two-parameter logistic item response theory model, the generalized partial credit model, as well as a variety of skill profile models, are special cases of this GDM. In addition to an introduction to this model, the paper presents a parameter recovery study using simulated data and an application to real data from the field test for TOEFL Internet-based testing.

  15. Automated Functional Testing based on the Navigation of Web Applications

    Directory of Open Access Journals (Sweden)

    Boni García

    2011-08-01

    Full Text Available Web applications are becoming more and more complex. Testing such applications is an intricate hard and time-consuming activity. Therefore, testing is often poorly performed or skipped by practitioners. Test automation can help to avoid this situation. Hence, this paper presents a novel approach to perform automated software testing for web applications based on its navigation. On the one hand, web navigation is the process of traversing a web application using a browser. On the other hand, functional requirements are actions that an application must do. Therefore, the evaluation of the correct navigation of web applications results in the assessment of the specified functional requirements. The proposed method to perform the automation is done in four levels: test case generation, test data derivation, test case execution, and test case reporting. This method is driven by three kinds of inputs: i UML models; ii Selenium scripts; iii XML files. We have implemented our approach in an open-source testing framework named Automatic Testing Platform. The validation of this work has been carried out by means of a case study, in which the target is a real invoice management system developed using a model-driven approach.

  16. Creating a simulation model of software testing using Simulink package

    Directory of Open Access Journals (Sweden)

    V. M. Dubovoi

    2016-12-01

    Full Text Available The determination of the solution model of software testing that allows prediction both the whole process and its specific stages is actual for IT-industry. The article focuses on solving this problem. The aim of the article is prediction the time and improvement the quality of software testing. The analysis of the software testing process shows that it can be attributed to the branched cyclic technological processes because it is cyclical with decision-making on control operations. The investigation uses authors' previous works andsoftware testing process method based on Markov model. The proposed method enables execution the prediction for each software module, which leads to better decision-making of each controlled suboperation of all processes. Simulink simulation model shows implementation and verification of results of proposed technique. Results of the research have practically implemented in the IT-industry.

  17. Thurstonian models for sensory discrimination tests as generalized linear models

    DEFF Research Database (Denmark)

    Brockhoff, Per B.; Christensen, Rune Haubo Bojesen

    2010-01-01

    as a so-called generalized linear model. The underlying sensory difference 6 becomes directly a parameter of the statistical model and the estimate d' and it's standard error becomes the "usual" output of the statistical analysis. The d' for the monadic A-NOT A method is shown to appear as a standard......Sensory discrimination tests such as the triangle, duo-trio, 2-AFC and 3-AFC tests produce binary data and the Thurstonian decision rule links the underlying sensory difference 6 to the observed number of correct responses. In this paper it is shown how each of these four situations can be viewed...

  18. HEV Test Bench Based on CAN Bus Sensor Communication

    Directory of Open Access Journals (Sweden)

    Shupeng ZHAO

    2014-02-01

    Full Text Available The HEV test bench based on Controller Area Network bus was studied and developed. Control system of HEV power test bench used the CAN bus technology. The application of CAN bus technology on control system development has opened up a new research direction for domestic automobile experimental platform. The HEV power control system development work was completed, including power master controller, electric throttle controller, driving simulation platform, CAN2.0 B communication protocol procedures for formulation, CAN communication monitoring system, the simulation model based on MATLAB code automatic generation technology research, etc. Maximum absorption power of the test bench is 90 kW, the test bench top speed is 6000 r/min, the CAN communication data baud rate is 10~500 k, the conventional electric measurement parameter part precision satisfies the requirement of development of HEV. On the HEV test bench the result of regenerative braking experiment shows that the result got by the test bench was closer to the results got by outdoor road test. And the fuel consumption experiment test results show that the HEV fuel consumption and the charge-discharge character are in linear relationship. The establishment of the test platform for the evaluation of the development of hybrid electric vehicle and power provides physical simulation and test platform.

  19. Testing and Inference in Nonlinear Cointegrating Vector Error Correction Models

    DEFF Research Database (Denmark)

    Kristensen, Dennis; Rahbæk, Anders

    In this paper, we consider a general class of vector error correction models which allow for asymmetric and non-linear error correction. We provide asymptotic results for (quasi-)maximum likelihood (QML) based estimators and tests. General hypothesis testing is considered, where testing...... of non-stationary non-linear time series models. Thus the paper provides a full asymptotic theory for estimators as well as standard and non-standard test statistics. The derived asymptotic results prove to be new compared to results found elsewhere in the literature due to the impact of the estimated...... symmetric non-linear error correction considered. A simulation study shows that the fi…nite sample properties of the bootstrapped tests are satisfactory with good size and power properties for reasonable sample sizes....

  20. Testing and Inference in Nonlinear Cointegrating Vector Error Correction Models

    DEFF Research Database (Denmark)

    Kristensen, Dennis; Rahbek, Anders

    In this paper, we consider a general class of vector error correction models which allow for asymmetric and non-linear error correction. We provide asymptotic results for (quasi-)maximum likelihood (QML) based estimators and tests. General hypothesis testing is considered, where testing...... of non-stationary non-linear time series models. Thus the paper provides a full asymptotic theory for estimators as well as standard and non-standard test statistics. The derived asymptotic results prove to be new compared to results found elsewhere in the literature due to the impact of the estimated...... symmetric non-linear error correction are considered. A simulation study shows that the finite sample properties of the bootstrapped tests are satisfactory with good size and power properties for reasonable sample sizes....

  1. Space Launch System Base Heating Test: Experimental Operations & Results

    Science.gov (United States)

    Dufrene, Aaron; Mehta, Manish; MacLean, Matthew; Seaford, Mark; Holden, Michael

    2016-01-01

    NASA's Space Launch System (SLS) uses four clustered liquid rocket engines along with two solid rocket boosters. The interaction between all six rocket exhaust plumes will produce a complex and severe thermal environment in the base of the vehicle. This work focuses on a recent 2% scale, hot-fire SLS base heating test. These base heating tests are short-duration tests executed with chamber pressures near the full-scale values with gaseous hydrogen/oxygen engines and RSRMV analogous solid propellant motors. The LENS II shock tunnel/Ludwieg tube tunnel was used at or near flight duplicated conditions up to Mach 5. Model development was based on the Space Shuttle base heating tests with several improvements including doubling of the maximum chamber pressures and duplication of freestream conditions. Test methodology and conditions are presented, and base heating results from 76 runs are reported in non-dimensional form. Regions of high heating are identified and comparisons of various configuration and conditions are highlighted. Base pressure and radiometer results are also reported.

  2. Thermal modelling of Advanced LIGO test masses

    International Nuclear Information System (INIS)

    Wang, H; Dovale Álvarez, M; Mow-Lowry, C M; Freise, A; Blair, C; Brooks, A; Kasprzack, M F; Ramette, J; Meyers, P M; Kaufer, S; O’Reilly, B

    2017-01-01

    High-reflectivity fused silica mirrors are at the epicentre of today’s advanced gravitational wave detectors. In these detectors, the mirrors interact with high power laser beams. As a result of finite absorption in the high reflectivity coatings the mirrors suffer from a variety of thermal effects that impact on the detectors’ performance. We propose a model of the Advanced LIGO mirrors that introduces an empirical term to account for the radiative heat transfer between the mirror and its surroundings. The mechanical mode frequency is used as a probe for the overall temperature of the mirror. The thermal transient after power build-up in the optical cavities is used to refine and test the model. The model provides a coating absorption estimate of 1.5–2.0 ppm and estimates that 0.3 to 1.3 ppm of the circulating light is scattered onto the ring heater. (paper)

  3. Model-based consensus

    NARCIS (Netherlands)

    Boumans, M.; Martini, C.; Boumans, M.

    2014-01-01

    The aim of the rational-consensus method is to produce "rational consensus", that is, "mathematical aggregation", by weighing the performance of each expert on the basis of his or her knowledge and ability to judge relevant uncertainties. The measurement of the performance of the experts is based on

  4. Model-based consensus

    NARCIS (Netherlands)

    Boumans, Marcel

    2014-01-01

    The aim of the rational-consensus method is to produce “rational consensus”, that is, “mathematical aggregation”, by weighing the performance of each expert on the basis of his or her knowledge and ability to judge relevant uncertainties. The measurement of the performance of the experts is based on

  5. A new fit-for-purpose model testing framework: Decision Crash Tests

    Science.gov (United States)

    Tolson, Bryan; Craig, James

    2016-04-01

    Decision-makers in water resources are often burdened with selecting appropriate multi-million dollar strategies to mitigate the impacts of climate or land use change. Unfortunately, the suitability of existing hydrologic simulation models to accurately inform decision-making is in doubt because the testing procedures used to evaluate model utility (i.e., model validation) are insufficient. For example, many authors have identified that a good standard framework for model testing called the Klemes Crash Tests (KCTs), which are the classic model validation procedures from Klemeš (1986) that Andréassian et al. (2009) rename as KCTs, have yet to become common practice in hydrology. Furthermore, Andréassian et al. (2009) claim that the progression of hydrological science requires widespread use of KCT and the development of new crash tests. Existing simulation (not forecasting) model testing procedures such as KCTs look backwards (checking for consistency between simulations and past observations) rather than forwards (explicitly assessing if the model is likely to support future decisions). We propose a fundamentally different, forward-looking, decision-oriented hydrologic model testing framework based upon the concept of fit-for-purpose model testing that we call Decision Crash Tests or DCTs. Key DCT elements are i) the model purpose (i.e., decision the model is meant to support) must be identified so that model outputs can be mapped to management decisions ii) the framework evaluates not just the selected hydrologic model but the entire suite of model-building decisions associated with model discretization, calibration etc. The framework is constructed to directly and quantitatively evaluate model suitability. The DCT framework is applied to a model building case study on the Grand River in Ontario, Canada. A hypothetical binary decision scenario is analysed (upgrade or not upgrade the existing flood control structure) under two different sets of model building

  6. Testing substellar models with dynamical mass measurements

    Directory of Open Access Journals (Sweden)

    Liu M.C.

    2011-07-01

    Full Text Available We have been using Keck laser guide star adaptive optics to monitor the orbits of ultracool binaries, providing dynamical masses at lower luminosities and temperatures than previously available and enabling strong tests of theoretical models. We have identified three specific problems with theory: (1 We find that model color–magnitude diagrams cannot be reliably used to infer masses as they do not accurately reproduce the colors of ultracool dwarfs of known mass. (2 Effective temperatures inferred from evolutionary model radii are typically inconsistent with temperatures derived from fitting atmospheric models to observed spectra by 100–300 K. (3 For the only known pair of field brown dwarfs with a precise mass (3% and age determination (≈25%, the measured luminosities are ~2–3× higher than predicted by model cooling rates (i.e., masses inferred from Lbol and age are 20–30% larger than measured. To make progress in understanding the observed discrepancies, more mass measurements spanning a wide range of luminosity, temperature, and age are needed, along with more accurate age determinations (e.g., via asteroseismology for primary stars with brown dwarf binary companions. Also, resolved optical and infrared spectroscopy are needed to measure lithium depletion and to characterize the atmospheres of binary components in order to better assess model deficiencies.

  7. Testing and reference model analysis of FTTH system

    Science.gov (United States)

    Feng, Xiancheng; Cui, Wanlong; Chen, Ying

    2009-08-01

    With rapid development of Internet and broadband access network, the technologies of xDSL, FTTx+LAN , WLAN have more applications, new network service emerges in endless stream, especially the increase of network game, meeting TV, video on demand, etc. FTTH supports all present and future service with enormous bandwidth, including traditional telecommunication service, traditional data service and traditional TV service, and the future digital TV and VOD. With huge bandwidth of FTTH, it wins the final solution of broadband network, becomes the final goal of development of optical access network.. Fiber to the Home (FTTH) will be the goal of telecommunications cable broadband access. In accordance with the development trend of telecommunication services, to enhance the capacity of integrated access network, to achieve triple-play (voice, data, image), based on the existing optical Fiber to the curb (FTTC), Fiber To The Zone (FTTZ), Fiber to the Building (FTTB) user optical cable network, the optical fiber can extend to the FTTH system of end-user by using EPON technology. The article first introduced the basic components of FTTH system; and then explain the reference model and reference point for testing of the FTTH system; Finally, by testing connection diagram, the testing process, expected results, primarily analyze SNI Interface Testing, PON interface testing, Ethernet performance testing, UNI interface testing, Ethernet functional testing, PON functional testing, equipment functional testing, telephone functional testing, operational support capability testing and so on testing of FTTH system. ...

  8. Technical bases for the DWPF testing program

    International Nuclear Information System (INIS)

    Plodinec, M.J.

    1990-01-01

    The Defense Waste Processing Facility (DWPF) at the Savannah River Site (SRS) will be the first production facility in the United States for the immobilization of high-level nuclear waste. Production of DWPF canistered wasteforms will begin prior to repository licensing, so decisions on facility startup will have to be made before the final decisions on repository design are made. The Department of Energy's Office of Civilian Radioactive Waste Management (RW) has addressed this discrepancy by defining a Waste Acceptance Process. This process provides assurance that the borosilicate-glass wasteform, in a stainless-steel canister, produced by the DWPF will be acceptable for permanent storage in a federal repository. As part of this process, detailed technical specifications have been developed for the DWPF product. SRS has developed detailed strategies for demonstrating compliance with each of the Waste Acceptance Process specifications. An important part of the compliance is the testing which will be carried out in the DWPF. In this paper, the bases for each of the tests to be performed in the DWPF to establish compliance with the specifications are described, and the tests are detailed. The results of initial tests relating to characterization of sealed canisters are reported

  9. Agent-Based Modeling in Systems Pharmacology.

    Science.gov (United States)

    Cosgrove, J; Butler, J; Alden, K; Read, M; Kumar, V; Cucurull-Sanchez, L; Timmis, J; Coles, M

    2015-11-01

    Modeling and simulation (M&S) techniques provide a platform for knowledge integration and hypothesis testing to gain insights into biological systems that would not be possible a priori. Agent-based modeling (ABM) is an M&S technique that focuses on describing individual components rather than homogenous populations. This tutorial introduces ABM to systems pharmacologists, using relevant case studies to highlight how ABM-specific strengths have yielded success in the area of preclinical mechanistic modeling.

  10. Computer Based Test Untuk Seleksi Masuk Politeknik Negeri Bengkalis

    Directory of Open Access Journals (Sweden)

    Agus Tedyyana

    2017-11-01

    Full Text Available AbstrakPenyeleksian calon mahasiswa baru dapat dilakukan dengan aplikasi Computer Based Test (CBT. Metode yang digunakan meliputi teknik pengumpulan data, analisis sistem, model perancangan, implementasi dan pengujian. Penelitian ini menghasilkan aplikasi CBT dimana soal yang dimunculkan dari bank soal melalui proses pengacakan dengan tidak akan memunculkan soal yang sama dengan menggunakan metoda Fisher-Yates Shuffle. Dalam proses pengamanan informasi soal saat terhubung ke jaringan maka diperlukan teknik untuk penyandian pesan agar soal tersebut sebeum dimunculkan melewati proses enkripsi dan deskripsi data terlebih dahulu maka digunakan algoritma kriptografi  RSA. Metode perancangan perangkat lunak menggunakan model waterfall, perancangan database menggunakan entity relationship diagram, perancangan antarmuka menggunakan hypertext markup language (HTML Cascading Style Sheet (CSS dan jQuery serta diimplementasikan berbasis web dengan menggunakan bahasa pemrograman PHP dan database MySQL, Arsitektur jaringan yang digunakan aplikasi Computer Based Test adalah model jaringan client-server dengan jaringan Local Area Network (LAN. Kata kunci: Computer Based Test, Fisher-Yates Shuffle, Criptography, Local Area Network AbstractSelection of new student candidates can be done with Computer Based Test (CBT application. The methods used include data collection techniques, system analysis, design model, implementation and testing. This study produces a CBT application where the questions raised from the question bank through randomization process will not bring up the same problem using the Fisher-Yates Shuffle method. In the process of securing information about the problem when connected to the network it is necessary techniques for encoding the message so that the problem before appear through the process of encryption and description of data first then used RSA cryptography algorithm. Software design method using waterfall model, database design

  11. Identification of walking human model using agent-based modelling

    Science.gov (United States)

    Shahabpoor, Erfan; Pavic, Aleksandar; Racic, Vitomir

    2018-03-01

    The interaction of walking people with large vibrating structures, such as footbridges and floors, in the vertical direction is an important yet challenging phenomenon to describe mathematically. Several different models have been proposed in the literature to simulate interaction of stationary people with vibrating structures. However, the research on moving (walking) human models, explicitly identified for vibration serviceability assessment of civil structures, is still sparse. In this study, the results of a comprehensive set of FRF-based modal tests were used, in which, over a hundred test subjects walked in different group sizes and walking patterns on a test structure. An agent-based model was used to simulate discrete traffic-structure interactions. The occupied structure modal parameters found in tests were used to identify the parameters of the walking individual's single-degree-of-freedom (SDOF) mass-spring-damper model using 'reverse engineering' methodology. The analysis of the results suggested that the normal distribution with the average of μ = 2.85Hz and standard deviation of σ = 0.34Hz can describe human SDOF model natural frequency. Similarly, the normal distribution with μ = 0.295 and σ = 0.047 can describe the human model damping ratio. Compared to the previous studies, the agent-based modelling methodology proposed in this paper offers significant flexibility in simulating multi-pedestrian walking traffics, external forces and simulating different mechanisms of human-structure and human-environment interaction at the same time.

  12. Tests results of skutterudite based thermoelectric unicouples

    International Nuclear Information System (INIS)

    Saber, Hamed H.; El-Genk, Mohamed S.; Caillat, Thierry

    2007-01-01

    Tests were performed of skutterudite based unicouples with (MAY-04) and without (MAR-03) metallic coating on the legs near the hot junction to quantify the effect on reducing performance degradation with operation time. The p-legs in the unicouples were made of CeFe 3.5 Co 0.5 Sb 12 and the n-legs of CoSb 3 . The MAY-04 test was performed in vacuum (∼9 x 10 -7 torr) for ∼2000 h at hot and cold junction temperatures of 892.1 ± 11.9 K and 316.1 ± 5.5 K, respectively, while the MAR-03 test was performed in argon cover gas (0.051-0.068 MPa) at 972.61 ± 10.0 K and 301.1 ± 5.1 K, respectively. The argon cover gas decreased antimony loss from the legs in the MAR-03 test, but marked degradation in performance occurred over time. Conversely, the metallic coating in the MAY-04 test was very effective in reducing performance degradation of the unicouple. Because the cross sectional areas of the legs in MAY-04 were larger than those in MAR-03, the measured electrical power of the former is much higher than that of the latter, but the Beginning of Test (BOT) open circuit voltages, V oc (204.2 mV) for both unicouples were almost the same. The peak electrical power of the MAY-04 unicouple decreased 12.35% from 1.62W e at BOT to 1.42W e after ∼2000 h of testing, while that of the MAR-03 unicouple decreased 25.37% from 0.67 to 0.5W e after 261 h of testing at the above temperatures. The estimated peak efficiency of the MAY-04 unicouple, shortly after BOT (10.65%), was only ∼0.37% points lower than the theoretical value, calculated assuming zero side heat losses and zero contact resistance per leg

  13. Polynomial model inversion control: numerical tests and applications

    OpenAIRE

    Novara, Carlo

    2015-01-01

    A novel control design approach for general nonlinear systems is described in this paper. The approach is based on the identification of a polynomial model of the system to control and on the on-line inversion of this model. Extensive simulations are carried out to test the numerical efficiency of the approach. Numerical examples of applicative interest are presented, concerned with control of the Duffing oscillator, control of a robot manipulator and insulin regulation in a type 1 diabetic p...

  14. The Science Camp Model based on maker movement and tinkering activity for developing concept of electricity in middle school students to meet standard evaluation of ordinary national educational test (O-NET)

    Science.gov (United States)

    Chamrat, Suthida

    2018-01-01

    The standard evaluation of Thai education relies excessively on the Ordinary National Educational Test, widely known as O-NET. However, a focus on O-Net results can lead to unsatisfactory teaching practices, especially in science subjects. Among the negative consequences, is that schools frequently engage in "cramming" practices in order to elevate their O-NET scores. Higher education, which is committed to generating and applying knowledge by socially engaged scholars, needs to take account of this situation. This research article portrays the collaboration between the faculty of education at Chiang Mai University and an educational service area to develop the model of science camp. The activities designed for the Science Camp Model were based on the Tinkering and Maker Movement. Specifically, the Science Camp Model was designed to enhance the conceptualization of electricity for Middle School Students in order to meet the standard evaluation of the Ordinary National Educational Test. The hands-on activities consisted of 5 modules which were simple electrical circuits, paper circuits, electrical measurement roleplay motor art robots and Force from Motor. The data were collected by 11 items of Electricity Socratic-based Test adapted from cumulative published O-NET tests focused on the concept of electricity concept. The qualitative data were also collected virtually via Flinga.com. The results indicated that students after participating in 5modules of science camp based on the Maker Movement and tinkering activity developed average percentage of test scores from 33.64 to 65.45. Gain score analysis using dependent t-test compared pretest and posttest mean scores. The p value was found to be statistically significant (less than 0.001). The posttest had a considerably higher mean score compared with the pretest. Qualitative data also indicated that students could explain the main concepts of electrical circuits, and the transformation of electrical energy to

  15. Activity-based DEVS modeling

    DEFF Research Database (Denmark)

    Alshareef, Abdurrahman; Sarjoughian, Hessam S.; Zarrin, Bahram

    2018-01-01

    architecture and the UML concepts. In this paper, we further this work by grounding Activity-based DEVS modeling and developing a fully-fledged modeling engine to demonstrate applicability. We also detail the relevant aspects of the created metamodel in terms of modeling and simulation. A significant number......Use of model-driven approaches has been increasing to significantly benefit the process of building complex systems. Recently, an approach for specifying model behavior using UML activities has been devised to support the creation of DEVS models in a disciplined manner based on the model driven...... of the artifacts of the UML 2.5 activities and actions, from the vantage point of DEVS behavioral modeling, is covered in details. Their semantics are discussed to the extent of time-accurate requirements for simulation. We characterize them in correspondence with the specification of the atomic model behavior. We...

  16. Antirandom Testing: A Distance-Based Approach

    Directory of Open Access Journals (Sweden)

    Shen Hui Wu

    2008-01-01

    Full Text Available Random testing requires each test to be selected randomly regardless of the tests previously applied. This paper introduces the concept of antirandom testing where each test applied is chosen such that its total distance from all previous tests is maximum. This spans the test vector space to the maximum extent possible for a given number of vectors. An algorithm for generating antirandom tests is presented. Compared with traditional pseudorandom testing, antirandom testing is found to be very effective when a high-fault coverage needs to be achieved with a limited number of test vectors. The superiority of the new approach is even more significant for testing bridging faults.

  17. Development of an evaluation method for fracture mechanical tests on small samples based on a cohesive zone model; Entwicklung einer Auswertemethode fuer bruchmechanische Versuche an kleinen Proben auf der Basis eines Kohaesivzonenmodells

    Energy Technology Data Exchange (ETDEWEB)

    Mahler, Michael

    2016-07-01

    The safety and reliability of nuclear power plants of the fourth generation is an important issue. It is based on a reliable interpretation of the components for which, among other fracture mechanical material properties are required. The existing irradiation in the power plants significantly affects the material properties which therefore need to be determined on irradiated material. Often only small amounts of irradiated material are available for characterization. In that case it is not possible to manufacture sufficiently large specimens, which are necessary for fracture mechanical testing in agreement with the standard. Small specimens must be used. From this follows the idea of this study, in which the fracture toughness can be predicted with the developed method based on tests of small specimens. For this purpose, the fracture process including the crack growth is described with a continuum mechanical approach using the finite element method and the cohesive zone model. The experiments on small specimens are used for parameter identification of the cohesive zone model. The two parameters of the cohesive zone model are determined by tensile tests on notched specimens (cohesive stress) and by parameter fitting to the fracture behavior of smalls specimens (cohesive energy). To account the different triaxialities of the specimens, the cohesive stress is used depending on the triaxiality. After parameter identification a large specimen can be simulated with the cohesive zone parameters derived from small specimens. The predicted fracture toughness of this big specimen fulfills the size requirements in the standard (ASTM E1820 or ASTM E399) in contrast to the small specimen. This method can be used for ductile and brittle material behavior and was validated in this work. In summary, this method offers the possibility to determine the fracture toughness indirectly based on small specimen testing. Main advantage is the low required specimen volume. Thereby massively

  18. Consistency test of the standard model

    International Nuclear Information System (INIS)

    Pawlowski, M.; Raczka, R.

    1997-01-01

    If the 'Higgs mass' is not the physical mass of a real particle but rather an effective ultraviolet cutoff then a process energy dependence of this cutoff must be admitted. Precision data from at least two energy scale experimental points are necessary to test this hypothesis. The first set of precision data is provided by the Z-boson peak experiments. We argue that the second set can be given by 10-20 GeV e + e - colliders. We pay attention to the special role of tau polarization experiments that can be sensitive to the 'Higgs mass' for a sample of ∼ 10 8 produced tau pairs. We argue that such a study may be regarded as a negative selfconsistency test of the Standard Model and of most of its extensions

  19. Delay model and performance testing for FPGA carry chain TDC

    International Nuclear Information System (INIS)

    Kang Xiaowen; Liu Yaqiang; Cui Junjian Yang Zhangcan; Jin Yongjie

    2011-01-01

    Time-of-flight (TOF) information would improve the performance of PET (position emission tomography). TDC design is a key technique. It proposed Carry Chain TDC Delay model. Through changing the significant delay parameter of model, paper compared the difference of TDC performance, and finally realized Time-to-Digital Convertor (TDC) based on Carry Chain Method using FPGA EP2C20Q240C8N with 69 ps LSB, max error below 2 LSB. Such result could meet the TOF demand. It also proposed a Coaxial Cable Measuring method for TDC testing, without High-precision test equipment. (authors)

  20. Simulation of thermohydraulic phenomena and model test for FBR

    International Nuclear Information System (INIS)

    Satoh, Kazuziro

    1994-01-01

    This paper summarizes the major thermohydraulic phenomena of FBRs and the conventional ways of their model tests, and introduces the recent findings regarding measurement technology and computational science. In the future commercial stage of FBRs, the design optimization will becomes important to improve economy and safety more and more. It is indispensable to use computational science to the plant design and safety evaluation. The most of the model tests will be replaced by the simulation analyses based on computational science. The measurement technology using ultrasonic and the numerical simulation with super parallel computing are considered to be the key technology to realize the design by analysis method. (author)

  1. Corneal topographer based on the Hartmann test.

    Science.gov (United States)

    Mejía, Yobani; Galeano, Janneth C

    2009-04-01

    The purpose of this article is to show the performance of a topographer based on the Hartmann test for convex surfaces of F/# approximately 1. This topographer, called "Hartmann Test topographer (HT topographer)," is a prototype developed in the Physics Department of the Universidad Nacional de Colombia. From the Hartmann pattern generated by the surface under test, and by the Fourier analysis and the optical aberration theory we obtain the sagitta (elevation map) of the surface. Then, taking the first and the second derivatives of the sagitta in the radial direction we obtain the meridional curvature map. The method is illustrated with an example. To check the performance of the HT topographer a toric surface, a revolution aspherical surface, and two human corneas were measured. Our results are compared with those obtained with a Placido ring topographer (Tomey TMS-4 videokeratoscope), and we show that our curvature maps are similar to those obtained with the Placido ring topographer. The HT topographer is able to reconstruct the corneal topography potentially eradicating the skew ray problem, therefore, corneal defects can be visualized more. The results are presented by elevation and meridional curvature maps.

  2. Testing Software Development Project Productivity Model

    Science.gov (United States)

    Lipkin, Ilya

    Software development is an increasingly influential factor in today's business environment, and a major issue affecting software development is how an organization estimates projects. If the organization underestimates cost, schedule, and quality requirements, the end results will not meet customer needs. On the other hand, if the organization overestimates these criteria, resources that could have been used more profitably will be wasted. There is no accurate model or measure available that can guide an organization in a quest for software development, with existing estimation models often underestimating software development efforts as much as 500 to 600 percent. To address this issue, existing models usually are calibrated using local data with a small sample size, with resulting estimates not offering improved cost analysis. This study presents a conceptual model for accurately estimating software development, based on an extensive literature review and theoretical analysis based on Sociotechnical Systems (STS) theory. The conceptual model serves as a solution to bridge organizational and technological factors and is validated using an empirical dataset provided by the DoD. Practical implications of this study allow for practitioners to concentrate on specific constructs of interest that provide the best value for the least amount of time. This study outlines key contributing constructs that are unique for Software Size E-SLOC, Man-hours Spent, and Quality of the Product, those constructs having the largest contribution to project productivity. This study discusses customer characteristics and provides a framework for a simplified project analysis for source selection evaluation and audit task reviews for the customers and suppliers. Theoretical contributions of this study provide an initial theory-based hypothesized project productivity model that can be used as a generic overall model across several application domains such as IT, Command and Control

  3. Tests on thirteen navy type model propellers

    Science.gov (United States)

    Durand, W F

    1927-01-01

    The tests on these model propellers were undertaken for the purpose of determining the performance coefficients and characteristics for certain selected series of propellers of form and type as commonly used in recent navy designs. The first series includes seven propellers of pitch ratio varying by 0.10 to 1.10, the area, form of blade, thickness, etc., representing an arbitrary standard propeller which had shown good results. The second series covers changes in thickness of blade section, other things equal, and the third series, changes in blade area, other things equal. These models are all of 36-inch diameter. Propellers A to G form the series on pitch ratio, C, N. I. J the series on thickness of section, and K, M, C, L the series on area. (author)

  4. Stochastic models for strength of wind turbine blades using tests

    DEFF Research Database (Denmark)

    Toft, H.S.; Sørensen, John Dalsgaard

    2008-01-01

    The structural cost of wind turbine blades is dependent on the values of the partial safety factors which reflect the uncertainties in the design values, including statistical uncertainty from a limited number of tests. This paper presents a probabilistic model for ultimate and fatigue strength...... of wind turbine blades especially considering the influence of prior knowledge and test results and how partial safety factors can be updated when additional full-scale tests are performed. This updating is performed by adopting a probabilistic design basis based on Bayesian statistical methods....

  5. Gradient-based model calibration with proxy-model assistance

    Science.gov (United States)

    Burrows, Wesley; Doherty, John

    2016-02-01

    Use of a proxy model in gradient-based calibration and uncertainty analysis of a complex groundwater model with large run times and problematic numerical behaviour is described. The methodology is general, and can be used with models of all types. The proxy model is based on a series of analytical functions that link all model outputs used in the calibration process to all parameters requiring estimation. In enforcing history-matching constraints during the calibration and post-calibration uncertainty analysis processes, the proxy model is run for the purposes of populating the Jacobian matrix, while the original model is run when testing parameter upgrades; the latter process is readily parallelized. Use of a proxy model in this fashion dramatically reduces the computational burden of complex model calibration and uncertainty analysis. At the same time, the effect of model numerical misbehaviour on calculation of local gradients is mitigated, this allowing access to the benefits of gradient-based analysis where lack of integrity in finite-difference derivatives calculation would otherwise have impeded such access. Construction of a proxy model, and its subsequent use in calibration of a complex model, and in analysing the uncertainties of predictions made by that model, is implemented in the PEST suite.

  6. Event-Based Conceptual Modeling

    DEFF Research Database (Denmark)

    Bækgaard, Lars

    The paper demonstrates that a wide variety of event-based modeling approaches are based on special cases of the same general event concept, and that the general event concept can be used to unify the otherwise unrelated fields of information modeling and process modeling. A set of event......-based modeling approaches are analyzed and the results are used to formulate a general event concept that can be used for unifying the seemingly unrelated event concepts. Events are characterized as short-duration processes that have participants, consequences, and properties, and that may be modeled in terms...... of information structures. The general event concept can be used to guide systems analysis and design and to improve modeling approaches....

  7. Space Launch System Base Heating Test: Environments and Base Flow Physics

    Science.gov (United States)

    Mehta, Manish; Knox, Kyle S.; Seaford, C. Mark; Dufrene, Aaron T.

    2016-01-01

    The NASA Space Launch System (SLS) vehicle is composed of four RS-25 liquid oxygen- hydrogen rocket engines in the core-stage and two 5-segment solid rocket boosters and as a result six hot supersonic plumes interact within the aft section of the vehicle during ight. Due to the complex nature of rocket plume-induced ows within the launch vehicle base during ascent and a new vehicle con guration, sub-scale wind tunnel testing is required to reduce SLS base convective environment uncertainty and design risk levels. This hot- re test program was conducted at the CUBRC Large Energy National Shock (LENS) II short-duration test facility to simulate ight from altitudes of 50 kft to 210 kft. The test program is a challenging and innovative e ort that has not been attempted in 40+ years for a NASA vehicle. This presentation discusses the various trends of base convective heat ux and pressure as a function of altitude at various locations within the core-stage and booster base regions of the two-percent SLS wind tunnel model. In-depth understanding of the base ow physics is presented using the test data, infrared high-speed imaging and theory. The normalized test design environments are compared to various NASA semi- empirical numerical models to determine exceedance and conservatism of the ight scaled test-derived base design environments. Brief discussion of thermal impact to the launch vehicle base components is also presented.

  8. Wedge-Splitting Test – Determination of Minimal Starting Notch Length for Various Cement Based Composites. Part I: Cohesive Crack Modelling

    Czech Academy of Sciences Publication Activity Database

    Veselý, V.; Řoutil, L.; Seitl, Stanislav

    2011-01-01

    Roč. 452-453, - (2011), s. 77-80 ISSN 1013-9826 R&D Projects: GA AV ČR KJB200410901 Institutional research plan: CEZ:AV0Z20410507 Keywords : wedge-splitting test * cementitious composites * quasi-brittle fracture * brittleness Subject RIV: JL - Materials Fatigue, Friction Mechanics

  9. SPSS and SAS programming for the testing of mediation models.

    Science.gov (United States)

    Dudley, William N; Benuzillo, Jose G; Carrico, Mineh S

    2004-01-01

    Mediation modeling can explain the nature of the relation among three or more variables. In addition, it can be used to show how a variable mediates the relation between levels of intervention and outcome. The Sobel test, developed in 1990, provides a statistical method for determining the influence of a mediator on an intervention or outcome. Although interactive Web-based and stand-alone methods exist for computing the Sobel test, SPSS and SAS programs that automatically run the required regression analyses and computations increase the accessibility of mediation modeling to nursing researchers. To illustrate the utility of the Sobel test and to make this programming available to the Nursing Research audience in both SAS and SPSS. The history, logic, and technical aspects of mediation testing are introduced. The syntax files sobel.sps and sobel.sas, created to automate the computation of the regression analysis and test statistic, are available from the corresponding author. The reported programming allows the user to complete mediation testing with the user's own data in a single-step fashion. A technical manual included with the programming provides instruction on program use and interpretation of the output. Mediation modeling is a useful tool for describing the relation between three or more variables. Programming and manuals for using this model are made available.

  10. HMM-based Trust Model

    DEFF Research Database (Denmark)

    ElSalamouny, Ehab; Nielsen, Mogens; Sassone, Vladimiro

    2010-01-01

    Probabilistic trust has been adopted as an approach to taking security sensitive decisions in modern global computing environments. Existing probabilistic trust frameworks either assume fixed behaviour for the principals or incorporate the notion of ‘decay' as an ad hoc approach to cope...... with their dynamic behaviour. Using Hidden Markov Models (HMMs) for both modelling and approximating the behaviours of principals, we introduce the HMM-based trust model as a new approach to evaluating trust in systems exhibiting dynamic behaviour. This model avoids the fixed behaviour assumption which is considered...... the major limitation of existing Beta trust model. We show the consistency of the HMM-based trust model and contrast it against the well known Beta trust model with the decay principle in terms of the estimation precision....

  11. Development of a fault test experimental facility model using Matlab

    Energy Technology Data Exchange (ETDEWEB)

    Pereira, Iraci Martinez; Moraes, Davi Almeida, E-mail: martinez@ipen.br, E-mail: dmoraes@dk8.com.br [Instituto de Pesquisas Energeticas e Nucleares (IPEN/CNEN-SP), Sao Paulo, SP (Brazil)

    2015-07-01

    The Fault Test Experimental Facility was developed to simulate a PWR nuclear power plant and is instrumented with temperature, level and pressure sensors. The Fault Test Experimental Facility can be operated to generate normal and fault data, and these failures can be added initially small, and their magnitude being increasing gradually. This work presents the Fault Test Experimental Facility model developed using the Matlab GUIDE (Graphical User Interface Development Environment) toolbox that consists of a set of functions designed to create interfaces in an easy and fast way. The system model is based on the mass and energy inventory balance equations. Physical as well as operational aspects are taken into consideration. The interface layout looks like a process flowchart and the user can set the input variables. Besides the normal operation conditions, there is the possibility to choose a faulty variable from a list. The program also allows the user to set the noise level for the input variables. Using the model, data were generated for different operational conditions, both under normal and fault conditions with different noise levels added to the input variables. Data generated by the model will be compared with Fault Test Experimental Facility data. The Fault Test Experimental Facility theoretical model results will be used for the development of a Monitoring and Fault Detection System. (author)

  12. Development of a fault test experimental facility model using Matlab

    International Nuclear Information System (INIS)

    Pereira, Iraci Martinez; Moraes, Davi Almeida

    2015-01-01

    The Fault Test Experimental Facility was developed to simulate a PWR nuclear power plant and is instrumented with temperature, level and pressure sensors. The Fault Test Experimental Facility can be operated to generate normal and fault data, and these failures can be added initially small, and their magnitude being increasing gradually. This work presents the Fault Test Experimental Facility model developed using the Matlab GUIDE (Graphical User Interface Development Environment) toolbox that consists of a set of functions designed to create interfaces in an easy and fast way. The system model is based on the mass and energy inventory balance equations. Physical as well as operational aspects are taken into consideration. The interface layout looks like a process flowchart and the user can set the input variables. Besides the normal operation conditions, there is the possibility to choose a faulty variable from a list. The program also allows the user to set the noise level for the input variables. Using the model, data were generated for different operational conditions, both under normal and fault conditions with different noise levels added to the input variables. Data generated by the model will be compared with Fault Test Experimental Facility data. The Fault Test Experimental Facility theoretical model results will be used for the development of a Monitoring and Fault Detection System. (author)

  13. The cycle use test of Pt based catalyst for the steam reforming of naphthalene / benzene as model tar compounds of biomass gasification

    Energy Technology Data Exchange (ETDEWEB)

    Furusawa, Takeshi; Saito, Katsuhiko; Sato, Masahide; Suzuki, Noboru [Utsunomiya Univ. (Japan). Graduate School of Engineering

    2010-07-01

    Although Pt/Al{sub 2}O{sub 3} catalyst showed high and stable activity (carbon conv. to gas: 90%) for steam reforming of naphthalene/benzene at 1073 K with S/C=3, this catalyst gradually lost its activity at 1023 K with S/C=3 due to deposition of carboneous species. Two kinds of regeneration treatment was conducted to enlongate the life time of Pt/Al{sub 2}O{sub 3} catalyst. Although regeneration treatment completely remove the carboneous species from catalyst, mild oxidation treatment led to decrease activity due to sintering of Pt particles. On the contrary, hydrogen treatment led to maintain activity until 5th cycle test. It was concluded from these obtained results that hydrogen treatment is suitable regeneration method during cycle test in the case of Pt/Al{sub 2}O{sub 3} catalyst. (orig.)

  14. Canine distemper virus (CDV) infection of ferrets as a model for testing Morbillivirus vaccine strategies: NYVAC- and ALVAC-based CDV recombinants protect against symptomatic infection.

    OpenAIRE

    Stephensen, C B; Welter, J; Thaker, S R; Taylor, J; Tartaglia, J; Paoletti, E

    1997-01-01

    Canine distemper virus (CDV) infection of ferrets causes an acute systemic disease involving multiple organ systems, including the respiratory tract, lymphoid system, and central nervous system (CNS). We have tested candidate CDV vaccines incorporating the fusion (F) and hemagglutinin (HA) proteins in the highly attenuated NYVAC strain of vaccinia virus and in the ALVAC strain of canarypox virus, which does not productively replicate in mammalian hosts. Juvenile ferrets were vaccinated twice ...

  15. A magnetorheological actuation system: test and model

    International Nuclear Information System (INIS)

    John, Shaju; Chaudhuri, Anirban; Wereley, Norman M

    2008-01-01

    Self-contained actuation systems, based on frequency rectification of the high frequency motion of an active material, can produce high force and stroke output. Magnetorheological (MR) fluids are active fluids whose rheological properties can be altered by the application of a magnetic field. By using MR fluids as the energy transmission medium in such hybrid devices, a valving system with no moving parts can be implemented and used to control the motion of an output cylinder shaft. The MR fluid based valves are configured in the form of an H-bridge to produce bi-directional motion in an output cylinder by alternately applying magnetic fields in the two opposite arms of the bridge. The rheological properties of the MR fluid are modeled using both Bingham plastic and bi-viscous models. In this study, the primary actuation is performed using a compact terfenol-D rod driven pump and frequency rectification of the rod motion is done using passive reed valves. The pump and reed valve configuration along with MR fluidic valves form a compact hydraulic actuation system. Actuator design, analysis and experimental results are presented in this paper. A time domain model of the actuator is developed and validated using experimental data

  16. Experimentally testing the standard cosmological model

    Energy Technology Data Exchange (ETDEWEB)

    Schramm, D.N. (Chicago Univ., IL (USA) Fermi National Accelerator Lab., Batavia, IL (USA))

    1990-11-01

    The standard model of cosmology, the big bang, is now being tested and confirmed to remarkable accuracy. Recent high precision measurements relate to the microwave background; and big bang nucleosynthesis. This paper focuses on the latter since that relates more directly to high energy experiments. In particular, the recent LEP (and SLC) results on the number of neutrinos are discussed as a positive laboratory test of the standard cosmology scenario. Discussion is presented on the improved light element observational data as well as the improved neutron lifetime data. alternate nucleosynthesis scenarios of decaying matter or of quark-hadron induced inhomogeneities are discussed. It is shown that when these scenarios are made to fit the observed abundances accurately, the resulting conclusions on the baryonic density relative to the critical density, {Omega}{sub b}, remain approximately the same as in the standard homogeneous case, thus, adding to the robustness of the standard model conclusion that {Omega}{sub b} {approximately} 0.06. This latter point is the deriving force behind the need for non-baryonic dark matter (assuming {Omega}{sub total} = 1) and the need for dark baryonic matter, since {Omega}{sub visible} < {Omega}{sub b}. Recent accelerator constraints on non-baryonic matter are discussed, showing that any massive cold dark matter candidate must now have a mass M{sub x} {approx gt} 20 GeV and an interaction weaker than the Z{sup 0} coupling to a neutrino. It is also noted that recent hints regarding the solar neutrino experiments coupled with the see-saw model for {nu}-masses may imply that the {nu}{sub {tau}} is a good hot dark matter candidate. 73 refs., 5 figs.

  17. Experimentally testing the standard cosmological model

    International Nuclear Information System (INIS)

    Schramm, D.N.

    1990-11-01

    The standard model of cosmology, the big bang, is now being tested and confirmed to remarkable accuracy. Recent high precision measurements relate to the microwave background; and big bang nucleosynthesis. This paper focuses on the latter since that relates more directly to high energy experiments. In particular, the recent LEP (and SLC) results on the number of neutrinos are discussed as a positive laboratory test of the standard cosmology scenario. Discussion is presented on the improved light element observational data as well as the improved neutron lifetime data. alternate nucleosynthesis scenarios of decaying matter or of quark-hadron induced inhomogeneities are discussed. It is shown that when these scenarios are made to fit the observed abundances accurately, the resulting conclusions on the baryonic density relative to the critical density, Ω b , remain approximately the same as in the standard homogeneous case, thus, adding to the robustness of the standard model conclusion that Ω b ∼ 0.06. This latter point is the deriving force behind the need for non-baryonic dark matter (assuming Ω total = 1) and the need for dark baryonic matter, since Ω visible b . Recent accelerator constraints on non-baryonic matter are discussed, showing that any massive cold dark matter candidate must now have a mass M x approx-gt 20 GeV and an interaction weaker than the Z 0 coupling to a neutrino. It is also noted that recent hints regarding the solar neutrino experiments coupled with the see-saw model for ν-masses may imply that the ν τ is a good hot dark matter candidate. 73 refs., 5 figs

  18. Modeling Guru: Knowledge Base for NASA Modelers

    Science.gov (United States)

    Seablom, M. S.; Wojcik, G. S.; van Aartsen, B. H.

    2009-05-01

    Modeling Guru is an on-line knowledge-sharing resource for anyone involved with or interested in NASA's scientific models or High End Computing (HEC) systems. Developed and maintained by the NASA's Software Integration and Visualization Office (SIVO) and the NASA Center for Computational Sciences (NCCS), Modeling Guru's combined forums and knowledge base for research and collaboration is becoming a repository for the accumulated expertise of NASA's scientific modeling and HEC communities. All NASA modelers and associates are encouraged to participate and provide knowledge about the models and systems so that other users may benefit from their experience. Modeling Guru is divided into a hierarchy of communities, each with its own set forums and knowledge base documents. Current modeling communities include those for space science, land and atmospheric dynamics, atmospheric chemistry, and oceanography. In addition, there are communities focused on NCCS systems, HEC tools and libraries, and programming and scripting languages. Anyone may view most of the content on Modeling Guru (available at http://modelingguru.nasa.gov/), but you must log in to post messages and subscribe to community postings. The site offers a full range of "Web 2.0" features, including discussion forums, "wiki" document generation, document uploading, RSS feeds, search tools, blogs, email notification, and "breadcrumb" links. A discussion (a.k.a. forum "thread") is used to post comments, solicit feedback, or ask questions. If marked as a question, SIVO will monitor the thread, and normally respond within a day. Discussions can include embedded images, tables, and formatting through the use of the Rich Text Editor. Also, the user can add "Tags" to their thread to facilitate later searches. The "knowledge base" is comprised of documents that are used to capture and share expertise with others. The default "wiki" document lets users edit within the browser so others can easily collaborate on the

  19. Brain regions engaged by part- and whole-task performance in a video game: a model-based test of the decomposition hypothesis.

    Science.gov (United States)

    Anderson, John R; Bothell, Daniel; Fincham, Jon M; Anderson, Abraham R; Poole, Ben; Qin, Yulin

    2011-12-01

    Part- and whole-task conditions were created by manipulating the presence of certain components of the Space Fortress video game. A cognitive model was created for two-part games that could be combined into a model that performed the whole game. The model generated predictions both for behavioral patterns and activation patterns in various brain regions. The activation predictions concerned both tonic activation that was constant in these regions during performance of the game and phasic activation that occurred when there was resource competition. The model's predictions were confirmed about how tonic and phasic activation in different regions would vary with condition. These results support the Decomposition Hypothesis that the execution of a complex task can be decomposed into a set of information-processing components and that these components combine unchanged in different task conditions. In addition, individual differences in learning gains were predicted by individual differences in phasic activation in those regions that displayed highest tonic activity. This individual difference pattern suggests that the rate of learning of a complex skill is determined by capacity limits.

  20. Overheating Anomalies during Flight Test Due to the Base Bleeding

    Science.gov (United States)

    Luchinsky, Dmitry; Hafiychuck, Halyna; Osipov, Slava; Ponizhovskaya, Ekaterina; Smelyanskiy, Vadim; Dagostino, Mark; Canabal, Francisco; Mobley, Brandon L.

    2012-01-01

    In this paper we present the results of the analytical and numerical studies of the plume interaction with the base flow in the presence of base out-gassing. The physics-based analysis and CFD modeling of the base heating for single solid rocket motor performed in this research addressed the following questions: what are the key factors making base flow so different from that in the Shuttle [1]; why CFD analysis of this problem reveals small plume recirculation; what major factors influence base temperature; and why overheating was initiated at a given time in the flight. To answer these questions topological analysis of the base flow was performed and Korst theory was used to estimate relative contributions of radiation, plume recirculation, and chemically reactive out-gassing to the base heating. It was shown that base bleeding and small base volume are the key factors contributing to the overheating, while plume recirculation is effectively suppressed by asymmetric configuration of the flow formed earlier in the flight. These findings are further verified using CFD simulations that include multi-species gas environment both in the plume and in the base. Solid particles in the exhaust plume (Al2O3) and char particles in the base bleeding were also included into the simulations and their relative contributions into the base temperature rise were estimated. The results of simulations are in good agreement with the temperature and pressure in the base measured during the test.

  1. NEIGHBORHOOD TEST DESIGN BASED ON HISTORIC PRECEDENTS

    Directory of Open Access Journals (Sweden)

    Besim S. Hakim

    2012-07-01

    Full Text Available There have been various attempts to emulate traditional architecture and to experiment with the form and aesthetics of building design. However, learning from precedents of urban morphology is rare. This design study is a test at the neighborhood level using the pattern of traditional courtyard housing that is prevalent in the majority of historic towns and cities of North Africa and the Middle East. The study is undertaken at five levels of design enquiry: dwelling types, dwelling groups, neighborhood segment and community center. All of which are synthesized into a full prototype neighborhood comprising of 428 dwelling units covering an area that includes circulation and the community center, of 17.6 hectares. The test demonstrates that the traditional pattern of neighborhoods that are based on the typology of the courtyard dwelling as the initial generator of urban form may be used to develop a contemporary settlement pattern that is compatible with current necessities of lifestyle, vehicular circulation,  including parking and infrastructure achieving an attractive livable environment with an overall gross density, that includes a community center, of about 24 dwelling units per hectare.

  2. Structure-Based Turbulence Model

    National Research Council Canada - National Science Library

    Reynolds, W

    2000-01-01

    .... Maire carried out this work as part of his Phi) research. During the award period we began to explore ways to simplify the structure-based modeling so that it could be used in repetitive engineering calculations...

  3. Shaking table test of a base isolated model in main control room of nuclear power plant using LRB (lead rubber bearing)

    International Nuclear Information System (INIS)

    Ham, K. W.; Lee, K. J.; Suh, Y. P.

    2005-01-01

    LRB(Lead Rubber Bearing) is a widely used isolation system which is installed between equipment and foundation to reduce seismic vibration from ground. LRB is consist of bearings which are resistant to lateral motion and torsion and has a high vertical stiffness. For that reason, several studies are conducted to apply LRB to the nuclear power plant. In this study, we designed two types of main control floor systems (type I, type II) and a number of shaking table tests with and without isolation system were conducted to evaluate floor isolation effectiveness of LRB

  4. Event-Based Conceptual Modeling

    DEFF Research Database (Denmark)

    Bækgaard, Lars

    2009-01-01

    The purpose of the paper is to obtain insight into and provide practical advice for event-based conceptual modeling. We analyze a set of event concepts and use the results to formulate a conceptual event model that is used to identify guidelines for creation of dynamic process models and static...... information models. We characterize events as short-duration processes that have participants, consequences, and properties, and that may be modeled in terms of information structures. The conceptual event model is used to characterize a variety of event concepts and it is used to illustrate how events can...... be used to integrate dynamic modeling of processes and static modeling of information structures. The results are unique in the sense that no other general event concept has been used to unify a similar broad variety of seemingly incompatible event concepts. The general event concept can be used...

  5. Testing an integral conceptual model of frailty.

    Science.gov (United States)

    Gobbens, Robbert J; van Assen, Marcel A; Luijkx, Katrien G; Schols, Jos M

    2012-09-01

    This paper is a report of a study conducted to test three hypotheses derived from an integral conceptual model of frailty.   The integral model of frailty describes the pathway from life-course determinants to frailty to adverse outcomes. The model assumes that life-course determinants and the three domains of frailty (physical, psychological, social) affect adverse outcomes, the effect of disease(s) on adverse outcomes is mediated by frailty, and the effect of frailty on adverse outcomes depends on the life-course determinants. In June 2008 a questionnaire was sent to a sample of community-dwelling people, aged 75 years and older (n = 213). Life-course determinants and frailty were assessed using the Tilburg frailty indicator. Adverse outcomes were measured using the Groningen activity restriction scale, the WHOQOL-BREF and questions regarding healthcare utilization. The effect of seven self-reported chronic diseases was examined. Life-course determinants, chronic disease(s), and frailty together explain a moderate to large part of the variance of the seven continuous adverse outcomes (26-57%). All these predictors together explained a significant part of each of the five dichotomous adverse outcomes. The effect of chronic disease(s) on all 12 adverse outcomes was mediated at least partly by frailty. The effect of frailty domains on adverse outcomes did not depend on life-course determinants. Our finding that the adverse outcomes are differently and uniquely affected by the three domains of frailty (physical, psychological, social), and life-course determinants and disease(s), emphasizes the importance of an integral conceptual model of frailty. © 2011 Blackwell Publishing Ltd.

  6. Computer Based Modelling and Simulation

    Indian Academy of Sciences (India)

    Home; Journals; Resonance – Journal of Science Education; Volume 6; Issue 3. Computer Based Modelling and Simulation - Modelling Deterministic Systems. N K Srinivasan. General Article Volume 6 Issue 3 March 2001 pp 46-54. Fulltext. Click here to view fulltext PDF. Permanent link:

  7. Springer handbook of model-based science

    CERN Document Server

    Bertolotti, Tommaso

    2017-01-01

    The handbook offers the first comprehensive reference guide to the interdisciplinary field of model-based reasoning. It highlights the role of models as mediators between theory and experimentation, and as educational devices, as well as their relevance in testing hypotheses and explanatory functions. The Springer Handbook merges philosophical, cognitive and epistemological perspectives on models with the more practical needs related to the application of this tool across various disciplines and practices. The result is a unique, reliable source of information that guides readers toward an understanding of different aspects of model-based science, such as the theoretical and cognitive nature of models, as well as their practical and logical aspects. The inferential role of models in hypothetical reasoning, abduction and creativity once they are constructed, adopted, and manipulated for different scientific and technological purposes is also discussed. Written by a group of internationally renowned experts in ...

  8. Updating the Duplex Design for Test-Based Accountability in the Twenty-First Century

    Science.gov (United States)

    Bejar, Isaac I.; Graf, E. Aurora

    2010-01-01

    The duplex design by Bock and Mislevy for school-based testing is revisited and evaluated as a potential platform in test-based accountability assessments today. We conclude that the model could be useful in meeting the many competing demands of today's test-based accountability assessments, although many research questions will need to be…

  9. Laboratory test of an APS-based sun sensor prototype

    Science.gov (United States)

    Rufino, Giancarlo; Perrotta, Alessandro; Grassi, Michele

    2017-11-01

    This paper deals with design and prototype development of an Active Pixel Sensor - based miniature sun sensor and a laboratory facility for its indoor test and calibration. The miniature sun sensor is described and the laboratory test facility is presented in detail. The major focus of the paper is on tests and calibration of the sensor. Two different calibration functions have been adopted. They are based, respectively, on a geometrical model, which has required least-squares optimisation of system physical parameters estimates, and on neural networks. Calibration results are presented for the above solutions, showing that accuracy in the order of 0.01° has been achieved. Neural calibration functions have attained better performance thanks to their intrinsic auto-adaptive structure.

  10. Experimental Tests of the Algebraic Cluster Model

    Science.gov (United States)

    Gai, Moshe

    2018-02-01

    The Algebraic Cluster Model (ACM) of Bijker and Iachello that was proposed already in 2000 has been recently applied to 12C and 16O with much success. We review the current status in 12C with the outstanding observation of the ground state rotational band composed of the spin-parity states of: 0+, 2+, 3-, 4± and 5-. The observation of the 4± parity doublet is a characteristic of (tri-atomic) molecular configuration where the three alpha- particles are arranged in an equilateral triangular configuration of a symmetric spinning top. We discuss future measurement with electron scattering, 12C(e,e’) to test the predicted B(Eλ) of the ACM.

  11. Social inequality and HIV-testing: Comparing home- and clinic-based testing in rural Malawi

    Directory of Open Access Journals (Sweden)

    Alexander A. Weinreb

    2009-10-01

    Full Text Available The plan to increase HIV testing is a cornerstone of the international health strategy against the HIV/AIDS epidemic, particularly in sub-Saharan Africa. This paper highlights a problematic aspect of that plan: the reliance on clinic- rather than home-based testing. First, drawing on DHS data from across Africa, we demonstrate the substantial differences in socio-demographic and economic profiles between those who report having ever had an HIV test, and those who report never having had one. Then, using data from a random household survey in rural Malawi, we show that substituting home-based for clinic-based testing may eliminate this source of inequality between those tested and those not tested. This result, which is stable across modeling frameworks, has important implications for accurately and equitably addressing the counseling and treatment programs that comprise the international health strategy against AIDS, and that promise to shape the future trajectory of the epidemic in Africa and beyond.

  12. SABATPG-A Structural Analysis Based Automatic Test Generation System

    Institute of Scientific and Technical Information of China (English)

    李忠诚; 潘榆奇; 闵应骅

    1994-01-01

    A TPG system, SABATPG, is given based on a generic structural model of large circuits. Three techniques of partial implication, aftereffect of identified undetectable faults and shared sensitization with new concepts of localization and aftereffect are employed in the system to improve FAN algorithm. Experiments for the 10 ISCAS benchmark circuits show that the computing time of SABATPG for test generation is 19.42% less than that of FAN algorithm.

  13. Predictors of HIV-test utilization in PMTCT among antenatal care attendees in government health centers: institution-based cross-sectional study using health belief model in Addis Ababa, Ethiopia, 2013

    Directory of Open Access Journals (Sweden)

    Workagegn F

    2015-07-01

    Full Text Available Fikremariam Workagegn, Getachew Kiros, Lakew Abebe Health Education and Behavioral Sciences Department, Public and Medical Sciences College, Jimma University, Jimma, Ethiopia Background: Human immunodeficiency virus (HIV/acquired immune deficiency syndrome (AIDS is the most dramatic epidemic of the century that has claimed over two decades more than 3 million deaths. Sub-Saharan Africa is heavily affected and accounts for nearly 70% of all cases. Mother-to-child transmission of HIV is responsible for 20% of all HIV transmissions. With no preventive interventions, 50% of HIV infections are transmitted from HIV-positive mothers to newborns. HIV-testing is central to prevent vertical transmission. Despite, awareness campaigns, prevention measures, and more recently, promotion of antiviral regimens, the prevalence of cases and deaths is still rising and the prevalence of prevention of mother-to-child transmission (PMTCT voluntary counseling test (VCT use remains low. This study identifies predictors and possible barriers of HIV-testing among antenatal care attendees based on the health belief model (HBM in Addis Ababa, Ethiopia. Methods: The study was an institution-based cross-sectional survey conducted from September 1 to September 30, 2013. A total of 308 individuals were interviewed using structured questionnaires adopted and modified from similar studies. Data were collected through face-to-face interviews. A logistic regression was used to identify factors associated with HIV-test use. Results: In spite of satisfactory knowledge on HIV/AIDS transmission, participants are still at high risk of contracting the infection, wherein only 51.8% tested for HIV; among the married, only 84.1% and among the gestational age of third trimester, 34.1% mothers tested for HIV. Based on the HBM, failure to use PMTCT-HIV-test was related to its perceived lack of net benefit (adjusted odds ratio [AOR] =0.34, confidence interval [CI] [0.19–0.58], P<0.001, but

  14. Two Bayesian tests of the GLOMOsys Model.

    Science.gov (United States)

    Field, Sarahanne M; Wagenmakers, Eric-Jan; Newell, Ben R; Zeelenberg, René; van Ravenzwaaij, Don

    2016-12-01

    Priming is arguably one of the key phenomena in contemporary social psychology. Recent retractions and failed replication attempts have led to a division in the field between proponents and skeptics and have reinforced the importance of confirming certain priming effects through replication. In this study, we describe the results of 2 preregistered replication attempts of 1 experiment by Förster and Denzler (2012). In both experiments, participants first processed letters either globally or locally, then were tested using a typicality rating task. Bayes factor hypothesis tests were conducted for both experiments: Experiment 1 (N = 100) yielded an indecisive Bayes factor of 1.38, indicating that the in-lab data are 1.38 times more likely to have occurred under the null hypothesis than under the alternative. Experiment 2 (N = 908) yielded a Bayes factor of 10.84, indicating strong support for the null hypothesis that global priming does not affect participants' mean typicality ratings. The failure to replicate this priming effect challenges existing support for the GLOMO sys model. (PsycINFO Database Record (c) 2016 APA, all rights reserved).

  15. Experimental tests of the standard model

    International Nuclear Information System (INIS)

    Nodulman, L.

    1998-01-01

    The title implies an impossibly broad field, as the Standard Model includes the fermion matter states, as well as the forces and fields of SU(3) x SU(2) x U(1). For practical purposes, I will confine myself to electroweak unification, as discussed in the lectures of M. Herrero. Quarks and mixing were discussed in the lectures of R. Aleksan, and leptons and mixing were discussed in the lectures of K. Nakamura. I will essentially assume universality, that is flavor independence, rather than discussing tests of it. I will not pursue tests of QED beyond noting the consistency and precision of measurements of α EM in various processes including the Lamb shift, the anomalous magnetic moment (g-2) of the electron, and the quantum Hall effect. The fantastic precision and agreement of these predictions and measurements is something that convinces people that there may be something to this science enterprise. Also impressive is the success of the ''Universal Fermi Interaction'' description of beta decay processes, or in more modern parlance, weak charged current interactions. With one coupling constant G F , most precisely determined in muon decay, a huge number of nuclear instabilities are described. The slightly slow rate for neutron beta decay was one of the initial pieces of evidence for Cabbibo mixing, now generalized so that all charged current decays of any flavor are covered

  16. Bayes Factor Covariance Testing in Item Response Models.

    Science.gov (United States)

    Fox, Jean-Paul; Mulder, Joris; Sinharay, Sandip

    2017-12-01

    Two marginal one-parameter item response theory models are introduced, by integrating out the latent variable or random item parameter. It is shown that both marginal response models are multivariate (probit) models with a compound symmetry covariance structure. Several common hypotheses concerning the underlying covariance structure are evaluated using (fractional) Bayes factor tests. The support for a unidimensional factor (i.e., assumption of local independence) and differential item functioning are evaluated by testing the covariance components. The posterior distribution of common covariance components is obtained in closed form by transforming latent responses with an orthogonal (Helmert) matrix. This posterior distribution is defined as a shifted-inverse-gamma, thereby introducing a default prior and a balanced prior distribution. Based on that, an MCMC algorithm is described to estimate all model parameters and to compute (fractional) Bayes factor tests. Simulation studies are used to show that the (fractional) Bayes factor tests have good properties for testing the underlying covariance structure of binary response data. The method is illustrated with two real data studies.

  17. Tower of London test: a comparison between conventional statistic approach and modelling based on artificial neural network in differentiating fronto-temporal dementia from Alzheimer's disease.

    Science.gov (United States)

    Franceschi, Massimo; Caffarra, Paolo; Savarè, Rita; Cerutti, Renata; Grossi, Enzo

    2011-01-01

    The early differentiation of Alzheimer's disease (AD) from frontotemporal dementia (FTD) may be difficult. The Tower of London (ToL), thought to assess executive functions such as planning and visuo-spatial working memory, could help in this purpose. Twentytwo Dementia Centers consecutively recruited patients with early FTD or AD. ToL performances of these groups were analyzed using both the conventional statistical approaches and the Artificial Neural Networks (ANNs) modelling. Ninety-four non aphasic FTD and 160 AD patients were recruited. ToL Accuracy Score (AS) significantly (p advanced ANNs developed by Semeion Institute. The best ANNs were selected and submitted to ROC curves. The non-linear model was able to discriminate FTD from AD with an average AUC for 7 independent trials of 0.82. The use of hidden information contained in the different items of ToL and the non linear processing of the data through ANNs allows a high discrimination between FTD and AD in individual patients.

  18. Tests of gravity with future space-based experiments

    Science.gov (United States)

    Sakstein, Jeremy

    2018-03-01

    Future space-based tests of relativistic gravitation—laser ranging to Phobos, accelerometers in orbit, and optical networks surrounding Earth—will constrain the theory of gravity with unprecedented precision by testing the inverse-square law, the strong and weak equivalence principles, and the deflection and time delay of light by massive bodies. In this paper, we estimate the bounds that could be obtained on alternative gravity theories that use screening mechanisms to suppress deviations from general relativity in the Solar System: chameleon, symmetron, and Galileon models. We find that space-based tests of the parametrized post-Newtonian parameter γ will constrain chameleon and symmetron theories to new levels, and that tests of the inverse-square law using laser ranging to Phobos will provide the most stringent constraints on Galileon theories to date. We end by discussing the potential for constraining these theories using upcoming tests of the weak equivalence principle, and conclude that further theoretical modeling is required in order to fully utilize the data.

  19. Model-based machine learning.

    Science.gov (United States)

    Bishop, Christopher M

    2013-02-13

    Several decades of research in the field of machine learning have resulted in a multitude of different algorithms for solving a broad range of problems. To tackle a new application, a researcher typically tries to map their problem onto one of these existing methods, often influenced by their familiarity with specific algorithms and by the availability of corresponding software implementations. In this study, we describe an alternative methodology for applying machine learning, in which a bespoke solution is formulated for each new application. The solution is expressed through a compact modelling language, and the corresponding custom machine learning code is then generated automatically. This model-based approach offers several major advantages, including the opportunity to create highly tailored models for specific scenarios, as well as rapid prototyping and comparison of a range of alternative models. Furthermore, newcomers to the field of machine learning do not have to learn about the huge range of traditional methods, but instead can focus their attention on understanding a single modelling environment. In this study, we show how probabilistic graphical models, coupled with efficient inference algorithms, provide a very flexible foundation for model-based machine learning, and we outline a large-scale commercial application of this framework involving tens of millions of users. We also describe the concept of probabilistic programming as a powerful software environment for model-based machine learning, and we discuss a specific probabilistic programming language called Infer.NET, which has been widely used in practical applications.

  20. Accelerated load testing of geosynthetic base reinforced pavement test sections.

    Science.gov (United States)

    2011-02-01

    The main objective of this research is to evaluate the benefits of geosynthetic stabilization and reinforcement of subgrade/base aggregate layers in flexible pavements built on weak subgrades and the effect of pre-rut pavement sections, prior to the ...

  1. Cernavoda NPP risk - Based test and maintenance planning - Methodology development

    International Nuclear Information System (INIS)

    Georgescu, G.; Popa, P.; Petrescu, A.; Naum, M.; Gutu, M.

    1997-01-01

    The Cernavoda Power Plant starts the commercial operation in November 1996. During operation of the nuclear power plant, several mandatory tests and maintenance are performed on stand-by safety system components to ensure their availability in case of accident. The basic purpose of such activities is the early detection of any failure and degradation, and timely correction of deteriorations. Because of the large number of such activities, emphasis on plant safety and allocation of resources becomes difficult. The probabilistic model and methodology can be effectively used to obtain the risk significance of these activities so that the resources are directed to the most important areas. The proposed Research Contract activity is strongly connected with other safety related areas under development. Since, the Cernavoda Probabilistic Safety Evaluation Level 1 PSA Study (CPSE) was performed and now the study is revised taking into account the as-built information, it is recommended to implement into the model the necessary modeling features to support further PSA application, especially related to Test and Maintenance optimization. Methods need to be developed in order to apply the PSA model including risk information together with other needed information for Test and Maintenance optimization. Also, in parallel with the CPSE study updating, the software interface for the PSA model is under development (Risk Monitor Software class), methods and models needing to be developed for the purpose of using it for qualified monitoring of Test and Maintenance Strategy efficiency. Similar, the Data Collection System need to be appropriate for the purpose of an ongoing implementation of a risk - based Test and Maintenance Strategy. (author). 4 refs, 1 fig

  2. GIS Modelling of Radionuclide Transport from the Semipalatinsk Test Site

    Science.gov (United States)

    Balakay, L.; Zakarin, E.; Mahura, A.; Baklanov, A.; Sorensen, J. H.

    2009-04-01

    In this study, the software complex GIS-project MigRad (Migration of Radionuclide) was developed, tested and applied for the territory of the Semipalatinsk test site/ polygon (Republic of Kazakhstan), where since 1961, in total 348 underground nuclear explosions were conducted. The MigRad is oriented on integration of large volumes of different information (mapping, ground-based, and satellite-based survey): and also includes modeling on its base local redistribution of radionuclides by precipitation and surface waters and by long-range transport of radioactive aerosols. The existing thermal anomaly on territory of the polygon was investigated in details, and the object-oriented analysis was applied for the studied area. Employing the RUNOFF model, the simulation of radionuclides migration with surface waters was performed. Employing the DERMA model, the simulation of long-term atmospheric transport, dispersion and deposition patterns for cesium was conducted from 3 selected locations (Balapan, Delegen, and Experimental Field). Employing geoinformation technology, the mapping of the of the high temperature zones and epicenters of radioactive aerosols transport for the territory of the test site was carried out with post-processing and integration of modelling results into GIS environment. Contamination levels of pollution due to former nuclear explosions for population and environment of the surrounding polygon territories of Kazakhstan as well as adjacent countries were analyzed and evaluated. The MigRad was designed as instrument for comprehensive analysis of complex territorial processes influenced by former nuclear explosions on the territory of Semipalatinsk test site. It provides possibilities in detailed analyses for (i) extensive cartographic material, remote sensing, and field measurements data collected in different level databases; (ii) radionuclide migration with flows using accumulation and redistribution of soil particles; (iii) thermal anomalies

  3. The use of scale models in impact testing

    International Nuclear Information System (INIS)

    Donelan, P.J.; Dowling, A.R.

    1985-01-01

    Theoretical analysis, component testing and model flask testing are employed to investigate the validity of scale models for demonstrating the behaviour of Magnox flasks under impact conditions. Model testing is shown to be a powerful and convenient tool provided adequate care is taken with detail design and manufacture of models and with experimental control. (author)

  4. Modeling and Testing of EVs - Preliminary Study and Laboratory Development

    DEFF Research Database (Denmark)

    Yang, Guang-Ya; Marra, Francesco; Nielsen, Arne Hejde

    2010-01-01

    Electric vehicles (EVs) are expected to play a key role in the future energy management system to stabilize both supply and consumption with the presence of high penetration of renewable generation. A reasonably accurate model of battery is a key element for the study of EVs behavior and the grid...... tests, followed by the suggestions towards a feasible battery model for further studies.......Electric vehicles (EVs) are expected to play a key role in the future energy management system to stabilize both supply and consumption with the presence of high penetration of renewable generation. A reasonably accurate model of battery is a key element for the study of EVs behavior and the grid...... impact at different geographical areas, as well as driving and charging patterns. Electric circuit model is deployed in this work to represent the electrical properties of a lithium-ion battery. This paper reports the preliminary modeling and validation work based on manufacturer data sheet and realistic...

  5. Development and Test of TQC models, LARP Technological Quadrupole Magnets

    Energy Technology Data Exchange (ETDEWEB)

    Bossert, R.C.; Ambrosio, G.; Andreev, N.; Barzi, E.; Carcagno, R.; Feher, S.; Kashikhin, V.S.; Kashikhin, V.V.; Nobrega, F.; Novitski, I.; Orris, D.; Tartaglia, M.; Zlobin, A.V.; Caspi, S.; Dietderich, D.; Ferracin, P.; Hafalia, A.R.; Sabbi, G.

    2008-06-01

    In support of the development of a large-aperture Nb3Sn superconducting quadrupole for the Large Hadron Collider (LHC) luminosity upgrade, two-layer quadrupole models (TQC and TQS) with 90mm aperture are being constructed at Fermilab and LBNL within the framework of the US LHC Accelerator Research Program (LARP). This paper describes the development and test of TQC01b, the second TQC model, and the experience during construction of TQE02 and TQC02, subsequent models in the series. ANSYS analysis of the mechanical structure, its underlying assumptions, and changes based on experience with TQC01 are presented and discussed. Construction experience, in-process measurements, and modifications to the assembly since TQC01 are described. The test results presented here include magnet strain and quench performance during training of TQC01b, as well as quench studies of current ramp rate dependence.

  6. Development and Test of TQC models, LARP Technological Quadrupole Magnets

    International Nuclear Information System (INIS)

    Bossert, R.C.; Ambrosio, G.; Andreev, N.; Barzi, E.; Carcagno, R.; Feher, S.; Kashikhin, V.S.; Kashikhin, V.V.; Nobrega, F.; Novitski, I.; Orris, D.; Tartaglia, M.; Zlobin, A.V.; Caspi, S.; Dietderich, D.; Ferracin, P.; Hafalia, A.R.; Sabbi, G.

    2008-01-01

    In support of the development of a large-aperture Nb3Sn superconducting quadrupole for the Large Hadron Collider (LHC) luminosity upgrade, two-layer quadrupole models (TQC and TQS) with 90mm aperture are being constructed at Fermilab and LBNL within the framework of the US LHC Accelerator Research Program (LARP). This paper describes the development and test of TQC01b, the second TQC model, and the experience during construction of TQE02 and TQC02, subsequent models in the series. ANSYS analysis of the mechanical structure, its underlying assumptions, and changes based on experience with TQC01 are presented and discussed. Construction experience, in-process measurements, and modifications to the assembly since TQC01 are described. The test results presented here include magnet strain and quench performance during training of TQC01b, as well as quench studies of current ramp rate dependence

  7. Some tests for parameter constancy in cointegrated VAR-models

    DEFF Research Database (Denmark)

    Hansen, Henrik; Johansen, Søren

    1999-01-01

    Some methods for the evaluation of parameter constancy in vector autoregressive (VAR) models are discussed. Two different ways of re-estimating the VAR model are proposed; one in which all parameters are estimated recursively based upon the likelihood function for the first observations, and anot...... be applied to test the constancy of the long-run parameters in the cointegrated VAR-model. All results are illustrated using a model for the term structure of interest rates on US Treasury securities. ......Some methods for the evaluation of parameter constancy in vector autoregressive (VAR) models are discussed. Two different ways of re-estimating the VAR model are proposed; one in which all parameters are estimated recursively based upon the likelihood function for the first observations......, and another in which the cointegrating relations are estimated recursively from a likelihood function, where the short-run parameters have been concentrated out. We suggest graphical procedures based on recursively estimated eigenvalues to evaluate the constancy of the long-run parameters in the model...

  8. Stimulating Scientific Reasoning with Drawing-Based Modeling

    Science.gov (United States)

    Heijnes, Dewi; van Joolingen, Wouter; Leenaars, Frank

    2018-01-01

    We investigate the way students' reasoning about evolution can be supported by drawing-based modeling. We modified the drawing-based modeling tool SimSketch to allow for modeling evolutionary processes. In three iterations of development and testing, students in lower secondary education worked on creating an evolutionary model. After each…

  9. A probabilistic graphical model based stochastic input model construction

    International Nuclear Information System (INIS)

    Wan, Jiang; Zabaras, Nicholas

    2014-01-01

    Model reduction techniques have been widely used in modeling of high-dimensional stochastic input in uncertainty quantification tasks. However, the probabilistic modeling of random variables projected into reduced-order spaces presents a number of computational challenges. Due to the curse of dimensionality, the underlying dependence relationships between these random variables are difficult to capture. In this work, a probabilistic graphical model based approach is employed to learn the dependence by running a number of conditional independence tests using observation data. Thus a probabilistic model of the joint PDF is obtained and the PDF is factorized into a set of conditional distributions based on the dependence structure of the variables. The estimation of the joint PDF from data is then transformed to estimating conditional distributions under reduced dimensions. To improve the computational efficiency, a polynomial chaos expansion is further applied to represent the random field in terms of a set of standard random variables. This technique is combined with both linear and nonlinear model reduction methods. Numerical examples are presented to demonstrate the accuracy and efficiency of the probabilistic graphical model based stochastic input models. - Highlights: • Data-driven stochastic input models without the assumption of independence of the reduced random variables. • The problem is transformed to a Bayesian network structure learning problem. • Examples are given in flows in random media

  10. Accelerated testing statistical models, test plans, and data analysis

    CERN Document Server

    Nelson, Wayne B

    2009-01-01

    The Wiley-Interscience Paperback Series consists of selected books that have been made more accessible to consumers in an effort to increase global appeal and general circulation. With these new unabridged softcover volumes, Wiley hopes to extend the lives of these works by making them available to future generations of statisticians, mathematicians, and scientists. "". . . a goldmine of knowledge on accelerated life testing principles and practices . . . one of the very few capable of advancing the science of reliability. It definitely belongs in every bookshelf on engineering.""-Dev G.

  11. Rule-based Test Generation with Mind Maps

    Directory of Open Access Journals (Sweden)

    Dimitry Polivaev

    2012-02-01

    Full Text Available This paper introduces basic concepts of rule based test generation with mind maps, and reports experiences learned from industrial application of this technique in the domain of smart card testing by Giesecke & Devrient GmbH over the last years. It describes the formalization of test selection criteria used by our test generator, our test generation architecture and test generation framework.

  12. Validity of selected cardiovascular field-based test among Malaysian ...

    African Journals Online (AJOL)

    Based on emerge obese problem among Malaysian, this research is formulated to validate published tests among healthy female adult. Selected test namely; 20 meter multi-stage shuttle run, 2.4km run test, 1 mile walk test and Harvard Step test were correlated with laboratory test (Bruce protocol) to find the criterion validity ...

  13. Test models for improving filtering with model errors through stochastic parameter estimation

    International Nuclear Information System (INIS)

    Gershgorin, B.; Harlim, J.; Majda, A.J.

    2010-01-01

    The filtering skill for turbulent signals from nature is often limited by model errors created by utilizing an imperfect model for filtering. Updating the parameters in the imperfect model through stochastic parameter estimation is one way to increase filtering skill and model performance. Here a suite of stringent test models for filtering with stochastic parameter estimation is developed based on the Stochastic Parameterization Extended Kalman Filter (SPEKF). These new SPEKF-algorithms systematically correct both multiplicative and additive biases and involve exact formulas for propagating the mean and covariance including the parameters in the test model. A comprehensive study is presented of robust parameter regimes for increasing filtering skill through stochastic parameter estimation for turbulent signals as the observation time and observation noise are varied and even when the forcing is incorrectly specified. The results here provide useful guidelines for filtering turbulent signals in more complex systems with significant model errors.

  14. Risk based test interval and maintenance optimisation - Application and uses

    International Nuclear Information System (INIS)

    Sparre, E.

    1999-10-01

    The project is part of an IAEA co-ordinated Research Project (CRP) on 'Development of Methodologies for Optimisation of Surveillance Testing and Maintenance of Safety Related Equipment at NPPs'. The purpose of the project is to investigate the sensitivity of the results obtained when performing risk based optimisation of the technical specifications. Previous projects have shown that complete LPSA models can be created and that these models allow optimisation of technical specifications. However, these optimisations did not include any in depth check of the result sensitivity with regards to methods, model completeness etc. Four different test intervals have been investigated in this study. Aside from an original, nominal, optimisation a set of sensitivity analyses has been performed and the results from these analyses have been compared to the original optimisation. The analyses indicate that the result of an optimisation is rather stable. However, it is not possible to draw any certain conclusions without performing a number of sensitivity analyses. Significant differences in the optimisation result were discovered when analysing an alternative configuration. Also deterministic uncertainties seem to affect the result of an optimisation largely. The sensitivity of failure data uncertainties is important to investigate in detail since the methodology is based on the assumption that the unavailability of a component is dependent on the length of the test interval

  15. User Context Aware Base Station Power Flow Model

    OpenAIRE

    Walsh, Barbara; Farrell, Ronan

    2005-01-01

    At present the testing of power amplifiers within base station transmitters is limited to testing at component level as opposed to testing at the system level. While the detection of catastrophic failure is possible, that of performance degradation is not. This paper proposes a base station model with respect to transmitter output power with the aim of introducing system level monitoring of the power amplifier behaviour within the base station. Our model reflects the expe...

  16. BWR regional instability model and verification on ringhals-1 test

    International Nuclear Information System (INIS)

    Hotta, Akitoshi; Suzawa, Yojiro

    1996-01-01

    Regional instability is known as one type of the coupled neutronic-thermohydraulic phenomena of boiling water reactors (BWRs), where the thermohydraulic density wave propagation mechanism is predominant. Historically, it has been simulated by the three-dimensional time domain code in spite of its significant computing time. On the other hand, there have been proposals to apply the frequency domain models in regional instability considering the subcriticality of the higher neutronic mode. However, their application still remains in corewide instability mainly because of the lack of more detailed methodological and empirical studies. In this study, the current version of the frequency domain model was extended and verified based on actual core regional instability measurement data. The mathematical model LAPUR, the well-known frequency domain stability code, was reviewed from the standpoint of pure thermohydraulics and neutronic-thermohydraulic interaction mechanisms. Based on the ex-core loop test data, the original LAPUR mixed friction and local pressure loss model was modified, taking into account the different dynamic behavior of these two pressure-loss mechanisms. The perturbation term of the two-phase friction multiplier, which is the sum of the derivative of void fraction and subcool enthalpy, was adjusted theoretically. The adequacy of the instability evaluation system was verified based on the Ringhals unit 1 test data, which were supplied to participants of the Organization for Economic Cooperation and Development/Nuclear Energy Agency BWR Stability Benchmark Project

  17. Standard Model theory calculations and experimental tests

    International Nuclear Information System (INIS)

    Cacciari, M.; Hamel de Monchenault, G.

    2015-01-01

    To present knowledge, all the physics at the Large Hadron Collider (LHC) can be described in the framework of the Standard Model (SM) of particle physics. Indeed the newly discovered Higgs boson with a mass close to 125 GeV seems to confirm the predictions of the SM. Thus, besides looking for direct manifestations of the physics beyond the SM, one of the primary missions of the LHC is to perform ever more stringent tests of the SM. This requires not only improved theoretical developments to produce testable predictions and provide experiments with reliable event generators, but also sophisticated analyses techniques to overcome the formidable experimental environment of the LHC and perform precision measurements. In the first section, we describe the state of the art of the theoretical tools and event generators that are used to provide predictions for the production cross sections of the processes of interest. In section 2, inclusive cross section measurements with jets, leptons and vector bosons are presented. Examples of differential cross sections, charge asymmetries and the study of lepton pairs are proposed in section 3. Finally, in section 4, we report studies on the multiple production of gauge bosons and constraints on anomalous gauge couplings

  18. Methods and models for the construction of weakly parallel tests

    NARCIS (Netherlands)

    Adema, J.J.; Adema, Jos J.

    1990-01-01

    Methods are proposed for the construction of weakly parallel tests, that is, tests with the same test information function. A mathematical programing model for constructing tests with a prespecified test information function and a heuristic for assigning items to tests such that their information

  19. Neutron Sources for Standard-Based Testing

    Energy Technology Data Exchange (ETDEWEB)

    Radev, Radoslav [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); McLean, Thomas [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2014-11-10

    The DHS TC Standards and the consensus ANSI Standards use 252Cf as the neutron source for performance testing because its energy spectrum is similar to the 235U and 239Pu fission sources used in nuclear weapons. An emission rate of 20,000 ± 20% neutrons per second is used for testing of the radiological requirements both in the ANSI standards and the TCS. Determination of the accurate neutron emission rate of the test source is important for maintaining consistency and agreement between testing results obtained at different testing facilities. Several characteristics in the manufacture and the decay of the source need to be understood and accounted for in order to make an accurate measurement of the performance of the neutron detection instrument. Additionally, neutron response characteristics of the particular instrument need to be known and taken into account as well as neutron scattering in the testing environment.

  20. FMIT test-end instrumentation development bases

    International Nuclear Information System (INIS)

    Fuller, J.L.

    1982-06-01

    FMIT test-end measurements proposed for deuteron beam control, target diagnostics, and irradiation sample dosimetry are listed. The test-end refers to the area inside the test cell, but includes measurements inside and outside the cell. Justification, categorization, and limits qualification are presented for each measurement. Methods are purposefully de-emphasized in order to clarify the measurement needs, not techniques. Some discussion of techniques currently under investigation is given in the last section of the report

  1. Atmospheric resuspension of radionuclides. Model testing using Chernobyl data

    International Nuclear Information System (INIS)

    Garger, E.; Lev, T.; Talerko, N.; Galeriu, D.; Garland, J.; Hoffman, O.; Nair, S.; Thiessen, K.; Miller, C.; Mueller, H.; Kryshev, A.

    1996-10-01

    Resuspension can be an important secondary source of contamination after a release has stopped, as well as a source of contamination for people and areas not exposed to the original release. The inhalation of resuspended radionuclides contributes to the overall dose received by exposed individuals. Based on measurements collected after the Chernobyl accident, Scenario R was developed to provide an opportunity to test existing mathematical models of contamination resuspension. In particular, this scenario provided the opportunity to examine data and test models for atmospheric resuspension of radionuclides at several different locations from the release, to investigate resuspension processes on both local and regional scales, and to investigate the importance of seasonal variations of these processes. Participants in the test exercise were provided with information for three different types of locations: (1) within the 30-km zone, where local resuspension processes are expected to dominate; (2) a large urban location (Kiev) 120 km from the release site, where vehicular traffic is expected to be the dominant mechanism for resuspension; and (3) an agricultural area 40-60 km from the release site, where highly contaminated upwind 'hot spots' are expected to be important. Input information included characteristics of the ground contamination around specific sites, climatological data for the sites, characteristics of the terrain and topography, and locations of the sampling sites. Participants were requested to predict the average (quarterly and yearly) concentrations of 137 Cs in air at specified locations due to resuspension of Chernobyl fallout; predictions for 90 Sr and 239 + 240 Pu were also requested for one location and time point. Predictions for specified resuspension factors and rates were also requested. Most participants used empirical models for the resuspension factor as a function of time K(t), as opposed to process-based models. While many of these

  2. An integrated tiered service delivery model (ITSDM based on local CD4 testing demands can improve turn-around times and save costs whilst ensuring accessible and scalable CD4 services across a national programme.

    Directory of Open Access Journals (Sweden)

    Deborah K Glencross

    Full Text Available The South African National Health Laboratory Service (NHLS responded to HIV treatment initiatives with two-tiered CD4 laboratory services in 2004. Increasing programmatic burden, as more patients access anti-retroviral therapy (ART, has demanded extending CD4 services to meet increasing clinical needs. The aim of this study was to review existing services and develop a service-model that integrated laboratory-based and point-of-care testing (POCT, to extend national coverage, improve local turn-around/(TAT and contain programmatic costs.NHLS Corporate Data Warehouse CD4 data, from 60-70 laboratories and 4756 referring health facilities was reviewed for referral laboratory workload, respective referring facility volumes and related TAT, from 2009-2012.An integrated tiered service delivery model (ITSDM is proposed. Tier-1/POCT delivers CD4 testing at single health-clinics providing ART in hard-to-reach areas (350-1500 tests/day, serving ≥ 200 health-clinics. Tier-6 provides national support for standardisation, harmonization and quality across the organization.The ITSDM offers improved local TAT by extending CD4 services into rural/remote areas with new Tier-3 or Tier-2/POC-Hub services installed in existing community laboratories, most with developed infrastructure. The advantage of lower laboratory CD4 costs and use of existing infrastructure enables subsidization of delivery of more expensive POC services, into hard-to-reach districts without reasonable access to a local CD4 laboratory. Full ITSDM implementation across 5 service tiers (as opposed to widespread implementation of POC testing to extend service can facilitate sustainable 'full service coverage' across South Africa, and save>than R125 million in HIV/AIDS programmatic costs. ITSDM hierarchical parental-support also assures laboratory/POC management, equipment maintenance, quality control and on-going training between tiers.

  3. The Linear Logistic Test Model (LLTM as the methodological foundation of item generating rules for a new verbal reasoning test

    Directory of Open Access Journals (Sweden)

    HERBERT POINSTINGL

    2009-06-01

    Full Text Available Based on the demand for new verbal reasoning tests to enrich psychological test inventory, a pilot version of a new test was analysed: the 'Family Relation Reasoning Test' (FRRT; Poinstingl, Kubinger, Skoda & Schechtner, forthcoming, in which several basic cognitive operations (logical rules have been embedded/implemented. Given family relationships of varying complexity embedded in short stories, testees had to logically conclude the correct relationship between two individuals within a family. Using empirical data, the linear logistic test model (LLTM; Fischer, 1972, a special case of the Rasch model, was used to test the construct validity of the test: The hypothetically assumed basic cognitive operations had to explain the Rasch model's item difficulty parameters. After being shaped in LLTM's matrices of weights ((qij, none of these operations were corroborated by means of the Andersen's Likelihood Ratio Test.

  4. TR-EDB: Test Reactor Embrittlement Data Base, Version 1

    Energy Technology Data Exchange (ETDEWEB)

    Stallmann, F.W.; Wang, J.A.; Kam, F.B.K. [Oak Ridge National Lab., TN (United States)

    1994-01-01

    The Test Reactor Embrittlement Data Base (TR-EDB) is a collection of results from irradiation in materials test reactors. It complements the Power Reactor Embrittlement Data Base (PR-EDB), whose data are restricted to the results from the analysis of surveillance capsules in commercial power reactors. The rationale behind their restriction was the assumption that the results of test reactor experiments may not be applicable to power reactors and could, therefore, be challenged if such data were included. For this very reason the embrittlement predictions in the Reg. Guide 1.99, Rev. 2, were based exclusively on power reactor data. However, test reactor experiments are able to cover a much wider range of materials and irradiation conditions that are needed to explore more fully a variety of models for the prediction of irradiation embrittlement. These data are also needed for the study of effects of annealing for life extension of reactor pressure vessels that are difficult to obtain from surveillance capsule results.

  5. TR-EDB: Test Reactor Embrittlement Data Base, Version 1

    International Nuclear Information System (INIS)

    Stallmann, F.W.; Wang, J.A.; Kam, F.B.K.

    1994-01-01

    The Test Reactor Embrittlement Data Base (TR-EDB) is a collection of results from irradiation in materials test reactors. It complements the Power Reactor Embrittlement Data Base (PR-EDB), whose data are restricted to the results from the analysis of surveillance capsules in commercial power reactors. The rationale behind their restriction was the assumption that the results of test reactor experiments may not be applicable to power reactors and could, therefore, be challenged if such data were included. For this very reason the embrittlement predictions in the Reg. Guide 1.99, Rev. 2, were based exclusively on power reactor data. However, test reactor experiments are able to cover a much wider range of materials and irradiation conditions that are needed to explore more fully a variety of models for the prediction of irradiation embrittlement. These data are also needed for the study of effects of annealing for life extension of reactor pressure vessels that are difficult to obtain from surveillance capsule results

  6. GPS Device Testing Based on User Performance Metrics

    Science.gov (United States)

    2015-10-02

    1. Rationale for a Test Program Based on User Performance Metrics ; 2. Roberson and Associates Test Program ; 3. Status of, and Revisions to, the Roberson and Associates Test Program ; 4. Comparison of Roberson and DOT/Volpe Programs

  7. Efficient p-value evaluation for resampling-based tests

    KAUST Repository

    Yu, K.; Liang, F.; Ciampa, J.; Chatterjee, N.

    2011-01-01

    The resampling-based test, which often relies on permutation or bootstrap procedures, has been widely used for statistical hypothesis testing when the asymptotic distribution of the test statistic is unavailable or unreliable. It requires repeated

  8. ExEP yield modeling tool and validation test results

    Science.gov (United States)

    Morgan, Rhonda; Turmon, Michael; Delacroix, Christian; Savransky, Dmitry; Garrett, Daniel; Lowrance, Patrick; Liu, Xiang Cate; Nunez, Paul

    2017-09-01

    EXOSIMS is an open-source simulation tool for parametric modeling of the detection yield and characterization of exoplanets. EXOSIMS has been adopted by the Exoplanet Exploration Programs Standards Definition and Evaluation Team (ExSDET) as a common mechanism for comparison of exoplanet mission concept studies. To ensure trustworthiness of the tool, we developed a validation test plan that leverages the Python-language unit-test framework, utilizes integration tests for selected module interactions, and performs end-to-end crossvalidation with other yield tools. This paper presents the test methods and results, with the physics-based tests such as photometry and integration time calculation treated in detail and the functional tests treated summarily. The test case utilized a 4m unobscured telescope with an idealized coronagraph and an exoplanet population from the IPAC radial velocity (RV) exoplanet catalog. The known RV planets were set at quadrature to allow deterministic validation of the calculation of physical parameters, such as working angle, photon counts and integration time. The observing keepout region was tested by generating plots and movies of the targets and the keepout zone over a year. Although the keepout integration test required the interpretation of a user, the test revealed problems in the L2 halo orbit and the parameterization of keepout applied to some solar system bodies, which the development team was able to address. The validation testing of EXOSIMS was performed iteratively with the developers of EXOSIMS and resulted in a more robust, stable, and trustworthy tool that the exoplanet community can use to simulate exoplanet direct-detection missions from probe class, to WFIRST, up to large mission concepts such as HabEx and LUVOIR.

  9. Perceived game realism: a test of three alternative models.

    Science.gov (United States)

    Ribbens, Wannes

    2013-01-01

    Perceived realism is considered a key concept in explaining the mental processing of media messages and the societal impact of media. Despite its importance, little is known about its conceptualization and dimensional structure, especially with regard to digital games. The aim of this study was to test a six-factor model of perceived game realism comprised of simulational realism, freedom of choice, perceptual pervasiveness, social realism, authenticity, and character involvement and to assess it against an alternative single- and five-factor model. Data were collected from 380 male digital game users who judged the realism of the first-person shooter Half-Life 2 based upon their previous experience with the game. Confirmatory factor analysis was applied to investigate which model fits the data best. The results support the six-factor model over the single- and five-factor solutions. The study contributes to our knowledge of perceived game realism by further developing its conceptualization and measurement.

  10. Putting hydrological modelling practice to the test

    NARCIS (Netherlands)

    Melsen, Lieke Anna

    2017-01-01

    Six steps can be distinguished in the process of hydrological modelling: the perceptual model (deciding on the processes), the conceptual model (deciding on the equations), the procedural model (get the code to run on a computer), calibration (identify the parameters), evaluation (confronting

  11. Thermohydraulic tests in nuclear fuel model

    International Nuclear Information System (INIS)

    Ladeira, L.C.D.; Navarro, M.A.

    1984-01-01

    The main experimental works performed in the Thermohydraulics Laboratory of the NUCLEBRAS Nuclear Technology Development Center, in the field of thermofluodynamics are briefly described. These works include the performing of steady-state flow tests in single tube test sections, and the design and construction of a rod bundle test section, which will be also used for those kind of testes. Mention is made of the works to be performed in the near future, related to steady-state and transient flow tests. (Author) [pt

  12. Graph Model Based Indoor Tracking

    DEFF Research Database (Denmark)

    Jensen, Christian Søndergaard; Lu, Hua; Yang, Bin

    2009-01-01

    The tracking of the locations of moving objects in large indoor spaces is important, as it enables a range of applications related to, e.g., security and indoor navigation and guidance. This paper presents a graph model based approach to indoor tracking that offers a uniform data management...

  13. Computer Based Modelling and Simulation

    Indian Academy of Sciences (India)

    GENERAL I ARTICLE. Computer Based ... universities, and later did system analysis, ... sonal computers (PC) and low cost software packages and tools. They can serve as useful learning experience through student projects. Models are .... Let us consider a numerical example: to calculate the velocity of a trainer aircraft ...

  14. Multi-objective Search-based Mobile Testing

    OpenAIRE

    Mao, K.

    2017-01-01

    Despite the tremendous popularity of mobile applications, mobile testing still relies heavily on manual testing. This thesis presents mobile test automation approaches based on multi-objective search. We introduce three approaches: Sapienz (for native Android app testing), Octopuz (for hybrid/web JavaScript app testing) and Polariz (for using crowdsourcing to support search-based mobile testing). These three approaches represent the primary scientific and technical contributions of the thesis...

  15. A test for the parameters of multiple linear regression models ...

    African Journals Online (AJOL)

    A test for the parameters of multiple linear regression models is developed for conducting tests simultaneously on all the parameters of multiple linear regression models. The test is robust relative to the assumptions of homogeneity of variances and absence of serial correlation of the classical F-test. Under certain null and ...

  16. Pescara benchmark: overview of modelling, testing and identification

    International Nuclear Information System (INIS)

    Bellino, A; Garibaldi, L; Marchesiello, S; Brancaleoni, F; Gabriele, S; Spina, D; Bregant, L; Carminelli, A; Catania, G; Sorrentino, S; Di Evangelista, A; Valente, C; Zuccarino, L

    2011-01-01

    The 'Pescara benchmark' is part of the national research project 'BriViDi' (BRIdge VIbrations and DIagnosis) supported by the Italian Ministero dell'Universita e Ricerca. The project is aimed at developing an integrated methodology for the structural health evaluation of railway r/c, p/c bridges. The methodology should provide for applicability in operating conditions, easy data acquisition through common industrial instrumentation, robustness and reliability against structural and environmental uncertainties. The Pescara benchmark consisted in lab tests to get a consistent and large experimental data base and subsequent data processing. Special tests were devised to simulate the train transit effects in actual field conditions. Prestressed concrete beams of current industrial production both sound and damaged at various severity corrosion levels were tested. The results were collected either in a deterministic setting and in a form suitable to deal with experimental uncertainties. Damage identification was split in two approaches: with or without a reference model. In the first case f.e. models were used in conjunction with non conventional updating techniques. In the second case, specialized output-only identification techniques capable to deal with time-variant and possibly non linear systems were developed. The lab tests allowed validating the above approaches and the performances of classical modal based damage indicators.

  17. Rate-control algorithms testing by using video source model

    DEFF Research Database (Denmark)

    Belyaev, Evgeny; Turlikov, Andrey; Ukhanova, Anna

    2008-01-01

    In this paper the method of rate control algorithms testing by the use of video source model is suggested. The proposed method allows to significantly improve algorithms testing over the big test set.......In this paper the method of rate control algorithms testing by the use of video source model is suggested. The proposed method allows to significantly improve algorithms testing over the big test set....

  18. Computer-Aided Test Flow in Core-Based Design

    NARCIS (Netherlands)

    Zivkovic, V.; Tangelder, R.J.W.T.; Kerkhoff, Hans G.

    2000-01-01

    This paper copes with the efficient test-pattern generation in a core-based design. A consistent Computer-Aided Test (CAT) flow is proposed based on the required core-test strategy. It generates a test-pattern set for the embedded cores with high fault coverage and low DfT area overhead. The CAT

  19. Oscillation-based test in mixed-signal circuits

    CERN Document Server

    Sánchez, Gloria Huertas; Rueda, Adoración Rueda

    2007-01-01

    This book presents the development and experimental validation of the structural test strategy called Oscillation-Based Test - OBT in short. The results presented here assert, not only from a theoretical point of view, but also based on a wide experimental support, that OBT is an efficient defect-oriented test solution, complementing the existing functional test techniques for mixed-signal circuits.

  20. Nonlinear system modeling based on bilinear Laguerre orthonormal bases.

    Science.gov (United States)

    Garna, Tarek; Bouzrara, Kais; Ragot, José; Messaoud, Hassani

    2013-05-01

    This paper proposes a new representation of discrete bilinear model by developing its coefficients associated to the input, to the output and to the crossed product on three independent Laguerre orthonormal bases. Compared to classical bilinear model, the resulting model entitled bilinear-Laguerre model ensures a significant parameter number reduction as well as simple recursive representation. However, such reduction still constrained by an optimal choice of Laguerre pole characterizing each basis. To do so, we develop a pole optimization algorithm which constitutes an extension of that proposed by Tanguy et al.. The bilinear-Laguerre model as well as the proposed pole optimization algorithm are illustrated and tested on a numerical simulations and validated on the Continuous Stirred Tank Reactor (CSTR) System. Copyright © 2012 ISA. Published by Elsevier Ltd. All rights reserved.