WorldWideScience

Sample records for explicit modeling tool

  1. Philosophical Reflections made explicit as a Tool for Mathematical Reasoning

    DEFF Research Database (Denmark)

    Frølund, Sune; Andresen, Mette

    2009-01-01

        A new construct, ‘multidiciplinarity', is prescribed in the curricula of Danish Upper Secondary Schools by governmental regulations since 2006. Multidisciplinarity offers a good chance to introduce philosophical tools or methods in mathematics with the aim to improve the students' learning...... of both subjects, and to study the students' reactions and signs of progressive mathematizing. Based on realistic mathematics education (RME) which is rooted in Hans Freudenthal's idea of mathematics as a human activity, we decided to centre our work on the concept of reflection and to build a model...... for making students reflections in the mathematics class explicit to themselves. In our paper, we present a combination of two stratifications of reflections which were developed recently in works by other authors. The paper outlines our model and exemplifies its use on the teaching of mathematical models...

  2. Integrating remote sensing and spatially explicit epidemiological modeling

    Science.gov (United States)

    Finger, Flavio; Knox, Allyn; Bertuzzo, Enrico; Mari, Lorenzo; Bompangue, Didier; Gatto, Marino; Rinaldo, Andrea

    2015-04-01

    Spatially explicit epidemiological models are a crucial tool for the prediction of epidemiological patterns in time and space as well as for the allocation of health care resources. In addition they can provide valuable information about epidemiological processes and allow for the identification of environmental drivers of the disease spread. Most epidemiological models rely on environmental data as inputs. They can either be measured in the field by the means of conventional instruments or using remote sensing techniques to measure suitable proxies of the variables of interest. The later benefit from several advantages over conventional methods, including data availability, which can be an issue especially in developing, and spatial as well as temporal resolution of the data, which is particularly crucial for spatially explicit models. Here we present the case study of a spatially explicit, semi-mechanistic model applied to recurring cholera outbreaks in the Lake Kivu area (Democratic Republic of the Congo). The model describes the cholera incidence in eight health zones on the shore of the lake. Remotely sensed datasets of chlorophyll a concentration in the lake, precipitation and indices of global climate anomalies are used as environmental drivers. Human mobility and its effect on the disease spread is also taken into account. Several model configurations are tested on a data set of reported cases. The best models, accounting for different environmental drivers, and selected using the Akaike information criterion, are formally compared via cross validation. The best performing model accounts for seasonality, El Niño Southern Oscillation, precipitation and human mobility.

  3. Explicitly represented polygon wall boundary model for the explicit MPS method

    Science.gov (United States)

    Mitsume, Naoto; Yoshimura, Shinobu; Murotani, Kohei; Yamada, Tomonori

    2015-05-01

    This study presents an accurate and robust boundary model, the explicitly represented polygon (ERP) wall boundary model, to treat arbitrarily shaped wall boundaries in the explicit moving particle simulation (E-MPS) method, which is a mesh-free particle method for strong form partial differential equations. The ERP model expresses wall boundaries as polygons, which are explicitly represented without using the distance function. These are derived so that for viscous fluids, and with less computational cost, they satisfy the Neumann boundary condition for the pressure and the slip/no-slip condition on the wall surface. The proposed model is verified and validated by comparing computed results with the theoretical solution, results obtained by other models, and experimental results. Two simulations with complex boundary movements are conducted to demonstrate the applicability of the E-MPS method to the ERP model.

  4. Explicit Nonlinear Model Predictive Control Theory and Applications

    CERN Document Server

    Grancharova, Alexandra

    2012-01-01

    Nonlinear Model Predictive Control (NMPC) has become the accepted methodology to solve complex control problems related to process industries. The main motivation behind explicit NMPC is that an explicit state feedback law avoids the need for executing a numerical optimization algorithm in real time. The benefits of an explicit solution, in addition to the efficient on-line computations, include also verifiability of the implementation and the possibility to design embedded control systems with low software and hardware complexity. This book considers the multi-parametric Nonlinear Programming (mp-NLP) approaches to explicit approximate NMPC of constrained nonlinear systems, developed by the authors, as well as their applications to various NMPC problem formulations and several case studies. The following types of nonlinear systems are considered, resulting in different NMPC problem formulations: Ø  Nonlinear systems described by first-principles models and nonlinear systems described by black-box models; �...

  5. The SOA/VOC/NOx system: an explicit model of secondary organic aerosol formation

    Directory of Open Access Journals (Sweden)

    S. Madronich

    2007-11-01

    Full Text Available Our current understanding of secondary organic aerosol (SOA formation is limited by our knowledge of gaseous secondary organics involved in gas/particle partitioning. The objective of this study is to explore (i the potential for products of multiple oxidation steps contributing to SOA, and (ii the evolution of the SOA/VOC/NOx system. We developed an explicit model based on the coupling of detailed gas-phase oxidation schemes with a thermodynamic condensation module. Such a model allows prediction of SOA mass and speciation on the basis of first principles. The SOA/VOC/NOx system is studied for the oxidation of 1-octene under atmospherically relevant concentrations. In this study, gaseous oxidation of octene is simulated to lead to SOA formation. Contributors to SOA formation are shown to be formed via multiple oxidation steps of the parent hydrocarbon. The behaviour of the SOA/VOC/NOx system simulated using the explicit model agrees with general tendencies observed during laboratory chamber experiments. This explicit modelling of SOA formation appears as a useful exploratory tool to (i support interpretations of SOA formation observed in laboratory chamber experiments, (ii give some insights on SOA formation under atmospherically relevant conditions and (iii investigate implications for the regional/global lifetimes of the SOA.

  6. SPATIALLY-EXPLICIT BAT IMPACT SCREENING TOOL FOR WIND TURBINE SITING

    Energy Technology Data Exchange (ETDEWEB)

    Versar, Inc.; Exponent, Inc.

    2013-10-28

    As the U.S. seeks to increase energy production from renewable energy sources, development of wind power resources continues to grow. One of the most important ecological issues restricting wind energy development, especially the siting of wind turbines, is the potential adverse effect on bats. High levels of bat fatality have been recorded at a number of wind energy facilities, especially in the eastern United States. The U.S. Department of Energy contracted with Versar, Inc., and Exponent to develop a spatially-explicit site screening tool to evaluate the mortality of bats resulting from interactions (collisions or barotrauma) with wind turbines. The resulting Bat Vulnerability Assessment Tool (BVAT) presented in this report integrates spatial information about turbine locations, bat habitat features, and bat behavior as it relates to possible interactions with turbines. A model demonstration was conducted that focuses on two bat species, the eastern red bat (Lasiurus borealis) and the Indiana bat (Myotis sodalis). The eastern red bat is a relatively common tree-roosting species that ranges broadly during migration in the Eastern U.S., whereas the Indiana bat is regional species that migrates between a summer range and cave hibernacula. Moreover, Indiana bats are listed as endangered, and so the impacts to this species are of particular interest. The model demonstration used conditions at the Mountaineer Wind Energy Center (MWEC), which consists of 44 wind turbines arranged in a linear array near Thomas, West Virginia (Tucker County), to illustrate model functions and not to represent actual or potential impacts of the facility. The turbines at MWEC are erected on the ridge of Backbone Mountain with a nacelle height of 70 meters and a collision area of 72 meters (blade height) or 4,071 meters square. The habitat surrounding the turbines is an Appalachian mixed mesophytic forest. Model sensitivity runs showed that bat mortality in the model was most sensitive to

  7. Spatially explicit modelling of cholera epidemics

    Science.gov (United States)

    Finger, F.; Bertuzzo, E.; Mari, L.; Knox, A. C.; Gatto, M.; Rinaldo, A.

    2013-12-01

    Epidemiological models can provide crucial understanding about the dynamics of infectious diseases. Possible applications range from real-time forecasting and allocation of health care resources to testing alternative intervention mechanisms such as vaccines, antibiotics or the improvement of sanitary conditions. We apply a spatially explicit model to the cholera epidemic that struck Haiti in October 2010 and is still ongoing. The dynamics of susceptibles as well as symptomatic and asymptomatic infectives are modelled at the scale of local human communities. Dissemination of Vibrio cholerae through hydrological transport and human mobility along the road network is explicitly taken into account, as well as the effect of rainfall as a driver of increasing disease incidence. The model is calibrated using a dataset of reported cholera cases. We further model the long term impact of several types of interventions on the disease dynamics by varying parameters appropriately. Key epidemiological mechanisms and parameters which affect the efficiency of treatments such as antibiotics are identified. Our results lead to conclusions about the influence of different intervention strategies on the overall epidemiological dynamics.

  8. Are mixed explicit/implicit solvation models reliable for studying phosphate hydrolysis? A comparative study of continuum, explicit and mixed solvation models.

    Energy Technology Data Exchange (ETDEWEB)

    Kamerlin, Shina C. L.; Haranczyk, Maciej; Warshel, Arieh

    2009-05-01

    Phosphate hydrolysis is ubiquitous in biology. However, despite intensive research on this class of reactions, the precise nature of the reaction mechanism remains controversial. In this work, we have examined the hydrolysis of three homologous phosphate diesters. The solvation free energy was simulated by means of either an implicit solvation model (COSMO), hybrid quantum mechanical / molecular mechanical free energy perturbation (QM/MM-FEP) or a mixed solvation model in which N water molecules were explicitly included in the ab initio description of the reacting system (where N=1-3), with the remainder of the solvent being implicitly modelled as a continuum. Here, both COSMO and QM/MM-FEP reproduce Delta Gobs within an error of about 2kcal/mol. However, we demonstrate that in order to obtain any form of reliable results from a mixed model, it is essential to carefully select the explicit water molecules from short QM/MM runs that act as a model for the true infinite system. Additionally, the mixed models tend to be increasingly inaccurate the more explicit water molecules are placed into the system. Thus, our analysis indicates that this approach provides an unreliable way for modelling phosphate hydrolysis in solution.

  9. An Efficient Explicit-time Description Method for Timed Model Checking

    Directory of Open Access Journals (Sweden)

    Hao Wang

    2009-12-01

    Full Text Available Timed model checking, the method to formally verify real-time systems, is attracting increasing attention from both the model checking community and the real-time community. Explicit-time description methods verify real-time systems using general model constructs found in standard un-timed model checkers. Lamport proposed an explicit-time description method using a clock-ticking process (Tick to simulate the passage of time together with a group of global variables to model time requirements. Two methods, the Sync-based Explicit-time Description Method using rendezvous synchronization steps and the Semaphore-based Explicit-time Description Method using only one global variable were proposed; they both achieve better modularity than Lamport's method in modeling the real-time systems. In contrast to timed automata based model checkers like UPPAAL, explicit-time description methods can access and store the current time instant for future calculations necessary for many real-time systems, especially those with pre-emptive scheduling. However, the Tick process in the above three methods increments the time by one unit in each tick; the state spaces therefore grow relatively fast as the time parameters increase, a problem when the system's time period is relatively long. In this paper, we propose a more efficient method which enables the Tick process to leap multiple time units in one tick. Preliminary experimental results in a high performance computing environment show that this new method significantly reduces the state space and improves both the time and memory efficiency.

  10. Verifying Real-Time Systems using Explicit-time Description Methods

    Directory of Open Access Journals (Sweden)

    Hao Wang

    2009-12-01

    Full Text Available Timed model checking has been extensively researched in recent years. Many new formalisms with time extensions and tools based on them have been presented. On the other hand, Explicit-Time Description Methods aim to verify real-time systems with general untimed model checkers. Lamport presented an explicit-time description method using a clock-ticking process (Tick to simulate the passage of time together with a group of global variables for time requirements. This paper proposes a new explicit-time description method with no reliance on global variables. Instead, it uses rendezvous synchronization steps between the Tick process and each system process to simulate time. This new method achieves better modularity and facilitates usage of more complex timing constraints. The two explicit-time description methods are implemented in DIVINE, a well-known distributed-memory model checker. Preliminary experiment results show that our new method, with better modularity, is comparable to Lamport's method with respect to time and memory efficiency.

  11. Depletion benchmarks calculation of random media using explicit modeling approach of RMC

    International Nuclear Information System (INIS)

    Liu, Shichang; She, Ding; Liang, Jin-gang; Wang, Kan

    2016-01-01

    Highlights: • Explicit modeling of RMC is applied to depletion benchmark for HTGR fuel element. • Explicit modeling can provide detailed burnup distribution and burnup heterogeneity. • The results would serve as a supplement for the HTGR fuel depletion benchmark. • The method of adjacent burnup regions combination is proposed for full-core problems. • The combination method can reduce memory footprint, keeping the computing accuracy. - Abstract: Monte Carlo method plays an important role in accurate simulation of random media, owing to its advantages of the flexible geometry modeling and the use of continuous-energy nuclear cross sections. Three stochastic geometry modeling methods including Random Lattice Method, Chord Length Sampling and explicit modeling approach with mesh acceleration technique, have been implemented in RMC to simulate the particle transport in the dispersed fuels, in which the explicit modeling method is regarded as the best choice. In this paper, the explicit modeling method is applied to the depletion benchmark for HTGR fuel element, and the method of combination of adjacent burnup regions has been proposed and investigated. The results show that the explicit modeling can provide detailed burnup distribution of individual TRISO particles, and this work would serve as a supplement for the HTGR fuel depletion benchmark calculations. The combination of adjacent burnup regions can effectively reduce the memory footprint while keeping the computational accuracy.

  12. A new parallelization algorithm of ocean model with explicit scheme

    Science.gov (United States)

    Fu, X. D.

    2017-08-01

    This paper will focus on the parallelization of ocean model with explicit scheme which is one of the most commonly used schemes in the discretization of governing equation of ocean model. The characteristic of explicit schema is that calculation is simple, and that the value of the given grid point of ocean model depends on the grid point at the previous time step, which means that one doesn’t need to solve sparse linear equations in the process of solving the governing equation of the ocean model. Aiming at characteristics of the explicit scheme, this paper designs a parallel algorithm named halo cells update with tiny modification of original ocean model and little change of space step and time step of the original ocean model, which can parallelize ocean model by designing transmission module between sub-domains. This paper takes the GRGO for an example to implement the parallelization of GRGO (Global Reduced Gravity Ocean model) with halo update. The result demonstrates that the higher speedup can be achieved at different problem size.

  13. Quantifying multiple telecouplings using an integrated suite of spatially-explicit tools

    Science.gov (United States)

    Tonini, F.; Liu, J.

    2016-12-01

    Telecoupling is an interdisciplinary research umbrella concept that enables natural and social scientists to understand and generate information for managing how humans and nature can sustainably coexist worldwide. To systematically study telecoupling, it is essential to build a comprehensive set of spatially-explicit tools for describing and quantifying multiple reciprocal socioeconomic and environmental interactions between a focal area and other areas. Here we introduce the Telecoupling Toolbox, a new free and open-source set of tools developed to map and identify the five major interrelated components of the telecoupling framework: systems, flows, agents, causes, and effects. The modular design of the toolbox allows the integration of existing tools and software (e.g. InVEST) to assess synergies and tradeoffs associated with policies and other local to global interventions. We show applications of the toolbox using a number of representative studies that address a variety of scientific and management issues related to telecouplings throughout the world. The results suggest that the toolbox can thoroughly map and quantify multiple telecouplings under various contexts while providing users with an easy-to-use interface. It provides a powerful platform to address globally important issues, such as land use and land cover change, species invasion, migration, flows of ecosystem services, and international trade of goods and products.

  14. Modelling the Hydraulic Behaviour of Growing Media with the Explicit Finite Volume Solution

    Directory of Open Access Journals (Sweden)

    Marco Carbone

    2015-02-01

    Full Text Available The increasing imperviousness of urban areas reduces the infiltration and evapotranspiration capacity of urban catchments and results in increased runoff. In the last few decades, several solutions and techniques have been proposed to prevent such impacts by restoring the hydrological cycle. A limiting factor in spreading the use of such systems is the lack of proper modelling tools for design, especially for the infiltration processes in a growing medium. In this research, a physically-based model, employing the explicit Finite Volume Method (FVM, is proposed for modelling infiltration into growing media. The model solves a modified version of the Richards equation using a formulation which takes into account the main characteristics of green infrastructure substrates. The proposed model was verified against the HYDRUS-1D software and the comparison of results confirmed the suitability of the proposed model for correctly describing the hydraulic behaviour of soil substrates.

  15. Recent Advances in Explicit Multiparametric Nonlinear Model Predictive Control

    KAUST Repository

    Domínguez, Luis F.

    2011-01-19

    In this paper we present recent advances in multiparametric nonlinear programming (mp-NLP) algorithms for explicit nonlinear model predictive control (mp-NMPC). Three mp-NLP algorithms for NMPC are discussed, based on which novel mp-NMPC controllers are derived. The performance of the explicit controllers are then tested and compared in a simulation example involving the operation of a continuous stirred-tank reactor (CSTR). © 2010 American Chemical Society.

  16. Computer system for identification of tool wear model in hot forging

    Directory of Open Access Journals (Sweden)

    Wilkus Marek

    2016-01-01

    Full Text Available The aim of the research was to create a methodology that will enable effective and reliable prediction of the tool wear. The idea of the hybrid model, which accounts for various mechanisms of tool material deterioration, is proposed in the paper. The mechanisms, which were considered, include abrasive wear, adhesive wear, thermal fatigue, mechanical fatigue, oxidation and plastic deformation. Individual models of various complexity were used for separate phenomena and strategy of combination of these models in one hybrid system was developed to account for the synergy of various mechanisms. The complex hybrid model was built on the basis of these individual models for various wear mechanisms. The individual models expanded from phenomenological ones for abrasive wear to multi-scale methods for modelling micro cracks initiation and propagation utilizing virtual representations of granular microstructures. The latter have been intensively developed recently and they form potentially a powerful tool that allows modelling of thermal and mechanical fatigue, accounting explicitly for the tool material microstructure.

  17. Novel application of explicit dynamics occupancy models to ongoing aquatic invasions

    Science.gov (United States)

    Sepulveda, Adam J.

    2018-01-01

    Identification of suitable habitats, where invasive species can establish, is an important step towards controlling their spread. Accurate identification is difficult for new or slow invaders because unoccupied habitats may be suitable, given enough time for dispersal, while occupied habitats may prove to be unsuitable for establishment.To identify the suitable habitat of a recent invader, I used an explicit dynamics occupancy modelling framework to evaluate habitat covariates related to successful and failed establishments of American bullfrogs (Lithobates catesbeianus) within the Yellowstone River floodplain of Montana, USA from 2012-2016.During this five-year period, bullfrogs failed to establish at most sites they colonized. Bullfrog establishment was most likely to occur and least likely to fail at sites closest to human-modified ponds and lakes and those with emergent vegetation. These habitat covariates were generally associated with the presence of permanent water.Suitable habitat for bullfrog establishment is abundant in the Yellowstone River floodplain, though many sites with suitable habitat remain uncolonized. Thus, the maximum distribution of bullfrogs is much greater than their current distribution.Synthesis and applications. Focused control efforts on habitats with or proximate to permanent waters are most likely to reduce the potential for invasive bullfrog establishment and spread in the Yellowstone River. The novel application of explicit dynamics occupancy models is a useful and widely applicable tool for guiding management efforts towards those habitats where new or slow invaders are most likely to establish and persist.

  18. Comparison of explicit and effective models for calculating ionic populations in argon plasmas

    International Nuclear Information System (INIS)

    Abdallah, J. Jr.; Clark, R.E.H.

    1994-01-01

    Calculations have been performed to model the state populations of argon plasmas at electron densities at and above those required for the validity of coronal equilibrium. Both effective and explicit models have been used, and both are based on the same set of atomic cross sections. The effective model includes ground and singly excited states explicitly, while the effect of autoionizing states is accounted for by branching factors which describe their depopulation into the various non-autoionizing states. The explicit model considers both autoionizing and non-autoionizing states explicitly. The effective model requires a significantly reduced amount of computer time and memory. Good agreement between the two models can be obtained through moderate densities if the branching factors include electron density dependent terms which describe the collisional stabilization of each autoionizing state. The effective model breaks down as density is increased because the population of individual autoionizing states become significant. Results for both ionization balance and radiated power loss are presented. (Author)

  19. Modeling Agricultural Watersheds with the Soil and Water Assessment Tool (SWAT): Calibration and Validation with a Novel Procedure for Spatially Explicit HRUs.

    Science.gov (United States)

    Teshager, Awoke Dagnew; Gassman, Philip W; Secchi, Silvia; Schoof, Justin T; Misgna, Girmaye

    2016-04-01

    Applications of the Soil and Water Assessment Tool (SWAT) model typically involve delineation of a watershed into subwatersheds/subbasins that are then further subdivided into hydrologic response units (HRUs) which are homogeneous areas of aggregated soil, landuse, and slope and are the smallest modeling units used within the model. In a given standard SWAT application, multiple potential HRUs (farm fields) in a subbasin are usually aggregated into a single HRU feature. In other words, the standard version of the model combines multiple potential HRUs (farm fields) with the same landuse/landcover, soil, and slope, but located at different places of a subbasin (spatially non-unique), and considers them as one HRU. In this study, ArcGIS pre-processing procedures were developed to spatially define a one-to-one match between farm fields and HRUs (spatially unique HRUs) within a subbasin prior to SWAT simulations to facilitate input processing, input/output mapping, and further analysis at the individual farm field level. Model input data such as landuse/landcover (LULC), soil, crop rotation, and other management data were processed through these HRUs. The SWAT model was then calibrated/validated for Raccoon River watershed in Iowa for 2002-2010 and Big Creek River watershed in Illinois for 2000-2003. SWAT was able to replicate annual, monthly, and daily streamflow, as well as sediment, nitrate and mineral phosphorous within recommended accuracy in most cases. The one-to-one match between farm fields and HRUs created and used in this study is a first step in performing LULC change, climate change impact, and other analyses in a more spatially explicit manner.

  20. Modelling conflict management in design: an explicit approach

    NARCIS (Netherlands)

    Brazier, F.M.; van Langen, P.H.G.; Treur, J.

    1995-01-01

    This paper focusses on how conflicts that arise during a design process and the management of conflicts can be modelled. A number of possible conflict types are distinguished and it is described how each of them can be detected during the design process, using an explicit meta-representation.

  1. Spatially explicit models, generalized reproduction numbers and the prediction of patterns of waterborne disease

    Science.gov (United States)

    Rinaldo, A.; Gatto, M.; Mari, L.; Casagrandi, R.; Righetto, L.; Bertuzzo, E.; Rodriguez-Iturbe, I.

    2012-12-01

    still lacking. Here, we show that the requirement that all the local reproduction numbers R0 be larger than unity is neither necessary nor sufficient for outbreaks to occur when local settlements are connected by networks of primary and secondary infection mechanisms. To determine onset conditions, we derive general analytical expressions for a reproduction matrix G0 explicitly accounting for spatial distributions of human settlements and pathogen transmission via hydrological and human mobility networks. At disease onset, a generalized reproduction number Λ0 (the dominant eigenvalue of G0) must be larger than unity. We also show that geographical outbreak patterns in complex environments are linked to the dominant eigenvector and to spectral properties of G0. Tests against data and computations for the 2010 Haiti and 2000 KwaZulu-Natal cholera outbreaks, as well as against computations for metapopulation networks, demonstrate that eigenvectors of G0 provide a synthetic and effective tool for predicting the disease course in space and time. Networked connectivity models, describing the interplay between hydrology, epidemiology and social behavior sustaining human mobility, thus prove to be key tools for emergency management of waterborne infections.

  2. Explicit all-atom modeling of realistically sized ligand-capped nanocrystals

    KAUST Repository

    Kaushik, Ananth P.

    2012-01-01

    We present a study of an explicit all-atom representation of nanocrystals of experimentally relevant sizes (up to 6 nm), capped with alkyl chain ligands, in vacuum. We employ all-atom molecular dynamics simulation methods in concert with a well-tested intermolecular potential model, MM3 (molecular mechanics 3), for the studies presented here. These studies include determining the preferred conformation of an isolated single nanocrystal (NC), pairs of isolated NCs, and (presaging studies of superlattice arrays) unit cells of NC superlattices. We observe that very small NCs (3 nm) behave differently in a superlattice as compared to larger NCs (6 nm and above) due to the conformations adopted by the capping ligands on the NC surface. Short ligands adopt a uniform distribution of orientational preferences, including some that lie against the face of the nanocrystal. In contrast, longer ligands prefer to interdigitate. We also study the effect of changing ligand length and ligand coverage on the NCs on the preferred ligand configurations. Since explicit all-atom modeling constrains the maximum system size that can be studied, we discuss issues related to coarse-graining the representation of the ligands, including a comparison of two commonly used coarse-grained models. We find that care has to be exercised in the choice of coarse-grained model. The data provided by these realistically sized ligand-capped NCs, determined using explicit all-atom models, should serve as a reference standard for future models of coarse-graining ligands using united atom models, especially for self-assembly processes. © 2012 American Institute of Physics.

  3. Modeling Active Aging and Explicit Memory: An Empirical Study.

    Science.gov (United States)

    Ponce de León, Laura Ponce; Lévy, Jean Pierre; Fernández, Tomás; Ballesteros, Soledad

    2015-08-01

    The rapid growth of the population of older adults and their concomitant psychological status and health needs have captured the attention of researchers and health professionals. To help fill the void of literature available to social workers interested in mental health promotion and aging, the authors provide a model for active aging that uses psychosocial variables. Structural equation modeling was used to examine the relationships among the latent variables of the state of explicit memory, the perception of social resources, depression, and the perception of quality of life in a sample of 184 older adults. The results suggest that explicit memory is not a direct indicator of the perception of quality of life, but it could be considered an indirect indicator as it is positively correlated with perception of social resources and negatively correlated with depression. These last two variables influenced the perception of quality of life directly, the former positively and the latter negatively. The main outcome suggests that the perception of social support improves explicit memory and quality of life and reduces depression in active older adults. The findings also suggest that gerontological professionals should design memory training programs, improve available social resources, and offer environments with opportunities to exercise memory.

  4. Memory Efficient Data Structures for Explicit Verification of Timed Systems

    DEFF Research Database (Denmark)

    Taankvist, Jakob Haahr; Srba, Jiri; Larsen, Kim Guldstrand

    2014-01-01

    Timed analysis of real-time systems can be performed using continuous (symbolic) or discrete (explicit) techniques. The explicit state-space exploration can be considerably faster for models with moderately small constants, however, at the expense of high memory consumption. In the setting of timed......-arc Petri nets, we explore new data structures for lowering the used memory: PTries for efficient storing of configurations and time darts for semi-symbolic description of the state-space. Both methods are implemented as a part of the tool TAPAAL and the experiments document at least one order of magnitude...... of memory savings while preserving comparable verification times....

  5. A unitary signal-detection model of implicit and explicit memory.

    Science.gov (United States)

    Berry, Christopher J; Shanks, David R; Henson, Richard N A

    2008-10-01

    Do dissociations imply independent systems? In the memory field, the view that there are independent implicit and explicit memory systems has been predominantly supported by dissociation evidence. Here, we argue that many of these dissociations do not necessarily imply distinct memory systems. We review recent work with a single-system computational model that extends signal-detection theory (SDT) to implicit memory. SDT has had a major influence on research in a variety of domains. The current work shows that it can be broadened even further in its range of application. Indeed, the single-system model that we present does surprisingly well in accounting for some key dissociations that have been taken as evidence for independent implicit and explicit memory systems.

  6. Scaling-up spatially-explicit ecological models using graphics processors

    NARCIS (Netherlands)

    Koppel, Johan van de; Gupta, Rohit; Vuik, Cornelis

    2011-01-01

    How the properties of ecosystems relate to spatial scale is a prominent topic in current ecosystem research. Despite this, spatially explicit models typically include only a limited range of spatial scales, mostly because of computing limitations. Here, we describe the use of graphics processors to

  7. Tacit to explicit knowledge conversion.

    Science.gov (United States)

    Cairó Battistutti, Osvaldo; Bork, Dominik

    2017-11-01

    The ability to create, use and transfer knowledge may allow the creation or improvement of new products or services. But knowledge is often tacit: It lives in the minds of individuals, and therefore, it is difficult to transfer it to another person by means of the written word or verbal expression. This paper addresses this important problem by introducing a methodology, consisting of a four-step process that facilitates tacit to explicit knowledge conversion. The methodology utilizes conceptual modeling, thus enabling understanding and reasoning through visual knowledge representation. This implies the possibility of understanding concepts and ideas, visualized through conceptual models, without using linguistic or algebraic means. The proposed methodology is conducted in a metamodel-based tool environment whose aim is efficient application and ease of use.

  8. Explicit estimating equations for semiparametric generalized linear latent variable models

    KAUST Repository

    Ma, Yanyuan

    2010-07-05

    We study generalized linear latent variable models without requiring a distributional assumption of the latent variables. Using a geometric approach, we derive consistent semiparametric estimators. We demonstrate that these models have a property which is similar to that of a sufficient complete statistic, which enables us to simplify the estimating procedure and explicitly to formulate the semiparametric estimating equations. We further show that the explicit estimators have the usual root n consistency and asymptotic normality. We explain the computational implementation of our method and illustrate the numerical performance of the estimators in finite sample situations via extensive simulation studies. The advantage of our estimators over the existing likelihood approach is also shown via numerical comparison. We employ the method to analyse a real data example from economics. © 2010 Royal Statistical Society.

  9. Flood vulnerability assessment of residential buildings by explicit damage process modelling

    DEFF Research Database (Denmark)

    Custer, Rocco; Nishijima, Kazuyoshi

    2015-01-01

    The present paper introduces a vulnerability modelling approach for residential buildings in flood. The modelling approach explicitly considers relevant damage processes, i.e. water infiltration into the building, mechanical failure of components in the building envelope and damage from water...

  10. DYNAMO-HIA--a Dynamic Modeling tool for generic Health Impact Assessments.

    Directory of Open Access Journals (Sweden)

    Stefan K Lhachimi

    Full Text Available BACKGROUND: Currently, no standard tool is publicly available that allows researchers or policy-makers to quantify the impact of policies using epidemiological evidence within the causal framework of Health Impact Assessment (HIA. A standard tool should comply with three technical criteria (real-life population, dynamic projection, explicit risk-factor states and three usability criteria (modest data requirements, rich model output, generally accessible to be useful in the applied setting of HIA. With DYNAMO-HIA (Dynamic Modeling for Health Impact Assessment, we introduce such a generic software tool specifically designed to facilitate quantification in the assessment of the health impacts of policies. METHODS AND RESULTS: DYNAMO-HIA quantifies the impact of user-specified risk-factor changes on multiple diseases and in turn on overall population health, comparing one reference scenario with one or more intervention scenarios. The Markov-based modeling approach allows for explicit risk-factor states and simulation of a real-life population. A built-in parameter estimation module ensures that only standard population-level epidemiological evidence is required, i.e. data on incidence, prevalence, relative risks, and mortality. DYNAMO-HIA provides a rich output of summary measures--e.g. life expectancy and disease-free life expectancy--and detailed data--e.g. prevalences and mortality/survival rates--by age, sex, and risk-factor status over time. DYNAMO-HIA is controlled via a graphical user interface and is publicly available from the internet, ensuring general accessibility. We illustrate the use of DYNAMO-HIA with two example applications: a policy causing an overall increase in alcohol consumption and quantifying the disease-burden of smoking. CONCLUSION: By combining modest data needs with general accessibility and user friendliness within the causal framework of HIA, DYNAMO-HIA is a potential standard tool for health impact assessment based

  11. Predicting continental-scale patterns of bird species richness with spatially explicit models

    DEFF Research Database (Denmark)

    Rahbek, Carsten; Gotelli, Nicholas J; Colwell, Robert K

    2007-01-01

    the extraordinary diversity of avian species in the montane tropics, the most species-rich region on Earth. Our findings imply that correlative climatic models substantially underestimate the importance of historical factors and small-scale niche-driven assembly processes in shaping contemporary species-richness......The causes of global variation in species richness have been debated for nearly two centuries with no clear resolution in sight. Competing hypotheses have typically been evaluated with correlative models that do not explicitly incorporate the mechanisms responsible for biotic diversity gradients....... Here, we employ a fundamentally different approach that uses spatially explicit Monte Carlo models of the placement of cohesive geographical ranges in an environmentally heterogeneous landscape. These models predict species richness of endemic South American birds (2248 species) measured...

  12. Explicit ions/implicit water generalized Born model for nucleic acids

    Science.gov (United States)

    Tolokh, Igor S.; Thomas, Dennis G.; Onufriev, Alexey V.

    2018-05-01

    The ion atmosphere around highly charged nucleic acid molecules plays a significant role in their dynamics, structure, and interactions. Here we utilized the implicit solvent framework to develop a model for the explicit treatment of ions interacting with nucleic acid molecules. The proposed explicit ions/implicit water model is based on a significantly modified generalized Born (GB) model and utilizes a non-standard approach to define the solute/solvent dielectric boundary. Specifically, the model includes modifications to the GB interaction terms for the case of multiple interacting solutes—disconnected dielectric boundary around the solute-ion or ion-ion pairs. A fully analytical description of all energy components for charge-charge interactions is provided. The effectiveness of the approach is demonstrated by calculating the potential of mean force for Na+-Cl- ion pair and by carrying out a set of Monte Carlo (MC) simulations of mono- and trivalent ions interacting with DNA and RNA duplexes. The monovalent (Na+) and trivalent (CoHex3+) counterion distributions predicted by the model are in close quantitative agreement with all-atom explicit water molecular dynamics simulations used as reference. Expressed in the units of energy, the maximum deviations of local ion concentrations from the reference are within kBT. The proposed explicit ions/implicit water GB model is able to resolve subtle features and differences of CoHex distributions around DNA and RNA duplexes. These features include preferential CoHex binding inside the major groove of the RNA duplex, in contrast to CoHex biding at the "external" surface of the sugar-phosphate backbone of the DNA duplex; these differences in the counterion binding patters were earlier shown to be responsible for the observed drastic differences in condensation propensities between short DNA and RNA duplexes. MC simulations of CoHex ions interacting with the homopolymeric poly(dA.dT) DNA duplex with modified (de

  13. Spatially explicit modeling of greater sage-grouse (Centrocercus urophasianus) habitat in Nevada and northeastern California: a decision-support tool for management

    Science.gov (United States)

    Coates, Peter S.; Casazza, Michael L.; Brussee, Brianne E.; Ricca, Mark A.; Gustafson, K. Benjamin; Overton, Cory T.; Sanchez-Chopitea, Erika; Kroger, Travis; Mauch, Kimberly; Niell, Lara; Howe, Kristy; Gardner, Scott; Espinosa, Shawn; Delehanty, David J.

    2014-01-01

    Greater sage-grouse (Centrocercus urophasianus, hereafter referred to as “sage-grouse”) populations are declining throughout the sagebrush (Artemisia spp.) ecosystem, including millions of acres of potential habitat across the West. Habitat maps derived from empirical data are needed given impending listing decisions that will affect both sage-grouse population dynamics and human land-use restrictions. This report presents the process for developing spatially explicit maps describing relative habitat suitability for sage-grouse in Nevada and northeastern California. Maps depicting habitat suitability indices (HSI) values were generated based on model-averaged resource selection functions informed by more than 31,000 independent telemetry locations from more than 1,500 radio-marked sage-grouse across 12 project areas in Nevada and northeastern California collected during a 15-year period (1998–2013). Modeled habitat covariates included land cover composition, water resources, habitat configuration, elevation, and topography, each at multiple spatial scales that were relevant to empirically observed sage-grouse movement patterns. We then present an example of how the HSI can be delineated into categories. Specifically, we demonstrate that the deviation from the mean can be used to classify habitat suitability into three categories of habitat quality (high, moderate, and low) and one non-habitat category. The classification resulted in an agreement of 93–97 percent for habitat versus non-habitat across a suite of independent validation datasets. Lastly, we provide an example of how space use models can be integrated with habitat models to help inform conservation planning. In this example, we combined probabilistic breeding density with a non-linear probability of occurrence relative to distance to nearest lek (traditional breeding ground) using count data to calculate a composite space use index (SUI). The SUI was then classified into two categories of use

  14. Skeeter Buster: a stochastic, spatially explicit modeling tool for studying Aedes aegypti population replacement and population suppression strategies.

    Directory of Open Access Journals (Sweden)

    Krisztian Magori

    2009-09-01

    Full Text Available Dengue is the most important mosquito-borne viral disease affecting humans. The only prevention measure currently available is the control of its vectors, primarily Aedes aegypti. Recent advances in genetic engineering have opened the possibility for a new range of control strategies based on genetically modified mosquitoes. Assessing the potential efficacy of genetic (and conventional strategies requires the availability of modeling tools that accurately describe the dynamics and genetics of Ae. aegypti populations.We describe in this paper a new modeling tool of Ae. aegypti population dynamics and genetics named Skeeter Buster. This model operates at the scale of individual water-filled containers for immature stages and individual properties (houses for adults. The biology of cohorts of mosquitoes is modeled based on the algorithms used in the non-spatial Container Inhabiting Mosquitoes Simulation Model (CIMSiM. Additional features incorporated into Skeeter Buster include stochasticity, spatial structure and detailed population genetics. We observe that the stochastic modeling of individual containers in Skeeter Buster is associated with a strongly reduced temporal variation in stage-specific population densities. We show that heterogeneity in container composition of individual properties has a major impact on spatial heterogeneity in population density between properties. We detail how adult dispersal reduces this spatial heterogeneity. Finally, we present the predicted genetic structure of the population by calculating F(ST values and isolation by distance patterns, and examine the effects of adult dispersal and container movement between properties.We demonstrate that the incorporated stochasticity and level of spatial detail have major impacts on the simulated population dynamics, which could potentially impact predictions in terms of control measures. The capacity to describe population genetics confers the ability to model the outcome

  15. Explicit Modeling of Ancestry Improves Polygenic Risk Scores and BLUP Prediction.

    Science.gov (United States)

    Chen, Chia-Yen; Han, Jiali; Hunter, David J; Kraft, Peter; Price, Alkes L

    2015-09-01

    Polygenic prediction using genome-wide SNPs can provide high prediction accuracy for complex traits. Here, we investigate the question of how to account for genetic ancestry when conducting polygenic prediction. We show that the accuracy of polygenic prediction in structured populations may be partly due to genetic ancestry. However, we hypothesized that explicitly modeling ancestry could improve polygenic prediction accuracy. We analyzed three GWAS of hair color (HC), tanning ability (TA), and basal cell carcinoma (BCC) in European Americans (sample size from 7,440 to 9,822) and considered two widely used polygenic prediction approaches: polygenic risk scores (PRSs) and best linear unbiased prediction (BLUP). We compared polygenic prediction without correction for ancestry to polygenic prediction with ancestry as a separate component in the model. In 10-fold cross-validation using the PRS approach, the R(2) for HC increased by 66% (0.0456-0.0755; P ancestry, which prevents ancestry effects from entering into each SNP effect and being overweighted. Surprisingly, explicitly modeling ancestry produces a similar improvement when using the BLUP approach, which fits all SNPs simultaneously in a single variance component and causes ancestry to be underweighted. We validate our findings via simulations, which show that the differences in prediction accuracy will increase in magnitude as sample sizes increase. In summary, our results show that explicitly modeling ancestry can be important in both PRS and BLUP prediction. © 2015 WILEY PERIODICALS, INC.

  16. Model of high-tech businesses management under the trends of explicit and implicit knowledge markets: classification and business model

    OpenAIRE

    Guzel Isayevna Gumerova; Elmira Shamilevna Shaimieva

    2015-01-01

    Objective to define the notion of ldquohightech businessrdquo to elaborate classification of hightech businesses to elaborate the business model for hightech business management. Methods general scientific methods of theoretical and empirical cognition. Results the research presents a business model of hightech businesses management basing on the trends of explicit and explicit knowledge market with the dominating implicit knowledge market classification of hightech business...

  17. Comparison of Explicitly Simulated and Downscaled Tropical Cyclone Activity in a High-Resolution Global Climate Model

    Directory of Open Access Journals (Sweden)

    Hirofumi Tomita

    2010-01-01

    Full Text Available The response of tropical cyclone activity to climate change is a matter of great inherent interest and practical importance. Most current global climate models are not, however, capable of adequately resolving tropical cyclones; this has led to the development of downscaling techniques designed to infer tropical cyclone activity from the large-scale fields produced by climate models. Here we compare the statistics of tropical cyclones simulated explicitly in a very high resolution (~14 km grid mesh global climate model to the results of one such downscaling technique driven by the same global model. This is done for a simulation of the current climate and also for a simulation of a climate warmed by the addition of carbon dioxide. The explicitly simulated and downscaled storms are similarly distributed in space, but the intensity distribution of the downscaled events has a somewhat longer high-intensity tail, owing to the higher resolution of the downscaling model. Both explicitly simulated and downscaled events show large increases in the frequency of events at the high-intensity ends of their respective intensity distributions, but the downscaled storms also show increases in low-intensity events, whereas the explicitly simulated weaker events decline in number. On the regional scale, there are large differences in the responses of the explicitly simulated and downscaled events to global warming. In particular, the power dissipation of downscaled events shows a 175% increase in the Atlantic, while the power dissipation of explicitly simulated events declines there.

  18. Dynamic modeling and explicit/multi-parametric MPC control of pressure swing adsorption systems

    KAUST Repository

    Khajuria, Harish

    2011-01-01

    Pressure swing adsorption (PSA) is a flexible, albeit complex gas separation system. Due to its inherent nonlinear nature and discontinuous operation, the design of a model based PSA controller, especially with varying operating conditions, is a challenging task. This work focuses on the design of an explicit/multi-parametric model predictive controller for a PSA system. Based on a system involving four adsorbent beds separating 70% H2, 30% CH4 mixture into high purity hydrogen, the key controller objective is to fast track H2 purity to a set point value of 99.99%. To perform this task, a rigorous and systematic framework is employed. First, a high fidelity detailed dynamic model is built to represent the system\\'s real operation, and understand its dynamic behavior. The model is then used to derive appropriate linear models by applying suitable system identification techniques. For the reduced models, a model predictive control (MPC) step is formulated, where latest developments in multi-parametric programming and control are applied to derive a novel explicit MPC controller. To test the performance of the designed controller, closed loop simulations are performed where the dynamic model is used as the virtual plant. Comparison studies of the derived explicit MPC controller are also performed with conventional PID controllers. © 2010 Elsevier Ltd. All rights reserved.

  19. High Performance Programming Using Explicit Shared Memory Model on Cray T3D1

    Science.gov (United States)

    Simon, Horst D.; Saini, Subhash; Grassi, Charles

    1994-01-01

    The Cray T3D system is the first-phase system in Cray Research, Inc.'s (CRI) three-phase massively parallel processing (MPP) program. This system features a heterogeneous architecture that closely couples DEC's Alpha microprocessors and CRI's parallel-vector technology, i.e., the Cray Y-MP and Cray C90. An overview of the Cray T3D hardware and available programming models is presented. Under Cray Research adaptive Fortran (CRAFT) model four programming methods (data parallel, work sharing, message-passing using PVM, and explicit shared memory model) are available to the users. However, at this time data parallel and work sharing programming models are not available to the user community. The differences between standard PVM and CRI's PVM are highlighted with performance measurements such as latencies and communication bandwidths. We have found that the performance of neither standard PVM nor CRI s PVM exploits the hardware capabilities of the T3D. The reasons for the bad performance of PVM as a native message-passing library are presented. This is illustrated by the performance of NAS Parallel Benchmarks (NPB) programmed in explicit shared memory model on Cray T3D. In general, the performance of standard PVM is about 4 to 5 times less than obtained by using explicit shared memory model. This degradation in performance is also seen on CM-5 where the performance of applications using native message-passing library CMMD on CM-5 is also about 4 to 5 times less than using data parallel methods. The issues involved (such as barriers, synchronization, invalidating data cache, aligning data cache etc.) while programming in explicit shared memory model are discussed. Comparative performance of NPB using explicit shared memory programming model on the Cray T3D and other highly parallel systems such as the TMC CM-5, Intel Paragon, Cray C90, IBM-SP1, etc. is presented.

  20. Prediction of strongly-heated gas flows in a vertical tube using explicit algebraic stress/heat-flux models

    International Nuclear Information System (INIS)

    Baek, Seong Gu; Park, Seung O.

    2003-01-01

    This paper provides the assessment of prediction performance of explicit algebraic stress and heat-flux models under conditions of mixed convective gas flows in a strongly-heated vertical tube. Two explicit algebraic stress models and four algebraic heat-flux models are selected for assessment. Eight combinations of explicit algebraic stress and heat-flux models are used in predicting the flows experimentally studied by Shehata and McEligot (IJHMT 41(1998) p.4333) in which property variation was significant. Among the various model combinations, the Wallin and Johansson (JFM 403(2000) p. 89) explicit algebraic stress model-Abe, Kondo, and Nagano (IJHFF 17(1996) p. 228) algebraic heat-flux model combination is found to perform best. We also found that the dimensionless wall distance y + should be calculated based on the local property rather than the property at the wall for property-variation flows. When the buoyancy or the property variation effects are so strong that the flow may relaminarize, the choice of the basic platform two-equation model is a most important factor in improving the predictions

  1. Precision tools and models to narrow in on the 750 GeV diphoton resonance

    International Nuclear Information System (INIS)

    Staub, Florian; Athron, Peter; Basso, Lorenzo

    2016-02-01

    The hints for a new resonance at 750 GeV from ATLAS and CMS have triggered a significant amount of attention. Since the simplest extensions of the standard model cannot accommodate the observation, many alternatives have been considered to explain the excess. Here we focus on several proposed renormalisable weakly-coupled models and revisit results given in the literature. We point out that physically important subtleties are often missed or neglected. To facilitate the study of the excess we have created a collection of 40 model files, selected from recent literature, for the Mathematica package SARAH. With SARAH one can generate files to perform numerical studies using the tailor-made spectrum generators FlexibleSUSY and SPheno. These have been extended to automatically include crucial higher order corrections to the diphoton and digluon decay rates for both CP-even and CP-odd scalars. Additionally, we have extended the UFO and CalcHep interfaces of SARAH, to pass the precise information about the effective vertices from the spectrum generator to a Monte-Carlo tool. Finally, as an example to demonstrate the power of the entire setup, we present a new supersymmetric model that accommodates the diphoton excess, explicitly demonstrating how a large width can be obtained. We explicitly show several steps in detail to elucidate the use of these public tools in the precision study of this model.

  2. Precision tools and models to narrow in on the 750 GeV diphoton resonance

    Energy Technology Data Exchange (ETDEWEB)

    Staub, Florian [CERN, Geneva (Switzerland). Theoretical Physics Dept.; Athron, Peter [Monash Univ., Melbourne (Australia). ARC Center of Excellence for Particle Physics at the Terascale; Basso, Lorenzo [Aix-Marseille Univ., CNRS-IN2P3, UMR 7346 (France). CPPM; and others

    2016-02-15

    The hints for a new resonance at 750 GeV from ATLAS and CMS have triggered a significant amount of attention. Since the simplest extensions of the standard model cannot accommodate the observation, many alternatives have been considered to explain the excess. Here we focus on several proposed renormalisable weakly-coupled models and revisit results given in the literature. We point out that physically important subtleties are often missed or neglected. To facilitate the study of the excess we have created a collection of 40 model files, selected from recent literature, for the Mathematica package SARAH. With SARAH one can generate files to perform numerical studies using the tailor-made spectrum generators FlexibleSUSY and SPheno. These have been extended to automatically include crucial higher order corrections to the diphoton and digluon decay rates for both CP-even and CP-odd scalars. Additionally, we have extended the UFO and CalcHep interfaces of SARAH, to pass the precise information about the effective vertices from the spectrum generator to a Monte-Carlo tool. Finally, as an example to demonstrate the power of the entire setup, we present a new supersymmetric model that accommodates the diphoton excess, explicitly demonstrating how a large width can be obtained. We explicitly show several steps in detail to elucidate the use of these public tools in the precision study of this model.

  3. Precision tools and models to narrow in on the 750 GeV diphoton resonance

    International Nuclear Information System (INIS)

    Staub, Florian; Athron, Peter; Basso, Lorenzo; Goodsell, Mark D.; Harries, Dylan; Krauss, Manuel E.; Nickel, Kilian; Opferkuch, Toby; Ubaldi, Lorenzo; Vicente, Avelino; Voigt, Alexander

    2016-01-01

    The hints for a new resonance at 750 GeV from ATLAS and CMS have triggered a significant amount of attention. Since the simplest extensions of the standard model cannot accommodate the observation, many alternatives have been considered to explain the excess. Here we focus on several proposed renormalisable weakly-coupled models and revisit results given in the literature. We point out that physically important subtleties are often missed or neglected. To facilitate the study of the excess we have created a collection of 40 model files, selected from recent literature, for the Mathematica package SARAH. With SARAH one can generate files to perform numerical studies using the tailor-made spectrum generators FlexibleSUSY and SPheno. These have been extended to automatically include crucial higher order corrections to the diphoton and digluon decay rates for both CP-even and CP-odd scalars. Additionally, we have extended the UFO and CalcHep interfaces of SARAH, to pass the precise information about the effective vertices from the spectrum generator to a Monte-Carlo tool. Finally, as an example to demonstrate the power of the entire setup, we present a new supersymmetric model that accommodates the diphoton excess, explicitly demonstrating how a large width can be obtained. We explicitly show several steps in detail to elucidate the use of these public tools in the precision study of this model. (orig.)

  4. Precision tools and models to narrow in on the 750 GeV diphoton resonance

    Energy Technology Data Exchange (ETDEWEB)

    Staub, Florian [CERN, Theoretical Physics Department, Geneva (Switzerland); Athron, Peter [Monash University, ARC Centre of Excellence for Particle Physics at the Terascale, School of Physics, Melbourne, VIC (Australia); Basso, Lorenzo [CPPM, Aix-Marseille Universite, CNRS-IN2P3, UMR 7346, Marseille Cedex 9 (France); Goodsell, Mark D. [Sorbonne Universites, LPTHE, UMR 7589, CNRS and Universite Pierre et Marie Curie, Paris Cedex 05 (France); Harries, Dylan [The University of Adelaide, Department of Physics, ARC Centre of Excellence for Particle Physics at the Terascale, Adelaide, SA (Australia); Krauss, Manuel E.; Nickel, Kilian; Opferkuch, Toby [Bethe Center for Theoretical Physics and Physikalisches Institut der Universitaet Bonn, Bonn (Germany); Ubaldi, Lorenzo [Tel-Aviv University, Raymond and Beverly Sackler School of Physics and Astronomy, Tel Aviv (Israel); Vicente, Avelino [Instituto de Fisica Corpuscular (CSIC-Universitat de Valencia), Valencia (Spain); Voigt, Alexander [Deutsches Elektronen-Synchrotron DESY, Hamburg (Germany)

    2016-09-15

    The hints for a new resonance at 750 GeV from ATLAS and CMS have triggered a significant amount of attention. Since the simplest extensions of the standard model cannot accommodate the observation, many alternatives have been considered to explain the excess. Here we focus on several proposed renormalisable weakly-coupled models and revisit results given in the literature. We point out that physically important subtleties are often missed or neglected. To facilitate the study of the excess we have created a collection of 40 model files, selected from recent literature, for the Mathematica package SARAH. With SARAH one can generate files to perform numerical studies using the tailor-made spectrum generators FlexibleSUSY and SPheno. These have been extended to automatically include crucial higher order corrections to the diphoton and digluon decay rates for both CP-even and CP-odd scalars. Additionally, we have extended the UFO and CalcHep interfaces of SARAH, to pass the precise information about the effective vertices from the spectrum generator to a Monte-Carlo tool. Finally, as an example to demonstrate the power of the entire setup, we present a new supersymmetric model that accommodates the diphoton excess, explicitly demonstrating how a large width can be obtained. We explicitly show several steps in detail to elucidate the use of these public tools in the precision study of this model. (orig.)

  5. Modeling the Explicit Chemistry of Anthropogenic and Biogenic Organic Aerosols

    Energy Technology Data Exchange (ETDEWEB)

    Madronich, Sasha [Univ. Corporation for Atmospheric Research, Boulder, CO (United States)

    2015-12-09

    The atmospheric burden of Secondary Organic Aerosols (SOA) remains one of the most important yet uncertain aspects of the radiative forcing of climate. This grant focused on improving our quantitative understanding of SOA formation and evolution, by developing, applying, and improving a highly detailed model of atmospheric organic chemistry, the Generation of Explicit Chemistry and Kinetics of Organics in the Atmosphere (GECKO-A) model. Eleven (11) publications have resulted from this grant.

  6. Spatially explicit multi-criteria decision analysis for managing vector-borne diseases

    Science.gov (United States)

    2011-01-01

    The complex epidemiology of vector-borne diseases creates significant challenges in the design and delivery of prevention and control strategies, especially in light of rapid social and environmental changes. Spatial models for predicting disease risk based on environmental factors such as climate and landscape have been developed for a number of important vector-borne diseases. The resulting risk maps have proven value for highlighting areas for targeting public health programs. However, these methods generally only offer technical information on the spatial distribution of disease risk itself, which may be incomplete for making decisions in a complex situation. In prioritizing surveillance and intervention strategies, decision-makers often also need to consider spatially explicit information on other important dimensions, such as the regional specificity of public acceptance, population vulnerability, resource availability, intervention effectiveness, and land use. There is a need for a unified strategy for supporting public health decision making that integrates available data for assessing spatially explicit disease risk, with other criteria, to implement effective prevention and control strategies. Multi-criteria decision analysis (MCDA) is a decision support tool that allows for the consideration of diverse quantitative and qualitative criteria using both data-driven and qualitative indicators for evaluating alternative strategies with transparency and stakeholder participation. Here we propose a MCDA-based approach to the development of geospatial models and spatially explicit decision support tools for the management of vector-borne diseases. We describe the conceptual framework that MCDA offers as well as technical considerations, approaches to implementation and expected outcomes. We conclude that MCDA is a powerful tool that offers tremendous potential for use in public health decision-making in general and vector-borne disease management in particular

  7. Spatially explicit multi-criteria decision analysis for managing vector-borne diseases

    Directory of Open Access Journals (Sweden)

    Hongoh Valerie

    2011-12-01

    Full Text Available Abstract The complex epidemiology of vector-borne diseases creates significant challenges in the design and delivery of prevention and control strategies, especially in light of rapid social and environmental changes. Spatial models for predicting disease risk based on environmental factors such as climate and landscape have been developed for a number of important vector-borne diseases. The resulting risk maps have proven value for highlighting areas for targeting public health programs. However, these methods generally only offer technical information on the spatial distribution of disease risk itself, which may be incomplete for making decisions in a complex situation. In prioritizing surveillance and intervention strategies, decision-makers often also need to consider spatially explicit information on other important dimensions, such as the regional specificity of public acceptance, population vulnerability, resource availability, intervention effectiveness, and land use. There is a need for a unified strategy for supporting public health decision making that integrates available data for assessing spatially explicit disease risk, with other criteria, to implement effective prevention and control strategies. Multi-criteria decision analysis (MCDA is a decision support tool that allows for the consideration of diverse quantitative and qualitative criteria using both data-driven and qualitative indicators for evaluating alternative strategies with transparency and stakeholder participation. Here we propose a MCDA-based approach to the development of geospatial models and spatially explicit decision support tools for the management of vector-borne diseases. We describe the conceptual framework that MCDA offers as well as technical considerations, approaches to implementation and expected outcomes. We conclude that MCDA is a powerful tool that offers tremendous potential for use in public health decision-making in general and vector

  8. Testing the cognitive catalyst model of rumination with explicit and implicit cognitive content.

    Science.gov (United States)

    Sova, Christopher C; Roberts, John E

    2018-06-01

    The cognitive catalyst model posits that rumination and negative cognitive content, such as negative schema, interact to predict depressive affect. Past research has found support for this model using explicit measures of negative cognitive content such as self-report measures of trait self-esteem and dysfunctional attitudes. The present study tested whether these findings would extend to implicit measures of negative cognitive content such as implicit self-esteem, and whether effects would depend on initial mood state and history of depression. Sixty-one undergraduate students selected on the basis of depression history (27 previously depressed; 34 never depressed) completed explicit and implicit measures of negative cognitive content prior to random assignment to a rumination induction followed by a distraction induction or vice versa. Dysphoric affect was measured both before and after these inductions. Analyses revealed that explicit measures, but not implicit measures, interacted with rumination to predict change in dysphoric affect, and these interactions were further moderated by baseline levels of dysphoria. Limitations include the small nonclinical sample and use of a self-report measure of depression history. These findings suggest that rumination amplifies the association between explicit negative cognitive content and depressive affect primarily among people who are already experiencing sad mood. Copyright © 2018 Elsevier Ltd. All rights reserved.

  9. Spatially explicit modeling of annual and seasonal habitat for greater sage-grouse (Centrocercus urophasianus) in Nevada and Northeastern California—An updated decision-support tool for management

    Science.gov (United States)

    Coates, Peter S.; Casazza, Michael L.; Brussee, Brianne E.; Ricca, Mark A.; Gustafson, K. Benjamin; Sanchez-Chopitea, Erika; Mauch, Kimberly; Niell, Lara; Gardner, Scott; Espinosa, Shawn; Delehanty, David J.

    2016-05-20

    Successful adaptive management hinges largely upon integrating new and improved sources of information as they become available. As a timely example of this tenet, we updated a management decision support tool that was previously developed for greater sage-grouse (Centrocercus urophasianus, hereinafter referred to as “sage-grouse”) populations in Nevada and California. Specifically, recently developed spatially explicit habitat maps derived from empirical data played a key role in the conservation of this species facing listing under the Endangered Species Act. This report provides an updated process for mapping relative habitat suitability and management categories for sage-grouse in Nevada and northeastern California (Coates and others, 2014, 2016). These updates include: (1) adding radio and GPS telemetry locations from sage-grouse monitored at multiple sites during 2014 to the original location dataset beginning in 1998; (2) integrating output from high resolution maps (1–2 m2) of sagebrush and pinyon-juniper cover as covariates in resource selection models; (3) modifying the spatial extent of the analyses to match newly available vegetation layers; (4) explicit modeling of relative habitat suitability during three seasons (spring, summer, winter) that corresponded to critical life history periods for sage-grouse (breeding, brood-rearing, over-wintering); (5) accounting for differences in habitat availability between more mesic sagebrush steppe communities in the northern part of the study area and drier Great Basin sagebrush in more southerly regions by categorizing continuous region-wide surfaces of habitat suitability index (HSI) with independent locations falling within two hydrological zones; (6) integrating the three seasonal maps into a composite map of annual relative habitat suitability; (7) deriving updated land management categories based on previously determined cut-points for intersections of habitat suitability and an updated index of sage

  10. Modeling single versus multiple systems in implicit and explicit memory.

    Science.gov (United States)

    Starns, Jeffrey J; Ratcliff, Roger; McKoon, Gail

    2012-04-01

    It is currently controversial whether priming on implicit tasks and discrimination on explicit recognition tests are supported by a single memory system or by multiple, independent systems. In a Psychological Review article, Berry and colleagues used mathematical modeling to address this question and provide compelling evidence against the independent-systems approach. Copyright © 2012 Elsevier Ltd. All rights reserved.

  11. Fire Propagation Tracing Model in the Explicit Treatment of Events of Fire PSA

    International Nuclear Information System (INIS)

    Lim, Ho Gon; Han, Sang Hoon; Yang, Jun Eon

    2010-01-01

    The fire propagation model in a fire PSA has not been considered analytically instead a simplified analyst's intuition was used to consider the fire propagation path. A fire propagation equation is developed to trace all the propagation paths in the fire area in which a zone is defined to identify various fire ignition sources. An initiation of fire is assumed to take place in a zone. Then, the propagation is modeled with a Boolean equation. Since the explicit fire PSA modeling requires an exclusive event set to sum up the..., exclusive event sets are derived from the fire propagation equation. As an example, we show the exclusive set for a 2x3 rectangular fire zone. Also, the applicability the developed fire equation is discussed when the number of zone increases including the limitation of the explicit fire PSA modeling method

  12. Model of high-tech businesses management under the trends of explicit and implicit knowledge markets: classification and business model

    Directory of Open Access Journals (Sweden)

    Guzel Isayevna Gumerova

    2015-03-01

    Full Text Available Objective to define the notion of ldquohightech businessrdquo to elaborate classification of hightech businesses to elaborate the business model for hightech business management. Methods general scientific methods of theoretical and empirical cognition. Results the research presents a business model of hightech businesses management basing on the trends of explicit and explicit knowledge market with the dominating implicit knowledge market classification of hightech businesses taking into consideration the three types of economic activity possibilities to manage hightech business basing on its market cost technological innovations costs and business indicators. Scientific novelty the interpretation of the notion of ldquohightech businessrdquo has been renewed the classification of hightech businesses has been elaborated for the first time allocating three groups of enterprises. Practical value theoretical significance ndash development of notional apparatus of hightech business management practical significancenbsp ndash grounding of the necessity to manage enterprises under development of explicit and explicit knowledge markets in Russia as a complex of capital and noncapital assets with dominating indicators of ldquomarket valuerdquo and ldquolife span of a companyrdquo. nbsp

  13. DEFINING RECOVERY GOALS AND STRATEGIES FOR ENDANGERED SPECIES USING SPATIALLY-EXPLICIT POPULATION MODELS

    Science.gov (United States)

    We used a spatially explicit population model of wolves (Canis lupus) to propose a framework for defining rangewide recovery priorities and finer-scale strategies for regional reintroductions. The model predicts that Yellowstone and central Idaho, where wolves have recently been ...

  14. Spatially-Explicit Bayesian Information Entropy Metrics for Calibrating Landscape Transformation Models

    Directory of Open Access Journals (Sweden)

    Kostas Alexandridis

    2013-06-01

    Full Text Available Assessing spatial model performance often presents challenges related to the choice and suitability of traditional statistical methods in capturing the true validity and dynamics of the predicted outcomes. The stochastic nature of many of our contemporary spatial models of land use change necessitate the testing and development of new and innovative methodologies in statistical spatial assessment. In many cases, spatial model performance depends critically on the spatially-explicit prior distributions, characteristics, availability and prevalence of the variables and factors under study. This study explores the statistical spatial characteristics of statistical model assessment of modeling land use change dynamics in a seven-county study area in South-Eastern Wisconsin during the historical period of 1963–1990. The artificial neural network-based Land Transformation Model (LTM predictions are used to compare simulated with historical land use transformations in urban/suburban landscapes. We introduce a range of Bayesian information entropy statistical spatial metrics for assessing the model performance across multiple simulation testing runs. Bayesian entropic estimates of model performance are compared against information-theoretic stochastic entropy estimates and theoretically-derived accuracy assessments. We argue for the critical role of informational uncertainty across different scales of spatial resolution in informing spatial landscape model assessment. Our analysis reveals how incorporation of spatial and landscape information asymmetry estimates can improve our stochastic assessments of spatial model predictions. Finally our study shows how spatially-explicit entropic classification accuracy estimates can work closely with dynamic modeling methodologies in improving our scientific understanding of landscape change as a complex adaptive system and process.

  15. Fuselage Versus Subcomponent Panel Response Correlation Based on ABAQUS Explicit Progressive Damage Analysis Tools

    Science.gov (United States)

    Gould, Kevin E.; Satyanarayana, Arunkumar; Bogert, Philip B.

    2016-01-01

    Analysis performed in this study substantiates the need for high fidelity vehicle level progressive damage analyses (PDA) structural models for use in the verification and validation of proposed sub-scale structural models and to support required full-scale vehicle level testing. PDA results are presented that capture and correlate the responses of sub-scale 3-stringer and 7-stringer panel models and an idealized 8-ft diameter fuselage model, which provides a vehicle level environment for the 7-stringer sub-scale panel model. Two unique skin-stringer attachment assumptions are considered and correlated in the models analyzed: the TIE constraint interface versus the cohesive element (COH3D8) interface. Evaluating different interfaces allows for assessing a range of predicted damage modes, including delamination and crack propagation responses. Damage models considered in this study are the ABAQUS built-in Hashin procedure and the COmplete STress Reduction (COSTR) damage procedure implemented through a VUMAT user subroutine using the ABAQUS/Explicit code.

  16. Probabilistic modelling in urban drainage – two approaches that explicitly account for temporal variation of model errors

    DEFF Research Database (Denmark)

    Löwe, Roland; Del Giudice, Dario; Mikkelsen, Peter Steen

    of input uncertainties observed in the models. The explicit inclusion of such variations in the modelling process will lead to a better fulfilment of the assumptions made in formal statistical frameworks, thus reducing the need to resolve to informal methods. The two approaches presented here...

  17. Analysis of explicit model predictive control for path-following control.

    Science.gov (United States)

    Lee, Junho; Chang, Hyuk-Jun

    2018-01-01

    In this paper, explicit Model Predictive Control(MPC) is employed for automated lane-keeping systems. MPC has been regarded as the key to handle such constrained systems. However, the massive computational complexity of MPC, which employs online optimization, has been a major drawback that limits the range of its target application to relatively small and/or slow problems. Explicit MPC can reduce this computational burden using a multi-parametric quadratic programming technique(mp-QP). The control objective is to derive an optimal front steering wheel angle at each sampling time so that autonomous vehicles travel along desired paths, including straight, circular, and clothoid parts, at high entry speeds. In terms of the design of the proposed controller, a method of choosing weighting matrices in an optimization problem and the range of horizons for path-following control are described through simulations. For the verification of the proposed controller, simulation results obtained using other control methods such as MPC, Linear-Quadratic Regulator(LQR), and driver model are employed, and CarSim, which reflects the features of a vehicle more realistically than MATLAB/Simulink, is used for reliable demonstration.

  18. Analysis of explicit model predictive control for path-following control

    Science.gov (United States)

    2018-01-01

    In this paper, explicit Model Predictive Control(MPC) is employed for automated lane-keeping systems. MPC has been regarded as the key to handle such constrained systems. However, the massive computational complexity of MPC, which employs online optimization, has been a major drawback that limits the range of its target application to relatively small and/or slow problems. Explicit MPC can reduce this computational burden using a multi-parametric quadratic programming technique(mp-QP). The control objective is to derive an optimal front steering wheel angle at each sampling time so that autonomous vehicles travel along desired paths, including straight, circular, and clothoid parts, at high entry speeds. In terms of the design of the proposed controller, a method of choosing weighting matrices in an optimization problem and the range of horizons for path-following control are described through simulations. For the verification of the proposed controller, simulation results obtained using other control methods such as MPC, Linear-Quadratic Regulator(LQR), and driver model are employed, and CarSim, which reflects the features of a vehicle more realistically than MATLAB/Simulink, is used for reliable demonstration. PMID:29534080

  19. A spatially explicit scenario-driven model of adaptive capacity to global change in Europe

    NARCIS (Netherlands)

    Acosta, L.; Klein, R.J.T.; Reidsma, P.; Metzger, M.J.; Rounsevell, M.D.A.; Leemans, R.

    2013-01-01

    Traditional impact models combine exposure in the form of scenarios and sensitivity in the form of parameters, providing potential impacts of global change as model outputs. However, adaptive capacity is rarely addressed in these models. This paper presents the first spatially explicit

  20. Implicit-explicit (IMEX) Runge-Kutta methods for non-hydrostatic atmospheric models

    Science.gov (United States)

    Gardner, David J.; Guerra, Jorge E.; Hamon, François P.; Reynolds, Daniel R.; Ullrich, Paul A.; Woodward, Carol S.

    2018-04-01

    The efficient simulation of non-hydrostatic atmospheric dynamics requires time integration methods capable of overcoming the explicit stability constraints on time step size arising from acoustic waves. In this work, we investigate various implicit-explicit (IMEX) additive Runge-Kutta (ARK) methods for evolving acoustic waves implicitly to enable larger time step sizes in a global non-hydrostatic atmospheric model. The IMEX formulations considered include horizontally explicit - vertically implicit (HEVI) approaches as well as splittings that treat some horizontal dynamics implicitly. In each case, the impact of solving nonlinear systems in each implicit ARK stage in a linearly implicit fashion is also explored. The accuracy and efficiency of the IMEX splittings, ARK methods, and solver options are evaluated on a gravity wave and baroclinic wave test case. HEVI splittings that treat some vertical dynamics explicitly do not show a benefit in solution quality or run time over the most implicit HEVI formulation. While splittings that implicitly evolve some horizontal dynamics increase the maximum stable step size of a method, the gains are insufficient to overcome the additional cost of solving a globally coupled system. Solving implicit stage systems in a linearly implicit manner limits the solver cost but this is offset by a reduction in step size to achieve the desired accuracy for some methods. Overall, the third-order ARS343 and ARK324 methods performed the best, followed by the second-order ARS232 and ARK232 methods.

  1. Implicit–explicit (IMEX Runge–Kutta methods for non-hydrostatic atmospheric models

    Directory of Open Access Journals (Sweden)

    D. J. Gardner

    2018-04-01

    Full Text Available The efficient simulation of non-hydrostatic atmospheric dynamics requires time integration methods capable of overcoming the explicit stability constraints on time step size arising from acoustic waves. In this work, we investigate various implicit–explicit (IMEX additive Runge–Kutta (ARK methods for evolving acoustic waves implicitly to enable larger time step sizes in a global non-hydrostatic atmospheric model. The IMEX formulations considered include horizontally explicit – vertically implicit (HEVI approaches as well as splittings that treat some horizontal dynamics implicitly. In each case, the impact of solving nonlinear systems in each implicit ARK stage in a linearly implicit fashion is also explored.The accuracy and efficiency of the IMEX splittings, ARK methods, and solver options are evaluated on a gravity wave and baroclinic wave test case. HEVI splittings that treat some vertical dynamics explicitly do not show a benefit in solution quality or run time over the most implicit HEVI formulation. While splittings that implicitly evolve some horizontal dynamics increase the maximum stable step size of a method, the gains are insufficient to overcome the additional cost of solving a globally coupled system. Solving implicit stage systems in a linearly implicit manner limits the solver cost but this is offset by a reduction in step size to achieve the desired accuracy for some methods. Overall, the third-order ARS343 and ARK324 methods performed the best, followed by the second-order ARS232 and ARK232 methods.

  2. A web-tool to find spatially explicit climate-smart solutions for the sector agriculture

    Science.gov (United States)

    Verzandvoort, Simone; Kuikman, Peter; Walvoort, Dennis

    2017-04-01

    Europe faces the challenge to produce more food and more biomass for the bio-economy, to adapt its agricultural sector to negative consequences of climate change, and to reduce greenhouse gas emissions from agriculture. Climate-smart agriculture (CSA) solutions and technologies improve agriculture's productivity and provide economic growth and stability, increase resilience, and help to reduce GHG emissions from agricultural activities. The Climate Smart Agriculture Booster (CSAb) (http://csabooster.climate-kic.org/) is a Flagship Program under Climate-KIC, aiming to facilitate the adoption of CSA solutions and technologies in the European agro-food sector. This adoption requires spatially explicit, contextual information on farming activities and risks and opportunities related to climate change in regions across Europe. Other spatial information supporting adoption includes Information on where successful implementations were already done, on where CSA would profit from enabling policy conditions, and where markets or business opportunities for selling or purchasing technology and knowledge are located or emerging. The Spatial Solution Finder is a web-based spatial tool aiming to help agri-food companies (supply and processing), authorities or agricultural organisations find CSA solutions and technologies that fit local farmers and regions, and to demonstrate examples of successful implementations as well as expected impact at the farm and regional level. The tool is based on state of the art (geo)datasets of environmental and socio-economic conditions (partly open access, partly derived from previous research) and open source web-technology. The philosophy of the tool is that combining existing datasets with contextual information on the region of interest with personalized information entered by the user provides a suitable basis for offering a basket of options for CSA solutions and technologies. Solutions and technologies are recommended to the user based on

  3. The Explicit-Cloud Parameterized-Pollutant hybrid approach for aerosol-cloud interactions in multiscale modeling framework models: tracer transport results

    International Nuclear Information System (INIS)

    Jr, William I Gustafson; Berg, Larry K; Easter, Richard C; Ghan, Steven J

    2008-01-01

    All estimates of aerosol indirect effects on the global energy balance have either completely neglected the influence of aerosol on convective clouds or treated the influence in a highly parameterized manner. Embedding cloud-resolving models (CRMs) within each grid cell of a global model provides a multiscale modeling framework for treating both the influence of aerosols on convective as well as stratiform clouds and the influence of clouds on the aerosol, but treating the interactions explicitly by simulating all aerosol processes in the CRM is computationally prohibitive. An alternate approach is to use horizontal statistics (e.g., cloud mass flux, cloud fraction, and precipitation) from the CRM simulation to drive a single-column parameterization of cloud effects on the aerosol and then use the aerosol profile to simulate aerosol effects on clouds within the CRM. Here, we present results from the first component of the Explicit-Cloud Parameterized-Pollutant parameterization to be developed, which handles vertical transport of tracers by clouds. A CRM with explicit tracer transport serves as a benchmark. We show that this parameterization, driven by the CRM's cloud mass fluxes, reproduces the CRM tracer transport significantly better than a single-column model that uses a conventional convective cloud parameterization

  4. The Explicit-Cloud Parameterized-Pollutant hybrid approach for aerosol-cloud interactions in multiscale modeling framework models: tracer transport results

    Energy Technology Data Exchange (ETDEWEB)

    Jr, William I Gustafson; Berg, Larry K; Easter, Richard C; Ghan, Steven J [Atmospheric Science and Global Change Division, Pacific Northwest National Laboratory, PO Box 999, MSIN K9-30, Richland, WA (United States)], E-mail: William.Gustafson@pnl.gov

    2008-04-15

    All estimates of aerosol indirect effects on the global energy balance have either completely neglected the influence of aerosol on convective clouds or treated the influence in a highly parameterized manner. Embedding cloud-resolving models (CRMs) within each grid cell of a global model provides a multiscale modeling framework for treating both the influence of aerosols on convective as well as stratiform clouds and the influence of clouds on the aerosol, but treating the interactions explicitly by simulating all aerosol processes in the CRM is computationally prohibitive. An alternate approach is to use horizontal statistics (e.g., cloud mass flux, cloud fraction, and precipitation) from the CRM simulation to drive a single-column parameterization of cloud effects on the aerosol and then use the aerosol profile to simulate aerosol effects on clouds within the CRM. Here, we present results from the first component of the Explicit-Cloud Parameterized-Pollutant parameterization to be developed, which handles vertical transport of tracers by clouds. A CRM with explicit tracer transport serves as a benchmark. We show that this parameterization, driven by the CRM's cloud mass fluxes, reproduces the CRM tracer transport significantly better than a single-column model that uses a conventional convective cloud parameterization.

  5. Dynamic optimization and robust explicit model predictive control of hydrogen storage tank

    KAUST Repository

    Panos, C.

    2010-09-01

    We present a general framework for the optimal design and control of a metal-hydride bed under hydrogen desorption operation. The framework features: (i) a detailed two-dimension dynamic process model, (ii) a design and operational dynamic optimization step, and (iii) an explicit/multi-parametric model predictive controller design step. For the controller design, a reduced order approximate model is obtained, based on which nominal and robust multi-parametric controllers are designed. © 2010 Elsevier Ltd.

  6. Dynamic optimization and robust explicit model predictive control of hydrogen storage tank

    KAUST Repository

    Panos, C.; Kouramas, K.I.; Georgiadis, M.C.; Pistikopoulos, E.N.

    2010-01-01

    We present a general framework for the optimal design and control of a metal-hydride bed under hydrogen desorption operation. The framework features: (i) a detailed two-dimension dynamic process model, (ii) a design and operational dynamic optimization step, and (iii) an explicit/multi-parametric model predictive controller design step. For the controller design, a reduced order approximate model is obtained, based on which nominal and robust multi-parametric controllers are designed. © 2010 Elsevier Ltd.

  7. Speech Enhancement Using Gaussian Mixture Models, Explicit Bayesian Estimation and Wiener Filtering

    Directory of Open Access Journals (Sweden)

    M. H. Savoji

    2014-09-01

    Full Text Available Gaussian Mixture Models (GMMs of power spectral densities of speech and noise are used with explicit Bayesian estimations in Wiener filtering of noisy speech. No assumption is made on the nature or stationarity of the noise. No voice activity detection (VAD or any other means is employed to estimate the input SNR. The GMM mean vectors are used to form sets of over-determined system of equations whose solutions lead to the first estimates of speech and noise power spectra. The noise source is also identified and the input SNR estimated in this first step. These first estimates are then refined using approximate but explicit MMSE and MAP estimation formulations. The refined estimates are then used in a Wiener filter to reduce noise and enhance the noisy speech. The proposed schemes show good results. Nevertheless, it is shown that the MAP explicit solution, introduced here for the first time, reduces the computation time to less than one third with a slight higher improvement in SNR and PESQ score and also less distortion in comparison to the MMSE solution.

  8. Latin hypercube sampling and geostatistical modeling of spatial uncertainty in a spatially explicit forest landscape model simulation

    Science.gov (United States)

    Chonggang Xu; Hong S. He; Yuanman Hu; Yu Chang; Xiuzhen Li; Rencang Bu

    2005-01-01

    Geostatistical stochastic simulation is always combined with Monte Carlo method to quantify the uncertainty in spatial model simulations. However, due to the relatively long running time of spatially explicit forest models as a result of their complexity, it is always infeasible to generate hundreds or thousands of Monte Carlo simulations. Thus, it is of great...

  9. Design and application of a technologically explicit hybrid energy-economy policy model with micro and macro economic dynamics

    Science.gov (United States)

    Bataille, Christopher G. F.

    2005-11-01

    autonomous energy efficiency indices (AEEI) from the model, parameters that could be used in long-run computable general equilibrium (CGE) analysis. The thesis concludes with a summary of the strengths and weakness of the new model as a policy tool, a work plan for its further improvement, and a discussion of the general potential for technologically explicit general equilibrium modelling.

  10. Finite Element Modelling of the effect of tool rake angle on tool temperature and cutting force during high speed machining of AISI 4340 steel

    International Nuclear Information System (INIS)

    Sulaiman, S; Roshan, A; Ariffin, M K A

    2013-01-01

    In this paper, a Finite Element Method (FEM) based on the ABAQUS explicit software which involves Johnson-Cook material model was used to simulate cutting force and tool temperature during high speed machining (HSM) of AISI 4340 steel. In this simulation work, a tool rake angle ranging from 0° to 20° and a range of cutting speeds between 300 to 550 m/min was investigated. The purpose of this simulation analysis was to find optimum tool rake angle where cutting force is smallest as well as tool temperature is lowest during high speed machining. It was found that cutting forces to have a decreasing trend as rake angle increased to positive direction. The optimum rake angle observed between 10° and 18° due to decrease of cutting force as 20% for all simulated cutting speeds. In addition, increasing cutting tool rake angle over its optimum value had negative influence on tool's performance and led to an increase in cutting temperature. The results give a better understanding and recognition of the cutting tool design for high speed machining processes

  11. An Explicit Approach Toward Modeling Thermo-Coupled Deformation Behaviors of SMPs

    Directory of Open Access Journals (Sweden)

    Hao Li

    2017-03-01

    Full Text Available A new elastoplastic J 2 -flow models with thermal effects is proposed toward simulating thermo-coupled finite deformation behaviors of shape memory polymers. In this new model, an elastic potential evolving with development of plastic flow is incorporated to characterize the stress-softening effect at unloading and, moreover, thermo-induced plastic flow is introduced to represent the strain recovery effect at heating. It is shown that any given test data for both effects may be accurately simulated by means of direct and explicit procedures. Numerical examples for model predictions compare well with test data in literature.

  12. Hybrid Neural Network Approach Based Tool for the Modelling of Photovoltaic Panels

    Directory of Open Access Journals (Sweden)

    Antonino Laudani

    2015-01-01

    Full Text Available A hybrid neural network approach based tool for identifying the photovoltaic one-diode model is presented. The generalization capabilities of neural networks are used together with the robustness of the reduced form of one-diode model. Indeed, from the studies performed by the authors and the works present in the literature, it was found that a direct computation of the five parameters via multiple inputs and multiple outputs neural network is a very difficult task. The reduced form consists in a series of explicit formulae for the support to the neural network that, in our case, is aimed at predicting just two parameters among the five ones identifying the model: the other three parameters are computed by reduced form. The present hybrid approach is efficient from the computational cost point of view and accurate in the estimation of the five parameters. It constitutes a complete and extremely easy tool suitable to be implemented in a microcontroller based architecture. Validations are made on about 10000 PV panels belonging to the California Energy Commission database.

  13. Methods used to parameterize the spatially-explicit components of a state-and-transition simulation model

    Directory of Open Access Journals (Sweden)

    Rachel R. Sleeter

    2015-06-01

    Full Text Available Spatially-explicit state-and-transition simulation models of land use and land cover (LULC increase our ability to assess regional landscape characteristics and associated carbon dynamics across multiple scenarios. By characterizing appropriate spatial attributes such as forest age and land-use distribution, a state-and-transition model can more effectively simulate the pattern and spread of LULC changes. This manuscript describes the methods and input parameters of the Land Use and Carbon Scenario Simulator (LUCAS, a customized state-and-transition simulation model utilized to assess the relative impacts of LULC on carbon stocks for the conterminous U.S. The methods and input parameters are spatially explicit and describe initial conditions (strata, state classes and forest age, spatial multipliers, and carbon stock density. Initial conditions were derived from harmonization of multi-temporal data characterizing changes in land use as well as land cover. Harmonization combines numerous national-level datasets through a cell-based data fusion process to generate maps of primary LULC categories. Forest age was parameterized using data from the North American Carbon Program and spatially-explicit maps showing the locations of past disturbances (i.e. wildfire and harvest. Spatial multipliers were developed to spatially constrain the location of future LULC transitions. Based on distance-decay theory, maps were generated to guide the placement of changes related to forest harvest, agricultural intensification/extensification, and urbanization. We analyze the spatially-explicit input parameters with a sensitivity analysis, by showing how LUCAS responds to variations in the model input. This manuscript uses Mediterranean California as a regional subset to highlight local to regional aspects of land change, which demonstrates the utility of LUCAS at many scales and applications.

  14. Methods used to parameterize the spatially-explicit components of a state-and-transition simulation model

    Science.gov (United States)

    Sleeter, Rachel; Acevedo, William; Soulard, Christopher E.; Sleeter, Benjamin M.

    2015-01-01

    Spatially-explicit state-and-transition simulation models of land use and land cover (LULC) increase our ability to assess regional landscape characteristics and associated carbon dynamics across multiple scenarios. By characterizing appropriate spatial attributes such as forest age and land-use distribution, a state-and-transition model can more effectively simulate the pattern and spread of LULC changes. This manuscript describes the methods and input parameters of the Land Use and Carbon Scenario Simulator (LUCAS), a customized state-and-transition simulation model utilized to assess the relative impacts of LULC on carbon stocks for the conterminous U.S. The methods and input parameters are spatially explicit and describe initial conditions (strata, state classes and forest age), spatial multipliers, and carbon stock density. Initial conditions were derived from harmonization of multi-temporal data characterizing changes in land use as well as land cover. Harmonization combines numerous national-level datasets through a cell-based data fusion process to generate maps of primary LULC categories. Forest age was parameterized using data from the North American Carbon Program and spatially-explicit maps showing the locations of past disturbances (i.e. wildfire and harvest). Spatial multipliers were developed to spatially constrain the location of future LULC transitions. Based on distance-decay theory, maps were generated to guide the placement of changes related to forest harvest, agricultural intensification/extensification, and urbanization. We analyze the spatially-explicit input parameters with a sensitivity analysis, by showing how LUCAS responds to variations in the model input. This manuscript uses Mediterranean California as a regional subset to highlight local to regional aspects of land change, which demonstrates the utility of LUCAS at many scales and applications.

  15. Introduction to the Explicit Finite Element Method for Nonlinear Transient Dynamics

    CERN Document Server

    Wu, Shen R

    2012-01-01

    A systematic introduction to the theories and formulations of the explicit finite element method As numerical technology continues to grow and evolve with industrial applications, understanding the explicit finite element method has become increasingly important, particularly in the areas of crashworthiness, metal forming, and impact engineering. Introduction to the Explicit FiniteElement Method for Nonlinear Transient Dynamics is the first book to address specifically what is now accepted as the most successful numerical tool for nonlinear transient dynamics. The book aids readers in master

  16. Building an explicit de Sitter

    Energy Technology Data Exchange (ETDEWEB)

    Louis, Jan [Hamburg Univ. (Germany). 2. Inst. fuer Theoretische Physik; Hamburg Univ. (Germany). Zentrum fuer Mathematische Physik; Rummel, Markus; Valandro, Roberto [Hamburg Univ. (Germany). 2. Inst. fuer Theoretische Physik; Westphal, Alexander [Deutsches Elektronen-Synchrotron (DESY), Hamburg (Germany). Gruppe Theorie

    2012-11-15

    We construct an explicit example of a de Sitter vacuum in type IIB string theory that realizes the proposal of Kaehler uplifting. As the large volume limit in this method depends on the rank of the largest condensing gauge group we carry out a scan of gauge group ranks over the Kreuzer-Skarke set of toric Calabi-Yau threefolds. We find large numbers of models with the largest gauge group factor easily exceeding a rank of one hundred. We construct a global model with Kaehler uplifting on a two-parameter model on CP{sup 4}{sub 11169}, by an explicit analysis from both the type IIB and F-theory point of view. The explicitness of the construction lies in the realization of a D7 brane configuration, gauge flux and RR and NS flux choices, such that all known consistency conditions are met and the geometric moduli are stabilized in a metastable de Sitter vacuum with spontaneous GUT scale supersymmetry breaking driven by an F-term of the Kaehler moduli.

  17. Building an explicit de Sitter

    International Nuclear Information System (INIS)

    Louis, Jan; Hamburg Univ.; Rummel, Markus; Valandro, Roberto; Westphal, Alexander

    2012-11-01

    We construct an explicit example of a de Sitter vacuum in type IIB string theory that realizes the proposal of Kaehler uplifting. As the large volume limit in this method depends on the rank of the largest condensing gauge group we carry out a scan of gauge group ranks over the Kreuzer-Skarke set of toric Calabi-Yau threefolds. We find large numbers of models with the largest gauge group factor easily exceeding a rank of one hundred. We construct a global model with Kaehler uplifting on a two-parameter model on CP 4 11169 , by an explicit analysis from both the type IIB and F-theory point of view. The explicitness of the construction lies in the realization of a D7 brane configuration, gauge flux and RR and NS flux choices, such that all known consistency conditions are met and the geometric moduli are stabilized in a metastable de Sitter vacuum with spontaneous GUT scale supersymmetry breaking driven by an F-term of the Kaehler moduli.

  18. Explicit estimating equations for semiparametric generalized linear latent variable models

    KAUST Repository

    Ma, Yanyuan; Genton, Marc G.

    2010-01-01

    which is similar to that of a sufficient complete statistic, which enables us to simplify the estimating procedure and explicitly to formulate the semiparametric estimating equations. We further show that the explicit estimators have the usual root n

  19. Cholera in the Lake Kivu region (DRC): Integrating remote sensing and spatially explicit epidemiological modeling

    Science.gov (United States)

    Finger, Flavio; Knox, Allyn; Bertuzzo, Enrico; Mari, Lorenzo; Bompangue, Didier; Gatto, Marino; Rodriguez-Iturbe, Ignacio; Rinaldo, Andrea

    2014-07-01

    Mathematical models of cholera dynamics can not only help in identifying environmental drivers and processes that influence disease transmission, but may also represent valuable tools for the prediction of the epidemiological patterns in time and space as well as for the allocation of health care resources. Cholera outbreaks have been reported in the Democratic Republic of the Congo since the 1970s. They have been ravaging the shore of Lake Kivu in the east of the country repeatedly during the last decades. Here we employ a spatially explicit, inhomogeneous Markov chain model to describe cholera incidence in eight health zones on the shore of the lake. Remotely sensed data sets of chlorophyll a concentration in the lake, precipitation and indices of global climate anomalies are used as environmental drivers in addition to baseline seasonality. The effect of human mobility is also modelled mechanistically. We test several models on a multiyear data set of reported cholera cases. The best fourteen models, accounting for different environmental drivers, and selected using the Akaike information criterion, are formally compared via proper cross validation. Among these, the one accounting for seasonality, El Niño Southern Oscillation, precipitation and human mobility outperforms the others in cross validation. Some drivers (such as human mobility and rainfall) are retained only by a few models, possibly indicating that the mechanisms through which they influence cholera dynamics in the area will have to be investigated further.

  20. Comparison of BrainTool to other UML modeling and model transformation tools

    Science.gov (United States)

    Nikiforova, Oksana; Gusarovs, Konstantins

    2017-07-01

    In the last 30 years there were numerous model generated software systems offered targeting problems with the development productivity and the resulting software quality. CASE tools developed due today's date are being advertised as having "complete code-generation capabilities". Nowadays the Object Management Group (OMG) is calling similar arguments in regards to the Unified Modeling Language (UML) models at different levels of abstraction. It is being said that software development automation using CASE tools enables significant level of automation. Actual today's CASE tools are usually offering a combination of several features starting with a model editor and a model repository for a traditional ones and ending with code generator (that could be using a scripting or domain-specific (DSL) language), transformation tool to produce the new artifacts from the manually created and transformation definition editor to define new transformations for the most advanced ones. Present paper contains the results of CASE tool (mainly UML editors) comparison against the level of the automation they are offering.

  1. An Explicit Formula for Symmetric Polynomials Related to the Eigenfunctions of Calogero-Sutherland Models

    Directory of Open Access Journals (Sweden)

    Martin Hallnäs

    2007-03-01

    Full Text Available We review a recent construction of an explicit analytic series representation for symmetric polynomials which up to a groundstate factor are eigenfunctions of Calogero-Sutherland type models. We also indicate a generalisation of this result to polynomials which give the eigenfunctions of so-called 'deformed' Calogero-Sutherland type models.

  2. An interactive modelling tool for understanding hydrological processes in lowland catchments

    Science.gov (United States)

    Brauer, Claudia; Torfs, Paul; Uijlenhoet, Remko

    2016-04-01

    Recently, we developed the Wageningen Lowland Runoff Simulator (WALRUS), a rainfall-runoff model for catchments with shallow groundwater (Brauer et al., 2014ab). WALRUS explicitly simulates processes which are important in lowland catchments, such as feedbacks between saturated and unsaturated zone and between groundwater and surface water. WALRUS has a simple model structure and few parameters with physical connotations. Some default functions (which can be changed easily for research purposes) are implemented to facilitate application by practitioners and students. The effect of water management on hydrological variables can be simulated explicitly. The model description and applications are published in open access journals (Brauer et al, 2014). The open source code (provided as R package) and manual can be downloaded freely (www.github.com/ClaudiaBrauer/WALRUS). We organised a short course for Dutch water managers and consultants to become acquainted with WALRUS. We are now adapting this course as a stand-alone tutorial suitable for a varied, international audience. In addition, simple models can aid teachers to explain hydrological principles effectively. We used WALRUS to generate examples for simple interactive tools, which we will present at the EGU General Assembly. C.C. Brauer, A.J. Teuling, P.J.J.F. Torfs, R. Uijlenhoet (2014a): The Wageningen Lowland Runoff Simulator (WALRUS): a lumped rainfall-runoff model for catchments with shallow groundwater, Geosci. Model Dev., 7, 2313-2332. C.C. Brauer, P.J.J.F. Torfs, A.J. Teuling, R. Uijlenhoet (2014b): The Wageningen Lowland Runoff Simulator (WALRUS): application to the Hupsel Brook catchment and Cabauw polder, Hydrol. Earth Syst. Sci., 18, 4007-4028.

  3. CDPOP: A spatially explicit cost distance population genetics program

    Science.gov (United States)

    Erin L. Landguth; S. A. Cushman

    2010-01-01

    Spatially explicit simulation of gene flow in complex landscapes is essential to explain observed population responses and provide a foundation for landscape genetics. To address this need, we wrote a spatially explicit, individual-based population genetics model (CDPOP). The model implements individual-based population modelling with Mendelian inheritance and k-allele...

  4. Multiscale modeling of a rectifying bipolar nanopore: explicit-water versus implicit-water simulations.

    Science.gov (United States)

    Ható, Zoltán; Valiskó, Mónika; Kristóf, Tamás; Gillespie, Dirk; Boda, Dezsö

    2017-07-21

    In a multiscale modeling approach, we present computer simulation results for a rectifying bipolar nanopore at two modeling levels. In an all-atom model, we use explicit water to simulate ion transport directly with the molecular dynamics technique. In a reduced model, we use implicit water and apply the Local Equilibrium Monte Carlo method together with the Nernst-Planck transport equation. This hybrid method makes the fast calculation of ion transport possible at the price of lost details. We show that the implicit-water model is an appropriate representation of the explicit-water model when we look at the system at the device (i.e., input vs. output) level. The two models produce qualitatively similar behavior of the electrical current for different voltages and model parameters. Looking at the details of concentration and potential profiles, we find profound differences between the two models. These differences, however, do not influence the basic behavior of the model as a device because they do not influence the z-dependence of the concentration profiles which are the main determinants of current. These results then address an old paradox: how do reduced models, whose assumptions should break down in a nanoscale device, predict experimental data? Our simulations show that reduced models can still capture the overall device physics correctly, even though they get some important aspects of the molecular-scale physics quite wrong; reduced models work because they include the physics that is necessary from the point of view of device function. Therefore, reduced models can suffice for general device understanding and device design, but more detailed models might be needed for molecular level understanding.

  5. Pedagogical Model for Explicit Teaching of Reading Comprehension to English Language Learners

    Directory of Open Access Journals (Sweden)

    Al Tiyb Al Khaiyali

    2017-09-01

    Full Text Available Reading comprehension instruction is considered one of the major challenges that most English language teachers and students encounter. Therefore, providing a systematic, explicit, and flexible model to teaching reading comprehension strategies could help resolve some of these challenges and increase the possibility of teaching reading comprehension, particularly in language learners’ classrooms. Consequently, the purpose of this paper is to provide a model to teach reading comprehension strategies in language learning classrooms. The proposed instructional model is divided into three systematic phases through which strategies are taught before reading, during reading, and after reading. Each phase is explained and elaborated using recommended models for teachers. Finally, suggested considerations to consolidate this model are provided.

  6. Explicit chiral symmetry breaking in Gross-Neveu type models

    Energy Technology Data Exchange (ETDEWEB)

    Boehmer, Christian

    2011-07-25

    This thesis is devoted to the study of a 1+1-dimensional, fermionic quantum field theory with Lagrangian L= anti {psi}i{gamma}{sup {mu}}{partial_derivative}{sub {mu}}{psi}-m{sub 0} anti {psi}{psi}+(g{sup 2})/(2)(anti {psi}{psi}){sup 2}+(G{sup 2})/(2)(anti {psi}i{gamma}{sub 5}{psi}){sup 2} in the limit of an infinite number of flavors, using semiclassical methods. The main goal of the present work was to see what changes if we allow for explicit chiral symmetry breaking, either by a bare mass term, or a splitting of the scalar and pseudo-scalar coupling constants, or both. In the first case, this becomes the massive NJL{sub 2} model. In the 2nd and 3rd cases we are dealing with a model largely unexplored so far. The first half of this thesis deals with the massive NJL{sub 2} model. Before attacking the phase diagram, it was necessary to determine the baryons of the model. We have carried out full numerical Hartree-Fock calculations including the Dirac sea. The most important result is the first complete phase diagram of the massive NJL{sub 2} model in ({mu},T,{gamma}) space, where {gamma} arises from m{sub 0} through mass renormalization. In the 2nd half of the thesis we have studied a generalization of the massless NJL{sub 2} model with two different (scalar and pseudoscalar) coupling constants, first in the massless version. Renormalization of the 2 coupling constants leads to the usual dynamical mass by dynamical transmutation, but in addition to a novel {xi} parameter interpreted as chiral quenching parameter. As far as baryon structure is concerned, the most interesting result is the fact that the new baryons interpolate between the kink of the GN model and the massless baryon of the NJL{sub 2} model, always carrying fractional baryon number 1/2. The phase diagram of the massless model with 2 coupling constants has again been determined numerically. At zero temperature we have also investigated the massive, generalized GN model with 3 parameters. It is well

  7. Explicit chiral symmetry breaking in Gross-Neveu type models

    International Nuclear Information System (INIS)

    Boehmer, Christian

    2011-01-01

    This thesis is devoted to the study of a 1+1-dimensional, fermionic quantum field theory with Lagrangian L= anti ψiγ μ ∂ μ ψ-m 0 anti ψψ+(g 2 )/(2)(anti ψψ) 2 +(G 2 )/(2)(anti ψiγ 5 ψ) 2 in the limit of an infinite number of flavors, using semiclassical methods. The main goal of the present work was to see what changes if we allow for explicit chiral symmetry breaking, either by a bare mass term, or a splitting of the scalar and pseudo-scalar coupling constants, or both. In the first case, this becomes the massive NJL 2 model. In the 2nd and 3rd cases we are dealing with a model largely unexplored so far. The first half of this thesis deals with the massive NJL 2 model. Before attacking the phase diagram, it was necessary to determine the baryons of the model. We have carried out full numerical Hartree-Fock calculations including the Dirac sea. The most important result is the first complete phase diagram of the massive NJL 2 model in (μ,T,γ) space, where γ arises from m 0 through mass renormalization. In the 2nd half of the thesis we have studied a generalization of the massless NJL 2 model with two different (scalar and pseudoscalar) coupling constants, first in the massless version. Renormalization of the 2 coupling constants leads to the usual dynamical mass by dynamical transmutation, but in addition to a novel ξ parameter interpreted as chiral quenching parameter. As far as baryon structure is concerned, the most interesting result is the fact that the new baryons interpolate between the kink of the GN model and the massless baryon of the NJL 2 model, always carrying fractional baryon number 1/2. The phase diagram of the massless model with 2 coupling constants has again been determined numerically. At zero temperature we have also investigated the massive, generalized GN model with 3 parameters. It is well-known that the massless NJL 2 model can be solved analytically. The same is true for the GN model, be it massless or massive. Here, the

  8. A Bidirectional Subsurface Remote Sensing Reflectance Model Explicitly Accounting for Particle Backscattering Shapes

    Science.gov (United States)

    He, Shuangyan; Zhang, Xiaodong; Xiong, Yuanheng; Gray, Deric

    2017-11-01

    The subsurface remote sensing reflectance (rrs, sr-1), particularly its bidirectional reflectance distribution function (BRDF), depends fundamentally on the angular shape of the volume scattering functions (VSFs, m-1 sr-1). Recent technological advancement has greatly expanded the collection, and the knowledge of natural variability, of the VSFs of oceanic particles. This allows us to test the Zaneveld's theoretical rrs model that explicitly accounts for particle VSF shapes. We parameterized the rrs model based on HydroLight simulations using 114 VSFs measured in three coastal waters around the United States and in oceanic waters of North Atlantic Ocean. With the absorption coefficient (a), backscattering coefficient (bb), and VSF shape as inputs, the parameterized model is able to predict rrs with a root mean square relative error of ˜4% for solar zenith angles from 0 to 75°, viewing zenith angles from 0 to 60°, and viewing azimuth angles from 0 to 180°. A test with the field data indicates the performance of our model, when using only a and bb as inputs and selecting the VSF shape using bb, is comparable to or slightly better than the currently used models by Morel et al. and Lee et al. Explicitly expressing VSF shapes in rrs modeling has great potential to further constrain the uncertainty in the ocean color studies as our knowledge on the VSFs of natural particles continues to improve. Our study represents a first effort in this direction.

  9. Spatially explicit modeling in ecology: A review

    Science.gov (United States)

    DeAngelis, Donald L.; Yurek, Simeon

    2017-01-01

    The use of spatially explicit models (SEMs) in ecology has grown enormously in the past two decades. One major advancement has been that fine-scale details of landscapes, and of spatially dependent biological processes, such as dispersal and invasion, can now be simulated with great precision, due to improvements in computer technology. Many areas of modeling have shifted toward a focus on capturing these fine-scale details, to improve mechanistic understanding of ecosystems. However, spatially implicit models (SIMs) have played a dominant role in ecology, and arguments have been made that SIMs, which account for the effects of space without specifying spatial positions, have an advantage of being simpler and more broadly applicable, perhaps contributing more to understanding. We address this debate by comparing SEMs and SIMs in examples from the past few decades of modeling research. We argue that, although SIMs have been the dominant approach in the incorporation of space in theoretical ecology, SEMs have unique advantages for addressing pragmatic questions concerning species populations or communities in specific places, because local conditions, such as spatial heterogeneities, organism behaviors, and other contingencies, produce dynamics and patterns that usually cannot be incorporated into simpler SIMs. SEMs are also able to describe mechanisms at the local scale that can create amplifying positive feedbacks at that scale, creating emergent patterns at larger scales, and therefore are important to basic ecological theory. We review the use of SEMs at the level of populations, interacting populations, food webs, and ecosystems and argue that SEMs are not only essential in pragmatic issues, but must play a role in the understanding of causal relationships on landscapes.

  10. An explicit solution of the mathematical model for osmotic desalination process

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Do Yeon; Gu, Boram; Yang, Dae Ryook [Korea University, Seoul (Korea, Republic of)

    2013-09-15

    Membrane processes such as reverse osmosis and forward osmosis for seawater desalination have gained attention in recent years. Mathematical models have been used to interpret the mechanism of membrane processes. The membrane process model, consisting of flux and concentration polarization (CP) models, is coupled with balance equations and solved simultaneously. This set of model equations is, however, implicit and nonlinear; consequently, the model must be solved iteratively and numerically, which is time- and cost-intensive. We suggest a method to transform implicit equations to their explicit form, in order to avoid an iterative procedure. In addition, the performance of five solving methods, including the method that we suggest, is tested and compared for accuracy, computation time, and robustness based on input conditions. Our proposed method shows the best performance based on the robustness of various simulation conditions, accuracy, and a cost-effective computation time.

  11. [Application of spatially explicit landscape model in soil loss study in Huzhong area].

    Science.gov (United States)

    Xu, Chonggang; Hu, Yuanman; Chang, Yu; Li, Xiuzhen; Bu, Renchang; He, Hongshi; Leng, Wenfang

    2004-10-01

    Universal Soil Loss Equation (USLE) has been widely used to estimate the average annual soil loss. In most of the previous work on soil loss evaluation on forestland, cover management factor was calculated from the static forest landscape. The advent of spatially explicit forest landscape model in the last decade, which explicitly simulates the forest succession dynamics under natural and anthropogenic disturbances (fire, wind, harvest and so on) on heterogeneous landscape, makes it possible to take into consideration the change of forest cover, and to dynamically simulate the soil loss in different year (e.g. 10 years and 20 years after current year). In this study, we linked a spatially explicit landscape model (LANDIS) with USLE to simulate the soil loss dynamics under two scenarios: fire and no harvest, fire and harvest. We also simulated the soil loss with no fire and no harvest as a control. The results showed that soil loss varied periodically with simulation year, and the amplitude of change was the lowest under the control scenario and the highest under the fire and no harvest scenario. The effect of harvest on soil loss could not be easily identified on the map; however, the cumulative effect of harvest on soil loss was larger than that of fire. Decreasing the harvest area and the percent of bare soil increased by harvest could significantly reduce soil loss, but had no significant effects on the dynamic of soil loss. Although harvest increased the annual soil loss, it tended to decrease the variability of soil loss between different simulation years.

  12. SPATIALLY EXPLICIT MICRO-LEVEL MODELLING OF LAND USE CHANGE AT THE RURAL-URBAN INTERFACE. (R828012)

    Science.gov (United States)

    This paper describes micro-economic models of land use change applicable to the rural–urban interface in the US. Use of a spatially explicit micro-level modelling approach permits the analysis of regional patterns of land use as the aggregate outcomes of many, disparate...

  13. Age-related variations of visuo-motor adaptation beyond explicit knowledge

    Directory of Open Access Journals (Sweden)

    Herbert eHeuer

    2014-07-01

    Full Text Available Visuo-motor adaptation suffers at older working age. The age-related decline of behavioural adjustments is accompanied by reduced explicit knowledge of the visuo-motor transformation. It disappears when explicit knowledge is kept constant across the age range, except for particularly high levels of explicit knowledge. According to these findings, at older adult age both the acquisition of explicit knowledge and its application for strategic corrections become poorer. Recently it has been posited that visuo-motor adaptation can involve model-free reinforcement mechanisms of learning in addition to model-based mechanisms. We tested whether age-related declines of reinforcement learning can also contribute to the age-related changes of visuo-motor adaptation. Therefore we enhanced the contribution of reinforcement learning to visuo-motor adaptation by way of introducing salient markers of success and failure during practice. With such modified practice conditions, there were residual age-related variations of behavioural adjustments at all levels of explicit knowledge, even when explicit knowledge was absent. The residual age-related variations were observed for practiced target directions only, but not for new target directions. These findings are consistent with an age-related decline of model-free reinforcement learning as a third factor in the age-related decline of visuo-motor adaptation. Under practice conditions, which spur model-free reward-based learning, this factor adds to the decrements of the acquisition of explicit knowledge and its use for strategic corrections.

  14. Explicit calculation of indirect global warming potentials for halons using atmospheric models

    Directory of Open Access Journals (Sweden)

    D. J. Wuebbles

    2009-11-01

    Full Text Available The concept of Global Warming Potentials (GWPs has been extensively used in policy consideration as a relative index for comparing the climate impact of an emitted greenhouse gas (GHG, relative to carbon dioxide with equal mass emissions. Ozone depletion due to emission of chlorinated or brominated halocarbons leads to cooling of the climate system in the opposite direction to the direct warming contribution by halocarbons as GHGs. This cooling is a key indirect effect of the halocarbons on climatic radiative forcing, which is accounted for by indirect GWPs. With respect to climate, it is critical to understand net influences considering direct warming and indirect cooling effects especially for Halons due to the greater ozone-depleting efficiency of bromine over chlorine. Until now, the indirect GWPs have been calculated using a parameterized approach based on the concept of Equivalent Effective Stratospheric Chlorine (EESC and the observed ozone depletion over the last few decades. As a step towards obtaining indirect GWPs through a more robust approach, we use atmospheric models to explicitly calculate the indirect GWPs of Halon-1211 and Halon-1301 for a 100-year time horizon. State-of-the-art global chemistry-transport models (CTMs were used as the computational tools to derive more realistic ozone depletion changes caused by an added pulse emission of the two major Halons at the surface. The radiative forcings on climate from the ozone changes have been calculated for indirect GWPs using an atmospheric radiative transfer model (RTM. The simulated temporal variations of global average total column Halons after a pulse perturbation follow an exponential decay with an e-folding time which is consistent with the expected chemical lifetimes of the Halons. Our calculated indirect GWPs for the two Halons are much smaller than those from past studies but are within a single standard deviation of WMO (2007 values and the direct GWP values derived

  15. Dissociation between implicit and explicit expectancies of cannabis use in adolescence.

    Science.gov (United States)

    Schmits, Emilie; Maurage, Pierre; Thirion, Romain; Quertemont, Etienne

    2015-12-30

    Cannabis is one of the most commonly drugs used by teenagers. Expectancies about its effects play a crucial role in cannabis consumption. Various tools have been used to assess expectancies, mainly self-report questionnaires measuring explicit expectancies, but implicit measures based on experimental tasks have also been developed, measuring implicit expectancies. The aim of this study was to simultaneously assess implicit/explicit expectancies related to cannabis among adolescent users and non-users. 130 teenagers attending school (55 girls) were enrolled (Age: M=16.40 years); 43.84% had never used cannabis ("non-users") and 56.16% had used cannabis ("users"). They completed self-report questionnaires evaluating cannabis use, cannabis-related problems, effect expectancies (explicit expectancies), alcohol use, social and trait anxiety, depression, as well as three Implicit Association Tests (IAT) assessing implicit expectancies. Adolescents manifested more implicit affective associations (relaxation, excitation, negative) than neutral ones regarding cannabis. These were not related to explicit expectancies. Cannabis users reported more implicit relaxation expectancies and less negative explicit expectancies than non-users. The frequency of use and related problems were positively associated with the explicit expectancies regarding relaxation and enhancement, and were negatively associated with negative explicit expectancies and negative implicit expectancies. Findings indicate that implicit and explicit expectancies play different roles in cannabis use by adolescents. The implications for experimentation and prevention are discussed. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  16. An online model composition tool for system biology models.

    Science.gov (United States)

    Coskun, Sarp A; Cicek, A Ercument; Lai, Nicola; Dash, Ranjan K; Ozsoyoglu, Z Meral; Ozsoyoglu, Gultekin

    2013-09-05

    There are multiple representation formats for Systems Biology computational models, and the Systems Biology Markup Language (SBML) is one of the most widely used. SBML is used to capture, store, and distribute computational models by Systems Biology data sources (e.g., the BioModels Database) and researchers. Therefore, there is a need for all-in-one web-based solutions that support advance SBML functionalities such as uploading, editing, composing, visualizing, simulating, querying, and browsing computational models. We present the design and implementation of the Model Composition Tool (Interface) within the PathCase-SB (PathCase Systems Biology) web portal. The tool helps users compose systems biology models to facilitate the complex process of merging systems biology models. We also present three tools that support the model composition tool, namely, (1) Model Simulation Interface that generates a visual plot of the simulation according to user's input, (2) iModel Tool as a platform for users to upload their own models to compose, and (3) SimCom Tool that provides a side by side comparison of models being composed in the same pathway. Finally, we provide a web site that hosts BioModels Database models and a separate web site that hosts SBML Test Suite models. Model composition tool (and the other three tools) can be used with little or no knowledge of the SBML document structure. For this reason, students or anyone who wants to learn about systems biology will benefit from the described functionalities. SBML Test Suite models will be a nice starting point for beginners. And, for more advanced purposes, users will able to access and employ models of the BioModels Database as well.

  17. An explicit formula for the interface tension of the 2D Potts model

    Science.gov (United States)

    Borgs, Christian; Janke, Wolfhard

    1992-11-01

    We consider the exact correlation length calculations for the two-dimensional Potts model at the transition point β_t by Klümper, Schadschneider and Zittartz, and by Buffenoir and Wallon. We argue that the correlation length calculated by the latter authors is the correlation length in the disordered phase and then combine their result with duality and the assumption of complete wetting to give an explicit formula for the order-disorder interface tension σ_od of this model. The result is used to clarify a controversy stemming from different numerical simulations of σ_od.

  18. Implicit and explicit ethnocentrism: revisiting the ideologies of prejudice.

    Science.gov (United States)

    Cunningham, William A; Nezlek, John B; Banaji, Mahzarin R

    2004-10-01

    Two studies investigated relationships among individual differences in implicit and explicit prejudice, right-wing ideology, and rigidity in thinking. The first study examined these relationships focusing on White Americans' prejudice toward Black Americans. The second study provided the first test of implicit ethnocentrism and its relationship to explicit ethnocentrism by studying the relationship between attitudes toward five social groups. Factor analyses found support for both implicit and explicit ethnocentrism. In both studies, mean explicit attitudes toward out groups were positive, whereas implicit attitudes were negative, suggesting that implicit and explicit prejudices are distinct; however, in both studies, implicit and explicit attitudes were related (r = .37, .47). Latent variable modeling indicates a simple structure within this ethnocentric system, with variables organized in order of specificity. These results lead to the conclusion that (a) implicit ethnocentrism exists and (b) it is related to and distinct from explicit ethnocentrism.

  19. TACIT, EXPLICIT, AND CULTURAL KNOWLEDGE IN DEMAND PLANNING

    Directory of Open Access Journals (Sweden)

    ANDRÉ EDUARDO MIRANDA DOS SANTOS

    2010-04-01

    Full Text Available Demand planning consists in a set of steps which evolves two main types of knowledge: tacit and explicit. It is under these types of knowledge that the present article was developed. It aimed at relating the concept of demand planning with knowledge management, but delimited to its tacit, explicit, and cultural knowledge components. An exploratory research was done with five companies. Therefore, the study was characterized as multicases and evolded enterviews, observations, and documents analysis. The analysis was made by content analysis. The results evidenced the presence of a set of components related to the three types of knowledge in demand planning like: information technology tools, professional experience, collective interation, social relations, and a corporative view, which leads the company´ business.

  20. Integrating a Decision Management Tool with UML Modeling Tools

    DEFF Research Database (Denmark)

    Könemann, Patrick

    by proposing potential subsequent design issues. In model-based software development, many decisions directly affect the structural and behavioral models used to describe and develop a software system and its architecture. However, these decisions are typically not connected to the models created during...... integration of formerly disconnected tools improves tool usability as well as decision maker productivity....

  1. Age effects on explicit and implicit memory

    Directory of Open Access Journals (Sweden)

    Emma eWard

    2013-09-01

    Full Text Available It is well documented that explicit memory (e.g., recognition declines with age. In contrast, many argue that implicit memory (e.g., priming is preserved in healthy aging. For example, priming on tasks such as perceptual identification is often not statistically different in groups of young and older adults. Such observations are commonly taken as evidence for distinct explicit and implicit learning/memory systems. In this article we discuss several lines of evidence that challenge this view. We describe how patterns of differential age-related decline may arise from differences in the ways in which the two forms of memory are commonly measured, and review recent research suggesting that under improved measurement methods, implicit memory is not age-invariant. Formal computational models are of considerable utility in revealing the nature of underlying systems. We report the results of applying single and multiple-systems models to data on age effects in implicit and explicit memory. Model comparison clearly favours the single-system view. Implications for the memory systems debate are discussed.

  2. Age effects on explicit and implicit memory.

    Science.gov (United States)

    Ward, Emma V; Berry, Christopher J; Shanks, David R

    2013-01-01

    It is well-documented that explicit memory (e.g., recognition) declines with age. In contrast, many argue that implicit memory (e.g., priming) is preserved in healthy aging. For example, priming on tasks such as perceptual identification is often not statistically different in groups of young and older adults. Such observations are commonly taken as evidence for distinct explicit and implicit learning/memory systems. In this article we discuss several lines of evidence that challenge this view. We describe how patterns of differential age-related decline may arise from differences in the ways in which the two forms of memory are commonly measured, and review recent research suggesting that under improved measurement methods, implicit memory is not age-invariant. Formal computational models are of considerable utility in revealing the nature of underlying systems. We report the results of applying single and multiple-systems models to data on age effects in implicit and explicit memory. Model comparison clearly favors the single-system view. Implications for the memory systems debate are discussed.

  3. Probing the role of interfacial waters in protein-DNA recognition using a hybrid implicit/explicit solvation model

    Science.gov (United States)

    Li, Shen; Bradley, Philip

    2013-01-01

    When proteins bind to their DNA target sites, ordered water molecules are often present at the protein-DNA interface bridging protein and DNA through hydrogen bonds. What is the role of these ordered interfacial waters? Are they important determinants of the specificity of DNA sequence recognition, or do they act in binding in a primarily non-specific manner, by improving packing of the interface, shielding unfavorable electrostatic interactions, and solvating unsatisfied polar groups that are inaccessible to bulk solvent? When modeling details of structure and binding preferences, can fully implicit solvent models be fruitfully applied to protein-DNA interfaces, or must the individualistic properties of these interfacial waters be accounted for? To address these questions, we have developed a hybrid implicit/explicit solvation model that specifically accounts for the locations and orientations of small numbers of DNA-bound water molecules while treating the majority of the solvent implicitly. Comparing the performance of this model to its fully implicit counterpart, we find that explicit treatment of interfacial waters results in a modest but significant improvement in protein sidechain placement and DNA sequence recovery. Base-by-base comparison of the performance of the two models highlights DNA sequence positions whose recognition may be dependent on interfacial water. Our study offers large-scale statistical evidence for the role of ordered water for protein DNA recognition, together with detailed examination of several well-characterized systems. In addition, our approach provides a template for modeling explicit water molecules at interfaces that should be extensible to other systems. PMID:23444044

  4. Explicit state representation and the ATLAS event data model: theory and practice

    International Nuclear Information System (INIS)

    Nowak, M; Snyder, S; Cranmer, K; Malon, D; Gemmeren, P v; Schaffer, A; Binet, S

    2008-01-01

    In anticipation of data taking, ATLAS has undertaken a program of work to develop an explicit state representation of the experiment's complex transient event data model. This effort has provided both an opportunity to consider explicitly the structure, organization, and content of the ATLAS persistent event store before writing tens of petabytes of data (replacing simple streaming, which uses the persistent store as a core dump of transient memory), and a locus for support of event data model evolution, including significant refactoring, beyond the automatic schema evolution capabilities of underlying persistence technologies. ATLAS has encountered the need for such non-trivial schema evolution on several occasions already. This paper describes the state representation strategy (transient/persistent separation) and its implementation, including both the payoffs that ATLAS has seen (significant and sometimes surprising space and performance improvements, the extra layer notwithstanding, and extremely general schema evolution support) and the costs (additional and relatively pervasive additional infrastructure development and maintenance). The paper further discusses how these costs are mitigated, and how ATLAS is able to implement this strategy without losing the ability to take advantage of the (improving!) automatic schema evolution capabilities of underlying technology layers when appropriate. Implications of state representations for direct ROOT browsability, and current strategies for associating physics analysis views with such state representations, are also described

  5. Short-Range Prediction of Monsoon Precipitation by NCMRWF Regional Unified Model with Explicit Convection

    Science.gov (United States)

    Mamgain, Ashu; Rajagopal, E. N.; Mitra, A. K.; Webster, S.

    2018-03-01

    There are increasing efforts towards the prediction of high-impact weather systems and understanding of related dynamical and physical processes. High-resolution numerical model simulations can be used directly to model the impact at fine-scale details. Improvement in forecast accuracy can help in disaster management planning and execution. National Centre for Medium Range Weather Forecasting (NCMRWF) has implemented high-resolution regional unified modeling system with explicit convection embedded within coarser resolution global model with parameterized convection. The models configurations are based on UK Met Office unified seamless modeling system. Recent land use/land cover data (2012-2013) obtained from Indian Space Research Organisation (ISRO) are also used in model simulations. Results based on short-range forecast of both the global and regional models over India for a month indicate that convection-permitting simulations by the high-resolution regional model is able to reduce the dry bias over southern parts of West Coast and monsoon trough zone with more intense rainfall mainly towards northern parts of monsoon trough zone. Regional model with explicit convection has significantly improved the phase of the diurnal cycle of rainfall as compared to the global model. Results from two monsoon depression cases during study period show substantial improvement in details of rainfall pattern. Many categories in rainfall defined for operational forecast purposes by Indian forecasters are also well represented in case of convection-permitting high-resolution simulations. For the statistics of number of days within a range of rain categories between `No-Rain' and `Heavy Rain', the regional model is outperforming the global model in all the ranges. In the very heavy and extremely heavy categories, the regional simulations show overestimation of rainfall days. Global model with parameterized convection have tendency to overestimate the light rainfall days and

  6. EPS Mid-Career Award 2011. Are there multiple memory systems? Tests of models of implicit and explicit memory.

    Science.gov (United States)

    Shanks, David R; Berry, Christopher J

    2012-01-01

    This article reviews recent work aimed at developing a new framework, based on signal detection theory, for understanding the relationship between explicit (e.g., recognition) and implicit (e.g., priming) memory. Within this framework, different assumptions about sources of memorial evidence can be framed. Application to experimental results provides robust evidence for a single-system model in preference to multiple-systems models. This evidence comes from several sources including studies of the effects of amnesia and ageing on explicit and implicit memory. The framework allows a range of concepts in current memory research, such as familiarity, recollection, fluency, and source memory, to be linked to implicit memory. More generally, this work emphasizes the value of modern computational modelling techniques in the study of learning and memory.

  7. Assessment of the Simulated Molecular Composition with the GECKO-A Modeling Tool Using Chamber Observations for α-Pinene.

    Science.gov (United States)

    Aumont, B.; Camredon, M.; Isaacman-VanWertz, G. A.; Karam, C.; Valorso, R.; Madronich, S.; Kroll, J. H.

    2016-12-01

    Gas phase oxidation of VOC is a gradual process leading to the formation of multifunctional organic compounds, i.e., typically species with higher oxidation state, high water solubility and low volatility. These species contribute to the formation of secondary organic aerosols (SOA) viamultiphase processes involving a myriad of organic species that evolve through thousands of reactions and gas/particle mass exchanges. Explicit chemical mechanisms reflect the understanding of these multigenerational oxidation steps. These mechanisms rely directly on elementary reactions to describe the chemical evolution and track the identity of organic carbon through various phases down to ultimate oxidation products. The development, assessment and improvement of such explicit schemes is a key issue, as major uncertainties remain on the chemical pathways involved during atmospheric oxidation of organic matter. An array of mass spectrometric techniques (CIMS, PTRMS, AMS) was recently used to track the composition of organic species during α-pinene oxidation in the MIT environmental chamber, providing an experimental database to evaluate and improve explicit mechanisms. In this study, the GECKO-A tool (Generator for Explicit Chemistry and Kinetics of Organics in the Atmosphere) is used to generate fully explicit oxidation schemes for α-pinene multiphase oxidation simulating the MIT experiment. The ability of the GECKO-A chemical scheme to explain the organic molecular composition in the gas and the condensed phases is explored. First results of this model/observation comparison at the molecular level will be presented.

  8. Simulation of a severe convective storm using a numerical model with explicitly incorporated aerosols

    Science.gov (United States)

    Lompar, Miloš; Ćurić, Mladjen; Romanic, Djordje

    2017-09-01

    Despite an important role the aerosols play in all stages of cloud lifecycle, their representation in numerical weather prediction models is often rather crude. This paper investigates the effects the explicit versus implicit inclusion of aerosols in a microphysics parameterization scheme in Weather Research and Forecasting (WRF) - Advanced Research WRF (WRF-ARW) model has on cloud dynamics and microphysics. The testbed selected for this study is a severe mesoscale convective system with supercells that struck west and central parts of Serbia in the afternoon of July 21, 2014. Numerical products of two model runs, i.e. one with aerosols explicitly (WRF-AE) included and another with aerosols implicitly (WRF-AI) assumed, are compared against precipitation measurements from surface network of rain gauges, as well as against radar and satellite observations. The WRF-AE model accurately captured the transportation of dust from the north Africa over the Mediterranean and to the Balkan region. On smaller scales, both models displaced the locations of clouds situated above west and central Serbia towards southeast and under-predicted the maximum values of composite radar reflectivity. Similar to satellite images, WRF-AE shows the mesoscale convective system as a merged cluster of cumulonimbus clouds. Both models over-predicted the precipitation amounts; WRF-AE over-predictions are particularly pronounced in the zones of light rain, while WRF-AI gave larger outliers. Unlike WRF-AI, the WRF-AE approach enables the modelling of time evolution and influx of aerosols into the cloud which could be of practical importance in weather forecasting and weather modification. Several likely causes for discrepancies between models and observations are discussed and prospects for further research in this field are outlined.

  9. Explicit and implicit springback simulation in sheet metal forming using fully coupled ductile damage and distortional hardening model

    Science.gov (United States)

    Yetna n'jock, M.; Houssem, B.; Labergere, C.; Saanouni, K.; Zhenming, Y.

    2018-05-01

    The springback is an important phenomenon which accompanies the forming of metallic sheets especially for high strength materials. A quantitative prediction of springback becomes very important for newly developed material with high mechanical characteristics. In this work, a numerical methodology is developed to quantify this undesirable phenomenon. This methodoly is based on the use of both explicit and implicit finite element solvers of Abaqus®. The most important ingredient of this methodology consists on the use of highly predictive mechanical model. A thermodynamically-consistent, non-associative and fully anisotropic elastoplastic constitutive model strongly coupled with isotropic ductile damage and accounting for distortional hardening is then used. An algorithm for local integration of the complete set of the constitutive equations is developed. This algorithm considers the rotated frame formulation (RFF) to ensure the incremental objectivity of the model in the framework of finite strains. This algorithm is implemented in both explicit (Abaqus/Explicit®) and implicit (Abaqus/Standard®) solvers of Abaqus® through the users routine VUMAT and UMAT respectively. The implicit solver of Abaqus® has been used to study spingback as it is generally a quasi-static unloading. In order to compare the methods `efficiency, the explicit method (Dynamic Relaxation Method) proposed by Rayleigh has been also used for springback prediction. The results obtained within U draw/bending benchmark are studied, discussed and compared with experimental results as reference. Finally, the purpose of this work is to evaluate the reliability of different methods predict efficiently springback in sheet metal forming.

  10. A Risk Assessment Example for Soil Invertebrates Using Spatially Explicit Agent-Based Models

    DEFF Research Database (Denmark)

    Reed, Melissa; Alvarez, Tania; Chelinho, Sonia

    2016-01-01

    Current risk assessment methods for measuring the toxicity of plant protection products (PPPs) on soil invertebrates use standardized laboratory conditions to determine acute effects on mortality and sublethal effects on reproduction. If an unacceptable risk is identified at the lower tier...... population models for ubiquitous soil invertebrates (collembolans and earthworms) as refinement options in current risk assessment. Both are spatially explicit agent-based models (ABMs), incorporating individual and landscape variability. The models were used to provide refined risk assessments for different...... application scenarios of a hypothetical pesticide applied to potato crops (full-field spray onto the soil surface [termed “overall”], in-furrow, and soil-incorporated pesticide applications). In the refined risk assessment, the population models suggest that soil invertebrate populations would likely recover...

  11. From explicit to implicit normal mode initialization of a limited-area model

    Energy Technology Data Exchange (ETDEWEB)

    Bijlsma, S.J.

    2013-02-15

    In this note the implicit normal mode initialization of a limited-area model is discussed from a different point of view. To that end it is shown that the equations describing the explicit normal mode initialization applied to the shallow water equations in differentiated form on the sphere can readily be derived in normal mode space if the model equations are separable, but only in the case of stationary Rossby modes can be transformed into the implicit equations in physical space. This is a consequence of the simple relations between the components of the different modes in that case. In addition a simple eigenvalue problem is given for the frequencies of the gravity waves. (orig.)

  12. Recent Advances in Explicit Multiparametric Nonlinear Model Predictive Control

    KAUST Repository

    Domínguez, Luis F.; Pistikopoulos, Efstratios N.

    2011-01-01

    are derived. The performance of the explicit controllers are then tested and compared in a simulation example involving the operation of a continuous stirred-tank reactor (CSTR). © 2010 American Chemical Society.

  13. Graph and model transformation tools for model migration : empirical results from the transformation tool contest

    NARCIS (Netherlands)

    Rose, L.M.; Herrmannsdoerfer, M.; Mazanek, S.; Van Gorp, P.M.E.; Buchwald, S.; Horn, T.; Kalnina, E.; Koch, A.; Lano, K.; Schätz, B.; Wimmer, M.

    2014-01-01

    We describe the results of the Transformation Tool Contest 2010 workshop, in which nine graph and model transformation tools were compared for specifying model migration. The model migration problem—migration of UML activity diagrams from version 1.4 to version 2.2—is non-trivial and practically

  14. Moderators of the Relationship between Implicit and Explicit Evaluation

    Science.gov (United States)

    Nosek, Brian A.

    2005-01-01

    Automatic and controlled modes of evaluation sometimes provide conflicting reports of the quality of social objects. This paper presents evidence for four moderators of the relationship between automatic (implicit) and controlled (explicit) evaluations. Implicit and explicit preferences were measured for a variety of object pairs using a large sample. The average correlation was r = .36, and 52 of the 57 object pairs showed a significant positive correlation. Results of multilevel modeling analyses suggested that: (a) implicit and explicit preferences are related, (b) the relationship varies as a function of the objects assessed, and (c) at least four variables moderate the relationship – self-presentation, evaluative strength, dimensionality, and distinctiveness. The variables moderated implicit-explicit correspondence across individuals and accounted for much of the observed variation across content domains. The resulting model of the relationship between automatic and controlled evaluative processes is grounded in personal experience with the targets of evaluation. PMID:16316292

  15. Rapid Response Tools and Datasets for Post-fire Hydrological Modeling

    Science.gov (United States)

    Miller, Mary Ellen; MacDonald, Lee H.; Billmire, Michael; Elliot, William J.; Robichaud, Pete R.

    2016-04-01

    Rapid response is critical following natural disasters. Flooding, erosion, and debris flows are a major threat to life, property and municipal water supplies after moderate and high severity wildfires. The problem is that mitigation measures must be rapidly implemented if they are to be effective, but they are expensive and cannot be applied everywhere. Fires, runoff, and erosion risks also are highly heterogeneous in space, so there is an urgent need for a rapid, spatially-explicit assessment. Past post-fire modeling efforts have usually relied on lumped, conceptual models because of the lack of readily available, spatially-explicit data layers on the key controls of topography, vegetation type, climate, and soil characteristics. The purpose of this project is to develop a set of spatially-explicit data layers for use in process-based models such as WEPP, and to make these data layers freely available. The resulting interactive online modeling database (http://geodjango.mtri.org/geowepp/) is now operational and publically available for 17 western states in the USA. After a fire, users only need to upload a soil burn severity map, and this is combined with the pre-existing data layers to generate the model inputs needed for spatially explicit models such as GeoWEPP (Renschler, 2003). The development of this online database has allowed us to predict post-fire erosion and various remediation scenarios in just 1-7 days for six fires ranging in size from 4-540 km2. These initial successes have stimulated efforts to further improve the spatial extent and amount of data, and add functionality to support the USGS debris flow model, batch processing for Disturbed WEPP (Elliot et al., 2004) and ERMiT (Robichaud et al., 2007), and to support erosion modeling for other land uses, such as agriculture or mining. The design and techniques used to create the database and the modeling interface are readily repeatable for any area or country that has the necessary topography

  16. Explicitation et structuration des connaissances pour la transformation de l’entreprise : les apports de la méthode Praxeme

    OpenAIRE

    Biard , Thierry; Bigand , Michel; Bourey , Jean-Pierre

    2013-01-01

    International audience; To transform itself, enterprises must explicit then structure their knowledge. Praxeme, an enterprise architecture method, proposes a framework for that. This article describes an approach for using Praxeme and advises some modelling tools. The aim of enterprise architecture is reminded. The concept of business architecture transformation, which takes care of the enterprise's knowledge, is indicated. An overview of methodological framework is provided. A transformation...

  17. Modeling mixed retention and early arrivals in multidimensional heterogeneous media using an explicit Lagrangian scheme

    Science.gov (United States)

    Zhang, Yong; Meerschaert, Mark M.; Baeumer, Boris; LaBolle, Eric M.

    2015-08-01

    This study develops an explicit two-step Lagrangian scheme based on the renewal-reward process to capture transient anomalous diffusion with mixed retention and early arrivals in multidimensional media. The resulting 3-D anomalous transport simulator provides a flexible platform for modeling transport. The first step explicitly models retention due to mass exchange between one mobile zone and any number of parallel immobile zones. The mobile component of the renewal process can be calculated as either an exponential random variable or a preassigned time step, and the subsequent random immobile time follows a Hyper-exponential distribution for finite immobile zones or a tempered stable distribution for infinite immobile zones with an exponentially tempered power-law memory function. The second step describes well-documented early arrivals which can follow streamlines due to mechanical dispersion using the method of subordination to regional flow. Applicability and implementation of the Lagrangian solver are further checked against transport observed in various media. Results show that, although the time-nonlocal model parameters are predictable for transport with retention in alluvial settings, the standard time-nonlocal model cannot capture early arrivals. Retention and early arrivals observed in porous and fractured media can be efficiently modeled by our Lagrangian solver, allowing anomalous transport to be incorporated into 2-D/3-D models with irregular flow fields. Extensions of the particle-tracking approach are also discussed for transport with parameters conditioned on local aquifer properties, as required by transient flow and nonstationary media.

  18. The explicit and implicit dance in psychoanalytic change.

    Science.gov (United States)

    Fosshage, James L

    2004-02-01

    How the implicit/non-declarative and explicit/declarative cognitive domains interact is centrally important in the consideration of effecting change within the psychoanalytic arena. Stern et al. (1998) declare that long-lasting change occurs in the domain of implicit relational knowledge. In the view of this author, the implicit and explicit domains are intricately intertwined in an interactive dance within a psychoanalytic process. The author views that a spirit of inquiry (Lichtenberg, Lachmann & Fosshage 2002) serves as the foundation of the psychoanalytic process. Analyst and patient strive to explore, understand and communicate and, thereby, create a 'spirit' of interaction that contributes, through gradual incremental learning, to new implicit relational knowledge. This spirit, as part of the implicit relational interaction, is a cornerstone of the analytic relationship. The 'inquiry' more directly brings explicit/declarative processing to the foreground in the joint attempt to explore and understand. The spirit of inquiry in the psychoanalytic arena highlights both the autobiographical scenarios of the explicit memory system and the mental models of the implicit memory system as each contributes to a sense of self, other, and self with other. This process facilitates the extrication and suspension of the old models, so that new models based on current relational experience can be gradually integrated into both memory systems for lasting change.

  19. A new method for explicit modelling of single failure event within different common cause failure groups

    International Nuclear Information System (INIS)

    Kančev, Duško; Čepin, Marko

    2012-01-01

    Redundancy and diversity are the main principles of the safety systems in the nuclear industry. Implementation of safety components redundancy has been acknowledged as an effective approach for assuring high levels of system reliability. The existence of redundant components, identical in most of the cases, implicates a probability of their simultaneous failure due to a shared cause—a common cause failure. This paper presents a new method for explicit modelling of single component failure event within multiple common cause failure groups simultaneously. The method is based on a modification of the frequently utilised Beta Factor parametric model. The motivation for development of this method lays in the fact that one of the most widespread softwares for fault tree and event tree modelling as part of the probabilistic safety assessment does not comprise the option for simultaneous assignment of single failure event to multiple common cause failure groups. In that sense, the proposed method can be seen as an advantage of the explicit modelling of common cause failures. A standard standby safety system is selected as a case study for application and study of the proposed methodology. The results and insights implicate improved, more transparent and more comprehensive models within probabilistic safety assessment.

  20. Numerical modelling of tool wear in turning with cemented carbide cutting tools

    Science.gov (United States)

    Franco, P.; Estrems, M.; Faura, F.

    2007-04-01

    A numerical model is proposed for analysing the flank and crater wear resulting from the loss of material on cutting tool surface in turning processes due to wear mechanisms of adhesion, abrasion and fracture. By means of this model, the material loss along cutting tool surface can be analysed, and the worn surface shape during the workpiece machining can be determined. The proposed model analyses the gradual degradation of cutting tool during turning operation, and tool wear can be estimated as a function of cutting time. Wear-land width (VB) and crater depth (KT) can be obtained for description of material loss on cutting tool surface, and the effects of the distinct wear mechanisms on surface shape can be studied. The parameters required for the tool wear model are obtained from bibliography and experimental observation for AISI 4340 steel turning with WC-Co cutting tools.

  1. Numerical modelling of tool wear in turning with cemented carbide cutting tools

    International Nuclear Information System (INIS)

    Franco, P.; Estrems, M.; Faura, F.

    2007-01-01

    A numerical model is proposed for analysing the flank and crater wear resulting from the loss of material on cutting tool surface in turning processes due to wear mechanisms of adhesion, abrasion and fracture. By means of this model, the material loss along cutting tool surface can be analysed, and the worn surface shape during the workpiece machining can be determined. The proposed model analyses the gradual degradation of cutting tool during turning operation, and tool wear can be estimated as a function of cutting time. Wear-land width (VB) and crater depth (KT) can be obtained for description of material loss on cutting tool surface, and the effects of the distinct wear mechanisms on surface shape can be studied. The parameters required for the tool wear model are obtained from bibliography and experimental observation for AISI 4340 steel turning with WC-Co cutting tools

  2. Making Culturally Responsive Mathematics Teaching Explicit: A Lesson Analysis Tool

    Science.gov (United States)

    Aguirre, Julia M.; Zavala, Maria del Rosario

    2013-01-01

    In the United States, there is a need for pedagogical tools that help teachers develop essential pedagogical content knowledge and practices to meet the mathematical education needs of a growing culturally and linguistically diverse student population. In this article, we introduce an innovative lesson analysis tool that focuses on integrating…

  3. A risk explicit interval linear programming model for uncertainty-based environmental economic optimization in the Lake Fuxian watershed, China.

    Science.gov (United States)

    Zhang, Xiaoling; Huang, Kai; Zou, Rui; Liu, Yong; Yu, Yajuan

    2013-01-01

    The conflict of water environment protection and economic development has brought severe water pollution and restricted the sustainable development in the watershed. A risk explicit interval linear programming (REILP) method was used to solve integrated watershed environmental-economic optimization problem. Interval linear programming (ILP) and REILP models for uncertainty-based environmental economic optimization at the watershed scale were developed for the management of Lake Fuxian watershed, China. Scenario analysis was introduced into model solution process to ensure the practicality and operability of optimization schemes. Decision makers' preferences for risk levels can be expressed through inputting different discrete aspiration level values into the REILP model in three periods under two scenarios. Through balancing the optimal system returns and corresponding system risks, decision makers can develop an efficient industrial restructuring scheme based directly on the window of "low risk and high return efficiency" in the trade-off curve. The representative schemes at the turning points of two scenarios were interpreted and compared to identify a preferable planning alternative, which has the relatively low risks and nearly maximum benefits. This study provides new insights and proposes a tool, which was REILP, for decision makers to develop an effectively environmental economic optimization scheme in integrated watershed management.

  4. A Risk Explicit Interval Linear Programming Model for Uncertainty-Based Environmental Economic Optimization in the Lake Fuxian Watershed, China

    Directory of Open Access Journals (Sweden)

    Xiaoling Zhang

    2013-01-01

    Full Text Available The conflict of water environment protection and economic development has brought severe water pollution and restricted the sustainable development in the watershed. A risk explicit interval linear programming (REILP method was used to solve integrated watershed environmental-economic optimization problem. Interval linear programming (ILP and REILP models for uncertainty-based environmental economic optimization at the watershed scale were developed for the management of Lake Fuxian watershed, China. Scenario analysis was introduced into model solution process to ensure the practicality and operability of optimization schemes. Decision makers’ preferences for risk levels can be expressed through inputting different discrete aspiration level values into the REILP model in three periods under two scenarios. Through balancing the optimal system returns and corresponding system risks, decision makers can develop an efficient industrial restructuring scheme based directly on the window of “low risk and high return efficiency” in the trade-off curve. The representative schemes at the turning points of two scenarios were interpreted and compared to identify a preferable planning alternative, which has the relatively low risks and nearly maximum benefits. This study provides new insights and proposes a tool, which was REILP, for decision makers to develop an effectively environmental economic optimization scheme in integrated watershed management.

  5. Explicit MDS Codes with Complementary Duals

    DEFF Research Database (Denmark)

    Beelen, Duals Peter; Jin, Lingfei

    2018-01-01

    In 1964, Massey introduced a class of codes with complementary duals which are called Linear Complimentary Dual (LCD for short) codes. He showed that LCD codes have applications in communication system, side-channel attack (SCA) and so on. LCD codes have been extensively studied in literature....... On the other hand, MDS codes form an optimal family of classical codes which have wide applications in both theory and practice. The main purpose of this paper is to give an explicit construction of several classes of LCD MDS codes, using tools from algebraic function fields. We exemplify this construction...

  6. Tools and Algorithms for the Construction and Analysis of Systems

    DEFF Research Database (Denmark)

    This book constitutes the refereed proceedings of the 10th International Conference on Tools and Algorithms for the Construction and Analysis of Systems, TACAS 2004, held in Barcelona, Spain in March/April 2004. The 37 revised full papers and 6 revised tool demonstration papers presented were...... carefully reviewed and selected from a total of 162 submissions. The papers are organized in topical sections on theorem proving, probabilistic model checking, testing, tools, explicit state and Petri nets, scheduling, constraint solving, timed systems, case studies, software, temporal logic, abstraction...

  7. Computer-Aided Modelling Methods and Tools

    DEFF Research Database (Denmark)

    Cameron, Ian; Gani, Rafiqul

    2011-01-01

    The development of models for a range of applications requires methods and tools. In many cases a reference model is required that allows the generation of application specific models that are fit for purpose. There are a range of computer aided modelling tools available that help to define the m...

  8. Face puzzle—two new video-based tasks for measuring explicit and implicit aspects of facial emotion recognition

    Science.gov (United States)

    Kliemann, Dorit; Rosenblau, Gabriela; Bölte, Sven; Heekeren, Hauke R.; Dziobek, Isabel

    2013-01-01

    Recognizing others' emotional states is crucial for effective social interaction. While most facial emotion recognition tasks use explicit prompts that trigger consciously controlled processing, emotional faces are almost exclusively processed implicitly in real life. Recent attempts in social cognition suggest a dual process perspective, whereby explicit and implicit processes largely operate independently. However, due to differences in methodology the direct comparison of implicit and explicit social cognition has remained a challenge. Here, we introduce a new tool to comparably measure implicit and explicit processing aspects comprising basic and complex emotions in facial expressions. We developed two video-based tasks with similar answer formats to assess performance in respective facial emotion recognition processes: Face Puzzle, implicit and explicit. To assess the tasks' sensitivity to atypical social cognition and to infer interrelationship patterns between explicit and implicit processes in typical and atypical development, we included healthy adults (NT, n = 24) and adults with autism spectrum disorder (ASD, n = 24). Item analyses yielded good reliability of the new tasks. Group-specific results indicated sensitivity to subtle social impairments in high-functioning ASD. Correlation analyses with established implicit and explicit socio-cognitive measures were further in favor of the tasks' external validity. Between group comparisons provide first hints of differential relations between implicit and explicit aspects of facial emotion recognition processes in healthy compared to ASD participants. In addition, an increased magnitude of between group differences in the implicit task was found for a speed-accuracy composite measure. The new Face Puzzle tool thus provides two new tasks to separately assess explicit and implicit social functioning, for instance, to measure subtle impairments as well as potential improvements due to social cognitive

  9. Nano-colloid electrophoretic transport: Fully explicit modelling via dissipative particle dynamics

    Science.gov (United States)

    Hassanzadeh Afrouzi, Hamid; Farhadi, Mousa; Sedighi, Kurosh; Moshfegh, Abouzar

    2018-02-01

    In present study, a novel fully explicit approach using dissipative particle dynamics (DPD) method is introduced for modelling electrophoretic transport of nano-colloids in an electrolyte solution. Slater type charge smearing function included in 3D Ewald summation method is employed to treat electrostatic interaction. Moreover, capability of different thermostats are challenged to control the system temperature and study the dynamic response of colloidal electrophoretic mobility under practical ranges of external electric field in nano scale application (0.072 600 in DPD units regardless of electric field intensity. Nosé-Hoover-Lowe-Andersen and Lowe-Andersen thermostats are found to function more effectively under high electric fields (E > 0.145 [ v / nm ]) while thermal equilibrium is maintained. Reasonable agreements are achieved by benchmarking the radial distribution function with available electrolyte structure modellings, as well as comparing reduced mobility against conventional Smoluchowski and Hückel theories, and numerical solution of Poisson-Boltzmann equation.

  10. Explicit/multi-parametric model predictive control (MPC) of linear discrete-time systems by dynamic and multi-parametric programming

    KAUST Repository

    Kouramas, K.I.; Faí sca, N.P.; Panos, C.; Pistikopoulos, E.N.

    2011-01-01

    This work presents a new algorithm for solving the explicit/multi- parametric model predictive control (or mp-MPC) problem for linear, time-invariant discrete-time systems, based on dynamic programming and multi-parametric programming techniques

  11. Modeling nitrous oxide production and reduction in soil through explicit representation of denitrification enzyme kinetics.

    Science.gov (United States)

    Zheng, Jianqiu; Doskey, Paul V

    2015-02-17

    An enzyme-explicit denitrification model with representations for pre- and de novo synthesized enzymes was developed to improve predictions of nitrous oxide (N2O) accumulations in soil and emissions from the surface. The metabolic model of denitrification is based on dual-substrate utilization and Monod growth kinetics. Enzyme synthesis/activation was incorporated into each sequential reduction step of denitrification to regulate dynamics of the denitrifier population and the active enzyme pool, which controlled the rate function. Parameterizations were developed from observations of the dynamics of N2O production and reduction in soil incubation experiments. The model successfully reproduced the dynamics of N2O and N2 accumulation in the incubations and revealed an important regulatory effect of denitrification enzyme kinetics on the accumulation of denitrification products. Pre-synthesized denitrification enzymes contributed 20, 13, 43, and 62% of N2O that accumulated in 48 h incubations of soil collected from depths of 0-5, 5-10, 10-15, and 15-25 cm, respectively. An enzyme activity function (E) was defined to estimate the relative concentration of active enzymes and variation in response to environmental conditions. The value of E allows for activities of pre-synthesized denitrification enzymes to be differentiated from de novo synthesized enzymes. Incorporating explicit representations of denitrification enzyme kinetics into biogeochemical models is a promising approach for accurately simulating dynamics of the production and reduction of N2O in soils.

  12. Model tool to describe chemical structures in XML format utilizing structural fragments and chemical ontology.

    Science.gov (United States)

    Sankar, Punnaivanam; Alain, Krief; Aghila, Gnanasekaran

    2010-05-24

    We have developed a model structure-editing tool, ChemEd, programmed in JAVA, which allows drawing chemical structures on a graphical user interface (GUI) by selecting appropriate structural fragments defined in a fragment library. The terms representing the structural fragments are organized in fragment ontology to provide a conceptual support. ChemEd describes the chemical structure in an XML document (ChemFul) with rich semantics explicitly encoding the details of the chemical bonding, the hybridization status, and the electron environment around each atom. The document can be further processed through suitable algorithms and with the support of external chemical ontologies to generate understandable reports about the functional groups present in the structure and their specific environment.

  13. Explicit Oral Narrative Intervention for Students with Williams Syndrome

    Directory of Open Access Journals (Sweden)

    Eliseo Diez-Itza

    2018-01-01

    Full Text Available Narrative skills play a crucial role in organizing experience, facilitating social interaction and building academic discourse and literacy. They are at the interface of cognitive, social, and linguistic abilities related to school engagement. Despite their relative strengths in social and grammatical skills, students with Williams syndrome (WS do not show parallel cognitive and pragmatic performance in narrative generation tasks. The aim of the present study was to assess retelling of a TV cartoon tale and the effect of an individualized explicit instruction of the narrative structure. Participants included eight students with WS who attended different special education levels. Narratives were elicited in two sessions (pre and post intervention, and were transcribed, coded and analyzed using the tools of the CHILDES Project. Narratives were coded for productivity and complexity at the microstructure and macrostructure levels. Microstructure productivity (i.e., length of narratives included number of utterances, clauses, and tokens. Microstructure complexity included mean length of utterances, lexical diversity and use of discourse markers as cohesive devices. Narrative macrostructure was assessed for textual coherence through the Pragmatic Evaluation Protocol for Speech Corpora (PREP-CORP. Macrostructure productivity and complexity included, respectively, the recall and sequential order of scenarios, episodes, events and characters. A total of four intervention sessions, lasting approximately 20 min, were delivered individually once a week. This brief intervention addressed explicit instruction about the narrative structure and the use of specific discourse markers to improve cohesion of story retellings. Intervention strategies included verbal scaffolding and modeling, conversational context for retelling the story and visual support with pictures printed from the cartoon. Results showed significant changes in WS students’ retelling of the

  14. Explicit Oral Narrative Intervention for Students with Williams Syndrome

    Science.gov (United States)

    Diez-Itza, Eliseo; Martínez, Verónica; Pérez, Vanesa; Fernández-Urquiza, Maite

    2018-01-01

    Narrative skills play a crucial role in organizing experience, facilitating social interaction and building academic discourse and literacy. They are at the interface of cognitive, social, and linguistic abilities related to school engagement. Despite their relative strengths in social and grammatical skills, students with Williams syndrome (WS) do not show parallel cognitive and pragmatic performance in narrative generation tasks. The aim of the present study was to assess retelling of a TV cartoon tale and the effect of an individualized explicit instruction of the narrative structure. Participants included eight students with WS who attended different special education levels. Narratives were elicited in two sessions (pre and post intervention), and were transcribed, coded and analyzed using the tools of the CHILDES Project. Narratives were coded for productivity and complexity at the microstructure and macrostructure levels. Microstructure productivity (i.e., length of narratives) included number of utterances, clauses, and tokens. Microstructure complexity included mean length of utterances, lexical diversity and use of discourse markers as cohesive devices. Narrative macrostructure was assessed for textual coherence through the Pragmatic Evaluation Protocol for Speech Corpora (PREP-CORP). Macrostructure productivity and complexity included, respectively, the recall and sequential order of scenarios, episodes, events and characters. A total of four intervention sessions, lasting approximately 20 min, were delivered individually once a week. This brief intervention addressed explicit instruction about the narrative structure and the use of specific discourse markers to improve cohesion of story retellings. Intervention strategies included verbal scaffolding and modeling, conversational context for retelling the story and visual support with pictures printed from the cartoon. Results showed significant changes in WS students’ retelling of the story, both at

  15. Energy-economy models and energy efficiency policy evaluation for the household sector. An analysis of modelling tools and analytical approaches

    Energy Technology Data Exchange (ETDEWEB)

    Mundaca, Luis; Neij, Lena

    2009-10-15

    -economy models, empirical literature shows that a larger variety of determinants need to be taken into account when analysing the process of adoption of efficient technologies. We then focus on the analysis of more than twenty case studies addressing the application of the reviewed modelling methodologies to the field of residential energy efficiency policy. Regarding policy instruments being evaluated, the majority of the cases focus on regulatory aspects, such as minimum performance standards and building codes. For the rest, evaluations focus on economically-driven policy instruments. The dominance of economic and engineering determinants for technology choice gives little room for the representation of informative policy instruments. In all cases, policy instruments are represented through technical factors and costs of measures for energy efficiency improvements. In addition, policy instruments tend to be modelled in an idealistic or oversimplified manner. The traditional but narrow single-criterion evaluation approach based on cost-effectiveness seems to dominate the limited number of evaluation studies. However, this criterion is inappropriate to comprehensively address the attributes of policy instruments and the institutional and market conditions in which they work. We then turn to identifying research areas that have the potential to further advance modelling tools. We first discuss modelling issues as such, including the importance of transparent modelling efforts; the explicit elaboration of methodologies to represent policies; the need to better translate modelling results into a set of concrete policy recommendations; and the use of complementary research methods to better comprehend the broad effects and attributes of policy instruments. Secondly, we approach techno-economic and environmental components of models. We discuss the integration of co-benefits as a key research element of modelling studies; the introduction of transaction costs to further improve the

  16. Spatially explicit spectral analysis of point clouds and geospatial data

    Science.gov (United States)

    Buscombe, Daniel D.

    2015-01-01

    The increasing use of spatially explicit analyses of high-resolution spatially distributed data (imagery and point clouds) for the purposes of characterising spatial heterogeneity in geophysical phenomena necessitates the development of custom analytical and computational tools. In recent years, such analyses have become the basis of, for example, automated texture characterisation and segmentation, roughness and grain size calculation, and feature detection and classification, from a variety of data types. In this work, much use has been made of statistical descriptors of localised spatial variations in amplitude variance (roughness), however the horizontal scale (wavelength) and spacing of roughness elements is rarely considered. This is despite the fact that the ratio of characteristic vertical to horizontal scales is not constant and can yield important information about physical scaling relationships. Spectral analysis is a hitherto under-utilised but powerful means to acquire statistical information about relevant amplitude and wavelength scales, simultaneously and with computational efficiency. Further, quantifying spatially distributed data in the frequency domain lends itself to the development of stochastic models for probing the underlying mechanisms which govern the spatial distribution of geological and geophysical phenomena. The software packagePySESA (Python program for Spatially Explicit Spectral Analysis) has been developed for generic analyses of spatially distributed data in both the spatial and frequency domains. Developed predominantly in Python, it accesses libraries written in Cython and C++ for efficiency. It is open source and modular, therefore readily incorporated into, and combined with, other data analysis tools and frameworks with particular utility for supporting research in the fields of geomorphology, geophysics, hydrography, photogrammetry and remote sensing. The analytical and computational structure of the toolbox is

  17. Spatially explicit spectral analysis of point clouds and geospatial data

    Science.gov (United States)

    Buscombe, Daniel

    2016-01-01

    The increasing use of spatially explicit analyses of high-resolution spatially distributed data (imagery and point clouds) for the purposes of characterising spatial heterogeneity in geophysical phenomena necessitates the development of custom analytical and computational tools. In recent years, such analyses have become the basis of, for example, automated texture characterisation and segmentation, roughness and grain size calculation, and feature detection and classification, from a variety of data types. In this work, much use has been made of statistical descriptors of localised spatial variations in amplitude variance (roughness), however the horizontal scale (wavelength) and spacing of roughness elements is rarely considered. This is despite the fact that the ratio of characteristic vertical to horizontal scales is not constant and can yield important information about physical scaling relationships. Spectral analysis is a hitherto under-utilised but powerful means to acquire statistical information about relevant amplitude and wavelength scales, simultaneously and with computational efficiency. Further, quantifying spatially distributed data in the frequency domain lends itself to the development of stochastic models for probing the underlying mechanisms which govern the spatial distribution of geological and geophysical phenomena. The software package PySESA (Python program for Spatially Explicit Spectral Analysis) has been developed for generic analyses of spatially distributed data in both the spatial and frequency domains. Developed predominantly in Python, it accesses libraries written in Cython and C++ for efficiency. It is open source and modular, therefore readily incorporated into, and combined with, other data analysis tools and frameworks with particular utility for supporting research in the fields of geomorphology, geophysics, hydrography, photogrammetry and remote sensing. The analytical and computational structure of the toolbox is described

  18. An explicit asymptotic model for the surface wave in a viscoelastic half-space based on applying Rabotnov's fractional exponential integral operators

    Science.gov (United States)

    Wilde, M. V.; Sergeeva, N. V.

    2018-05-01

    An explicit asymptotic model extracting the contribution of a surface wave to the dynamic response of a viscoelastic half-space is derived. Fractional exponential Rabotnov's integral operators are used for describing of material properties. The model is derived by extracting the principal part of the poles corresponding to the surface waves after applying Laplace and Fourier transforms. The simplified equations for the originals are written by using power series expansions. Padè approximation is constructed to unite short-time and long-time models. The form of this approximation allows to formulate the explicit model using a fractional exponential Rabotnov's integral operator with parameters depending on the properties of surface wave. The applicability of derived models is studied by comparing with the exact solutions of a model problem. It is revealed that the model based on Padè approximation is highly effective for all the possible time domains.

  19. Efficient explicit formulation for practical fuzzy structural analysis

    Indian Academy of Sciences (India)

    This paper presents a practical approach based on High Dimensional Model Representation (HDMR) for analysing the response of structures with fuzzy parameters. The proposed methodology involves integrated finite element modelling, HDMR based response surface generation, and explicit fuzzy analysis procedures.

  20. Cockpit System Situational Awareness Modeling Tool

    Science.gov (United States)

    Keller, John; Lebiere, Christian; Shay, Rick; Latorella, Kara

    2004-01-01

    This project explored the possibility of predicting pilot situational awareness (SA) using human performance modeling techniques for the purpose of evaluating developing cockpit systems. The Improved Performance Research Integration Tool (IMPRINT) was combined with the Adaptive Control of Thought-Rational (ACT-R) cognitive modeling architecture to produce a tool that can model both the discrete tasks of pilots and the cognitive processes associated with SA. The techniques for using this tool to predict SA were demonstrated using the newly developed Aviation Weather Information (AWIN) system. By providing an SA prediction tool to cockpit system designers, cockpit concepts can be assessed early in the design process while providing a cost-effective complement to the traditional pilot-in-the-loop experiments and data collection techniques.

  1. A spatially explicit hydro-ecological modeling framework (BEPS-TerrainLab V2.0): Model description and test in a boreal ecosystem in Eastern North America

    Science.gov (United States)

    Govind, Ajit; Chen, Jing Ming; Margolis, Hank; Ju, Weimin; Sonnentag, Oliver; Giasson, Marc-André

    2009-04-01

    SummaryA spatially explicit, process-based hydro-ecological model, BEPS-TerrainLab V2.0, was developed to improve the representation of ecophysiological, hydro-ecological and biogeochemical processes of boreal ecosystems in a tightly coupled manner. Several processes unique to boreal ecosystems were implemented including the sub-surface lateral water fluxes, stratification of vegetation into distinct layers for explicit ecophysiological representation, inclusion of novel spatial upscaling strategies and biogeochemical processes. To account for preferential water fluxes common in humid boreal ecosystems, a novel scheme was introduced based on laboratory analyses. Leaf-scale ecophysiological processes were upscaled to canopy-scale by explicitly considering leaf physiological conditions as affected by light and water stress. The modified model was tested with 2 years of continuous measurements taken at the Eastern Old Black Spruce Site of the Fluxnet-Canada Research Network located in a humid boreal watershed in eastern Canada. Comparison of the simulated and measured ET, water-table depth (WTD), volumetric soil water content (VSWC) and gross primary productivity (GPP) revealed that BEPS-TerrainLab V2.0 simulates hydro-ecological processes with reasonable accuracy. The model was able to explain 83% of the ET, 92% of the GPP variability and 72% of the WTD dynamics. The model suggests that in humid ecosystems such as eastern North American boreal watersheds, topographically driven sub-surface baseflow is the main mechanism of soil water partitioning which significantly affects the local-scale hydrological conditions.

  2. Software Engineering Tools for Scientific Models

    Science.gov (United States)

    Abrams, Marc; Saboo, Pallabi; Sonsini, Mike

    2013-01-01

    Software tools were constructed to address issues the NASA Fortran development community faces, and they were tested on real models currently in use at NASA. These proof-of-concept tools address the High-End Computing Program and the Modeling, Analysis, and Prediction Program. Two examples are the NASA Goddard Earth Observing System Model, Version 5 (GEOS-5) atmospheric model in Cell Fortran on the Cell Broadband Engine, and the Goddard Institute for Space Studies (GISS) coupled atmosphere- ocean model called ModelE, written in fixed format Fortran.

  3. An improved risk-explicit interval linear programming model for pollution load allocation for watershed management.

    Science.gov (United States)

    Xia, Bisheng; Qian, Xin; Yao, Hong

    2017-11-01

    Although the risk-explicit interval linear programming (REILP) model has solved the problem of having interval solutions, it has an equity problem, which can lead to unbalanced allocation between different decision variables. Therefore, an improved REILP model is proposed. This model adds an equity objective function and three constraint conditions to overcome this equity problem. In this case, pollution reduction is in proportion to pollutant load, which supports balanced development between different regional economies. The model is used to solve the problem of pollution load allocation in a small transboundary watershed. Compared with the REILP original model result, our model achieves equity between the upstream and downstream pollutant loads; it also overcomes the problem of greatest pollution reduction, where sources are nearest to the control section. The model provides a better solution to the problem of pollution load allocation than previous versions.

  4. Explicit all-atom modeling of realistically sized ligand-capped nanocrystals

    KAUST Repository

    Kaushik, Ananth P.; Clancy, Paulette

    2012-01-01

    We present a study of an explicit all-atom representation of nanocrystals of experimentally relevant sizes (up to 6 nm), capped with alkyl chain ligands, in vacuum. We employ all-atom molecular dynamics simulation methods in concert with a well

  5. An evaluation of BPMN modeling tools

    NARCIS (Netherlands)

    Yan, Z.; Reijers, H.A.; Dijkman, R.M.; Mendling, J.; Weidlich, M.

    2010-01-01

    Various BPMN modeling tools are available and it is close to impossible to understand their functional differences without simply trying them out. This paper presents an evaluation framework and presents the outcomes of its application to a set of five BPMN modeling tools. We report on various

  6. Explicit Nonlinear Model Predictive Control for a Saucer-Shaped Unmanned Aerial Vehicle

    Directory of Open Access Journals (Sweden)

    Zhihui Xing

    2013-01-01

    Full Text Available A lifting body unmanned aerial vehicle (UAV generates lift by its body and shows many significant advantages due to the particular shape, such as huge loading space, small wetted area, high-strength fuselage structure, and large lifting area. However, designing the control law for a lifting body UAV is quite challenging because it has strong nonlinearity and coupling, and usually lacks it rudders. In this paper, an explicit nonlinear model predictive control (ENMPC strategy is employed to design a control law for a saucer-shaped UAV which can be adequately modeled with a rigid 6-degrees-of-freedom (DOF representation. In the ENMPC, control signal is calculated by approximation of the tracking error in the receding horizon by its Taylor-series expansion to any specified order. It enhances the advantages of the nonlinear model predictive control and eliminates the time-consuming online optimization. The simulation results show that ENMPC is a propriety strategy for controlling lifting body UAVs and can compensate the insufficient control surface area.

  7. Proc. of the Workshop on Agent Simulation : Applications, Models, and Tools, Oct. 15-16, 1999

    International Nuclear Information System (INIS)

    Macal, C. M.; Sallach, D.

    2000-01-01

    The many motivations for employing agent-based computation in the social sciences are reviewed. It is argued that there exist three distinct uses of agent modeling techniques. One such use-the simplest-is conceptually quite close to traditional simulation in operations research. This use arises when equations can be formulated that completely describe a social process, and these equations are explicitly soluble, either analytically or numerically. In the former case, the agent model is merely a tool for presenting results, while in the latter it is a novel kind of Monte Carlo analysis. A second, more commonplace usage of computational agent models arises when mathematical models can be written down but not completely solved. In this case the agent-based model can shed significant light on the solution structure, illustrate dynamical properties of the model, serve to test the dependence of results on parameters and assumptions, and be a source of counter-examples. Finally, there are important classes of problems for which writing down equations is not a useful activity. In such circumstances, resort to agent-based computational models may be the only way available to explore such processes systematically, and constitute a third distinct usage of such models

  8. A new approach to spatially explicit modelling of forest dynamics: spacing, ageing and neighbourhood competition of mangrove trees

    NARCIS (Netherlands)

    Berger, U.; Hildenbrandt, H.

    2000-01-01

    This paper presents a new approach to spatially explicit modelling that enables the influence of neighbourhood effects on the dynamics of forests and plant communities to be analysed. We refer to this approach as 'field of neighbourhood' (FON). It combines the 'neighbourhood philosophy' of

  9. Alien wavelength modeling tool and field trial

    DEFF Research Database (Denmark)

    Sambo, N.; Sgambelluri, A.; Secondini, M.

    2015-01-01

    A modeling tool is presented for pre-FEC BER estimation of PM-QPSK alien wavelength signals. A field trial is demonstrated and used as validation of the tool's correctness. A very close correspondence between the performance of the field trial and the one predicted by the modeling tool has been...

  10. Improve Student Understanding Ability Through Gamification in Instructional Media Based Explicit Instruction

    Science.gov (United States)

    Firdausi, N.; Prabawa, H. W.; Sutarno, H.

    2017-02-01

    In an effort to maximize a student’s academic growth, one of the tools available to educators is the explicit instruction. Explicit instruction is marked by a series of support or scaffold, where the students will be guided through the learning process with a clear statement of purpose and a reason for learning new skills, a clear explanation and demonstration of learning targets, supported and practiced with independent feedback until mastery has been achieved. The technology development trend of todays, requires an adjustment in the development of learning object that supports the achievement of explicit instruction targets. This is where the gamification position is. In the role as a pedagogical strategy, the use of gamification preformance study class is still relatively new. Gamification not only use the game elements and game design techniques in non-game contexts, but also to empower and engage learners with the ability of motivation on learning approach and maintains a relaxed atmosphere. With using Reseach and Development methods, this paper presents the integration of technology (which in this case using the concept of gamification) in explicit instruction settings and the impact on the improvement of students’ understanding.

  11. Development and Validation of Spatially Explicit Habitat Models for Cavity-nesting Birds in Fishlake National Forest, Utah

    Science.gov (United States)

    Randall A., Jr. Schultz; Thomas C., Jr. Edwards; Gretchen G. Moisen; Tracey S. Frescino

    2005-01-01

    The ability of USDA Forest Service Forest Inventory and Analysis (FIA) generated spatial products to increase the predictive accuracy of spatially explicit, macroscale habitat models was examined for nest-site selection by cavity-nesting birds in Fishlake National Forest, Utah. One FIA-derived variable (percent basal area of aspen trees) was significant in the habitat...

  12. Explicit modeling of volatile organic compounds partitioning in the atmospheric aqueous phase

    Directory of Open Access Journals (Sweden)

    C. Mouchel-Vallon

    2013-01-01

    Full Text Available The gas phase oxidation of organic species is a multigenerational process involving a large number of secondary compounds. Most secondary organic species are water-soluble multifunctional oxygenated molecules. The fully explicit chemical mechanism GECKO-A (Generator of Explicit Chemistry and Kinetics of Organics in the Atmosphere is used to describe the oxidation of organics in the gas phase and their mass transfer to the aqueous phase. The oxidation of three hydrocarbons of atmospheric interest (isoprene, octane and α-pinene is investigated for various NOx conditions. The simulated oxidative trajectories are examined in a new two dimensional space defined by the mean oxidation state and the solubility. The amount of dissolved organic matter was found to be very low (yield less than 2% on carbon atom basis under a water content typical of deliquescent aerosols. For cloud water content, 50% (isoprene oxidation to 70% (octane oxidation of the carbon atoms are found in the aqueous phase after the removal of the parent hydrocarbons for low NOx conditions. For high NOx conditions, this ratio is only 5% in the isoprene oxidation case, but remains large for α-pinene and octane oxidation cases (40% and 60%, respectively. Although the model does not yet include chemical reactions in the aqueous phase, much of this dissolved organic matter should be processed in cloud drops and modify both oxidation rates and the speciation of organic species.

  13. Explicit Instruction Elements in Core Reading Programs

    Science.gov (United States)

    Child, Angela R.

    2012-01-01

    Classroom teachers are provided instructional recommendations for teaching reading from their adopted core reading programs (CRPs). Explicit instruction elements or what is also called instructional moves, including direct explanation, modeling, guided practice, independent practice, discussion, feedback, and monitoring, were examined within CRP…

  14. Modeling mind-wandering: a tool to better understand distraction

    NARCIS (Netherlands)

    van Vugt, Marieke; Taatgen, Niels; Sackur, Jerome; Bastian, Mikael; Taatgen, Niels; van Vugt, Marieke; Borst, Jelmer; Mehlhorn, Katja

    2015-01-01

    When we get distracted, we may engage in mind-wandering, or task-unrelated thinking, which impairs performance on cognitive tasks. Yet, we do not have cognitive models that make this process explicit. On the basis of both recent experiments that have started to investigate mind-wandering and

  15. Modeling and Tool Wear in Routing of CFRP

    International Nuclear Information System (INIS)

    Iliescu, D.; Fernandez, A.; Gutierrez-Orrantia, M. E.; Lopez de Lacalle, L. N.; Girot, F.

    2011-01-01

    This paper presents the prediction and evaluation of feed force in routing of carbon composite material. In order to extend tool life and improve quality of the machined surface, a better understanding of uncoated and coated tool behaviors is required. This work describes (1) the optimization of the geometry of multiple teeth tools minimizing the tool wear and the feed force, (2) the optimization of tool coating and (3) the development of a phenomenological model between the feed force, the routing parameters and the tool wear. The experimental results indicate that the feed rate, the cutting speed and the tool wear are the most significant factors affecting the feed force. In the case of multiple teeth tools, a particular geometry with 14 teeth right helix right cut and 11 teeth left helix right cut gives the best results. A thick AlTiN coating or a diamond coating can dramatically improve the tool life while minimizing the axial force, roughness and delamination. A wear model has then been developed based on an abrasive behavior of the tool. The model links the feed rate to the tool geometry parameters (tool diameter), to the process parameters (feed rate, cutting speed and depth of cut) and to the wear. The model presented has been verified by experimental tests.

  16. Hybrid fur rendering: combining volumetric fur with explicit hair strands

    DEFF Research Database (Denmark)

    Andersen, Tobias Grønbeck; Falster, Viggo; Frisvad, Jeppe Revall

    2016-01-01

    Hair is typically modeled and rendered using either explicitly defined hair strand geometry or a volume texture of hair densities. Taken each on their own, these two hair representations have difficulties in the case of animal fur as it consists of very dense and thin undercoat hairs in combination...... with coarse guard hairs. Explicit hair strand geometry is not well-suited for the undercoat hairs, while volume textures are not well-suited for the guard hairs. To efficiently model and render both guard hairs and undercoat hairs, we present a hybrid technique that combines rasterization of explicitly...... defined guard hairs with ray marching of a prismatic shell volume with dynamic resolution. The latter is the key to practical combination of the two techniques, and it also enables a high degree of detail in the undercoat. We demonstrate that our hybrid technique creates a more detailed and soft fur...

  17. Control of boiler temperature with explicit MPC; Panntemperaturreglering med explicit MPC

    Energy Technology Data Exchange (ETDEWEB)

    Slaetteke, Ola; Velut, Stefan; Raaberg, Martin

    2012-02-15

    MPC is the multivariable controller that has been most successful in the process industry and particularly the petrochemical industry. It has been described as one of the most significant developments in process control and the main reasons for this are: 1. It handles multivariable control problems in a natural manner. 2. It is relative easy to understand the structure of the controller, which is the same whether it is a simple loop or a multivariable system. 3. It handles limitations of both the process and other practical constraints in a systematic way. Examples of this is that a valve can only work between 0 and 100 %, but also that the CO-level in the flue gas must not exceed a certain level. 4. It allows for operating conditions near critical process boundaries, which in many cases is synonymous with increased production rates, reduced raw material consumption, better energy utilization, and faster process transitions. The aim of the project is to evaluate the potential of multivariable control in the form of explicit MPC in a boiler at Stora Enso Hylte Bruk. This research task can be divided into two sub-tasks: 1. General evaluation of explicit MPC. 2. Evaluation of multivariable control of boiler temperature The purpose of subtask one is to evaluate what is required of a facility owner to implement explicit MPC in a control system. This includes everything from available calculation tools, what is important to consider during the design phase of the controller, different pitfalls that exist, management of different operating modes, to how the controller should be implemented and commissioned. Subtask two is intended to evaluate the multivariable control of a boiler of CFB type (circulating fluidized bed). MPC controller will regulate the temperature in the boiler. In order to maintain the waste incineration directive, the temperature in the upper part of the boiler is controlled. This is done by means of changes in the flow of natural gas injection and

  18. Studies of implicit and explicit solution techniques in transient thermal analysis of structures

    International Nuclear Information System (INIS)

    Adelman, H.M.; Haftka, R.T.; Robinson, J.C.

    1982-08-01

    Studies aimed at an increase in the efficiency of calculating transient temperature fields in complex aerospace vehicle structures are reported. The advantages and disadvantages of explicit and implicit algorithms are discussed and a promising set of implicit algorithms with variable time steps, known as GEARIB, is described. Test problems, used for evaluating and comparing various algorithms, are discussed and finite element models of the configurations are described. These problems include a coarse model of the Space Shuttle wing, an insulated frame test article, a metallic panel for a thermal protection system, and detailed models of sections of the Space Shuttle wing. Results generally indicate a preference for implicit over explicit algorithms for transient structural heat transfer problems when the governing equations are stiff (typical of many practical problems such as insulated metal structures). The effects on algorithm performance of different models of an insulated cylinder are demonstrated. The stiffness of the problem is highly sensitive to modeling details and careful modeling can reduce the stiffness of the equations to the extent that explicit methods may become the best choice. Preliminary applications of a mixed implicit-explicit algorithm and operator splitting techniques for speeding up the solution of the algebraic equations are also described

  19. Studies of implicit and explicit solution techniques in transient thermal analysis of structures

    Science.gov (United States)

    Adelman, H. M.; Haftka, R. T.; Robinson, J. C.

    1982-01-01

    Studies aimed at an increase in the efficiency of calculating transient temperature fields in complex aerospace vehicle structures are reported. The advantages and disadvantages of explicit and implicit algorithms are discussed and a promising set of implicit algorithms with variable time steps, known as GEARIB, is described. Test problems, used for evaluating and comparing various algorithms, are discussed and finite element models of the configurations are described. These problems include a coarse model of the Space Shuttle wing, an insulated frame tst article, a metallic panel for a thermal protection system, and detailed models of sections of the Space Shuttle wing. Results generally indicate a preference for implicit over explicit algorithms for transient structural heat transfer problems when the governing equations are stiff (typical of many practical problems such as insulated metal structures). The effects on algorithm performance of different models of an insulated cylinder are demonstrated. The stiffness of the problem is highly sensitive to modeling details and careful modeling can reduce the stiffness of the equations to the extent that explicit methods may become the best choice. Preliminary applications of a mixed implicit-explicit algorithm and operator splitting techniques for speeding up the solution of the algebraic equations are also described.

  20. Explicit simulation of ice particle habits in a Numerical Weather Prediction Model

    Science.gov (United States)

    Hashino, Tempei

    2007-05-01

    This study developed a scheme for explicit simulation of ice particle habits in Numerical Weather Prediction (NWP) Models. The scheme is called Spectral Ice Habit Prediction System (SHIPS), and the goal is to retain growth history of ice particles in the Eulerian dynamics framework. It diagnoses characteristics of ice particles based on a series of particle property variables (PPVs) that reflect history of microphysieal processes and the transport between mass bins and air parcels in space. Therefore, categorization of ice particles typically used in bulk microphysical parameterization and traditional bin models is not necessary, so that errors that stem from the categorization can be avoided. SHIPS predicts polycrystals as well as hexagonal monocrystals based on empirically derived habit frequency and growth rate, and simulates the habit-dependent aggregation and riming processes by use of the stochastic collection equation with predicted PPVs. Idealized two dimensional simulations were performed with SHIPS in a NWP model. The predicted spatial distribution of ice particle habits and types, and evolution of particle size distributions showed good quantitative agreement with observation This comprehensive model of ice particle properties, distributions, and evolution in clouds can be used to better understand problems facing wide range of research disciplines, including microphysics processes, radiative transfer in a cloudy atmosphere, data assimilation, and weather modification.

  1. AUTOMATED GEOSPATIAL WATERSHED ASSESSMENT: A GIS-BASED HYDROLOGIC MODELING TOOL

    Science.gov (United States)

    Planning and assessment in land and water resource management are evolving toward complex, spatially explicit regional assessments. These problems have to be addressed with distributed models that can compute runoff and erosion at different spatial and temporal scales. The extens...

  2. Comparison of two different modelling tools

    DEFF Research Database (Denmark)

    Brix, Wiebke; Elmegaard, Brian

    2009-01-01

    In this paper a test case is solved using two different modelling tools, Engineering Equation Solver (EES) and WinDali, in order to compare the tools. The system of equations solved, is a static model of an evaporator used for refrigeration. The evaporator consists of two parallel channels......, and it is investigated how a non-uniform airflow influences the refrigerant mass flow rate distribution and the total cooling capacity of the heat exchanger. It is shown that the cooling capacity decreases significantly with increasing maldistribution of the airflow. Comparing the two simulation tools it is found...

  3. Modeling fuels and fire effects in 3D: Model description and applications

    Science.gov (United States)

    Francois Pimont; Russell Parsons; Eric Rigolot; Francois de Coligny; Jean-Luc Dupuy; Philippe Dreyfus; Rodman R. Linn

    2016-01-01

    Scientists and managers critically need ways to assess how fuel treatments alter fire behavior, yet few tools currently exist for this purpose.We present a spatially-explicit-fuel-modeling system, FuelManager, which models fuels, vegetation growth, fire behavior (using a physics-based model, FIRETEC), and fire effects. FuelManager's flexible approach facilitates...

  4. Spatially explicit modeling of conflict zones between wildlife and snow sports: prioritizing areas for winter refuges.

    Science.gov (United States)

    Braunisch, Veronika; Patthey, Patrick; Arlettaz, Raphaël

    2011-04-01

    Outdoor winter recreation exerts an increasing pressure upon mountain ecosystems, with unpredictable, free-ranging activities (e.g., ski mountaineering, snowboarding, and snowshoeing) representing a major source of stress for wildlife. Mitigating anthropogenic disturbance requires the spatially explicit prediction of the interference between the activities of humans and wildlife. We applied spatial modeling to localize conflict zones between wintering Black Grouse (Tetrao tetrix), a declining species of Alpine timberline ecosystems, and two free-ranging winter sports (off-piste skiing [including snow-boarding] and snowshoeing). Track data (snow-sports and birds' traces) obtained from aerial photographs taken over a 585-km transect running along the timberline, implemented within a maximum entropy model, were used to predict the occurrence of snow sports and Black Grouse as a function of landscape characteristics. By modeling Black Grouse presence in the theoretical absence of free-ranging activities and ski infrastructure, we first estimated the amount of habitat reduction caused by these two factors. The models were then extrapolated to the altitudinal range occupied by Black Grouse, while the spatial extent and intensity of potential conflict were assessed by calculating the probability of human-wildlife co-occurrence. The two snow-sports showed different distribution patterns. Skiers' occurrence was mainly determined by ski-lift presence and a smooth terrain, while snowshoers' occurrence was linked to hiking or skiing routes and moderate slopes. Wintering Black Grouse avoided ski lifts and areas frequented by free-ranging snow sports. According to the models, Black Grouse have faced a substantial reduction of suitable wintering habitat along the timberline transect: 12% due to ski infrastructure and another 16% when adding free-ranging activities. Extrapolating the models over the whole study area results in an overall habitat loss due to ski infrastructure of

  5. Modeling of fatigue crack induced nonlinear ultrasonics using a highly parallelized explicit local interaction simulation approach

    Science.gov (United States)

    Shen, Yanfeng; Cesnik, Carlos E. S.

    2016-04-01

    This paper presents a parallelized modeling technique for the efficient simulation of nonlinear ultrasonics introduced by the wave interaction with fatigue cracks. The elastodynamic wave equations with contact effects are formulated using an explicit Local Interaction Simulation Approach (LISA). The LISA formulation is extended to capture the contact-impact phenomena during the wave damage interaction based on the penalty method. A Coulomb friction model is integrated into the computation procedure to capture the stick-slip contact shear motion. The LISA procedure is coded using the Compute Unified Device Architecture (CUDA), which enables the highly parallelized supercomputing on powerful graphic cards. Both the explicit contact formulation and the parallel feature facilitates LISA's superb computational efficiency over the conventional finite element method (FEM). The theoretical formulations based on the penalty method is introduced and a guideline for the proper choice of the contact stiffness is given. The convergence behavior of the solution under various contact stiffness values is examined. A numerical benchmark problem is used to investigate the new LISA formulation and results are compared with a conventional contact finite element solution. Various nonlinear ultrasonic phenomena are successfully captured using this contact LISA formulation, including the generation of nonlinear higher harmonic responses. Nonlinear mode conversion of guided waves at fatigue cracks is also studied.

  6. Open Tools for Integrated Modelling to Understand SDG development - The OPTIMUS program

    Science.gov (United States)

    Howells, Mark; Zepeda, Eduardo; Rogner, H. Holger; Sanchez, Marco; Roehrl, Alexander; Cicowiez, Matrin; Mentis, Dimitris; Korkevelos, Alexandros; Taliotis, Constantinos; Broad, Oliver; Alfstad, Thomas

    2016-04-01

    The recently adopted Sustainable Development Goals (SDGs) - a set of 17 measurable and time-bound goals with 169 associated targets for 2030 - are highly inclusive challenges before the world community ranging from eliminating poverty to human rights, inequality, a secure world and protection of the environment. Each individual goal or target by themselves present enormous tasks, taken together they are overwhelming. There strong and weak interlinkages, hence trade-offs and complementarities among goals and targets. Some targets may affect several goals while other goals and targets may conflict or be mutually exclusive (Ref). Meeting each of these requires the judicious exploitation of resource, with energy playing an important role. Such complexity demands to be addressed in an integrated way using systems analysis tools to support informed policy formulation, planning, allocation of scarce resources, monitoring progress, effectiveness and review at different scales. There is no one size fits all methodology that conceivably could include all goal and targets simultaneously. But there are methodologies encapsulating critical subsets of the goal and targets with strong interlinkages with a 'soft' reflection on the weak interlinkages. Universal food security or sustainable energy for all inherently support goals and targets on human rights and equality but possibly at the cost of biodiversity or desertification. Integrated analysis and planning tools are not yet commonplace at national universities - or indeed in many policy making organs. What is needed is a fundamental realignment of institutions and integrations of their planning processes and decision making. We introduce a series of open source tools to support the SDG planning and implementation process. The Global User-friendly CLEW Open Source (GLUCOSE) tool optimizes resource interactions and constraints; The Global Electrification Tool kit (GETit) provides the first global spatially explicit

  7. Predictive Validity of Explicit and Implicit Threat Overestimation in Contamination Fear

    Science.gov (United States)

    Green, Jennifer S.; Teachman, Bethany A.

    2012-01-01

    We examined the predictive validity of explicit and implicit measures of threat overestimation in relation to contamination-fear outcomes using structural equation modeling. Undergraduate students high in contamination fear (N = 56) completed explicit measures of contamination threat likelihood and severity, as well as looming vulnerability cognitions, in addition to an implicit measure of danger associations with potential contaminants. Participants also completed measures of contamination-fear symptoms, as well as subjective distress and avoidance during a behavioral avoidance task, and state looming vulnerability cognitions during an exposure task. The latent explicit (but not implicit) threat overestimation variable was a significant and unique predictor of contamination fear symptoms and self-reported affective and cognitive facets of contamination fear. On the contrary, the implicit (but not explicit) latent measure predicted behavioral avoidance (at the level of a trend). Results are discussed in terms of differential predictive validity of implicit versus explicit markers of threat processing and multiple fear response systems. PMID:24073390

  8. A neurocomputational theory of how explicit learning bootstraps early procedural learning.

    Science.gov (United States)

    Paul, Erick J; Ashby, F Gregory

    2013-01-01

    It is widely accepted that human learning and memory is mediated by multiple memory systems that are each best suited to different requirements and demands. Within the domain of categorization, at least two systems are thought to facilitate learning: an explicit (declarative) system depending largely on the prefrontal cortex, and a procedural (non-declarative) system depending on the basal ganglia. Substantial evidence suggests that each system is optimally suited to learn particular categorization tasks. However, it remains unknown precisely how these systems interact to produce optimal learning and behavior. In order to investigate this issue, the present research evaluated the progression of learning through simulation of categorization tasks using COVIS, a well-known model of human category learning that includes both explicit and procedural learning systems. Specifically, the model's parameter space was thoroughly explored in procedurally learned categorization tasks across a variety of conditions and architectures to identify plausible interaction architectures. The simulation results support the hypothesis that one-way interaction between the systems occurs such that the explicit system "bootstraps" learning early on in the procedural system. Thus, the procedural system initially learns a suboptimal strategy employed by the explicit system and later refines its strategy. This bootstrapping could be from cortical-striatal projections that originate in premotor or motor regions of cortex, or possibly by the explicit system's control of motor responses through basal ganglia-mediated loops.

  9. Higgs pair production in the MSSM with explicit CP violation

    International Nuclear Information System (INIS)

    Demir, D.A.

    1999-07-01

    In the minimal supersymmetric standard model with explicit CP violation, associated production of the lightest Higgs boson with heavier ones is analyzed. Due to explicit CP violation, the Higgs bosons are no longer CP eigenstates so that both of the heavy Higgs bosons contribute to the process. While the radiative corrections in the Higgs sector turn out to be quite important, the vertex radiative corrections remain small as in the CP conserving theory. (author)

  10. Transparent Model Transformation: Turning Your Favourite Model Editor into a Transformation Tool

    DEFF Research Database (Denmark)

    Acretoaie, Vlad; Störrle, Harald; Strüber, Daniel

    2015-01-01

    Current model transformation languages are supported by dedicated editors, often closely coupled to a single execution engine. We introduce Transparent Model Transformation, a paradigm enabling modelers to specify transformations using a familiar tool: their model editor. We also present VMTL, th...... model transformation tool sharing the model editor’s benefits, transparently....

  11. Finite Element Simulation of Sheet Metal Forming Process Using Local Interpolation for Tool Surfaces

    International Nuclear Information System (INIS)

    Hama, Takayuki; Takuda, Hirohiko; Takamura, Masato; Makinouchi, Akitake; Teodosiu, Cristian

    2005-01-01

    Treatment of contact between a sheet and tools is one of the most difficult problems to deal with in finite-element simulations of sheet forming processes. In order to obtain more accurate tool models without increasing the number of elements, this paper describes a new formulation for contact problems using interpolation proposed by Nagata for tool surfaces. A contact search algorithm between sheet nodes and the interpolated tool surfaces was developed and was introduced into the static-explicit elastoplastic finite-element method code STAMP3D. Simulations of a square cup deep drawing process with a very coarsely discretized punch model were carried out. The simulated results showed that the proposed algorithm gave the proper drawn shape, demonstrating the validity of the proposed algorithm

  12. Pre-Processing and Modeling Tools for Bigdata

    Directory of Open Access Journals (Sweden)

    Hashem Hadi

    2016-09-01

    Full Text Available Modeling tools and operators help the user / developer to identify the processing field on the top of the sequence and to send into the computing module only the data related to the requested result. The remaining data is not relevant and it will slow down the processing. The biggest challenge nowadays is to get high quality processing results with a reduced computing time and costs. To do so, we must review the processing sequence, by adding several modeling tools. The existing processing models do not take in consideration this aspect and focus on getting high calculation performances which will increase the computing time and costs. In this paper we provide a study of the main modeling tools for BigData and a new model based on pre-processing.

  13. BETR-World: a geographically explicit model of chemical fate: application to transport of α-HCH to the Arctic

    International Nuclear Information System (INIS)

    Toose, L.; Woodfine, D.G.; MacLeod, M.; Mackay, D.; Gouin, J.

    2004-01-01

    The Berkeley-Trent (BETR)-World model, a 25 compartment, geographically explicit fugacity-based model is described and applied to evaluate the transport of chemicals from temperate source regions to receptor regions (such as the Arctic). The model was parameterized using GIS and an array of digital data on weather, oceans, freshwater, vegetation and geo-political boundaries. This version of the BETR model framework includes modification of atmospheric degradation rates by seasonally variable hydroxyl radical concentrations and temperature. Degradation rates in all other compartments vary with seasonally changing temperature. Deposition to the deep ocean has been included as a loss mechanism. A case study was undertaken for α-HCH. Dynamic emission scenarios were estimated for each of the 25 regions. Predicted environmental concentrations showed good agreement with measured values for the northern regions in air, and fresh and oceanic water and with the results from a previous model of global chemical fate. Potential for long-range transport and deposition to the Arctic region was assessed using a Transfer Efficiency combined with estimated emissions. European regions and the Orient including China have a high potential to contribute α-HCH contamination in the Arctic due to high rates of emission in these regions despite low Transfer Efficiencies. Sensitivity analyses reveal that the performance and reliability of the model is strongly influenced by parameters controlling degradation rates. - A geographically explicit multi-compartment model is applied to the transport of α-HCH to the Arctic, showing Europe and the Orient are key sources

  14. Hybrid design tools for conceptual design and design engineering processes: bridging the design gap: towards an intuitive design tool

    NARCIS (Netherlands)

    Wendrich, Robert E.

    2016-01-01

    Hybrid Design Tools; Representation; Computational Synthesis. Non-linear, non-explicit, non-standard thinking and ambiguity in design tools has a great impact on enhancement of creativity during ideation and conceptualization. Tacit-tangible representation based on a mere idiosyncratic and

  15. Predicting drought propagation within peat layers using a three dimensionally explicit voxel based model

    Science.gov (United States)

    Condro, A. A.; Pawitan, H.; Risdiyanto, I.

    2018-05-01

    Peatlands are very vulnerable to widespread fires during dry seasons, due to availability of aboveground fuel biomass on the surface and belowground fuel biomass on the sub-surface. Hence, understanding drought propagation occurring within peat layers is crucial with regards to disaster mitigation activities on peatlands. Using a three dimensionally explicit voxel-based model of peatland hydrology, this study predicted drought propagation time lags into sub-surface peat layers after drought events occurrence on the surface of about 1 month during La-Nina and 2.5 months during El-Nino. The study was carried out on a high-conservation-value area of oil palm plantation in West Kalimantan. Validity of the model was evaluated and its applicability for disaster mitigation was discussed. The animations of simulated voxels are available at: goo.gl/HDRMYN (El-Nino 2015 episode) and goo.gl/g1sXPl (La-Nina 2016 episode). The model is available at: goo.gl/RiuMQz.

  16. An unified framework to integrate biotic, abiotic processes and human activities in spatially explicit models of agricultural landscapes

    Directory of Open Access Journals (Sweden)

    Fabrice eVinatier

    2016-02-01

    Full Text Available Recent concern over possible ways to sustain ecosystem services has triggered important research worldwide on ecosystem processes at the landscape scale. Understanding this complexity of landscape functioning calls for coupled and spatially-explicit modelling approaches. However, disciplinary boundaries have limited the number of multi-process studies at the landscape scale, and current progress in coupling processes at this scale often reveals strong imbalance between biotic and abiotic processes, depending on the core discipline of the modellers. We propose a spatially-explicit, unified conceptual framework that allows researchers from different fields to develop a shared view of agricultural landscapes. In particular,we distinguish landscape elements that are mobile in space and represent biotic or abiotic objects (for example water, fauna or flora populations, and elements that are immobile and represent fixed landscape elements with a given geometry (for example ditch section or plot. The shared representation of these elements allows setting common objects and spatio-temporal process boundaries that may otherwise differ between disciplines. We present guidelines and an assessment of the applicability of this framework to a virtual landscape system with realistic properties. This framework allows the complex system to be represented with a limited set of concepts but leaves the possibility to include current modelling strategies specific to biotic or abiotic disciplines. Future operational challenges include model design, space and time discretization, and the availability of both landscape modelling platforms and data.

  17. Inferring the past and present connectivity across the range of a North American leaf beetle: combining ecological niche modeling and a geographically explicit model of coalescence.

    Science.gov (United States)

    Dellicour, Simon; Fearnley, Shannon; Lombal, Anicée; Heidl, Sarah; Dahlhoff, Elizabeth P; Rank, Nathan E; Mardulyn, Patrick

    2014-08-01

    The leaf beetle Chrysomela aeneicollis occurs across Western North America, either at high elevation or in small, isolated populations along the coast, and thus has a highly fragmented distribution. DNA sequence data (three loci) were collected from five regions across the species range. Population connectivity was examined using traditional ecological niche modeling, which suggested that gene flow could occur among regions now and in the past. We developed geographically explicit coalescence models of sequence evolution that incorporated a two-dimensional representation of the hypothesized ranges suggested by the niche-modeling estimates. We simulated sequence data according to these models and compared them to observed sequences to identify most probable scenarios regarding the migration history of C. aeneicollis. Our results disagreed with initial niche-modeling estimates by clearly rejecting recent connectivity among regions, and were instead most consistent with a long period of range fragmentation, extending well beyond the last glacial maximum. This application of geographically explicit models of coalescence has highlighted some limitations of the use of climatic variables for predicting the present and past range of a species and has explained aspects of the Pleistocene evolutionary history of a cold-adapted organism in Western North America. © 2014 The Author(s). Evolution © 2014 The Society for the Study of Evolution.

  18. Interactivity and Explicit Memory Formation of Consumer Undergraduate Male Students on Internet Environment

    Directory of Open Access Journals (Sweden)

    George Bedinelli Rossi

    2016-01-01

    Full Text Available ABSTRACT This research aims to integrate the theories of Explicit Memory and Interactivity, contributing to the theoretical development of both. We investigated whether the interactivity precedes the explicit consumer memory. Data collection was carried on by sending online questionnaire to 876 undergraduate male students, with a return of 453 valid questionnaires. Data were analyzed using Structural Equation Modeling of the constructs Explicit Memory and Interactivity. The analyzes indicate that interactivity increases explicit consumer memory, filling a theoretical gap of this concept about its effects. Moreover, it is a concept related to the future, not only to the past and to present, as shown by the classical definitions. As for explicit memory, its formation results from the individual's interactions with the environment, which was not explained by classical theories. The results indicated that interactivity and explicit memory are almost independent of each other, having low correlation or almost nil.

  19. An Explicit Structural Model of Root Hair and Soil Interactions Parameterised by Synchrotron X-ray Computed Tomography.

    Science.gov (United States)

    Keyes, Samuel David; Zygalakis, Konstantinos C; Roose, Tiina

    2017-12-01

    The rhizosphere is a zone of fundamental importance for understanding the dynamics of nutrient acquisition by plant roots. The canonical difficulty of experimentally investigating the rhizosphere led long ago to the adoption of mathematical models, the most sophisticated of which now incorporate explicit representations of root hairs and rhizosphere soil. Mathematical upscaling regimes, such as homogenisation, offer the possibility of incorporating into larger-scale models the important mechanistic processes occurring at the rhizosphere scale. However, we lack concrete descriptions of all the features required to fully parameterise models at the rhizosphere scale. By combining synchrotron X-ray computed tomography (SRXCT) and a novel root growth assay, we derive a three-dimensional description of rhizosphere soil structure suitable for use in multi-scale modelling frameworks. We describe an approach to mitigate sub-optimal root hair detection via structural root hair growth modelling. The growth model is explicitly parameterised with SRXCT data and simulates three-dimensional root hair ideotypes in silico, which are suitable for both ideotypic analysis and parameterisation of 3D geometry in mathematical models. The study considers different hypothetical conditions governing root hair interactions with soil matrices, with their respective effects on hair morphology being compared between idealised and image-derived soil/root geometries. The studies in idealised geometries suggest that packing arrangement of soil affects hair tortuosity more than the particle diameter. Results in field-derived soil suggest that hair access to poorly mobile nutrients is particularly sensitive to the physical interaction between the growing hairs and the phase of the soil in which soil water is present (i.e. the hydrated textural phase). The general trends in fluid-coincident hair length with distance from the root, and their dependence on hair/soil interaction mechanisms, are

  20. The explicit treatment of model uncertainties in the presence of aleatory and epistemic parameter uncertainties in risk and reliability analysis

    International Nuclear Information System (INIS)

    Ahn, Kwang Il; Yang, Joon Eon

    2003-01-01

    In the risk and reliability analysis of complex technological systems, the primary concern of formal uncertainty analysis is to understand why uncertainties arise, and to evaluate how they impact the results of the analysis. In recent times, many of the uncertainty analyses have focused on parameters of the risk and reliability analysis models, whose values are uncertain in an aleatory or an epistemic way. As the field of parametric uncertainty analysis matures, however, more attention is being paid to the explicit treatment of uncertainties that are addressed in the predictive model itself as well as the accuracy of the predictive model. The essential steps for evaluating impacts of these model uncertainties in the presence of parameter uncertainties are to determine rigorously various sources of uncertainties to be addressed in an underlying model itself and in turn model parameters, based on our state-of-knowledge and relevant evidence. Answering clearly the question of how to characterize and treat explicitly the forgoing different sources of uncertainty is particularly important for practical aspects such as risk and reliability optimization of systems as well as more transparent risk information and decision-making under various uncertainties. The main purpose of this paper is to provide practical guidance for quantitatively treating various model uncertainties that would often be encountered in the risk and reliability modeling process of complex technological systems

  1. The Importance of Representing Certain Key Vegetation Canopy Processes Explicitly in a Land Surface Model

    Science.gov (United States)

    Napoly, A.; Boone, A. A.; Martin, E.; Samuelsson, P.

    2015-12-01

    Land surface models are moving to more detailed vegetation canopy descriptions in order to better represent certain key processes, such as Carbon dynamics and snowpack evolution. Since such models are usually applied within coupled numerical weather prediction or spatially distributed hydrological models, these improvements must strike a balance between computational cost and complexity. The consequences of simplified or composite canopy approaches can be manifested in terms of increased errors with respect to soil temperatures, estimates of the diurnal cycle of the turbulent fluxes or snow canopy interception and melt. Vegetated areas and particularly forests are modeled in a quite simplified manner in the ISBA land surface model. However, continuous developments of surface processes now require a more accurate description of the canopy. A new version of the the model now includes a multi energy balance (MEB) option to explicitly represent the canopy and the forest floor. It will be shown that certain newly included processes such as the shading effect of the vegetation, the explicit heat capacity of the canopy, and the insulating effect of the forest floor turn out to be essential. A detailed study has been done for four French forested sites. It was found that the MEB option significantly improves the ground heat flux (RMSE decrease from 50W/m2 to 10W/m2 on average) and soil temperatures when compared against measurements. Also the sensible heat flux calculation was improved primarily owing to a better phasing with the solar insulation owing to a lower vegetation heat capacity. However, the total latent heat flux is less modified compared to the classical ISBA simulation since it is more related to water uptake and the formulation of the stomatal resistance (which are unchanged). Next, a benchmark over 40 Fluxnet sites (116 cumulated years) was performed and compared with results from the default composite soil-vegetation version of ISBA. The results show

  2. A Neurocomputational Theory of how Explicit Learning Bootstraps Early Procedural Learning

    Directory of Open Access Journals (Sweden)

    Erick Joseph Paul

    2013-12-01

    Full Text Available It is widely accepted that human learning and memory is mediated by multiple memory systems that are each best suited to different requirements and demands. Within the domain of categorization, at least two systems are thought to facilitate learning: an explicit (declarative system depending largely on the prefrontal cortex, and a procedural (non-declarative system depending on the basal ganglia. Substantial evidence suggests that each system is optimally suited to learn particular categorization tasks. However, it remains unknown precisely how these systems interact to produce optimal learning and behavior. In order to investigate this issue, the present research evaluated the progression of learning through simulation of categorization tasks using COVIS, a well-known model of human category learning that includes both explicit and procedural learning systems. Specifically, the model's parameter space was thoroughly explored in procedurally learned categorization tasks across a variety of conditions and architectures to identify plausible interaction architectures. The simulation results support the hypothesis that one-way interaction between the systems occurs such that the explicit system "bootstraps" learning early on in the procedural system. Thus, the procedural system initially learns a suboptimal strategy employed by the explicit system and later refines its strategy. This bootstrapping could be from cortical-striatal projections that originate in premotor or motor regions of cortex, or possibly by the explicit system’s control of motor responses through basal ganglia-mediated loops.

  3. Explicit Solution of Reinsurance-Investment Problem for an Insurer with Dynamic Income under Vasicek Model

    Directory of Open Access Journals (Sweden)

    De-Lei Sheng

    2016-01-01

    Full Text Available Unlike traditionally used reserves models, this paper focuses on a reserve process with dynamic income to study the reinsurance-investment problem for an insurer under Vasicek stochastic interest rate model. The insurer’s dynamic income is given by the remainder after a dynamic reward budget being subtracted from the insurer’s net premium which is calculated according to expected premium principle. Applying stochastic control technique, a Hamilton-Jacobi-Bellman equation is established and the explicit solution is obtained under the objective of maximizing the insurer’s power utility of terminal wealth. Some economic interpretations of the obtained results are explained in detail. In addition, numerical analysis and several graphics are given to illustrate our results more meticulous.

  4. Explicit learning in Act-R

    NARCIS (Netherlands)

    Taatgen, N.A.; Schmid, U; Krems, J; Wysotzky, F

    1999-01-01

    A popular distinction in the learning literature is the distinction between implicit and explicit learning. Although many studies elaborate on the nature of implicit learning, little attention is left for explicit learning. The unintentional aspect of implicit learning corresponds well to the

  5. Implicit, explicit and speculative knowledge

    NARCIS (Netherlands)

    van Ditmarsch, H.; French, T.; Velázquez-Quesada, F.R.; Wáng, Y.N.

    We compare different epistemic notions in the presence of awareness of propositional variables: the logic of implicit knowledge (in which explicit knowledge is definable), the logic of explicit knowledge, and the logic of speculative knowledge. Speculative knowledge is a novel epistemic notion that

  6. Explicit Interaction

    DEFF Research Database (Denmark)

    Löwgren, Jonas; Eriksen, Mette Agger; Linde, Per

    2006-01-01

    We report an ongoing study of palpable computing to support surgical rehabilitation, in the general field of interaction design for ubiquitous computing. Through explorative design, fieldwork and participatory design techniques, we explore the design principle of explicit interaction as an interp...

  7. The Cryosphere Model Comparison Tool (CmCt): Ice Sheet Model Validation and Comparison Tool for Greenland and Antarctica

    Science.gov (United States)

    Simon, E.; Nowicki, S.; Neumann, T.; Tyahla, L.; Saba, J. L.; Guerber, J. R.; Bonin, J. A.; DiMarzio, J. P.

    2017-12-01

    The Cryosphere model Comparison tool (CmCt) is a web based ice sheet model validation tool that is being developed by NASA to facilitate direct comparison between observational data and various ice sheet models. The CmCt allows the user to take advantage of several decades worth of observations from Greenland and Antarctica. Currently, the CmCt can be used to compare ice sheet models provided by the user with remotely sensed satellite data from ICESat (Ice, Cloud, and land Elevation Satellite) laser altimetry, GRACE (Gravity Recovery and Climate Experiment) satellite, and radar altimetry (ERS-1, ERS-2, and Envisat). One or more models can be uploaded through the CmCt website and compared with observational data, or compared to each other or other models. The CmCt calculates statistics on the differences between the model and observations, and other quantitative and qualitative metrics, which can be used to evaluate the different model simulations against the observations. The qualitative metrics consist of a range of visual outputs and the quantitative metrics consist of several whole-ice-sheet scalar values that can be used to assign an overall score to a particular simulation. The comparison results from CmCt are useful in quantifying improvements within a specific model (or within a class of models) as a result of differences in model dynamics (e.g., shallow vs. higher-order dynamics approximations), model physics (e.g., representations of ice sheet rheological or basal processes), or model resolution (mesh resolution and/or changes in the spatial resolution of input datasets). The framework and metrics could also be used for use as a model-to-model intercomparison tool, simply by swapping outputs from another model as the observational datasets. Future versions of the tool will include comparisons with other datasets that are of interest to the modeling community, such as ice velocity, ice thickness, and surface mass balance.

  8. Charged patchy particle models in explicit salt: Ion distributions, electrostatic potentials, and effective interactions.

    Science.gov (United States)

    Yigit, Cemil; Heyda, Jan; Dzubiella, Joachim

    2015-08-14

    We introduce a set of charged patchy particle models (CPPMs) in order to systematically study the influence of electrostatic charge patchiness and multipolarity on macromolecular interactions by means of implicit-solvent, explicit-ion Langevin dynamics simulations employing the Gromacs software. We consider well-defined zero-, one-, and two-patched spherical globules each of the same net charge and (nanometer) size which are composed of discrete atoms. The studied mono- and multipole moments of the CPPMs are comparable to those of globular proteins with similar size. We first characterize ion distributions and electrostatic potentials around a single CPPM. Although angle-resolved radial distribution functions reveal the expected local accumulation and depletion of counter- and co-ions around the patches, respectively, the orientation-averaged electrostatic potential shows only a small variation among the various CPPMs due to space charge cancellations. Furthermore, we study the orientation-averaged potential of mean force (PMF), the number of accumulated ions on the patches, as well as the CPPM orientations along the center-to-center distance of a pair of CPPMs. We compare the PMFs to the classical Derjaguin-Verwey-Landau-Overbeek theory and previously introduced orientation-averaged Debye-Hückel pair potentials including dipolar interactions. Our simulations confirm the adequacy of the theories in their respective regimes of validity, while low salt concentrations and large multipolar interactions remain a challenge for tractable theoretical descriptions.

  9. Pre-service teachers' implicit and explicit attitudes toward obesity influence their judgments of students

    NARCIS (Netherlands)

    Glock, S.; Oude Groote Beverborg, A.; Müller, B.C.N.

    2016-01-01

    Obese children experience disadvantages in school and discrimination from their teachers. Teachers' implicit and explicit attitudes have been identified as contributing to these disadvantages. Drawing on dual process models, we investigated the nature of pre-service teachers' implicit and explicit

  10. Pre-service teachers’ implicit and explicit attitudes toward obesity influence their judgments of students

    NARCIS (Netherlands)

    Glock, Sabine; Oude Groote Beverborg, Arnoud; Müller, Barbara C.N.

    2016-01-01

    Obese children experience disadvantages in school and discrimination from their teachers. Teachers’ implicit and explicit attitudes have been identified as contributing to these disadvantages. Drawing on dual process models, we investigated the nature of pre-service teachers’ implicit and explicit

  11. Modelling thermomechanical conditions at the tool/matrix interface in Friction Stir Welding

    DEFF Research Database (Denmark)

    Schmidt, Henrik Nikolaj Blich; Hattel, Jesper

    2004-01-01

    is obtained. A fully coupled thermo-mechanical 3D FE model has been developed in ABAQUS/Explicit using the ALE formulation and the Johnson-Cook material law. The contact forces are modelled by Coulomb’s law of friction making the contact condition highly solution dependent. The heat is generated by both...

  12. Designing tools for oil exploration using nuclear modeling

    Science.gov (United States)

    Mauborgne, Marie-Laure; Allioli, Françoise; Manclossi, Mauro; Nicoletti, Luisa; Stoller, Chris; Evans, Mike

    2017-09-01

    When designing nuclear tools for oil exploration, one of the first steps is typically nuclear modeling for concept evaluation and initial characterization. Having an accurate model, including the availability of accurate cross sections, is essential to reduce or avoid time consuming and costly design iterations. During tool response characterization, modeling is benchmarked with experimental data and then used to complement and to expand the database to make it more detailed and inclusive of more measurement environments which are difficult or impossible to reproduce in the laboratory. We present comparisons of our modeling results obtained using the ENDF/B-VI and ENDF/B-VII cross section data bases, focusing on the response to a few elements found in the tool, borehole and subsurface formation. For neutron-induced inelastic and capture gamma ray spectroscopy, major obstacles may be caused by missing or inaccurate cross sections for essential materials. We show examples of the benchmarking of modeling results against experimental data obtained during tool characterization and discuss observed discrepancies.

  13. Explicit equilibria in a kinetic model of gambling

    Science.gov (United States)

    Bassetti, F.; Toscani, G.

    2010-06-01

    We introduce and discuss a nonlinear kinetic equation of Boltzmann type which describes the evolution of wealth in a pure gambling process, where the entire sum of wealths of two agents is up for gambling, and randomly shared between the agents. For this equation the analytical form of the steady states is found for various realizations of the random fraction of the sum which is shared to the agents. Among others, the exponential distribution appears as steady state in case of a uniformly distributed random fraction, while Gamma distribution appears for a random fraction which is Beta distributed. The case in which the gambling game is only conservative-in-the-mean is shown to lead to an explicit heavy tailed distribution.

  14. Three Dimensional Explicit Model for Cometary Tail Ions Interactions with Solar Wind

    Science.gov (United States)

    Al Bermani, M. J. F.; Alhamed, S. A.; Khalaf, S. Z.; Ali, H. Sh.; Selman, A. A.

    2009-06-01

    The different interactions between cometary tail and solar wind ions are studied in the present paper based on three-dimensional Lax explicit method. The model used in this research is based on the continuity equations describing the cometary tail-solar wind interactions. Three dimensional system was considered in this paper. Simulation of the physical system was achieved using computer code written using Matlab 7.0. The parameters studied here assumed Halley comet type and include the particle density rho, the particles velocity v, the magnetic field strength B, dynamic pressure p and internal energy E. The results of the present research showed that the interaction near the cometary nucleus is mainly affected by the new ions added to the plasma of the solar wind, which increases the average molecular weight and result in many unique characteristics of the cometary tail. These characteristics were explained in the presence of the IMF.

  15. Explicitly-correlated ring-coupled-cluster-doubles theory: Including exchange for computations on closed-shell systems

    Energy Technology Data Exchange (ETDEWEB)

    Hehn, Anna-Sophia; Holzer, Christof; Klopper, Wim, E-mail: klopper@kit.edu

    2016-11-10

    Highlights: • Ring-coupled-cluster-doubles approach now implemented with exchange terms. • Ring-coupled-cluster-doubles approach now implemented with F12 functions. • Szabo–Ostlund scheme (SO2) implemented for use in SAPT. • Fast convergence to the limit of a complete basis. • Implementation in the TURBOMOLE program system. - Abstract: Random-phase-approximation (RPA) methods have proven to be powerful tools in electronic-structure theory, being non-empirical, computationally efficient and broadly applicable to a variety of molecular systems including small-gap systems, transition-metal compounds and dispersion-dominated complexes. Applications are however hindered due to the slow basis-set convergence of the electron-correlation energy with the one-electron basis. As a remedy, we present approximate explicitly-correlated RPA approaches based on the ring-coupled-cluster-doubles formulation including exchange contributions. Test calculations demonstrate that the basis-set convergence of correlation energies is drastically accelerated through the explicitly-correlated approach, reaching 99% of the basis-set limit with triple-zeta basis sets. When implemented in close analogy to early work by Szabo and Ostlund [36], the new explicitly-correlated ring-coupled-cluster-doubles approach including exchange has the perspective to become a valuable tool in the framework of symmetry-adapted perturbation theory (SAPT) for the computation of dispersion energies of molecular complexes of weakly interacting closed-shell systems.

  16. Shape: A 3D Modeling Tool for Astrophysics.

    Science.gov (United States)

    Steffen, Wolfgang; Koning, Nicholas; Wenger, Stephan; Morisset, Christophe; Magnor, Marcus

    2011-04-01

    We present a flexible interactive 3D morpho-kinematical modeling application for astrophysics. Compared to other systems, our application reduces the restrictions on the physical assumptions, data type, and amount that is required for a reconstruction of an object's morphology. It is one of the first publicly available tools to apply interactive graphics to astrophysical modeling. The tool allows astrophysicists to provide a priori knowledge about the object by interactively defining 3D structural elements. By direct comparison of model prediction with observational data, model parameters can then be automatically optimized to fit the observation. The tool has already been successfully used in a number of astrophysical research projects.

  17. Generalized Heteroskedasticity ACF for Moving Average Models in Explicit Forms

    Directory of Open Access Journals (Sweden)

    Samir Khaled Safi

    2014-02-01

    Full Text Available Normal 0 false false false MicrosoftInternetExplorer4 The autocorrelation function (ACF measures the correlation between observations at different   distances apart. We derive explicit equations for generalized heteroskedasticity ACF for moving average of order q, MA(q. We consider two cases: Firstly: when the disturbance term follow the general covariance matrix structure Cov(wi, wj=S with si,j ¹ 0 " i¹j . Secondly: when the diagonal elements of S are not all identical but sij = 0 " i¹j, i.e. S=diag(s11, s22,…,stt. The forms of the explicit equations depend essentially on the moving average coefficients and covariance structure of the disturbance terms.   /* Style Definitions */ table.MsoNormalTable {mso-style-name:"جدول عادي"; mso-tstyle-rowband-size:0; mso-tstyle-colband-size:0; mso-style-noshow:yes; mso-style-parent:""; mso-padding-alt:0cm 5.4pt 0cm 5.4pt; mso-para-margin:0cm; mso-para-margin-bottom:.0001pt; mso-pagination:widow-orphan; font-size:10.0pt; font-family:"Times New Roman"; mso-ansi-language:#0400; mso-fareast-language:#0400; mso-bidi-language:#0400;}

  18. Explicit/multi-parametric model predictive control (MPC) of linear discrete-time systems by dynamic and multi-parametric programming

    KAUST Repository

    Kouramas, K.I.

    2011-08-01

    This work presents a new algorithm for solving the explicit/multi- parametric model predictive control (or mp-MPC) problem for linear, time-invariant discrete-time systems, based on dynamic programming and multi-parametric programming techniques. The algorithm features two key steps: (i) a dynamic programming step, in which the mp-MPC problem is decomposed into a set of smaller subproblems in which only the current control, state variables, and constraints are considered, and (ii) a multi-parametric programming step, in which each subproblem is solved as a convex multi-parametric programming problem, to derive the control variables as an explicit function of the states. The key feature of the proposed method is that it overcomes potential limitations of previous methods for solving multi-parametric programming problems with dynamic programming, such as the need for global optimization for each subproblem of the dynamic programming step. © 2011 Elsevier Ltd. All rights reserved.

  19. Modelling of Tool Wear and Residual Stress during Machining of AISI H13 Tool Steel

    Science.gov (United States)

    Outeiro, José C.; Umbrello, Domenico; Pina, José C.; Rizzuti, Stefania

    2007-05-01

    Residual stresses can enhance or impair the ability of a component to withstand loading conditions in service (fatigue, creep, stress corrosion cracking, etc.), depending on their nature: compressive or tensile, respectively. This poses enormous problems in structural assembly as this affects the structural integrity of the whole part. In addition, tool wear issues are of critical importance in manufacturing since these affect component quality, tool life and machining cost. Therefore, prediction and control of both tool wear and the residual stresses in machining are absolutely necessary. In this work, a two-dimensional Finite Element model using an implicit Lagrangian formulation with an automatic remeshing was applied to simulate the orthogonal cutting process of AISI H13 tool steel. To validate such model the predicted and experimentally measured chip geometry, cutting forces, temperatures, tool wear and residual stresses on the machined affected layers were compared. The proposed FE model allowed us to investigate the influence of tool geometry, cutting regime parameters and tool wear on residual stress distribution in the machined surface and subsurface of AISI H13 tool steel. The obtained results permit to conclude that in order to reduce the magnitude of surface residual stresses, the cutting speed should be increased, the uncut chip thickness (or feed) should be reduced and machining with honed tools having large cutting edge radii produce better results than chamfered tools. Moreover, increasing tool wear increases the magnitude of surface residual stresses.

  20. A spatially explicit model for an Allee effect: why wolves recolonize so slowly in Greater Yellowstone.

    Science.gov (United States)

    Hurford, Amy; Hebblewhite, Mark; Lewis, Mark A

    2006-11-01

    A reduced probability of finding mates at low densities is a frequently hypothesized mechanism for a component Allee effect. At low densities dispersers are less likely to find mates and establish new breeding units. However, many mathematical models for an Allee effect do not make a distinction between breeding group establishment and subsequent population growth. Our objective is to derive a spatially explicit mathematical model, where dispersers have a reduced probability of finding mates at low densities, and parameterize the model for wolf recolonization in the Greater Yellowstone Ecosystem (GYE). In this model, only the probability of establishing new breeding units is influenced by the reduced probability of finding mates at low densities. We analytically and numerically solve the model to determine the effect of a decreased probability in finding mates at low densities on population spread rate and density. Our results suggest that a reduced probability of finding mates at low densities may slow recolonization rate.

  1. Pre-Service Teachers' Implicit and Explicit Attitudes toward Obesity Influence Their Judgments of Students

    Science.gov (United States)

    Glock, Sabine; Beverborg, Arnoud Oude Groote; Müller, Barbara C. N.

    2016-01-01

    Obese children experience disadvantages in school and discrimination from their teachers. Teachers' implicit and explicit attitudes have been identified as contributing to these disadvantages. Drawing on dual process models, we investigated the nature of pre-service teachers' implicit and explicit attitudes, their motivation to respond without…

  2. Numerical study on two-point contact by an explicit integration finite element method : A contribution to the modeling of flange squeal

    NARCIS (Netherlands)

    Yang, Z.; Li, Z.; Dollevoet, R.P.B.J.; Tournay, H; Grassie, S

    2015-01-01

    The precise mechanism which activates squeal, especially flange squeal has not been fully explained. The complex non-Hertzian contact and the broad-band high frequency feature bring great challenges to the modelling work of flange squeal. In this paper, an explicit integration finite element method

  3. Designing tools for oil exploration using nuclear modeling

    Directory of Open Access Journals (Sweden)

    Mauborgne Marie-Laure

    2017-01-01

    Full Text Available When designing nuclear tools for oil exploration, one of the first steps is typically nuclear modeling for concept evaluation and initial characterization. Having an accurate model, including the availability of accurate cross sections, is essential to reduce or avoid time consuming and costly design iterations. During tool response characterization, modeling is benchmarked with experimental data and then used to complement and to expand the database to make it more detailed and inclusive of more measurement environments which are difficult or impossible to reproduce in the laboratory. We present comparisons of our modeling results obtained using the ENDF/B-VI and ENDF/B-VII cross section data bases, focusing on the response to a few elements found in the tool, borehole and subsurface formation. For neutron-induced inelastic and capture gamma ray spectroscopy, major obstacles may be caused by missing or inaccurate cross sections for essential materials. We show examples of the benchmarking of modeling results against experimental data obtained during tool characterization and discuss observed discrepancies.

  4. An implicit-explicit approach for atmospheric transport-chemistry problems

    NARCIS (Netherlands)

    J.G. Verwer (Jan); J.G. Blom (Joke); W. Hundsdorfer (Willem)

    1995-01-01

    textabstractWe investigate numerical algorithms for use in air pollution models. The emphasis lies on time integration aspects in connection with advection, vertical turbulent diffusion and stiff chemical transformations. The time integration scheme considered is a 2nd-order implicit-explicit BDF

  5. Human Mobility Patterns and Cholera Epidemics: a Spatially Explicit Modeling Approach

    Science.gov (United States)

    Mari, L.; Bertuzzo, E.; Righetto, L.; Casagrandi, R.; Gatto, M.; Rodriguez-Iturbe, I.; Rinaldo, A.

    2010-12-01

    Cholera is an acute enteric disease caused by the ingestion of water or food contaminated by the bacterium Vibrio cholerae. Although most infected individuals do not develop severe symptoms, their stool may contain huge quantities of V.~cholerae cells. Therefore, while traveling or commuting, asymptomatic carriers can be responsible for the long-range dissemination of the disease. As a consequence, human mobility is an alternative and efficient driver for the spread of cholera, whose primary propagation pathway is hydrological transport through river networks. We present a multi-layer network model that accounts for the interplay between epidemiological dynamics, hydrological transport and long-distance dissemination of V.~cholerae due to human movement. In particular, building on top of state-of-the-art spatially explicit models for cholera spread through surface waters, we describe human movement and its effects on the propagation of the disease by means of a gravity-model approach borrowed from transportation theory. Gravity-like contact processes have been widely used in epidemiology, because they can satisfactorily depict human movement when data on actual mobility patterns are not available. We test our model against epidemiological data recorded during the cholera outbreak occurred in the KwaZulu-Natal province of South Africa during years 2000--2001. We show that human mobility does actually play an important role in the formation of the spatiotemporal patterns of cholera epidemics. In particular, long-range human movement may determine inter-catchment dissemination of V.~cholerae cells, thus in turn explaining the emergence of epidemic patterns that cannot be produced by hydrological transport alone. We also show that particular attention has to be devoted to study how heterogeneously distributed drinking water supplies and sanitation conditions may affect cholera transmission.

  6. OISI dynamic end-to-end modeling tool

    Science.gov (United States)

    Kersten, Michael; Weidler, Alexander; Wilhelm, Rainer; Johann, Ulrich A.; Szerdahelyi, Laszlo

    2000-07-01

    The OISI Dynamic end-to-end modeling tool is tailored to end-to-end modeling and dynamic simulation of Earth- and space-based actively controlled optical instruments such as e.g. optical stellar interferometers. `End-to-end modeling' is meant to denote the feature that the overall model comprises besides optical sub-models also structural, sensor, actuator, controller and disturbance sub-models influencing the optical transmission, so that the system- level instrument performance due to disturbances and active optics can be simulated. This tool has been developed to support performance analysis and prediction as well as control loop design and fine-tuning for OISI, Germany's preparatory program for optical/infrared spaceborne interferometry initiated in 1994 by Dornier Satellitensysteme GmbH in Friedrichshafen.

  7. Hybrid Design Tools Intuit Interaction

    NARCIS (Netherlands)

    Wendrich, Robert E.; Kyvsgaard Hansen, P.; Rasmussen, J.; Jorgensen, K.A.; Tollestrup, C.

    2012-01-01

    Non-linear, non-explicit, non-standard thinking and ambiguity in design tools has a great impact on enhancement of creativity during ideation and conceptualization. Tacit-tangible representation based on a mere idiosyncratic and individual approach combined with computational assistance allows the

  8. Prospects of implicit measures in assessment of defensive function of explicit homophobia in heterosexuals

    Directory of Open Access Journals (Sweden)

    Lazarević Ljiljana B.

    2015-01-01

    Full Text Available Some authors state that one of the functions of explicit homophobia is defensive, and that it has roots in latent attraction towards same sex. Large body of evidence suggests that implicit techniques enable assessment of those cognitions hidden from conscious awareness. Sample of 277 heterosexuals completed several implicit (Implicit Association Test-IAT and Affective Priming-AP and explicit measures of attitude (Test of homophobia and Connotative differential. Multi-group SEM analysis was done to investigate cross-sample stability of the model postulating influence of IAT and AP factors on latent explicit factor. Analyses suggest that both in males and females IAT latent factor predicts negative explicit attitude. Results revealed that explicit homophobia is related to implicit, negative attitude toward homosexuals and not to implicit attraction towards same sex. [Projekat Ministarstva nauke Republike Srbije, br. 179018 i br. 179033

  9. Model Checking Markov Chains: Techniques and Tools

    NARCIS (Netherlands)

    Zapreev, I.S.

    2008-01-01

    This dissertation deals with four important aspects of model checking Markov chains: the development of efficient model-checking tools, the improvement of model-checking algorithms, the efficiency of the state-space reduction techniques, and the development of simulation-based model-checking

  10. Effects of an explicit problem-solving skills training program using a metacomponential approach for outpatients with acquired brain injury.

    Science.gov (United States)

    Fong, Kenneth N K; Howie, Dorothy R

    2009-01-01

    We investigated the effects of an explicit problem-solving skills training program using a metacomponential approach with 33 outpatients with moderate acquired brain injury, in the Hong Kong context. We compared an experimental training intervention with this explicit problem-solving approach, which taught metacomponential strategies, with a conventional cognitive training approach that did not have this explicit metacognitive training. We found significant advantages for the experimental group on the Metacomponential Interview measure in association with the explicit metacomponential training, but transfer to the real-life problem-solving measures was not evidenced in statistically significant findings. Small sample size, limited time of intervention, and some limitations with these tools may have been contributing factors to these results. The training program was demonstrated to have a significantly greater effect than the conventional training approach on metacomponential functioning and the component of problem representation. However, these benefits were not transferable to real-life situations.

  11. Performability Modelling Tools, Evaluation Techniques and Applications

    NARCIS (Netherlands)

    Haverkort, Boudewijn R.H.M.

    1990-01-01

    This thesis deals with three aspects of quantitative evaluation of fault-tolerant and distributed computer and communication systems: performability evaluation techniques, performability modelling tools, and performability modelling applications. Performability modelling is a relatively new

  12. Low Cloud Feedback to Surface Warming in the World's First Global Climate Model with Explicit Embedded Boundary Layer Turbulence

    Science.gov (United States)

    Parishani, H.; Pritchard, M. S.; Bretherton, C. S.; Wyant, M. C.; Khairoutdinov, M.; Singh, B.

    2017-12-01

    Biases and parameterization formulation uncertainties in the representation of boundary layer clouds remain a leading source of possible systematic error in climate projections. Here we show the first results of cloud feedback to +4K SST warming in a new experimental climate model, the ``Ultra-Parameterized (UP)'' Community Atmosphere Model, UPCAM. We have developed UPCAM as an unusually high-resolution implementation of cloud superparameterization (SP) in which a global set of cloud resolving arrays is embedded in a host global climate model. In UP, the cloud-resolving scale includes sufficient internal resolution to explicitly generate the turbulent eddies that form marine stratocumulus and trade cumulus clouds. This is computationally costly but complements other available approaches for studying low clouds and their climate interaction, by avoiding parameterization of the relevant scales. In a recent publication we have shown that UP, while not without its own complexity trade-offs, can produce encouraging improvements in low cloud climatology in multi-month simulations of the present climate and is a promising target for exascale computing (Parishani et al. 2017). Here we show results of its low cloud feedback to warming in multi-year simulations for the first time. References: Parishani, H., M. S. Pritchard, C. S. Bretherton, M. C. Wyant, and M. Khairoutdinov (2017), Toward low-cloud-permitting cloud superparameterization with explicit boundary layer turbulence, J. Adv. Model. Earth Syst., 9, doi:10.1002/2017MS000968.

  13. Analytical explicit formulas of average run length for long memory process with ARFIMA model on CUSUM control chart

    Directory of Open Access Journals (Sweden)

    Wilasinee Peerajit

    2017-12-01

    Full Text Available This paper proposes the explicit formulas for the derivation of exact formulas from Average Run Lengths (ARLs using integral equation on CUSUM control chart when observations are long memory processes with exponential white noise. The authors compared efficiency in terms of the percentage of absolute difference to a similar method to verify the accuracy of the ARLs between the values obtained by the explicit formulas and numerical integral equation (NIE method. The explicit formulas were based on Banach fixed point theorem which was used to guarantee the existence and uniqueness of the solution for ARFIMA(p,d,q. Results showed that the two methods are similar in good agreement with the percentage of absolute difference at less than 0.23%. Therefore, the explicit formulas are an efficient alternative for implementation in real applications because the computational CPU time for ARLs from the explicit formulas are 1 second preferable over the NIE method.

  14. Dynamic wind turbine models in power system simulation tool

    DEFF Research Database (Denmark)

    Hansen, A.; Jauch, Clemens; Soerensen, P.

    The present report describes the dynamic wind turbine models implemented in the power system simulation tool DIgSILENT. The developed models are a part of the results of a national research project, whose overall objective is to create a model database in different simulation tools. The report...

  15. Connecting free energy surfaces in implicit and explicit solvent: an efficient method to compute conformational and solvation free energies.

    Science.gov (United States)

    Deng, Nanjie; Zhang, Bin W; Levy, Ronald M

    2015-06-09

    The ability to accurately model solvent effects on free energy surfaces is important for understanding many biophysical processes including protein folding and misfolding, allosteric transitions, and protein–ligand binding. Although all-atom simulations in explicit solvent can provide an accurate model for biomolecules in solution, explicit solvent simulations are hampered by the slow equilibration on rugged landscapes containing multiple basins separated by barriers. In many cases, implicit solvent models can be used to significantly speed up the conformational sampling; however, implicit solvent simulations do not fully capture the effects of a molecular solvent, and this can lead to loss of accuracy in the estimated free energies. Here we introduce a new approach to compute free energy changes in which the molecular details of explicit solvent simulations are retained while also taking advantage of the speed of the implicit solvent simulations. In this approach, the slow equilibration in explicit solvent, due to the long waiting times before barrier crossing, is avoided by using a thermodynamic cycle which connects the free energy basins in implicit solvent and explicit solvent using a localized decoupling scheme. We test this method by computing conformational free energy differences and solvation free energies of the model system alanine dipeptide in water. The free energy changes between basins in explicit solvent calculated using fully explicit solvent paths agree with the corresponding free energy differences obtained using the implicit/explicit thermodynamic cycle to within 0.3 kcal/mol out of ∼3 kcal/mol at only ∼8% of the computational cost. We note that WHAM methods can be used to further improve the efficiency and accuracy of the implicit/explicit thermodynamic cycle.

  16. Light Higgs boson in THDM with explicit CP violation

    International Nuclear Information System (INIS)

    Akhmetzyanova, Eh.N.; Dolgopolov, M.V.; Smirnov, I.A.; Dubinin, M.N.

    2005-01-01

    The effective Lagrangian of the two-doublet Higgs sector with complex parameters is considered in the case of Minimal Supersymmetric Model with explicit CP violation. Light Higgs boson decay widths are calculated for the scenario with maximal mixing of CP even and CP odd states [ru

  17. Interaction of the model alkyltrimethylammonium ions with alkali halide salts: an explicit water molecular dynamics study

    Directory of Open Access Journals (Sweden)

    M. Druchok

    2013-01-01

    Full Text Available We present an explicit water molecular dynamics simulation of dilute solutions of model alkyltrimethylammonium surfactant ions (number of methylene groups in the tail is 3, 5, 8, 10, and 12 in mixture with NaF, NaCl, NaBr, and NaI salts, respectively. The SPC/E model is used to describe water molecules. Results of the simulation at 298 K are presented in form of the radial distribution functions between nitrogen and carbon atoms of CH2 groups on the alkyltrimethylammonium ion, and the counterion species in the solution. The running coordination numbers between carbon atoms of surfactants and counterions are also calculated. We show that I- counterion exhibits the highest, and F- the lowest affinity to "bind" to the model surfactants. The results are discussed in view of the available experimental and simulation data for this and similar solutions.

  18. The Explicit Wake Parametrisation V1.0: a wind farm parametrisation in the mesoscale model WRF

    Directory of Open Access Journals (Sweden)

    P. J. H. Volker

    2015-11-01

    Full Text Available We describe the theoretical basis, implementation, and validation of a new parametrisation that accounts for the effect of large offshore wind farms on the atmosphere and can be used in mesoscale and large-scale atmospheric models. This new parametrisation, referred to as the Explicit Wake Parametrisation (EWP, uses classical wake theory to describe the unresolved wake expansion. The EWP scheme is validated for a neutral atmospheric boundary layer against filtered in situ measurements from two meteorological masts situated a few kilometres away from the Danish offshore wind farm Horns Rev I. The simulated velocity deficit in the wake of the wind farm compares well to that observed in the measurements, and the velocity profile is qualitatively similar to that simulated with large eddy simulation models and from wind tunnel studies. At the same time, the validation process highlights the challenges in verifying such models with real observations.

  19. Immersive Data Comprehension: Visualizing Uncertainty in Measurable Models

    Directory of Open Access Journals (Sweden)

    Pere eBrunet

    2015-09-01

    Full Text Available Recent advances in 3D scanning technologies have opened new possibilities in a broad range of applications includingcultural heritage, medicine, civil engineering and urban planning. Virtual Reality systems can provide new tools toprofessionals that want to understand acquired 3D models. In this paper, we review the concept of data comprehension with an emphasis on visualization and inspection tools on immersive setups. We claim that in most application fields, data comprehension requires model measurements which in turn should be based on the explicit visualization of uncertainty. As 3D digital representations are not faithful, information on their fidelity at local level should be included in the model itself as uncertainty bounds. We propose the concept of Measurable 3D Models as digital models that explicitly encode local uncertainty bounds related to their quality. We claim that professionals and experts can strongly benefit from immersive interaction through new specific, fidelity-aware measurement tools which can facilitate 3D data comprehension. Since noise and processing errors are ubiquitous in acquired datasets, we discuss the estimation, representation and visualization of data uncertainty. We show that, based on typical user requirements in Cultural Heritage and other domains, application-oriented measuring tools in 3D models must consider uncertainty and local error bounds. We also discuss the requirements of immersive interaction tools for the comprehension of huge 3D and nD datasets acquired from real objects.

  20. GPGPU-based explicit finite element computations for applications in biomechanics: the performance of material models, element technologies, and hardware generations.

    Science.gov (United States)

    Strbac, V; Pierce, D M; Vander Sloten, J; Famaey, N

    2017-12-01

    Finite element (FE) simulations are increasingly valuable in assessing and improving the performance of biomedical devices and procedures. Due to high computational demands such simulations may become difficult or even infeasible, especially when considering nearly incompressible and anisotropic material models prevalent in analyses of soft tissues. Implementations of GPGPU-based explicit FEs predominantly cover isotropic materials, e.g. the neo-Hookean model. To elucidate the computational expense of anisotropic materials, we implement the Gasser-Ogden-Holzapfel dispersed, fiber-reinforced model and compare solution times against the neo-Hookean model. Implementations of GPGPU-based explicit FEs conventionally rely on single-point (under) integration. To elucidate the expense of full and selective-reduced integration (more reliable) we implement both and compare corresponding solution times against those generated using underintegration. To better understand the advancement of hardware, we compare results generated using representative Nvidia GPGPUs from three recent generations: Fermi (C2075), Kepler (K20c), and Maxwell (GTX980). We explore scaling by solving the same boundary value problem (an extension-inflation test on a segment of human aorta) with progressively larger FE meshes. Our results demonstrate substantial improvements in simulation speeds relative to two benchmark FE codes (up to 300[Formula: see text] while maintaining accuracy), and thus open many avenues to novel applications in biomechanics and medicine.

  1. Generalized Heteroskedasticity ACF for Moving Average Models in Explicit Forms

    OpenAIRE

    Samir Khaled Safi

    2014-01-01

    The autocorrelation function (ACF) measures the correlation between observations at different   distances apart. We derive explicit equations for generalized heteroskedasticity ACF for moving average of order q, MA(q). We consider two cases: Firstly: when the disturbance term follow the general covariance matrix structure Cov(wi, wj)=S with si,j ¹ 0 " i¹j . Secondly: when the diagonal elements of S are not all identical but sij = 0 " i¹j, i.e. S=diag(s11, s22,&hellip...

  2. A comparison of tools for modeling freshwater ecosystem services.

    Science.gov (United States)

    Vigerstol, Kari L; Aukema, Juliann E

    2011-10-01

    Interest in ecosystem services has grown tremendously among a wide range of sectors, including government agencies, NGO's and the business community. Ecosystem services entailing freshwater (e.g. flood control, the provision of hydropower, and water supply), as well as carbon storage and sequestration, have received the greatest attention in both scientific and on-the-ground applications. Given the newness of the field and the variety of tools for predicting water-based services, it is difficult to know which tools to use for different questions. There are two types of freshwater-related tools--traditional hydrologic tools and newer ecosystem services tools. Here we review two of the most prominent tools of each type and their possible applications. In particular, we compare the data requirements, ease of use, questions addressed, and interpretability of results among the models. We discuss the strengths, challenges and most appropriate applications of the different models. Traditional hydrological tools provide more detail whereas ecosystem services tools tend to be more accessible to non-experts and can provide a good general picture of these ecosystem services. We also suggest gaps in the modeling toolbox that would provide the greatest advances by improving existing tools. Copyright © 2011 Elsevier Ltd. All rights reserved.

  3. Spatially-explicit LCIA model for marine eutrophication as a tool for sustainability assessment

    DEFF Research Database (Denmark)

    Cosme, Nuno Miguel Dias; Hauschild, Michael Zwicky

    2014-01-01

    The increasing emissions from human activities are overrunning the ecosystems’ natural capacity to absorb them. Nutrient emissions, mostly nitrogen- and phosphorus-forms (N, P) from e.g. agricultural runoff and combustion processes, may lead to social-economic impacts and environmental quality......-enrichment to impacts on marine ecosystems. Emitted nitrogen reaches marine coastal waters where it promotes the growth of phytoplankton biomass in the surface photic zone from where it eventually sinks to bottom waters. This downward flux of organic matter is respired there by bacteria resulting in the consumption...... of dissolved oxygen. An excessive depletion of oxygen affects the exposed organisms and loss of species diversity may be expected. A model framework was built to estimate the potential impacts arising from N-emissions (see figure). It combines the fate of N in rivers and coastal waters, the exposure...

  4. Green Infrastructure Models and Tools

    Science.gov (United States)

    The objective of this project is to modify and refine existing models and develop new tools to support decision making for the complete green infrastructure (GI) project lifecycle, including the planning and implementation of stormwater control in urban and agricultural settings,...

  5. Implicit and explicit host effects on excitons in pentacene derivatives

    Science.gov (United States)

    Charlton, R. J.; Fogarty, R. M.; Bogatko, S.; Zuehlsdorff, T. J.; Hine, N. D. M.; Heeney, M.; Horsfield, A. P.; Haynes, P. D.

    2018-03-01

    An ab initio study of the effects of implicit and explicit hosts on the excited state properties of pentacene and its nitrogen-based derivatives has been performed using ground state density functional theory (DFT), time-dependent DFT, and ΔSCF. We observe a significant solvatochromic redshift in the excitation energy of the lowest singlet state (S1) of pentacene from inclusion in a p-terphenyl host compared to vacuum; for an explicit host consisting of six nearest neighbour p-terphenyls, we obtain a redshift of 65 meV while a conductor-like polarisable continuum model (CPCM) yields a 78 meV redshift. Comparison is made between the excitonic properties of pentacene and four of its nitrogen-based analogs, 1,8-, 2,9-, 5,12-, and 6,13-diazapentacene with the latter found to be the most distinct due to local distortions in the ground state electronic structure. We observe that a CPCM is insufficient to fully understand the impact of the host due to the presence of a mild charge-transfer (CT) coupling between the chromophore and neighbouring p-terphenyls, a phenomenon which can only be captured using an explicit model. The strength of this CT interaction increases as the nitrogens are brought closer to the central acene ring of pentacene.

  6. Convergence studies of deterministic methods for LWR explicit reflector methodology

    International Nuclear Information System (INIS)

    Canepa, S.; Hursin, M.; Ferroukhi, H.; Pautz, A.

    2013-01-01

    The standard approach in modem 3-D core simulators, employed either for steady-state or transient simulations, is to use Albedo coefficients or explicit reflectors at the core axial and radial boundaries. In the latter approach, few-group homogenized nuclear data are a priori produced with lattice transport codes using 2-D reflector models. Recently, the explicit reflector methodology of the deterministic CASMO-4/SIMULATE-3 code system was identified to potentially constitute one of the main sources of errors for core analyses of the Swiss operating LWRs, which are all belonging to GII design. Considering that some of the new GIII designs will rely on very different reflector concepts, a review and assessment of the reflector methodology for various LWR designs appeared as relevant. Therefore, the purpose of this paper is to first recall the concepts of the explicit reflector modelling approach as employed by CASMO/SIMULATE. Then, for selected reflector configurations representative of both GII and GUI designs, a benchmarking of the few-group nuclear data produced with the deterministic lattice code CASMO-4 and its successor CASMO-5, is conducted. On this basis, a convergence study with regards to geometrical requirements when using deterministic methods with 2-D homogenous models is conducted and the effect on the downstream 3-D core analysis accuracy is evaluated for a typical GII deflector design in order to assess the results against available plant measurements. (authors)

  7. Multidisciplinary Views on Applying Explicit and Implicit Motor Learning in Practice: An International Survey.

    Directory of Open Access Journals (Sweden)

    Melanie Kleynen

    Full Text Available A variety of options and techniques for causing implicit and explicit motor learning have been described in the literature. The aim of the current paper was to provide clearer guidance for practitioners on how to apply motor learning in practice by exploring experts' opinions and experiences, using the distinction between implicit and explicit motor learning as a conceptual departure point.A survey was designed to collect and aggregate informed opinions and experiences from 40 international respondents who had demonstrable expertise related to motor learning in practice and/or research. The survey was administered through an online survey tool and addressed potential options and learning strategies for applying implicit and explicit motor learning. Responses were analysed in terms of consensus (≥ 70% and trends (≥ 50%. A summary figure was developed to illustrate a taxonomy of the different learning strategies and options indicated by the experts in the survey.Answers of experts were widely distributed. No consensus was found regarding the application of implicit and explicit motor learning. Some trends were identified: Explicit motor learning can be promoted by using instructions and various types of feedback, but when promoting implicit motor learning, instructions and feedback should be restricted. Further, for implicit motor learning, an external focus of attention should be considered, as well as practicing the entire skill. Experts agreed on three factors that influence motor learning choices: the learner's abilities, the type of task, and the stage of motor learning (94.5%; n = 34/36. Most experts agreed with the summary figure (64.7%; n = 22/34.The results provide an overview of possible ways to cause implicit or explicit motor learning, signposting examples from practice and factors that influence day-to-day motor learning decisions.

  8. Model based methods and tools for process systems engineering

    DEFF Research Database (Denmark)

    Gani, Rafiqul

    need to be integrated with work-flows and data-flows for specific product-process synthesis-design problems within a computer-aided framework. The framework therefore should be able to manage knowledge-data, models and the associated methods and tools needed by specific synthesis-design work...... of model based methods and tools within a computer aided framework for product-process synthesis-design will be highlighted.......Process systems engineering (PSE) provides means to solve a wide range of problems in a systematic and efficient manner. This presentation will give a perspective on model based methods and tools needed to solve a wide range of problems in product-process synthesis-design. These methods and tools...

  9. The european Trans-Tools transport model

    NARCIS (Netherlands)

    Rooijen, T. van; Burgess, A.

    2008-01-01

    The paper presents the use of ArcGIS in the Transtools Transport Model, TRANS-TOOLS, created by an international consortium for the European Commission. The model describe passenger as well as freight transport in Europe with all medium and long distance modes (cars, vans, trucks, train, inland

  10. Moderators of Implicit-Explicit Exercise Cognition Concordance.

    Science.gov (United States)

    Berry, Tanya R; Rodgers, Wendy M; Markland, David; Hall, Craig R

    2016-12-01

    Investigating implicit-explicit concordance can aid in understanding underlying mechanisms and possible intervention effects. This research examined the concordance between implicit associations of exercise with health or appearance and related explicit motives. Variables considered as possible moderators were behavioral regulations, explicit attitudes, and social desirability. Participants (N = 454) completed measures of implicit associations of exercise with health and appearance and questionnaire measures of health and appearance motives, attitudes, social desirability, and behavioral regulations. Attitudes significantly moderated the relationship between implicit associations of exercise with health and health motives. Identified regulations significantly moderated implicit-explicit concordance with respect to associations with appearance. These results suggest that implicit and explicit exercise-related cognitions are not necessarily independent and their relationship to each other may be moderated by attitudes or some forms of behavioral regulation. Future research that takes a dual-processing approach to exercise behavior should consider potential theoretical moderators of concordance.

  11. Modeling Tools for Drilling, Reservoir Navigation, and Formation Evaluation

    Directory of Open Access Journals (Sweden)

    Sushant Dutta

    2012-06-01

    Full Text Available The oil and gas industry routinely uses borehole tools for measuring or logging rock and fluid properties of geologic formations to locate hydrocarbons and maximize their production. Pore fluids in formations of interest are usually hydrocarbons or water. Resistivity logging is based on the fact that oil and gas have a substantially higher resistivity than water. The first resistivity log was acquired in 1927, and resistivity logging is still the foremost measurement used for drilling and evaluation. However, the acquisition and interpretation of resistivity logging data has grown in complexity over the years. Resistivity logging tools operate in a wide range of frequencies (from DC to GHz and encounter extremely high (several orders of magnitude conductivity contrast between the metal mandrel of the tool and the geologic formation. Typical challenges include arbitrary angles of tool inclination, full tensor electric and magnetic field measurements, and interpretation of complicated anisotropic formation properties. These challenges combine to form some of the most intractable computational electromagnetic problems in the world. Reliable, fast, and convenient numerical modeling of logging tool responses is critical for tool design, sensor optimization, virtual prototyping, and log data inversion. This spectrum of applications necessitates both depth and breadth of modeling software—from blazing fast one-dimensional (1-D modeling codes to advanced threedimensional (3-D modeling software, and from in-house developed codes to commercial modeling packages. In this paper, with the help of several examples, we demonstrate our approach for using different modeling software to address different drilling and evaluation applications. In one example, fast 1-D modeling provides proactive geosteering information from a deep-reading azimuthal propagation resistivity measurement. In the second example, a 3-D model with multiple vertical resistive fractures

  12. The spatial limitations of current neutral models of biodiversity.

    Directory of Open Access Journals (Sweden)

    Rampal S Etienne

    Full Text Available The unified neutral theory of biodiversity and biogeography is increasingly accepted as an informative null model of community composition and dynamics. It has successfully produced macro-ecological patterns such as species-area relationships and species abundance distributions. However, the models employed make many unrealistic auxiliary assumptions. For example, the popular spatially implicit version assumes a local plot exchanging migrants with a large panmictic regional source pool. This simple structure allows rigorous testing of its fit to data. In contrast, spatially explicit models assume that offspring disperse only limited distances from their parents, but one cannot as yet test the significance of their fit to data. Here we compare the spatially explicit and the spatially implicit model, fitting the most-used implicit model (with two levels, local and regional to data simulated by the most-used spatially explicit model (where offspring are distributed about their parent on a grid according to either a radially symmetric Gaussian or a 'fat-tailed' distribution. Based on these fits, we express spatially implicit parameters in terms of spatially explicit parameters. This suggests how we may obtain estimates of spatially explicit parameters from spatially implicit ones. The relationship between these parameters, however, makes no intuitive sense. Furthermore, the spatially implicit model usually fits observed species-abundance distributions better than those calculated from the spatially explicit model's simulated data. Current spatially explicit neutral models therefore have limited descriptive power. However, our results suggest that a fatter tail of the dispersal kernel seems to improve the fit, suggesting that dispersal kernels with even fatter tails should be studied in future. We conclude that more advanced spatially explicit models and tools to analyze them need to be developed.

  13. Explicit Versus Implicit Income Insurance

    OpenAIRE

    Thomas J. Kniesner; James P. Z‎iliak

    2001-01-01

    October 2001 (Revised from July 2001). Abstract: By supplementing income explicitly through payments or implicitly through taxes collected, income-based taxes and transfers make disposable income less variable. Because disposable income determines consumption, policies that smooth disposable income also create welfare improving consumption insurance. With data from the Panel Study of Income Dynamics we find that annual consumption variation is reduced by almost 20 percent due to explicit and ...

  14. Explicit solution of the quantum three-body Calogero-Sutherland model

    CERN Document Server

    Perelomov, A.M.; Zaugg, P.

    1998-01-01

    Quantum integrable systems generalizing Calogero-Sutherland systems were introduced by Olshanetsky and Perelomov (1977). Recently, it was proved that for systems with trigonometric potential, the series in the product of two wave functions is a deformation of the Clebsch-Gordan series. This yields recursion relations for the wave functions of those systems. In this note, this approach is used to compute the explicit expressions for the three-body Calogero-Sutherland wave functions, which are the Jack polynomials. We conjecture that similar results are also valid for the more general two-parameters deformation introduced by Macdonald.

  15. Requirements Validation: Execution of UML Models with CPN Tools

    DEFF Research Database (Denmark)

    Machado, Ricardo J.; Lassen, Kristian Bisgaard; Oliveira, Sérgio

    2007-01-01

    Requirements validation is a critical task in any engineering project. The confrontation of stakeholders with static requirements models is not enough, since stakeholders with non-computer science education are not able to discover all the inter-dependencies between the elicited requirements. Eve...... requirements, where the system to be built must explicitly support the interaction between people within a pervasive cooperative workflow execution. A case study from a real project is used to illustrate the proposed approach....

  16. Tool-Body Assimilation Model Based on Body Babbling and Neurodynamical System

    Directory of Open Access Journals (Sweden)

    Kuniyuki Takahashi

    2015-01-01

    Full Text Available We propose the new method of tool use with a tool-body assimilation model based on body babbling and a neurodynamical system for robots to use tools. Almost all existing studies for robots to use tools require predetermined motions and tool features; the motion patterns are limited and the robots cannot use novel tools. Other studies fully search for all available parameters for novel tools, but this leads to massive amounts of calculations. To solve these problems, we took the following approach: we used a humanoid robot model to generate random motions based on human body babbling. These rich motion experiences were used to train recurrent and deep neural networks for modeling a body image. Tool features were self-organized in parametric bias, modulating the body image according to the tool in use. Finally, we designed a neural network for the robot to generate motion only from the target image. Experiments were conducted with multiple tools for manipulating a cylindrical target object. The results show that the tool-body assimilation model is capable of motion generation.

  17. Large scale experiments as a tool for numerical model development

    DEFF Research Database (Denmark)

    Kirkegaard, Jens; Hansen, Erik Asp; Fuchs, Jesper

    2003-01-01

    Experimental modelling is an important tool for study of hydrodynamic phenomena. The applicability of experiments can be expanded by the use of numerical models and experiments are important for documentation of the validity of numerical tools. In other cases numerical tools can be applied...

  18. Storm Water Management Model Climate Adjustment Tool (SWMM-CAT)

    Science.gov (United States)

    The US EPA’s newest tool, the Stormwater Management Model (SWMM) – Climate Adjustment Tool (CAT) is meant to help municipal stormwater utilities better address potential climate change impacts affecting their operations. SWMM, first released in 1971, models hydrology and hydrauli...

  19. Utility assessment of a map-based online geo-collaboration tool.

    Science.gov (United States)

    Sidlar, Christopher L; Rinner, Claus

    2009-05-01

    Spatial group decision-making processes often include both informal and analytical components. Discussions among stakeholders or planning experts are an example of an informal component. When participants discuss spatial planning projects they typically express concerns and comments by pointing to places on a map. The Argumentation Map model provides a conceptual basis for collaborative tools that enable explicit linkages of arguments to the places to which they refer. These tools allow for the input of explicitly geo-referenced arguments as well as the visual access to arguments through a map interface. In this paper, we will review previous utility studies in geo-collaboration and evaluate a case study of a Web-based Argumentation Map application. The case study was conducted in the summer of 2005 when student participants discussed planning issues on the University of Toronto St. George campus. During a one-week unmoderated discussion phase, 11 participants wrote 60 comments on issues such as safety, facilities, parking, and building aesthetics. By measuring the participants' use of geographic references, we draw conclusions on how well the software tool supported the potential of the underlying concept. This research aims to contribute to a scientific approach to geo-collaboration in which the engineering of novel spatial decision support methods is complemented by a critical assessment of their utility in controlled, realistic experiments.

  20. Integrating decision management with UML modeling concepts and tools

    DEFF Research Database (Denmark)

    Könemann, Patrick

    2009-01-01

    , but also for guiding the user by proposing subsequent decisions. In model-based software development, many decisions directly affect the structural and behavioral models used to describe and develop a software system and its architecture. However, the decisions are typically not connected to these models...... of formerly disconnected tools could improve tool usability as well as decision maker productivity....

  1. Modeling the oxidation of ebselen and other organoselenium compounds using explicit solvent networks.

    Science.gov (United States)

    Bayse, Craig A; Antony, Sonia

    2009-05-14

    The oxidation of dimethylselenide, dimethyldiselenide, S-methylselenenyl-methylmercaptan, and truncated and full models of ebselen (N-phenyl-1,2-benzisoselenazol-3(2H)-one) by methyl hydrogen peroxide has been modeled using density functional theory (DFT) and solvent-assisted proton exchange (SAPE), a method of microsolvation that employs explicit solvent networks to facilitate proton transfer reactions. The calculated activation barriers for these systems were substantially lower in energy (DeltaG(double dagger) + DeltaG(solv) = 13 to 26 kcal/mol) than models that neglect the participation of solvent in proton exchange. The comparison of two- and three-water SAPE networks showed a reduction in the strain in the model system but without a substantial reduction in the activation barriers. Truncating the ebselen model to N-methylisoselenazol-3(2H)-one gave a larger activation barrier than ebselen or N-methyl-1,2-benzisoselenazol-3(2H)-one but provided an efficient means of determining an initial guess for larger transition-state models. The similar barriers obtained for ebselen and Me(2)Se(2) (DeltaG(double dagger) + DeltaG(solv) = 20.65 and 20.40 kcal/mol, respectively) were consistent with experimentally determined rate constants. The activation barrier for MeSeSMe (DeltaG(double dagger) + DeltaG(solv) = 21.25 kcal/mol) was similar to that of ebselen and Me(2)Se(2) despite its significantly lower experimental rate for oxidation of an ebselen selenenyl sulfide by hydrogen peroxide relative to ebselen and ebselen diselenide. The disparity is attributed to intramolecular Se-O interactions, which decrease the nucleophilicity of the selenium center of the selenenyl sulfide.

  2. Advanced reach tool (ART) : Development of the mechanistic model

    NARCIS (Netherlands)

    Fransman, W.; Tongeren, M. van; Cherrie, J.W.; Tischer, M.; Schneider, T.; Schinkel, J.; Kromhout, H.; Warren, N.; Goede, H.; Tielemans, E.

    2011-01-01

    This paper describes the development of the mechanistic model within a collaborative project, referred to as the Advanced REACH Tool (ART) project, to develop a tool to model inhalation exposure for workers sharing similar operational conditions across different industries and locations in Europe.

  3. Spatial Modeling Tools for Cell Biology

    National Research Council Canada - National Science Library

    Przekwas, Andrzej; Friend, Tom; Teixeira, Rodrigo; Chen, Z. J; Wilkerson, Patrick

    2006-01-01

    .... Scientific potentials and military relevance of computational biology and bioinformatics have inspired DARPA/IPTO's visionary BioSPICE project to develop computational framework and modeling tools for cell biology...

  4. Developmental Differences in Implicit and Explicit Memory Performance.

    Science.gov (United States)

    Perez, Lori A.; Peynircioglu, Zehra F.; Blaxton, Teresa A.

    1998-01-01

    Compared perceptual and conceptual implicit and explicit memory performance of preschool, elementary, and college students. Found that conceptual explicit memory improved with age. Perceptual explicit memory and implicit memory showed no developmental change. Perceptual processing during study led to better performance than conceptual processing…

  5. Fish habitat simulation models and integrated assessment tools

    International Nuclear Information System (INIS)

    Harby, A.; Alfredsen, K.

    1999-01-01

    Because of human development water use increases in importance, and this worldwide trend is leading to an increasing number of user conflicts with a strong need for assessment tools to measure the impacts both on the ecosystem and the different users and user groups. The quantitative tools must allow a comparison of alternatives, different user groups, etc., and the tools must be integrated while impact assessments includes different disciplines. Fish species, especially young ones, are indicators of the environmental state of a riverine system and monitoring them is a way to follow environmental changes. The direct and indirect impacts on the ecosystem itself are measured, and impacts on user groups is not included. Fish habitat simulation models are concentrated on, and methods and examples are considered from Norway. Some ideas on integrated modelling tools for impact assessment studies are included. One dimensional hydraulic models are rapidly calibrated and do not require any expert knowledge in hydraulics. Two and three dimensional models require a bit more skilled users, especially if the topography is very heterogeneous. The advantages of using two and three dimensional models include: they do not need any calibration, just validation; they are predictive; and they can be more cost effective than traditional habitat hydraulic models when combined with modern data acquisition systems and tailored in a multi-disciplinary study. Suitable modelling model choice should be based on available data and possible data acquisition, available manpower, computer, and software resources, and needed output and accuracy in the output. 58 refs

  6. Environmental context effects in conceptual explicit and implicit memory.

    Science.gov (United States)

    Parker, Andrew; Dagnall, Neil; Coyle, Anne-Marie

    2007-05-01

    Previous research has found environmental context effects for both conceptual explicit and conceptual implicit memory (Parker, Gellatly, & Waterman, 1999). The research presented here challenges these findings on methodological grounds. Experiment 1 assessed the effects of context change on category-exemplar generation (conceptual implicit memory test) and category-cued recall (conceptual explicit memory test). Experiment 2 assessed the effects of context change on word association (conceptual implicit memory test) and word associate cued recall (conceptual explicit memory test). In both experiments, study-test changes in environmental context were found to influence performance only on tests of explicit memory. It is concluded that when retrieval cues across explicit and implicit tests are matched, and the probability of explicit contamination is reduced, then only conceptual explicit test performance is reduced by study-test changes in environmental context.

  7. The Guided System Development Framework: Modeling and Verifying Communication Systems

    DEFF Research Database (Denmark)

    Carvalho Quaresma, Jose Nuno; Probst, Christian W.; Nielson, Flemming

    2014-01-01

    the verified specification. The refinement process carries thus security properties from the model to the implementation. Our approach also supports verification of systems previously developed and deployed. Internally, the reasoning in our framework is based on the Beliefs and Knowledge tool, a verification...... tool based on belief logics and explicit attacker knowledge....

  8. Direct versus Indirect Explicit Methods of Enhancing EFL Students' English Grammatical Competence: A Concept Checking-Based Consciousness-Raising Tasks Model

    Science.gov (United States)

    Dang, Trang Thi Doan; Nguyen, Huong Thu

    2013-01-01

    Two approaches to grammar instruction are often discussed in the ESL literature: direct explicit grammar instruction (DEGI) (deduction) and indirect explicit grammar instruction (IEGI) (induction). This study aims to explore the effects of indirect explicit grammar instruction on EFL learners' mastery of English tenses. Ninety-four…

  9. Model-based setup assistant for progressive tools

    Science.gov (United States)

    Springer, Robert; Gräler, Manuel; Homberg, Werner; Henke, Christian; Trächtler, Ansgar

    2018-05-01

    In the field of production systems, globalization and technological progress lead to increasing requirements regarding part quality, delivery time and costs. Hence, today's production is challenged much more than a few years ago: it has to be very flexible and produce economically small batch sizes to satisfy consumer's demands and avoid unnecessary stock. Furthermore, a trend towards increasing functional integration continues to lead to an ongoing miniaturization of sheet metal components. In the industry of electric connectivity for example, the miniaturized connectors are manufactured by progressive tools, which are usually used for very large batches. These tools are installed in mechanical presses and then set up by a technician, who has to manually adjust a wide range of punch-bending operations. Disturbances like material thickness, temperatures, lubrication or tool wear complicate the setup procedure. In prospect of the increasing demand of production flexibility, this time-consuming process has to be handled more and more often. In this paper, a new approach for a model-based setup assistant is proposed as a solution, which is exemplarily applied in combination with a progressive tool. First, progressive tools, more specifically, their setup process is described and based on that, the challenges are pointed out. As a result, a systematic process to set up the machines is introduced. Following, the process is investigated with an FE-Analysis regarding the effects of the disturbances. In the next step, design of experiments is used to systematically develop a regression model of the system's behaviour. This model is integrated within an optimization in order to calculate optimal machine parameters and the following necessary adjustment of the progressive tool due to the disturbances. Finally, the assistant is tested in a production environment and the results are discussed.

  10. 33 CFR 385.33 - Revisions to models and analytical tools.

    Science.gov (United States)

    2010-07-01

    ... on a case-by-case basis what documentation is appropriate for revisions to models and analytic tools... analytical tools. 385.33 Section 385.33 Navigation and Navigable Waters CORPS OF ENGINEERS, DEPARTMENT OF THE... Incorporating New Information Into the Plan § 385.33 Revisions to models and analytical tools. (a) In carrying...

  11. Novel multiscale modeling tool applied to Pseudomonas aeruginosa biofilm formation.

    Directory of Open Access Journals (Sweden)

    Matthew B Biggs

    Full Text Available Multiscale modeling is used to represent biological systems with increasing frequency and success. Multiscale models are often hybrids of different modeling frameworks and programming languages. We present the MATLAB-NetLogo extension (MatNet as a novel tool for multiscale modeling. We demonstrate the utility of the tool with a multiscale model of Pseudomonas aeruginosa biofilm formation that incorporates both an agent-based model (ABM and constraint-based metabolic modeling. The hybrid model correctly recapitulates oxygen-limited biofilm metabolic activity and predicts increased growth rate via anaerobic respiration with the addition of nitrate to the growth media. In addition, a genome-wide survey of metabolic mutants and biofilm formation exemplifies the powerful analyses that are enabled by this computational modeling tool.

  12. Novel multiscale modeling tool applied to Pseudomonas aeruginosa biofilm formation.

    Science.gov (United States)

    Biggs, Matthew B; Papin, Jason A

    2013-01-01

    Multiscale modeling is used to represent biological systems with increasing frequency and success. Multiscale models are often hybrids of different modeling frameworks and programming languages. We present the MATLAB-NetLogo extension (MatNet) as a novel tool for multiscale modeling. We demonstrate the utility of the tool with a multiscale model of Pseudomonas aeruginosa biofilm formation that incorporates both an agent-based model (ABM) and constraint-based metabolic modeling. The hybrid model correctly recapitulates oxygen-limited biofilm metabolic activity and predicts increased growth rate via anaerobic respiration with the addition of nitrate to the growth media. In addition, a genome-wide survey of metabolic mutants and biofilm formation exemplifies the powerful analyses that are enabled by this computational modeling tool.

  13. Implementing Explicit and Finding Implicit Sharing in Embedded DSLs

    Directory of Open Access Journals (Sweden)

    Oleg Kiselyov

    2011-09-01

    Full Text Available Aliasing, or sharing, is prominent in many domains, denoting that two differently-named objects are in fact identical: a change in one object (memory cell, circuit terminal, disk block is instantly reflected in the other. Languages for modelling such domains should let the programmer explicitly define the sharing among objects or expressions. A DSL compiler may find other identical expressions and share them, implicitly. Such common subexpression elimination is crucial to the efficient implementation of DSLs. Sharing is tricky in embedded DSL, since host aliasing may correspond to copying of the underlying objects rather than their sharing. This tutorial summarizes discussions of implementing sharing in Haskell DSLs for automotive embedded systems and hardware description languages. The technique has since been used in a Haskell SAT solver and the DSL for music synthesis. We demonstrate the embedding in pure Haskell of a simple DSL with a language form for explicit sharing. The DSL also has implicit sharing, implemented via hash-consing. Explicit sharing greatly speeds up hash-consing. The seemingly imperative nature of hash-consing is hidden beneath a simple combinator language. The overall implementation remains pure functional and easy to reason about.

  14. Predictions of titanium alloy properties using thermodynamic modeling tools

    Science.gov (United States)

    Zhang, F.; Xie, F.-Y.; Chen, S.-L.; Chang, Y. A.; Furrer, D.; Venkatesh, V.

    2005-12-01

    Thermodynamic modeling tools have become essential in understanding the effect of alloy chemistry on the final microstructure of a material. Implementation of such tools to improve titanium processing via parameter optimization has resulted in significant cost savings through the elimination of shop/laboratory trials and tests. In this study, a thermodynamic modeling tool developed at CompuTherm, LLC, is being used to predict β transus, phase proportions, phase chemistries, partitioning coefficients, and phase boundaries of multicomponent titanium alloys. This modeling tool includes Pandat, software for multicomponent phase equilibrium calculations, and PanTitanium, a thermodynamic database for titanium alloys. Model predictions are compared with experimental results for one α-β alloy (Ti-64) and two near-β alloys (Ti-17 and Ti-10-2-3). The alloying elements, especially the interstitial elements O, N, H, and C, have been shown to have a significant effect on the β transus temperature, and are discussed in more detail herein.

  15. Toward transient finite element simulation of thermal deformation of machine tools in real-time

    Science.gov (United States)

    Naumann, Andreas; Ruprecht, Daniel; Wensch, Joerg

    2018-01-01

    Finite element models without simplifying assumptions can accurately describe the spatial and temporal distribution of heat in machine tools as well as the resulting deformation. In principle, this allows to correct for displacements of the Tool Centre Point and enables high precision manufacturing. However, the computational cost of FE models and restriction to generic algorithms in commercial tools like ANSYS prevents their operational use since simulations have to run faster than real-time. For the case where heat diffusion is slow compared to machine movement, we introduce a tailored implicit-explicit multi-rate time stepping method of higher order based on spectral deferred corrections. Using the open-source FEM library DUNE, we show that fully coupled simulations of the temperature field are possible in real-time for a machine consisting of a stock sliding up and down on rails attached to a stand.

  16. Stochastic models for time series

    CERN Document Server

    Doukhan, Paul

    2018-01-01

    This book presents essential tools for modelling non-linear time series. The first part of the book describes the main standard tools of probability and statistics that directly apply to the time series context to obtain a wide range of modelling possibilities. Functional estimation and bootstrap are discussed, and stationarity is reviewed. The second part describes a number of tools from Gaussian chaos and proposes a tour of linear time series models. It goes on to address nonlinearity from polynomial or chaotic models for which explicit expansions are available, then turns to Markov and non-Markov linear models and discusses Bernoulli shifts time series models. Finally, the volume focuses on the limit theory, starting with the ergodic theorem, which is seen as the first step for statistics of time series. It defines the distributional range to obtain generic tools for limit theory under long or short-range dependences (LRD/SRD) and explains examples of LRD behaviours. More general techniques (central limit ...

  17. Explicit simulation of a midlatitude Mesoscale Convective System

    Energy Technology Data Exchange (ETDEWEB)

    Alexander, G.D.; Cotton, W.R. [Colorado State Univ., Fort Collins, CO (United States)

    1996-04-01

    We have explicitly simulated the mesoscale convective system (MCS) observed on 23-24 June 1985 during PRE-STORM, the Preliminary Regional Experiment for the Stormscale Operational and Research and Meterology Program. Stensrud and Maddox (1988), Johnson and Bartels (1992), and Bernstein and Johnson (1994) are among the researchers who have investigated various aspects of this MCS event. We have performed this MCS simulation (and a similar one of a tropical MCS; Alexander and Cotton 1994) in the spirit of the Global Energy and Water Cycle Experiment Cloud Systems Study (GCSS), in which cloud-resolving models are used to assist in the formulation and testing of cloud parameterization schemes for larger-scale models. In this paper, we describe (1) the nature of our 23-24 June MCS dimulation and (2) our efforts to date in using our explicit MCS simulations to assist in the development of a GCM parameterization for mesoscale flow branches. The paper is organized as follows. First, we discuss the synoptic situation surrounding the 23-24 June PRE-STORM MCS followed by a discussion of the model setup and results of our simulation. We then discuss the use of our MCS simulation. We then discuss the use of our MCS simulations in developing a GCM parameterization for mesoscale flow branches and summarize our results.

  18. A spatially explicit model of functional connectivity for the endangered Przewalski's gazelle (Procapra przewalskii in a patchy landscape.

    Directory of Open Access Journals (Sweden)

    Chunlin Li

    Full Text Available Habitat fragmentation, associated with human population expansion, impedes dispersal, reduces gene flow and aggravates inbreeding in species on the brink of extinction. Both scientific and conservation communities increasingly realize that maintaining and restoring landscape connectivity is of vital importance in biodiversity conservation. Prior to any conservation initiatives, it is helpful to present conservation practitioners with a spatially explicit model of functional connectivity for the target species or landscape.Using Przewalski's gazelle (Procapra przewalskii as a model of endangered ungulate species in highly fragmented landscape, we present a model providing spatially explicit information to inform the long-term preservation of well-connected metapopulations. We employed a Geographic Information System (GIS and expert-literature method to create a habitat suitability map, to identify potential habitats and to delineate a functional connectivity network (least-cost movement corridors and paths for the gazelle. Results indicated that there were limited suitable habitats for the gazelle, mainly found to the north and northwest of the Qinghai Lake where four of five potential habitat patches were identified. Fifteen pairs of least-cost corridors and paths were mapped connecting eleven extant populations and two neighboring potential patches. The least-cost paths ranged from 0.2 km to 26.8 km in length (averaging 12.4 km and were all longer than corresponding Euclidean distances.The model outputs were validated and supported by the latest findings in landscape genetics of the species, and may provide impetus for connectivity conservation programs. Dispersal barriers were examined and appropriate mitigation strategies were suggested. This study provides conservation practitioners with thorough and visualized information to reserve the landscape connectivity for Przewalski's gazelle. In a general sense, we proposed a heuristic framework

  19. Explicit and Implicit Processes Constitute the Fast and Slow Processes of Sensorimotor Learning.

    Science.gov (United States)

    McDougle, Samuel D; Bond, Krista M; Taylor, Jordan A

    2015-07-01

    A popular model of human sensorimotor learning suggests that a fast process and a slow process work in parallel to produce the canonical learning curve (Smith et al., 2006). Recent evidence supports the subdivision of sensorimotor learning into explicit and implicit processes that simultaneously subserve task performance (Taylor et al., 2014). We set out to test whether these two accounts of learning processes are homologous. Using a recently developed method to assay explicit and implicit learning directly in a sensorimotor task, along with a computational modeling analysis, we show that the fast process closely resembles explicit learning and the slow process approximates implicit learning. In addition, we provide evidence for a subdivision of the slow/implicit process into distinct manifestations of motor memory. We conclude that the two-state model of motor learning is a close approximation of sensorimotor learning, but it is unable to describe adequately the various implicit learning operations that forge the learning curve. Our results suggest that a wider net be cast in the search for the putative psychological mechanisms and neural substrates underlying the multiplicity of processes involved in motor learning. Copyright © 2015 the authors 0270-6474/15/359568-12$15.00/0.

  20. Spatially explicit shallow landslide susceptibility mapping over large areas

    Science.gov (United States)

    Dino Bellugi; William E. Dietrich; Jonathan Stock; Jim McKean; Brian Kazian; Paul Hargrove

    2011-01-01

    Recent advances in downscaling climate model precipitation predictions now yield spatially explicit patterns of rainfall that could be used to estimate shallow landslide susceptibility over large areas. In California, the United States Geological Survey is exploring community emergency response to the possible effects of a very large simulated storm event and to do so...

  1. Constructing an explicit AdS/CFT correspondence with Cartan geometry

    Directory of Open Access Journals (Sweden)

    Jeffrey S. Hazboun

    2018-04-01

    Full Text Available An explicit AdS/CFT correspondence is shown for the Lie group SO(4,2. The Lie symmetry structures allow for the construction of two physical theories through the tools of Cartan geometry. One is a gravitational theory that has anti-de Sitter symmetry. The other is also a gravitational theory but is conformally symmetric and lives on 8-dimensional biconformal space. These “extra” four dimensions have the degrees of freedom used to construct a Yang–Mills theory. The two theories, based on AdS or conformal symmetry, have a natural correspondence in the context of their Lie algebras alone where neither SUSY, nor holography, is necessary.

  2. Modeling the dielectric logging tool at high frequency

    International Nuclear Information System (INIS)

    Chew, W.C.

    1987-01-01

    The high frequency dielectric logging tool has been used widely in electromagnetic well logging, because by measuring the dielectric constant at high frequencies (1 GHz), the water saturation of rocks could be known without measuring the water salinity in the rocks. As such, it could be used to delineate fresh water bearing zones, as the dielectric constant of fresh water is much higher than that of oil while they may have the same resistivity. The authors present a computer model, though electromagnetic field analysis, the response of such a measurement tool in a well logging environment. As the measurement is performed at high frequency, usually with small separation between the transmitter and receivers, some small geological features could be measured by such a tool. They use the computer model to study the behavior of such a tool across geological bed boundaries, and also across thin geological beds. Such a study could be very useful in understanding the limitation on the resolution of the tool. Furthermore, they could study the standoff effect and the depth of investigation of such a tool. This could delineate the range of usefulness of the measurement

  3. A novel explicit approach to model bromide and pesticide transport in soils containing macropores

    Science.gov (United States)

    Klaus, J.; Zehe, E.

    2011-01-01

    The present study tests whether an explicit treatment of worm burrows is feasible for simulating water flow, bromide and pesticide transport in structured heterogeneous soils. The essence is to represent worm burrows as morphologically connected paths of low flow resistance in the spatially highly resolved model domain. A recent Monte Carlo study (Klaus and Zehe, 2010) revealed that this approach allowed successful reproduction of tile drain event discharge recorded during an irrigation experiment at a tile drained field site. However, several "hillslope architectures" that were all consistent with the available extensive data base allowed a good reproduction of tile drain flow response. Our second objective was thus to find out whether this "equifinality" in spatial model setups may be reduced when including bromide tracer data in the model falsification process. We thus simulated transport of bromide and Isoproturon (IPU) for the 13 spatial model setups, which performed best with respect to reproduce tile drain event discharge, without any further calibration. All model setups allowed a very good prediction of the temporal dynamics of cumulated bromide leaching into the tile drain, while only four of them matched the accumulated water balance and accumulated bromide loss into the tile drain. The number of behavioural model architectures could thus be reduced to four. One of those setups was used for simulating transport of IPU, using different parameter combinations to characterise adsorption according to the Footprint data base. Simulations could, however, only reproduce the observed leaching behaviour, when we allowed for retardation coefficients that were very close to one.

  4. Novel Multiscale Modeling Tool Applied to Pseudomonas aeruginosa Biofilm Formation

    OpenAIRE

    Biggs, Matthew B.; Papin, Jason A.

    2013-01-01

    Multiscale modeling is used to represent biological systems with increasing frequency and success. Multiscale models are often hybrids of different modeling frameworks and programming languages. We present the MATLAB-NetLogo extension (MatNet) as a novel tool for multiscale modeling. We demonstrate the utility of the tool with a multiscale model of Pseudomonas aeruginosa biofilm formation that incorporates both an agent-based model (ABM) and constraint-based metabolic modeling. The hybrid mod...

  5. Explicit formulae for the generalized Hermite polynomials in superspace

    International Nuclear Information System (INIS)

    Desrosiers, Patrick; Lapointe, Luc; Mathieu, Pierre

    2004-01-01

    We provide explicit formulae for the orthogonal eigenfunctions of the supersymmetric extension of the rational Calogero-Moser-Sutherland model with harmonic confinement, i.e., the generalized Hermite (or Hi-Jack) polynomials in superspace. The construction relies on the triangular action of the Hamiltonian on the supermonomial basis. This translates into determinantal expressions for the Hamiltonian's eigenfunctions

  6. MODELING THE EFFECT OF STREAM NETWORK CHARACTERISTICS AND JUVENILE MOVEMENT ON COHO SALMON

    Science.gov (United States)

    Simulation modeling can be a valuable tool for improving our scientific understanding of the mechanisms that affect fish abundance and sustainability. Spatially explicit models, in particular, can be used to study interactions between fish biology and spatiotemporal habitat patt...

  7. A stage-structured, spatially explicit migration model for Myotis bats: mortality location affects system dynamics

    Science.gov (United States)

    Erickson, Richard A.; Thogmartin, Wayne E.; Russell, Robin E.; Diffendorfer, James E.; Szymanski, Jennifer A.

    2014-01-01

    Bats are ecologically and economically important species because they consume insects, transport nutrients, and pollinate flowers.  Many species of bats, including those in the Myotis genus, are facing population decline and increased extinction risk.  Despite these conservation concerns, few models exist for providing insight into the population dynamics of bats in a spatially explicit context.  We developed a model for bats by considering the stage-structured colonial life history of Myotis bats with their annual migration behavior.  This model provided insight into network dynamics.  We specifically focused on two Myotis species living in the eastern United States: the Indiana bat (M. sodalis), which is a Federally listed endangered species, and the little brown bat (M. lucifugus), which is under consideration for listing as an endangered species.  We found that multiple equilibria exist for the local, migratory subpopulations even though the total population was constant.  These equilibria suggest the location and magnitude of stressors such as White-nose Syndrome, meteorological phenomena, or impacts of wind turbines on survival influence system dynamics and risk of population extirpation in difficult to predict ways.

  8. A Conceptual Model for the Design and Delivery of Explicit Thinking Skills Instruction

    Science.gov (United States)

    Kassem, Cherrie L.

    2005-01-01

    Developing student thinking skills is an important goal for most educators. However, due to time constraints and weighty content standards, thinking skills instruction is often embedded in subject matter, implicit and incidental. For best results, thinking skills instruction requires a systematic design and explicit teaching strategies. The…

  9. EdgeMaps: visualizing explicit and implicit relations

    Science.gov (United States)

    Dörk, Marian; Carpendale, Sheelagh; Williamson, Carey

    2011-01-01

    In this work, we introduce EdgeMaps as a new method for integrating the visualization of explicit and implicit data relations. Explicit relations are specific connections between entities already present in a given dataset, while implicit relations are derived from multidimensional data based on shared properties and similarity measures. Many datasets include both types of relations, which are often difficult to represent together in information visualizations. Node-link diagrams typically focus on explicit data connections, while not incorporating implicit similarities between entities. Multi-dimensional scaling considers similarities between items, however, explicit links between nodes are not displayed. In contrast, EdgeMaps visualize both implicit and explicit relations by combining and complementing spatialization and graph drawing techniques. As a case study for this approach we chose a dataset of philosophers, their interests, influences, and birthdates. By introducing the limitation of activating only one node at a time, interesting visual patterns emerge that resemble the aesthetics of fireworks and waves. We argue that the interactive exploration of these patterns may allow the viewer to grasp the structure of a graph better than complex node-link visualizations.

  10. Mixed deterministic statistical modelling of regional ozone air pollution

    KAUST Repository

    Kalenderski, Stoitchko; Steyn, Douw G.

    2011-01-01

    formalism, and explicitly accounts for advection of pollutants, using the advection equation. We apply the model to a specific case of regional ozone pollution-the Lower Fraser valley of British Columbia, Canada. As a predictive tool, we demonstrate

  11. Making dilemmas explicit through the use of a cognitive mapping collaboration tool

    NARCIS (Netherlands)

    Matos Castano, Julieta; van Amstel, Frederick; Hartmann, Timo; Dewulf, Geert

    2017-01-01

    Dilemmas are pervasive in decision making. Although they offer the potential of reflecting on issues at stake from different perspectives, dilemmas often lead to paralysis for those encountering them. This study presents a three dimensional collaboration tool specifically developed to surface

  12. Web tools for predictive toxicology model building.

    Science.gov (United States)

    Jeliazkova, Nina

    2012-07-01

    The development and use of web tools in chemistry has accumulated more than 15 years of history already. Powered by the advances in the Internet technologies, the current generation of web systems are starting to expand into areas, traditional for desktop applications. The web platforms integrate data storage, cheminformatics and data analysis tools. The ease of use and the collaborative potential of the web is compelling, despite the challenges. The topic of this review is a set of recently published web tools that facilitate predictive toxicology model building. The focus is on software platforms, offering web access to chemical structure-based methods, although some of the frameworks could also provide bioinformatics or hybrid data analysis functionalities. A number of historical and current developments are cited. In order to provide comparable assessment, the following characteristics are considered: support for workflows, descriptor calculations, visualization, modeling algorithms, data management and data sharing capabilities, availability of GUI or programmatic access and implementation details. The success of the Web is largely due to its highly decentralized, yet sufficiently interoperable model for information access. The expected future convergence between cheminformatics and bioinformatics databases provides new challenges toward management and analysis of large data sets. The web tools in predictive toxicology will likely continue to evolve toward the right mix of flexibility, performance, scalability, interoperability, sets of unique features offered, friendly user interfaces, programmatic access for advanced users, platform independence, results reproducibility, curation and crowdsourcing utilities, collaborative sharing and secure access.

  13. Intimate partner violence and preschoolers' explicit memory functioning.

    Science.gov (United States)

    Jouriles, Ernest N; Brown, Alan S; McDonald, Renee; Rosenfield, David; Leahy, Matthew M; Silver, Cheryl

    2008-06-01

    This research examines whether parents' intimate partner physical violence (IPV) relates to their preschoolers' explicit memory functioning, whether children's symptoms of hyperarousal mediate this relation, and whether mothers' positive parenting moderates this relation. Participants were 69 mothers and their 4- or 5-year-old child (34 girls). Mothers completed measures of IPV, children's hyperarousal symptoms, parent-child aggression, and positive parenting. Measures of explicit memory functioning were administered to preschoolers. As expected, IPV correlated negatively with preschoolers' performance on explicit memory tasks, even after controlling for parent-child aggression and demographic variables related to preschoolers' memory functioning. Preschoolers' hyperarousal symptoms did not mediate the relation between IPV and explicit memory functioning, but mothers' positive parenting moderated this relation. Specifically, the negative relation between IPV and preschoolers' performance on 2 of the 3 explicit memory tasks was weaker when mothers engaged in higher levels of positive parenting. These findings extend research on IPV and children's adjustment difficulties to explicit memory functioning in preschoolers and suggest that mothers can ameliorate the influence of IPV on preschoolers' memory functioning via their parenting. (c) 2008 APA, all rights reserved

  14. Implicit and explicit self-esteem and their reciprocal relationship with symptoms of depression and social anxiety: a longitudinal study in adolescents.

    Science.gov (United States)

    van Tuijl, Lonneke A; de Jong, Peter J; Sportel, B Esther; de Hullu, Eva; Nauta, Maaike H

    2014-03-01

    A negative self-view is a prominent factor in most cognitive vulnerability models of depression and anxiety. Recently, there has been increased attention to differentiate between the implicit (automatic) and the explicit (reflective) processing of self-related evaluations. This longitudinal study aimed to test the association between implicit and explicit self-esteem and symptoms of adolescent depression and social anxiety disorder. Two complementary models were tested: the vulnerability model and the scarring effect model. Participants were 1641 first and second year pupils of secondary schools in the Netherlands. The Rosenberg Self-Esteem Scale, self-esteem Implicit Association Test and Revised Child Anxiety and Depression Scale were completed to measure explicit self-esteem, implicit self-esteem and symptoms of social anxiety disorder (SAD) and major depressive disorder (MDD), respectively, at baseline and two-year follow-up. Explicit self-esteem at baseline was associated with symptoms of MDD and SAD at follow-up. Symptomatology at baseline was not associated with explicit self-esteem at follow-up. Implicit self-esteem was not associated with symptoms of MDD or SAD in either direction. We relied on self-report measures of MDD and SAD symptomatology. Also, findings are based on a non-clinical sample. Our findings support the vulnerability model, and not the scarring effect model. The implications of these findings suggest support of an explicit self-esteem intervention to prevent increases in MDD and SAD symptomatology in non-clinical adolescents. Copyright © 2013 Elsevier Ltd. All rights reserved.

  15. Landscape capability models as a tool to predict fine-scale forest bird occupancy and abundance

    Science.gov (United States)

    Loman, Zachary G.; DeLuca, William; Harrison, Daniel J.; Loftin, Cynthia S.; Rolek, Brian W.; Wood, Petra B.

    2018-01-01

    ContextSpecies-specific models of landscape capability (LC) can inform landscape conservation design. Landscape capability is “the ability of the landscape to provide the environment […] and the local resources […] needed for survival and reproduction […] in sufficient quantity, quality and accessibility to meet the life history requirements of individuals and local populations.” Landscape capability incorporates species’ life histories, ecologies, and distributions to model habitat for current and future landscapes and climates as a proactive strategy for conservation planning.ObjectivesWe tested the ability of a set of LC models to explain variation in point occupancy and abundance for seven bird species representative of spruce-fir, mixed conifer-hardwood, and riparian and wooded wetland macrohabitats.MethodsWe compiled point count data sets used for biological inventory, species monitoring, and field studies across the northeastern United States to create an independent validation data set. Our validation explicitly accounted for underestimation in validation data using joint distance and time removal sampling.ResultsBlackpoll warbler (Setophaga striata), wood thrush (Hylocichla mustelina), and Louisiana (Parkesia motacilla) and northern waterthrush (P. noveboracensis) models were validated as predicting variation in abundance, although this varied from not biologically meaningful (1%) to strongly meaningful (59%). We verified all seven species models [including ovenbird (Seiurus aurocapilla), blackburnian (Setophaga fusca) and cerulean warbler (Setophaga cerulea)], as all were positively related to occupancy data.ConclusionsLC models represent a useful tool for conservation planning owing to their predictive ability over a regional extent. As improved remote-sensed data become available, LC layers are updated, which will improve predictions.

  16. What Does It Take for an Infant to Learn How to Use a Tool by Observation?

    Science.gov (United States)

    Fagard, Jacqueline; Rat-Fischer, Lauriane; Esseily, Rana; Somogyi, Eszter; O'Regan, J K

    2016-01-01

    Observational learning is probably one of the most powerful factors determining progress during child development. When learning a new skill, infants rely on their own exploration; but they also frequently benefit from an adult's verbal support or from demonstration by an adult modeling the action. At what age and under what conditions does adult demonstration really help the infant to learn a novel behavior? In this review, we summarize recently published work we have conducted on the acquisition of tool use during the second year of life. In particular, we consider under what conditions and to what extent seeing a demonstration from an adult advances an infant's understanding of how to use a tool to obtain an out-of-reach object. Our results show that classic demonstration starts being helpful at 18 months of age. When adults explicitly show their intention prior to demonstration, even 16-month-old infants learn from the demonstration. On the other hand, providing an explicit demonstration ("look at how I do it") is not very useful before infants are ready to succeed by themselves anyway. In contrast, repeated observations of the required action in a social context, without explicit reference to this action, considerably advances the age of success and the usefulness of providing a demonstration. We also show that the effect of demonstration can be enhanced if the demonstration makes the baby laugh. Taken together, the results from this series of studies on observational learning of tool use in infants suggest, first, that when observing a demonstration, infants do not know what to pay attention to: demonstration must be accompanied by rich social cues to be effective; second, infants' attention is inhibited rather than enhanced by an explicit demand of "look at what I do"; and finally a humorous situation considerably helps infants understand the demonstration.

  17. Evaluating spatially explicit burn probabilities for strategic fire management planning

    Science.gov (United States)

    C. Miller; M.-A. Parisien; A. A. Ager; M. A. Finney

    2008-01-01

    Spatially explicit information on the probability of burning is necessary for virtually all strategic fire and fuels management planning activities, including conducting wildland fire risk assessments, optimizing fuel treatments, and prevention planning. Predictive models providing a reliable estimate of the annual likelihood of fire at each point on the landscape have...

  18. The “Destabilizing” Effect of Cannibalism in a Spatially Explicit Three-Species Age Structured Predator-Prey Model

    Directory of Open Access Journals (Sweden)

    Aladeen Al Basheer

    2017-01-01

    Full Text Available Cannibalism, the act of killing and consumption of conspecifics, is generally considered to be a stabilising process in ODE models of predator-prey systems. On the other hand, Sun et al. were the first to show that cannibalism can cause Turing instability, in the classical Rosenzweig-McArthur two-species PDE model, which is an impossibility without cannibalism. Magnússon’s classic work is the first to show that cannibalism in a structured three-species predator-prey ODE model can actually be destabilising. In the current manuscript we consider the PDE form of the three-species model proposed in Magnússon’s classic work. We prove that, in the absence of cannibalism, Turing instability is an impossibility in this model, for any range of parameters. However, the inclusion of cannibalism can cause Turing instability. Thus, to the best of our knowledge, we report the first cannibalism induced Turing instability result, in spatially explicit three-species age structured predator-prey systems. We also show that, in the classical ODE model proposed by Magnússon, cannibalism can act as a life boat mechanism, for the prey.

  19. Analysis of LYSA-calculus with explicit confidentiality annotations

    DEFF Research Database (Denmark)

    Gao, Han; Nielson, Hanne Riis

    2006-01-01

    Recently there has been an increased research interest in applying process calculi in the verification of cryptographic protocols due to their ability to formally model protocols. This work presents LYSA with explicit confidentiality annotations for indicating the expected behavior of target...... malicious activities performed by attackers as specified by the confidentiality annotations. The proposed analysis approach is fully automatic without the need of human intervention and has been applied successfully to a number of protocols....

  20. A Regional Model for Malaria Vector Developmental Habitats Evaluated Using Explicit, Pond-Resolving Surface Hydrology Simulations.

    Directory of Open Access Journals (Sweden)

    Ernest Ohene Asare

    Full Text Available Dynamical malaria models can relate precipitation to the availability of vector breeding sites using simple models of surface hydrology. Here, a revised scheme is developed for the VECTRI malaria model, which is evaluated alongside the default scheme using a two year simulation by HYDREMATS, a 10 metre resolution, village-scale model that explicitly simulates individual ponds. Despite the simplicity of the two VECTRI surface hydrology parametrization schemes, they can reproduce the sub-seasonal evolution of fractional water coverage. Calibration of the model parameters is required to simulate the mean pond fraction correctly. The default VECTRI model tended to overestimate water fraction in periods subject to light rainfall events and underestimate it during periods of intense rainfall. This systematic error was improved in the revised scheme by including the a parametrization for surface run-off, such that light rainfall below the initial abstraction threshold does not contribute to ponds. After calibration of the pond model, the VECTRI model was able to simulate vector densities that compared well to the detailed agent based model contained in HYDREMATS without further parameter adjustment. Substituting local rain-gauge data with satellite-retrieved precipitation gave a reasonable approximation, raising the prospects for regional malaria simulations even in data sparse regions. However, further improvements could be made if a method can be derived to calibrate the key hydrology parameters of the pond model in each grid cell location, possibly also incorporating slope and soil texture.

  1. A Regional Model for Malaria Vector Developmental Habitats Evaluated Using Explicit, Pond-Resolving Surface Hydrology Simulations.

    Science.gov (United States)

    Asare, Ernest Ohene; Tompkins, Adrian Mark; Bomblies, Arne

    2016-01-01

    Dynamical malaria models can relate precipitation to the availability of vector breeding sites using simple models of surface hydrology. Here, a revised scheme is developed for the VECTRI malaria model, which is evaluated alongside the default scheme using a two year simulation by HYDREMATS, a 10 metre resolution, village-scale model that explicitly simulates individual ponds. Despite the simplicity of the two VECTRI surface hydrology parametrization schemes, they can reproduce the sub-seasonal evolution of fractional water coverage. Calibration of the model parameters is required to simulate the mean pond fraction correctly. The default VECTRI model tended to overestimate water fraction in periods subject to light rainfall events and underestimate it during periods of intense rainfall. This systematic error was improved in the revised scheme by including the a parametrization for surface run-off, such that light rainfall below the initial abstraction threshold does not contribute to ponds. After calibration of the pond model, the VECTRI model was able to simulate vector densities that compared well to the detailed agent based model contained in HYDREMATS without further parameter adjustment. Substituting local rain-gauge data with satellite-retrieved precipitation gave a reasonable approximation, raising the prospects for regional malaria simulations even in data sparse regions. However, further improvements could be made if a method can be derived to calibrate the key hydrology parameters of the pond model in each grid cell location, possibly also incorporating slope and soil texture.

  2. Designer Modeling for Personalized Game Content Creation Tools

    DEFF Research Database (Denmark)

    Liapis, Antonios; Yannakakis, Georgios N.; Togelius, Julian

    2013-01-01

    preferences, goals and processes from their interaction with a computer-aided design tool, and suggests methods and domains within game development where such a model can be applied. We describe how designer modeling could be integrated with current work on automated and mixed-initiative content creation......With the growing use of automated content creation and computer-aided design tools in game development, there is potential for enhancing the design process through personalized interactions between the software and the game developer. This paper proposes designer modeling for capturing the designer’s......, and envision future directions which focus on personalizing the processes to a designer’s particular wishes....

  3. Towards a theoretical foundation for explicitation and implicitation

    OpenAIRE

    De Metsenaere, Hinde; Vandepitte, Sonia

    2017-01-01

    Explicitation and implicitation are two translation studies concepts that have given rise to a vast array of studies. These studies are, however, often difficult to compare, primarily because explicitation and implicitation have been interpreted differently, not rarely intuitively, by many translation studies researchers. This is due to the fact that the underlying concepts of explicitness and implicitness have not yet been satisfactorily defined for translation studies purposes. It is there...

  4. A novel explicit approach to model bromide and pesticide transport in connected soil structures

    Directory of Open Access Journals (Sweden)

    J. Klaus

    2011-07-01

    Full Text Available The present study tests whether an explicit treatment of worm burrows and tile drains as connected structures is feasible for simulating water flow, bromide and pesticide transport in structured heterogeneous soils at hillslope scale. The essence is to represent worm burrows as morphologically connected paths of low flow resistance in a hillslope model. A recent Monte Carlo study (Klaus and Zehe, 2010, Hydrological Processes, 24, p. 1595–1609 revealed that this approach allowed successful reproduction of tile drain event discharge recorded during an irrigation experiment at a tile drained field site. However, several "hillslope architectures" that were all consistent with the available extensive data base allowed a good reproduction of tile drain flow response. Our second objective was thus to find out whether this "equifinality" in spatial model setups may be reduced when including bromide tracer data in the model falsification process. We thus simulated transport of bromide for the 13 spatial model setups that performed best with respect to reproduce tile drain event discharge, without any further calibration. All model setups allowed a very good prediction of the temporal dynamics of cumulated bromide leaching into the tile drain, while only four of them matched the accumulated water balance and accumulated bromide loss into the tile drain. The number of behavioural model architectures could thus be reduced to four. One of those setups was used for simulating transport of Isoproturon, using different parameter combinations to characterise adsorption according to the Footprint data base. Simulations could, however, only reproduce the observed leaching behaviour, when we allowed for retardation coefficients that were very close to one.

  5. A novel explicit approach to model bromide and pesticide transport in connected soil structures

    Science.gov (United States)

    Klaus, J.; Zehe, E.

    2011-07-01

    The present study tests whether an explicit treatment of worm burrows and tile drains as connected structures is feasible for simulating water flow, bromide and pesticide transport in structured heterogeneous soils at hillslope scale. The essence is to represent worm burrows as morphologically connected paths of low flow resistance in a hillslope model. A recent Monte Carlo study (Klaus and Zehe, 2010, Hydrological Processes, 24, p. 1595-1609) revealed that this approach allowed successful reproduction of tile drain event discharge recorded during an irrigation experiment at a tile drained field site. However, several "hillslope architectures" that were all consistent with the available extensive data base allowed a good reproduction of tile drain flow response. Our second objective was thus to find out whether this "equifinality" in spatial model setups may be reduced when including bromide tracer data in the model falsification process. We thus simulated transport of bromide for the 13 spatial model setups that performed best with respect to reproduce tile drain event discharge, without any further calibration. All model setups allowed a very good prediction of the temporal dynamics of cumulated bromide leaching into the tile drain, while only four of them matched the accumulated water balance and accumulated bromide loss into the tile drain. The number of behavioural model architectures could thus be reduced to four. One of those setups was used for simulating transport of Isoproturon, using different parameter combinations to characterise adsorption according to the Footprint data base. Simulations could, however, only reproduce the observed leaching behaviour, when we allowed for retardation coefficients that were very close to one.

  6. Explicit and Implicit Stigma of Mental Illness as Predictors of the Recovery Attitudes of Assertive Community Treatment Practitioners.

    Science.gov (United States)

    Stull, Laura G; McConnell, Haley; McGrew, John; Salyers, Michelle P

    2017-01-01

    While explicit negative stereotypes of mental illness are well established as barriers to recovery, implicit attitudes also may negatively impact outcomes. The current study is unique in its focus on both explicit and implicit stigma as predictors of recovery attitudes of mental health practitioners. Assertive Community Treatment practitioners (n = 154) from 55 teams completed online measures of stigma, recovery attitudes, and an Implicit Association Test (IAT). Three of four explicit stigma variables (perceptions of blameworthiness, helplessness, and dangerousness) and all three implicit stigma variables were associated with lower recovery attitudes. In a multivariate, hierarchical model, however, implicit stigma did not explain additional variance in recovery attitudes. In the overall model, perceptions of dangerousness and implicitly associating mental illness with "bad" were significant individual predictors of lower recovery attitudes. The current study demonstrates a need for interventions to lower explicit stigma, particularly perceptions of dangerousness, to increase mental health providers' expectations for recovery. The extent to which implicit and explicit stigma differentially predict outcomes, including recovery attitudes, needs further research.

  7. Spatially explicit models for inference about density in unmarked or partially marked populations

    Science.gov (United States)

    Chandler, Richard B.; Royle, J. Andrew

    2013-01-01

    Recently developed spatial capture–recapture (SCR) models represent a major advance over traditional capture–recapture (CR) models because they yield explicit estimates of animal density instead of population size within an unknown area. Furthermore, unlike nonspatial CR methods, SCR models account for heterogeneity in capture probability arising from the juxtaposition of animal activity centers and sample locations. Although the utility of SCR methods is gaining recognition, the requirement that all individuals can be uniquely identified excludes their use in many contexts. In this paper, we develop models for situations in which individual recognition is not possible, thereby allowing SCR concepts to be applied in studies of unmarked or partially marked populations. The data required for our model are spatially referenced counts made on one or more sample occasions at a collection of closely spaced sample units such that individuals can be encountered at multiple locations. Our approach includes a spatial point process for the animal activity centers and uses the spatial correlation in counts as information about the number and location of the activity centers. Camera-traps, hair snares, track plates, sound recordings, and even point counts can yield spatially correlated count data, and thus our model is widely applicable. A simulation study demonstrated that while the posterior mean exhibits frequentist bias on the order of 5–10% in small samples, the posterior mode is an accurate point estimator as long as adequate spatial correlation is present. Marking a subset of the population substantially increases posterior precision and is recommended whenever possible. We applied our model to avian point count data collected on an unmarked population of the northern parula (Parula americana) and obtained a density estimate (posterior mode) of 0.38 (95% CI: 0.19–1.64) birds/ha. Our paper challenges sampling and analytical conventions in ecology by demonstrating

  8. Water transport through tall trees: A vertically-explicit, analytical model of xylem hydraulic conductance in stems.

    Science.gov (United States)

    Couvreur, Valentin; Ledder, Glenn; Manzoni, Stefano; Way, Danielle A; Muller, Erik B; Russo, Sabrina E

    2018-05-08

    Trees grow by vertically extending their stems, so accurate stem hydraulic models are fundamental to understanding the hydraulic challenges faced by tall trees. Using a literature survey, we showed that many tree species exhibit continuous vertical variation in hydraulic traits. To examine the effects of this variation on hydraulic function, we developed a spatially-explicit, analytical water transport model for stems. Our model allows Huber ratio, stem-saturated conductivity, pressure at 50% loss of conductivity, leaf area, and transpiration rate to vary continuously along the hydraulic path. Predictions from our model differ from a matric flux potential model parameterized with uniform traits. Analyses show that cavitation is a whole-stem emergent property resulting from nonlinear pressure-conductivity feedbacks that, with gravity, cause impaired water transport to accumulate along the path. Because of the compounding effects of vertical trait variation on hydraulic function, growing proportionally more sapwood and building tapered xylem with height, as well as reducing xylem vulnerability only at branch tips while maintaining transport capacity at the stem base, can compensate for these effects. We therefore conclude that the adaptive significance of vertical variation in stem hydraulic traits is to allow trees to grow tall and tolerate operating near their hydraulic limits. This article is protected by copyright. All rights reserved.

  9. A masked negative self-esteem? Implicit and explicit self-esteem in patients with Narcissistic Personality Disorder.

    Science.gov (United States)

    Marissen, Marlies A E; Brouwer, Marlies E; Hiemstra, Annemarie M F; Deen, Mathijs L; Franken, Ingmar H A

    2016-08-30

    The mask model of narcissism states that the narcissistic traits of patients with NPD are the result of a compensatory reaction to underlying ego fragility. This model assumes that high explicit self-esteem masks low implicit self-esteem. However, research on narcissism has predominantly focused on non-clinical participants and data derived from patients diagnosed with Narcissistic Personality Disorder (NPD) remain scarce. Therefore, the goal of the present study was to test the mask model hypothesis of narcissism among patients with NPD. Male patients with NPD were compared to patients with other PD's and healthy participants on implicit and explicit self-esteem. NPD patients did not differ in levels of explicit and implicit self-esteem compared to both the psychiatric and the healthy control group. Overall, the current study found no evidence in support of the mask model of narcissism among a clinical group. This implicates that it might not be relevant for clinicians to focus treatment of NPD on an underlying negative self-esteem. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.

  10. Computer-based tools for decision support at the Hanford Site

    International Nuclear Information System (INIS)

    Doctor, P.G.; Mahaffey, J.A.; Cowley, P.J.; Freshley, M.D.; Hassig, N.L.; Brothers, J.W.; Glantz, C.S.; Strachan, D.M.

    1992-11-01

    To help integrate activities in the environmental restoration and waste management mission of the Hanford Site, the Hanford Integrated Planning Project (HIPP) was established and funded by the US Department of Energy. The project is divided into three key program elements, the first focusing on an explicit, defensible and comprehensive method for evaluating technical options. Based on the premise that computer technology can be used to support the decision-making process and facilitate integration among programs and activities, the Decision Support Tools Task was charged with assessing the status of computer technology for those purposes at the Site. The task addressed two types of tools: tools need to provide technical information and management support tools. Technical tools include performance and risk assessment models, information management systems, data and the computer infrastructure to supports models, data, and information management systems. Management decision support tools are used to synthesize information at a high' level to assist with making decisions. The major conclusions resulting from the assessment are that there is much technical information available, but it is not reaching the decision-makers in a form to be used. Many existing tools provide components that are needed to integrate site activities; however, some components are missing and, more importantly, the ''glue'' or connections to tie the components together to answer decision-makers questions is largely absent. Top priority should be given to decision support tools that support activities given in the TPA. Other decision tools are needed to facilitate and support the environmental restoration and waste management mission

  11. Computer-based tools for decision support at the Hanford Site

    Energy Technology Data Exchange (ETDEWEB)

    Doctor, P.G.; Mahaffey, J.A.; Cowley, P.J.; Freshley, M.D.; Hassig, N.L.; Brothers, J.W.; Glantz, C.S.; Strachan, D.M.

    1992-11-01

    To help integrate activities in the environmental restoration and waste management mission of the Hanford Site, the Hanford Integrated Planning Project (HIPP) was established and funded by the US Department of Energy. The project is divided into three key program elements, the first focusing on an explicit, defensible and comprehensive method for evaluating technical options. Based on the premise that computer technology can be used to support the decision-making process and facilitate integration among programs and activities, the Decision Support Tools Task was charged with assessing the status of computer technology for those purposes at the Site. The task addressed two types of tools: tools need to provide technical information and management support tools. Technical tools include performance and risk assessment models, information management systems, data and the computer infrastructure to supports models, data, and information management systems. Management decision support tools are used to synthesize information at a high' level to assist with making decisions. The major conclusions resulting from the assessment are that there is much technical information available, but it is not reaching the decision-makers in a form to be used. Many existing tools provide components that are needed to integrate site activities; however, some components are missing and, more importantly, the glue'' or connections to tie the components together to answer decision-makers questions is largely absent. Top priority should be given to decision support tools that support activities given in the TPA. Other decision tools are needed to facilitate and support the environmental restoration and waste management mission.

  12. Computer-based tools for decision support at the Hanford Site

    Energy Technology Data Exchange (ETDEWEB)

    Doctor, P.G.; Mahaffey, J.A.; Cowley, P.J.; Freshley, M.D.; Hassig, N.L.; Brothers, J.W.; Glantz, C.S.; Strachan, D.M.

    1992-11-01

    To help integrate activities in the environmental restoration and waste management mission of the Hanford Site, the Hanford Integrated Planning Project (HIPP) was established and funded by the US Department of Energy. The project is divided into three key program elements, the first focusing on an explicit, defensible and comprehensive method for evaluating technical options. Based on the premise that computer technology can be used to support the decision-making process and facilitate integration among programs and activities, the Decision Support Tools Task was charged with assessing the status of computer technology for those purposes at the Site. The task addressed two types of tools: tools need to provide technical information and management support tools. Technical tools include performance and risk assessment models, information management systems, data and the computer infrastructure to supports models, data, and information management systems. Management decision support tools are used to synthesize information at a high` level to assist with making decisions. The major conclusions resulting from the assessment are that there is much technical information available, but it is not reaching the decision-makers in a form to be used. Many existing tools provide components that are needed to integrate site activities; however, some components are missing and, more importantly, the ``glue`` or connections to tie the components together to answer decision-makers questions is largely absent. Top priority should be given to decision support tools that support activities given in the TPA. Other decision tools are needed to facilitate and support the environmental restoration and waste management mission.

  13. Explicit representation and parametrised impacts of under ice shelf seas in the z∗ coordinate ocean model NEMO 3.6

    Directory of Open Access Journals (Sweden)

    P. Mathiot

    2017-07-01

    Full Text Available Ice-shelf–ocean interactions are a major source of freshwater on the Antarctic continental shelf and have a strong impact on ocean properties, ocean circulation and sea ice. However, climate models based on the ocean–sea ice model NEMO (Nucleus for European Modelling of the Ocean currently do not include these interactions in any detail. The capability of explicitly simulating the circulation beneath ice shelves is introduced in the non-linear free surface model NEMO. Its implementation into the NEMO framework and its assessment in an idealised and realistic circum-Antarctic configuration is described in this study. Compared with the current prescription of ice shelf melting (i.e. at the surface, inclusion of open sub-ice-shelf cavities leads to a decrease in sea ice thickness along the coast, a weakening of the ocean stratification on the shelf, a decrease in salinity of high-salinity shelf water on the Ross and Weddell sea shelves and an increase in the strength of the gyres that circulate within the over-deepened basins on the West Antarctic continental shelf. Mimicking the overturning circulation under the ice shelves by introducing a prescribed meltwater flux over the depth range of the ice shelf base, rather than at the surface, is also assessed. It yields similar improvements in the simulated ocean properties and circulation over the Antarctic continental shelf to those from the explicit ice shelf cavity representation. With the ice shelf cavities opened, the widely used three equation ice shelf melting formulation, which enables an interactive computation of melting, is tested. Comparison with observational estimates of ice shelf melting indicates realistic results for most ice shelves. However, melting rates for the Amery, Getz and George VI ice shelves are considerably overestimated.

  14. Explicit dissipative structures

    International Nuclear Information System (INIS)

    Roessler, O.E.

    1987-01-01

    Dissipative structures consisting of a few macrovariables arise out of a sea of reversible microvariables. Unexpected residual effects of the massive underlying reversibility, on the macrolevel, cannot therefore be excluded. In the age of molecular-dynamics simulations, explicit dissipative structures like excitable systems (explicit observers) can be generated in a computer from first reversible principles. A class of classical, 1-D Hamiltonian systems of chaotic type is considered which has the asset that the trajectorial behavior in phase space can be understood geometrically. If, as nuatural, the number of particle types is much smaller than that of particles, the Gibbs symmetry must be taken into account. The permutation invariance drastically changes the behavior in phase space (quasi-periodization). The explicity observer becomes effectively reversible on a short time scale. In consequence, his ability to measure microscopic motions is suspended in a characteristic fashion. Unlike quantum mechanics whose holistic nature cannot be transcended, the present holistic (internal-interface) effects - mimicking the former to some extent - can be understood fully in principle

  15. Comparing implicit and explicit semantic access of direct and indirect word pairs in schizophrenia to evaluate models of semantic memory.

    Science.gov (United States)

    Neill, Erica; Rossell, Susan Lee

    2013-02-28

    Semantic memory deficits in schizophrenia (SZ) are profound, yet there is no research comparing implicit and explicit semantic processing in the same participant sample. In the current study, both implicit and explicit priming are investigated using direct (LION-TIGER) and indirect (LION-STRIPES; where tiger is not displayed) stimuli comparing SZ to healthy controls. Based on a substantive review (Rossell and Stefanovic, 2007) and meta-analysis (Pomarol-Clotet et al., 2008), it was predicted that SZ would be associated with increased indirect priming implicitly. Further, it was predicted that SZ would be associated with abnormal indirect priming explicitly, replicating earlier work (Assaf et al., 2006). No specific hypotheses were made for implicit direct priming due to the heterogeneity of the literature. It was hypothesised that explicit direct priming would be intact based on the structured nature of this task. The pattern of results suggests (1) intact reaction time (RT) and error performance implicitly in the face of abnormal direct priming and (2) impaired RT and error performance explicitly. This pattern confirms general findings regarding implicit/explicit memory impairments in SZ whilst highlighting the unique pattern of performance specific to semantic priming. Finally, priming performance is discussed in relation to thought disorder and length of illness. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.

  16. Explicit and implicit cognition: a preliminary test of a dual-process theory of cognitive vulnerability to depression.

    Science.gov (United States)

    Haeffel, Gerald J; Abramson, Lyn Y; Brazy, Paige C; Shah, James Y; Teachman, Bethany A; Nosek, Brian A

    2007-06-01

    Two studies were conducted to test a dual-process theory of cognitive vulnerability to depression. According to this theory, implicit and explicit cognitive processes have differential effects on depressive reactions to stressful life events. Implicit processes are hypothesized to be critical in determining an individual's immediate affective reaction to stress whereas explicit cognitions are thought to be more involved in long-term depressive reactions. Consistent with hypotheses, the results of study 1 (cross-sectional; N=237) showed that implicit, but not explicit, cognitions predicted immediate affective reactions to a lab stressor. Study 2 (longitudinal; N=251) also supported the dual-process model of cognitive vulnerability to depression. Results showed that both the implicit and explicit measures interacted with life stress to predict prospective changes in depressive symptoms, respectively. However, when both implicit and explicit predictors were entered into a regression equation simultaneously, only the explicit measure interacted with stress to remain a unique predictor of depressive symptoms over the five-week prospective interval.

  17. Tool wear modeling using abductive networks

    Science.gov (United States)

    Masory, Oren

    1992-09-01

    A tool wear model based on Abductive Networks, which consists of a network of `polynomial' nodes, is described. The model relates the cutting parameters, components of the cutting force, and machining time to flank wear. Thus real time measurements of the cutting force can be used to monitor the machining process. The model is obtained by a training process in which the connectivity between the network's nodes and the polynomial coefficients of each node are determined by optimizing a performance criteria. Actual wear measurements of coated and uncoated carbide inserts were used for training and evaluating the established model.

  18. Animal models: an important tool in mycology.

    Science.gov (United States)

    Capilla, Javier; Clemons, Karl V; Stevens, David A

    2007-12-01

    Animal models of fungal infections are, and will remain, a key tool in the advancement of the medical mycology. Many different types of animal models of fungal infection have been developed, with murine models the most frequently used, for studies of pathogenesis, virulence, immunology, diagnosis, and therapy. The ability to control numerous variables in performing the model allows us to mimic human disease states and quantitatively monitor the course of the disease. However, no single model can answer all questions and different animal species or different routes of infection can show somewhat different results. Thus, the choice of which animal model to use must be made carefully, addressing issues of the type of human disease to mimic, the parameters to follow and collection of the appropriate data to answer those questions being asked. This review addresses a variety of uses for animal models in medical mycology. It focuses on the most clinically important diseases affecting humans and cites various examples of the different types of studies that have been performed. Overall, animal models of fungal infection will continue to be valuable tools in addressing questions concerning fungal infections and contribute to our deeper understanding of how these infections occur, progress and can be controlled and eliminated.

  19. EXPLICIT PLANNING FOR PARAGRAPH WRITING CLASS

    Directory of Open Access Journals (Sweden)

    Lestari Setyowati

    2017-11-01

    Full Text Available The purpose of the study is to improve the students writing ability for paragraph writing class. The subjects of the study were 37 students of English Education Study Program who joined the paragraph writing class. The design of the study was Classroom Action Research with two cycles. Cycle 1 consisted of three meetings, and cycle 2 consisted of two meetings. The types of explicit planning used in the action research were word listing and word mapping with phrases and sentence for detail.  The instruments used were direct writing test, observation, and  documentation of students’ reflective essay. To score the students’ writing, two raters  were asked to rate the composition by using Jacobs ESL Composition profile scoring rubric. The finding shows that the use of explicit planning was able to improve the students’ paragraph writing performance, indicated with the achievement of the criteria of success. The students’ mean improved from cycle 1 (74.62  to cycle2 (76.78. Although explicit planning instruction was able to help the students to write better, data from their self-reflection essay showed that many of the students preferred to use free writing instead of explicit planning instruction.

  20. Explicit mentalizing mechanisms and their adaptive role in memory conformity.

    Science.gov (United States)

    Wheeler, Rebecca; Allan, Kevin; Tsivilis, Dimitris; Martin, Douglas; Gabbert, Fiona

    2013-01-01

    Memory conformity occurs when an individual endorses what other individuals remember about past events. Research on memory conformity is currently dominated by a 'forensic' perspective, which views the phenomenon as inherently undesirable. This is because conformity not only distorts the accuracy of an individual's memory, but also produces false corroboration between individuals, effects that act to undermine criminal justice systems. There is growing awareness, however, that memory conformity may be interpreted more generally as an adaptive social behavior regulated by explicit mentalizing mechanisms. Here, we provide novel evidence in support of this emerging alternative theoretical perspective. We carried out a memory conformity experiment which revealed that explicit belief-simulation (i.e. using one's own beliefs to model what other people believe) systematically biases conformity towards like-minded individuals, even when there is no objective evidence that they have a more accurate memory than dissimilar individuals. We suggest that this bias is functional, i.e. adaptive, to the extent that it fosters trust, and hence cooperation, between in-group versus out-group individuals. We conclude that memory conformity is, in more fundamental terms, a highly desirable product of explicit mentalizing mechanisms that promote adaptive forms of social learning and cooperation.

  1. A spatially explicit model for the future progression of the current Haiti cholera epidemic

    Science.gov (United States)

    Bertuzzo, E.; Mari, L.; Righetto, L.; Gatto, M.; Casagrandi, R.; Rodriguez-Iturbe, I.; Rinaldo, A.

    2011-12-01

    As a major cholera epidemic progresses in Haiti, and the figures of the infection, up to July 2011, climb to 385,000 cases and 5,800 deaths, the development of general models to track and predict the evolution of the outbreak, so as to guide the allocation of medical supplies and staff, is gaining notable urgency. We propose here a spatially explicit epidemic model that accounts for the dynamics of susceptible and infected individuals as well as the redistribution of textit{Vibrio cholera}, the causative agent of the disease, among different human communities. In particular, we model two spreading pathways: the advection of pathogens through hydrologic connections and the dissemination due to human mobility described by means of a gravity-like model. To this end the country has been divided into hydrologic units based on drainage directions derived from a digital terrain model. Moreover the population of each unit has been estimated from census data downscaled to 1 km x 1 km resolution via remotely sensed geomorphological information (LandScan texttrademark project). The model directly account for the role of rainfall patterns in driving the seasonality of cholera outbreaks. The two main outbreaks in fact occurred during the rainy seasons (October and May) when extensive floodings severely worsened the sanitation conditions and, in turn, raised the risk of infection. The model capability to reproduce the spatiotemporal features of the epidemic up to date grants robustness to the foreseen future development. In this context, the duration of acquired immunity, a hotly debated topic in the scientific community, emerges as a controlling factor for progression of the epidemic in the near future. The framework presented here can straightforwardly be used to evaluate the effectiveness of alternative intervention strategies like mass vaccinations, clean water supply and educational campaigns, thus emerging as an essential component of the control of future cholera

  2. Using the IEA ETSAP modelling tools for Denmark

    DEFF Research Database (Denmark)

    Grohnheit, Poul Erik

    signed the agreement and contributed to some early annexes. This project is motivated by an invitation to participate in ETSAP Annex X, "Global Energy Systems and Common Analyses: Climate friendly, Secure and Productive Energy Systems" for the period 2005 to 2007. The main activity is semi......-annual workshops focusing on presentations of model analyses and use of the ETSAP' tools (the MARKAL/TIMES family of models). The project was also planned to benefit from the EU project ”NEEDS - New Energy Externalities Developments for Sustainability. ETSAP is contributing to a part of NEEDS that develops......, Environment and Health (CEEH), starting from January 2007. This report summarises the activities under ETSAP Annex X and related project, emphasising the development of modelling tools that will be useful for modelling the Danish energy system. It is also a status report for the development of a model...

  3. A Model-Driven Visualization Tool for Use with Model-Based Systems Engineering Projects

    Science.gov (United States)

    Trase, Kathryn; Fink, Eric

    2014-01-01

    Model-Based Systems Engineering (MBSE) promotes increased consistency between a system's design and its design documentation through the use of an object-oriented system model. The creation of this system model facilitates data presentation by providing a mechanism from which information can be extracted by automated manipulation of model content. Existing MBSE tools enable model creation, but are often too complex for the unfamiliar model viewer to easily use. These tools do not yet provide many opportunities for easing into the development and use of a system model when system design documentation already exists. This study creates a Systems Modeling Language (SysML) Document Traceability Framework (SDTF) for integrating design documentation with a system model, and develops an Interactive Visualization Engine for SysML Tools (InVEST), that exports consistent, clear, and concise views of SysML model data. These exported views are each meaningful to a variety of project stakeholders with differing subjects of concern and depth of technical involvement. InVEST allows a model user to generate multiple views and reports from a MBSE model, including wiki pages and interactive visualizations of data. System data can also be filtered to present only the information relevant to the particular stakeholder, resulting in a view that is both consistent with the larger system model and other model views. Viewing the relationships between system artifacts and documentation, and filtering through data to see specialized views improves the value of the system as a whole, as data becomes information

  4. Implicit and Explicit Instruction of Spelling Rules

    Science.gov (United States)

    Kemper, M. J.; Verhoeven, L.; Bosman, A. M. T.

    2012-01-01

    The study aimed to compare the differential effectiveness of explicit and implicit instruction of two Dutch spelling rules. Students with and without spelling disabilities were instructed a spelling rule either implicitly or explicitly in two experiments. Effects were tested in a pretest-intervention-posttest control group design. Experiment 1…

  5. Explicit TE/TM scheme for particle beam simulations

    International Nuclear Information System (INIS)

    Dohlus, M.; Zagorodnov, I.

    2008-10-01

    In this paper we propose an explicit two-level conservative scheme based on a TE/TM like splitting of the field components in time. Its dispersion properties are adjusted to accelerator problems. It is simpler and faster than the implicit version. It does not have dispersion in the longitudinal direction and the dispersion properties in the transversal plane are improved. The explicit character of the new scheme allows a uniformly stable conformal method without iterations and the scheme can be parallelized easily. It assures energy and charge conservation. A version of this explicit scheme for rotationally symmetric structures is free from the progressive time step reducing for higher order azimuthal modes as it takes place for Yee's explicit method used in the most popular electrodynamics codes. (orig.)

  6. A fuzzy-logic tool for multi-criteria decision making in fisheries: the case of the South African pelagic fishery

    DEFF Research Database (Denmark)

    Paterson, B.; Jarre, Astrid; Moloney, C.L.

    2007-01-01

    The present study presents an electronic decision-support tool that uses a fuzzy-logic model of expert knowledge to assist in multi-criteria decision-making in the context of an Ecosystem Approach to Fisheries (EAF). The prototype model integrates the multiple goals and objectives related...... to the evaluation of the ecosystem performance of the South African sardine Sardinops sagax fishery into a NetWeaver knowledge base and provides intuitive visual outputs to communicate results to managers and stakeholders. The software tool was developed in a consultative process with key experts and follows...... on expert opinion. Results show that the model is robust and conservative. The strength of the approach lies in the ability to include variables that are difficult to measure. It provides a means of rendering value judgements explicit and transparent. The tool synthesises a large amount of information...

  7. Spatially explicit shallow landslide susceptibility mapping over large areas

    Science.gov (United States)

    Bellugi, Dino; Dietrich, William E.; Stock, Jonathan D.; McKean, Jim; Kazian, Brian; Hargrove, Paul

    2011-01-01

    Recent advances in downscaling climate model precipitation predictions now yield spatially explicit patterns of rainfall that could be used to estimate shallow landslide susceptibility over large areas. In California, the United States Geological Survey is exploring community emergency response to the possible effects of a very large simulated storm event and to do so it has generated downscaled precipitation maps for the storm. To predict the corresponding pattern of shallow landslide susceptibility across the state, we have used the model Shalstab (a coupled steady state runoff and infinite slope stability model) which susceptibility spatially explicit estimates of relative potential instability. Such slope stability models that include the effects of subsurface runoff on potentially destabilizing pore pressure evolution require water routing and hence the definition of upslope drainage area to each potential cell. To calculate drainage area efficiently over a large area we developed a parallel framework to scale-up Shalstab and specifically introduce a new efficient parallel drainage area algorithm which produces seamless results. The single seamless shallow landslide susceptibility map for all of California was accomplished in a short run time, and indicates that much larger areas can be efficiently modelled. As landslide maps generally over predict the extent of instability for any given storm. Local empirical data on the fraction of predicted unstable cells that failed for observed rainfall intensity can be used to specify the likely extent of hazard for a given storm. This suggests that campaigns to collect local precipitation data and detailed shallow landslide location maps after major storms could be used to calibrate models and improve their use in hazard assessment for individual storms.

  8. System level modelling with open source tools

    DEFF Research Database (Denmark)

    Jakobsen, Mikkel Koefoed; Madsen, Jan; Niaki, Seyed Hosein Attarzadeh

    , called ForSyDe. ForSyDe is available under the open Source approach, which allows small and medium enterprises (SME) to get easy access to advanced modeling capabilities and tools. We give an introduction to the design methodology through the system level modeling of a simple industrial use case, and we...

  9. The effect of explicit financial incentives on physician behavior.

    Science.gov (United States)

    Armour, B S; Pitts, M M; Maclean, R; Cangialose, C; Kishel, M; Imai, H; Etchason, J

    2001-05-28

    Managed care organizations use explicit financial incentives to influence physicians' use of resources. This has contributed to concerns regarding conflicts of interest for physicians and adverse effects on the quality of patient care. In light of recent publicized legislative and legal battles about this issue, we reviewed the literature and analyzed studies that examine the effect of these explicit financial incentives on the behavior of physicians. The method used to undertake the literature review followed the approach set forth in the Cochrane Collaboration handbook. Our literature review revealed a paucity of data on the effect of explicit financial incentives. Based on this limited evidence, explicit incentives that place individual physicians at financial risk appear to be effective in reducing physician resource use. However, the empirical evidence regarding the effectiveness of bonus payments on physician resource use is mixed. Similarly, our review revealed mixed effects of the influence of explicit financial incentives on the quality of patient care. The effect of explicit financial incentives on physician behavior is complicated by a lack of understanding of the incentive structure by the managed care organization and the physician. The lack of a universally acceptable definition of quality renders it important that future researchers identify the term explicitly.

  10. Generating transverse response explicitly from harmonic oscillators

    Science.gov (United States)

    Yao, Yuan; Tang, Ying; Ao, Ping

    2017-10-01

    We obtain stochastic dynamics from a system-plus-bath mechanism as an extension of the Caldeira-Leggett (CL) model in the classical regime. An effective magnetic field and response functions with both longitudinal and transverse parts are exactly generated from the bath of harmonic oscillators. The effective magnetic field and transverse response are antisymmetric matrices: the former is explicitly time-independent corresponding to the geometric magnetism, while the latter can have memory. The present model can be reduced to previous representative examples of stochastic dynamics describing nonequilibrium processes. Our results demonstrate that a system coupled with a bath of harmonic oscillators is a general approach to studying stochastic dynamics, and provides a method to experimentally implement an effective magnetic field from coupling to the environment.

  11. Modeling biochemical transformation processes and information processing with Narrator.

    Science.gov (United States)

    Mandel, Johannes J; Fuss, Hendrik; Palfreyman, Niall M; Dubitzky, Werner

    2007-03-27

    Software tools that model and simulate the dynamics of biological processes and systems are becoming increasingly important. Some of these tools offer sophisticated graphical user interfaces (GUIs), which greatly enhance their acceptance by users. Such GUIs are based on symbolic or graphical notations used to describe, interact and communicate the developed models. Typically, these graphical notations are geared towards conventional biochemical pathway diagrams. They permit the user to represent the transport and transformation of chemical species and to define inhibitory and stimulatory dependencies. A critical weakness of existing tools is their lack of supporting an integrative representation of transport, transformation as well as biological information processing. Narrator is a software tool facilitating the development and simulation of biological systems as Co-dependence models. The Co-dependence Methodology complements the representation of species transport and transformation together with an explicit mechanism to express biological information processing. Thus, Co-dependence models explicitly capture, for instance, signal processing structures and the influence of exogenous factors or events affecting certain parts of a biological system or process. This combined set of features provides the system biologist with a powerful tool to describe and explore the dynamics of life phenomena. Narrator's GUI is based on an expressive graphical notation which forms an integral part of the Co-dependence Methodology. Behind the user-friendly GUI, Narrator hides a flexible feature which makes it relatively easy to map models defined via the graphical notation to mathematical formalisms and languages such as ordinary differential equations, the Systems Biology Markup Language or Gillespie's direct method. This powerful feature facilitates reuse, interoperability and conceptual model development. Narrator is a flexible and intuitive systems biology tool. It is

  12. Large scale spatially explicit modeling of blue and green water dynamics in a temperate mid-latitude basin

    Science.gov (United States)

    Du, Liuying; Rajib, Adnan; Merwade, Venkatesh

    2018-07-01

    Looking only at climate change impacts provides partial information about a changing hydrologic regime. Understanding the spatio-temporal nature of change in hydrologic processes, and the explicit contributions from both climate and land use drivers, holds more practical value for water resources management and policy intervention. This study presents a comprehensive assessment on the spatio-temporal trend of Blue Water (BW) and Green Water (GW) in a 490,000 km2 temperate mid-latitude basin (Ohio River Basin) over the past 80 years (1935-2014), and from thereon, quantifies the combined as well as relative contributions of climate and land use changes. The Soil and Water Assessment Tool (SWAT) is adopted to simulate hydrologic fluxes. Mann-Kendall and Theil-Sen statistical tests are performed on the modeled outputs to detect respectively the trend and magnitude of changes at three different spatial scales - the entire basin, regional level, and sub-basin level. Despite the overall volumetric increase of both BW and GW in the entire basin, changes in their annual average values during the period of simulation reveal a distinctive spatial pattern. GW has increased significantly in the upper and lower parts of the basin, which can be related to the prominent land use change in those areas. BW has increased significantly only in the lower part, likely being associated with the notable precipitation change there. Furthermore, the simulation under a time-varying climate but constant land use scenario identifies climate change in the Ohio River Basin to be influential on BW, while the impact is relatively nominal on GW; whereas, land use change increases GW remarkably, but is counterproductive on BW. The approach to quantify combined/relative effects of climate and land use change as shown in this study can be replicated to understand BW-GW dynamics in similar large basins around the globe.

  13. Effect of explicit representation of detailed stratigraphy on brine and gas flow at the Waste Isolation Pilot Plant

    International Nuclear Information System (INIS)

    Christian-Frear, T.L.; Webb, S.W.

    1996-04-01

    Stratigraphic units of the Salado Formation at the Waste Isolation Pilot Plant (WIPP) disposal room horizon includes various layers of halite, polyhalitic halite, argillaceous halite, clay, and anhydrite. Current models, including those used in the WIPP Performance Assessment calculations, employ a ''composite stratigraphy'' approach in modeling. This study was initiated to evaluate the impact that an explicit representation of detailed stratigraphy around the repository may have on fluid flow compared to the simplified ''composite stratigraphy'' models currently employed. Sensitivity of model results to intrinsic permeability anisotropy, interbed fracturing, two-phase characteristic curves, and gas-generation rates were studied. The results of this study indicate that explicit representation of the stratigraphy maintains higher pressures and does not allow as much fluid to leave the disposal room as compared to the ''composite stratigraphy'' approach. However, the differences are relatively small. Gas migration distances are also different between the two approaches. However, for the two cases in which explicit layering results were considerably different than the composite model (anisotropic and vapor-limited), the gas-migration distances for both models were negligible. For the cases in which gas migration distances were considerable, van Genuchten/Parker and interbed fracture, the differences between the two models were fairly insignificant. Overall, this study suggests that explicit representation of the stratigraphy in the WIPP PA models is not required for the parameter variations modeled if ''global quantities'' (e.g., disposal room pressures, net brine and gas flux into and out of disposal rooms) are the only concern

  14. Effect of explicit representation of detailed stratigraphy on brine and gas flow at the Waste Isolation Pilot Plant

    Energy Technology Data Exchange (ETDEWEB)

    Christian-Frear, T.L.; Webb, S.W. [Sandia National Labs., Albuquerque, NM (United States). Geohydrology Dept.

    1996-04-01

    Stratigraphic units of the Salado Formation at the Waste Isolation Pilot Plant (WIPP) disposal room horizon includes various layers of halite, polyhalitic halite, argillaceous halite, clay, and anhydrite. Current models, including those used in the WIPP Performance Assessment calculations, employ a ``composite stratigraphy`` approach in modeling. This study was initiated to evaluate the impact that an explicit representation of detailed stratigraphy around the repository may have on fluid flow compared to the simplified ``composite stratigraphy`` models currently employed. Sensitivity of model results to intrinsic permeability anisotropy, interbed fracturing, two-phase characteristic curves, and gas-generation rates were studied. The results of this study indicate that explicit representation of the stratigraphy maintains higher pressures and does not allow as much fluid to leave the disposal room as compared to the ``composite stratigraphy`` approach. However, the differences are relatively small. Gas migration distances are also different between the two approaches. However, for the two cases in which explicit layering results were considerably different than the composite model (anisotropic and vapor-limited), the gas-migration distances for both models were negligible. For the cases in which gas migration distances were considerable, van Genuchten/Parker and interbed fracture, the differences between the two models were fairly insignificant. Overall, this study suggests that explicit representation of the stratigraphy in the WIPP PA models is not required for the parameter variations modeled if ``global quantities`` (e.g., disposal room pressures, net brine and gas flux into and out of disposal rooms) are the only concern.

  15. Assessment of flood susceptible areas using spatially explicit, probabilistic multi-criteria decision analysis

    Science.gov (United States)

    Tang, Zhongqian; Zhang, Hua; Yi, Shanzhen; Xiao, Yangfan

    2018-03-01

    GIS-based multi-criteria decision analysis (MCDA) is increasingly used to support flood risk assessment. However, conventional GIS-MCDA methods fail to adequately represent spatial variability and are accompanied with considerable uncertainty. It is, thus, important to incorporate spatial variability and uncertainty into GIS-based decision analysis procedures. This research develops a spatially explicit, probabilistic GIS-MCDA approach for the delineation of potentially flood susceptible areas. The approach integrates the probabilistic and the local ordered weighted averaging (OWA) methods via Monte Carlo simulation, to take into account the uncertainty related to criteria weights, spatial heterogeneity of preferences and the risk attitude of the analyst. The approach is applied to a pilot study for the Gucheng County, central China, heavily affected by the hazardous 2012 flood. A GIS database of six geomorphological and hydrometeorological factors for the evaluation of susceptibility was created. Moreover, uncertainty and sensitivity analysis were performed to investigate the robustness of the model. The results indicate that the ensemble method improves the robustness of the model outcomes with respect to variation in criteria weights and identifies which criteria weights are most responsible for the variability of model outcomes. Therefore, the proposed approach is an improvement over the conventional deterministic method and can provides a more rational, objective and unbiased tool for flood susceptibility evaluation.

  16. Evaluation and comparison of models and modelling tools simulating nitrogen processes in treatment wetlands

    DEFF Research Database (Denmark)

    Edelfeldt, Stina; Fritzson, Peter

    2008-01-01

    with Modelica 2.1 (Wiley-IEEE Press, USA, 2004).] and an associated tool. The differences and similarities between the MathModelica Model Editor and three other ecological modelling tools have also been evaluated. The results show that the models can well be modelled and simulated in the MathModelica Model...... Editor, and that nitrogen decrease in a constructed treatment wetland should be described and simulated using the Nitrification/Denitrification model as this model has the highest overall quality score and provides a more variable environment.......In this paper, two ecological models of nitrogen processes in treatment wetlands have been evaluated and compared. These models were implemented, simulated, and visualized using the Modelica modelling and simulation language [P. Fritzson, Principles of Object-Oriented Modelling and Simulation...

  17. Everyday conceptions of object fall: explicit and tacit understanding during middle childhood.

    Science.gov (United States)

    Howe, Christine; Taylor Tavares, Joana; Devine, Amy

    2012-03-01

    Adults make erroneous predictions about object fall despite recognizing when observed displays are correct or incorrect. Prediction requires explicit engagement with conceptual knowledge, whereas recognition can be achieved through tacit processing. Therefore, it has been suggested that the greater challenge imposed by explicit engagement leads to elements of conceptual understanding being omitted from prediction that are included in recognition. Acknowledging that research with children provides a significant context for exploring this "omission hypothesis" further, this article reports two studies with 6- to 10-year-olds, each of which used prediction and recognition tasks. Study 1 (N=137) focused on understanding of direction of fall, and Study 2 (N=133) addressed speed. Although performance on the recognition tasks was generally superior to performance on the prediction tasks, qualitative differences also emerged. These differences argue against interpreting explicit level understanding purely in terms of omission of tacit constructs, and the article outlines alternative models that may account for the data. Copyright © 2011 Elsevier Inc. All rights reserved.

  18. Diagnosis of dynamic systems based on explicit and implicit behavioural models: an application to gas turbines in Esprit Project Tiger

    Energy Technology Data Exchange (ETDEWEB)

    Trave-Massuyes, L. [Centre National de la Recherche Scientifique (CNRS), 31 - Toulouse (France); Milne, R.

    1995-12-31

    We are interested in the monitoring and diagnosis of dynamic systems. In our work, we are combining explicit temporal models of the behaviour of a dynamic system with implicit behavioural models supporting model based approaches. This work is drive by the needs of and applied to, two gas turbines of very different size and power. In this paper we describe the problems of building systems for these domains and illustrate how we have developed a system where these two approaches complement each other to provide a comprehensive fault detection and diagnosis system. We also explore the strengths and weaknesses of each approach. The work described here is currently working continuously, on line to a gas turbine in a major chemical plant. (author) 24 refs.

  19. Diagnosis of dynamic systems based on explicit and implicit behavioural models: an application to gas turbines in Esprit Project Tiger

    Energy Technology Data Exchange (ETDEWEB)

    Trave-Massuyes, L [Centre National de la Recherche Scientifique (CNRS), 31 - Toulouse (France); Milne, R

    1996-12-31

    We are interested in the monitoring and diagnosis of dynamic systems. In our work, we are combining explicit temporal models of the behaviour of a dynamic system with implicit behavioural models supporting model based approaches. This work is drive by the needs of and applied to, two gas turbines of very different size and power. In this paper we describe the problems of building systems for these domains and illustrate how we have developed a system where these two approaches complement each other to provide a comprehensive fault detection and diagnosis system. We also explore the strengths and weaknesses of each approach. The work described here is currently working continuously, on line to a gas turbine in a major chemical plant. (author) 24 refs.

  20. Tailored graph ensembles as proxies or null models for real networks I: tools for quantifying structure

    International Nuclear Information System (INIS)

    Annibale, A; Coolen, A C C; Fernandes, L P; Fraternali, F; Kleinjung, J

    2009-01-01

    We study the tailoring of structured random graph ensembles to real networks, with the objective of generating precise and practical mathematical tools for quantifying and comparing network topologies macroscopically, beyond the level of degree statistics. Our family of ensembles can produce graphs with any prescribed degree distribution and any degree-degree correlation function; its control parameters can be calculated fully analytically, and as a result we can calculate (asymptotically) formulae for entropies and complexities and for information-theoretic distances between networks, expressed directly and explicitly in terms of their measured degree distribution and degree correlations.

  1. Constant pH molecular dynamics of proteins in explicit solvent with proton tautomerism.

    Science.gov (United States)

    Goh, Garrett B; Hulbert, Benjamin S; Zhou, Huiqing; Brooks, Charles L

    2014-07-01

    pH is a ubiquitous regulator of biological activity, including protein-folding, protein-protein interactions, and enzymatic activity. Existing constant pH molecular dynamics (CPHMD) models that were developed to address questions related to the pH-dependent properties of proteins are largely based on implicit solvent models. However, implicit solvent models are known to underestimate the desolvation energy of buried charged residues, increasing the error associated with predictions that involve internal ionizable residue that are important in processes like hydrogen transport and electron transfer. Furthermore, discrete water and ions cannot be modeled in implicit solvent, which are important in systems like membrane proteins and ion channels. We report on an explicit solvent constant pH molecular dynamics framework based on multi-site λ-dynamics (CPHMD(MSλD)). In the CPHMD(MSλD) framework, we performed seamless alchemical transitions between protonation and tautomeric states using multi-site λ-dynamics, and designed novel biasing potentials to ensure that the physical end-states are predominantly sampled. We show that explicit solvent CPHMD(MSλD) simulations model realistic pH-dependent properties of proteins such as the Hen-Egg White Lysozyme (HEWL), binding domain of 2-oxoglutarate dehydrogenase (BBL) and N-terminal domain of ribosomal protein L9 (NTL9), and the pKa predictions are in excellent agreement with experimental values, with a RMSE ranging from 0.72 to 0.84 pKa units. With the recent development of the explicit solvent CPHMD(MSλD) framework for nucleic acids, accurate modeling of pH-dependent properties of both major class of biomolecules-proteins and nucleic acids is now possible. © 2013 Wiley Periodicals, Inc.

  2. Towards Linking 3D SAR and Lidar Models with a Spatially Explicit Individual Based Forest Model

    Science.gov (United States)

    Osmanoglu, B.; Ranson, J.; Sun, G.; Armstrong, A. H.; Fischer, R.; Huth, A.

    2017-12-01

    In this study, we present a parameterization of the FORMIND individual-based gap model (IBGM)for old growth Atlantic lowland rainforest in La Selva, Costa Rica for the purpose of informing multisensor remote sensing techniques for above ground biomass techniques. The model was successfully parameterized and calibrated for the study site; results show that the simulated forest reproduces the structural complexity of Costa Rican rainforest based on comparisons with CARBONO inventory plot data. Though the simulated stem numbers (378) slightly underestimated the plot data (418), particularly for canopy dominant intermediate shade tolerant trees and shade tolerant understory trees, overall there was a 9.7% difference. Aboveground biomass (kg/ha) showed a 0.1% difference between the simulated forest and inventory plot dataset. The Costa Rica FORMIND simulation was then used to parameterize a spatially explicit (3D) SAR and lidar backscatter models. The simulated forest stands were used to generate a Look Up Table as a tractable means to estimate aboveground forest biomass for these complex forests. Various combinations of lidar and radar variables were evaluated in the LUT inversion. To test the capability of future data for estimation of forest height and biomass, we considered data of 1) L- (or P-) band polarimetric data (backscattering coefficients of HH, HV and VV); 2) L-band dual-pol repeat-pass InSAR data (HH/HV backscattering coefficients and coherences, height of scattering phase center at HH and HV using DEM or surface height from lidar data as reference); 3) P-band polarimetric InSAR data (canopy height from inversion of PolInSAR data or use the coherences and height of scattering phase center at HH, HV and VV); 4) various height indices from waveform lidar data); and 5) surface and canopy top height from photon-counting lidar data. The methods for parameterizing the remote sensing models with the IBGM and developing Look Up Tables will be discussed. Results

  3. Effects of stand composition and thinning in mixed-species forests : a modeling approach applied to Douglas-fir and beech

    NARCIS (Netherlands)

    Bartelink, H.H.

    2000-01-01

    Models estimating growth and yield of forest stands provide important tools for forest management. Pure stands have been modeled extensively and successfully for decades; however, relatively few models for mixed-species stands have been developed. A spatially explicit, mechanistic model (COMMIX) is

  4. Three Software Tools for Viewing Sectional Planes, Volume Models, and Surface Models of a Cadaver Hand.

    Science.gov (United States)

    Chung, Beom Sun; Chung, Min Suk; Shin, Byeong Seok; Kwon, Koojoo

    2018-02-19

    The hand anatomy, including the complicated hand muscles, can be grasped by using computer-assisted learning tools with high quality two-dimensional images and three-dimensional models. The purpose of this study was to present up-to-date software tools that promote learning of stereoscopic morphology of the hand. On the basis of horizontal sectioned images and outlined images of a male cadaver, vertical planes, volume models, and surface models were elaborated. Software to browse pairs of the sectioned and outlined images in orthogonal planes and software to peel and rotate the volume models, as well as a portable document format (PDF) file to select and rotate the surface models, were produced. All of the software tools were downloadable free of charge and usable off-line. The three types of tools for viewing multiple aspects of the hand could be adequately employed according to individual needs. These new tools involving the realistic images of a cadaver and the diverse functions are expected to improve comprehensive knowledge of the hand shape. © 2018 The Korean Academy of Medical Sciences.

  5. Modeling Implicit and Explicit Memory.

    NARCIS (Netherlands)

    Raaijmakers, J.G.W.; Ohta, N.; Izawa, C.

    2005-01-01

    Mathematical models of memory are useful for describing basic processes of memory in a way that enables generalization across a number of experimental paradigms. Models that have these characteristics do not just engage in empirical curve-fitting, but may also provide explanations for puzzling

  6. Explicit and implicit reinforcement learning across the psychosis spectrum.

    Science.gov (United States)

    Barch, Deanna M; Carter, Cameron S; Gold, James M; Johnson, Sheri L; Kring, Ann M; MacDonald, Angus W; Pizzagalli, Diego A; Ragland, J Daniel; Silverstein, Steven M; Strauss, Milton E

    2017-07-01

    Motivational and hedonic impairments are core features of a variety of types of psychopathology. An important aspect of motivational function is reinforcement learning (RL), including implicit (i.e., outside of conscious awareness) and explicit (i.e., including explicit representations about potential reward associations) learning, as well as both positive reinforcement (learning about actions that lead to reward) and punishment (learning to avoid actions that lead to loss). Here we present data from paradigms designed to assess both positive and negative components of both implicit and explicit RL, examine performance on each of these tasks among individuals with schizophrenia, schizoaffective disorder, and bipolar disorder with psychosis, and examine their relative relationships to specific symptom domains transdiagnostically. None of the diagnostic groups differed significantly from controls on the implicit RL tasks in either bias toward a rewarded response or bias away from a punished response. However, on the explicit RL task, both the individuals with schizophrenia and schizoaffective disorder performed significantly worse than controls, but the individuals with bipolar did not. Worse performance on the explicit RL task, but not the implicit RL task, was related to worse motivation and pleasure symptoms across all diagnostic categories. Performance on explicit RL, but not implicit RL, was related to working memory, which accounted for some of the diagnostic group differences. However, working memory did not account for the relationship of explicit RL to motivation and pleasure symptoms. These findings suggest transdiagnostic relationships across the spectrum of psychotic disorders between motivation and pleasure impairments and explicit RL. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  7. Co-occurrence of social anxiety and depression symptoms in adolescence : differential links with implicit and explicit self-esteem?

    NARCIS (Netherlands)

    de Jong, P. J.; Sportel, B. E.; de Hullu, E.; Nauta, M. H.

    Background. Social anxiety and depression often co-occur. As low self-esteem has been identified as a risk factor for both types of symptoms, it may help to explain their co-morbidity. Current dual process models of psychopathology differentiate between explicit and implicit self-esteem. Explicit

  8. Implicit and explicit prosocial motivation as antecedents of volunteering : The moderating role of parenthood

    NARCIS (Netherlands)

    Aydinli, A.; Bender, M.; Chasiotis, A.; van de Vijver, F.J.R.; Cemalcilar, Z.

    2015-01-01

    We test a model proposing that having children influences motivational pathways of volunteering in the same way across two different cultures. The model posits that parents’ engagement in volunteering is driven by implicit, whereas non-parents’ engagement in volunteering is related to explicit

  9. Tool Efficiency Analysis model research in SEMI industry

    Directory of Open Access Journals (Sweden)

    Lei Ma

    2018-01-01

    Full Text Available One of the key goals in SEMI industry is to improve equipment through put and ensure equipment production efficiency maximization. This paper is based on SEMI standards in semiconductor equipment control, defines the transaction rules between different tool states,and presents a TEA system model which is to analysis tool performance automatically based on finite state machine. The system was applied to fab tools and verified its effectiveness successfully, and obtained the parameter values used to measure the equipment performance, also including the advices of improvement.

  10. Spatially explicit models of full-season productivity and implications for landscape management of Golden-winged Warblers in the western Great Lakes Region: Chapter 9

    Science.gov (United States)

    Peterson, Sean M.; Streby, Henry M.; Andersen, David E.

    2016-01-01

    The relationship between landscape structure and composition and full-season productivity (FSP) is poorly understood for most birds. For species of high conservation concern, insight into how productivity is related to landscape structure and composition can be used to develop more effective conservation strategies that increase recruitment. We monitored nest productivity and fledgling survival of Golden-winged Warblers (Vermivora chrysoptera), a species of high conservation concern, in managed forest landscapes at two sites in northern Minnesota, and one site in southeastern Manitoba, Canada from 2010 to 2012. We used logistic exposure models to identify the influence of landscape structure and composition on nest productivity and fledgling survival. We used the models to predict spatially explicit, FSP across our study sites to identify areas of low relative productivity that could be targeted for management. We then used our models of spatially explicit, FSP to simulate the impact of potential management actions on our study sites with the goal of increasing total population productivity. Unlike previous studies that suggested wetland cover types provide higher quality breeding habitat for Golden-winged Warblers, our models predicted 14% greater productivity in upland cover types. Simulated succession of a 9-ha grassland patch to a shrubby upland suitable for nesting increased the total number of fledglings produced by that patch and adjacent upland shrublands by 30%, despite decreasing individual productivity by 13%. Further simulated succession of the same patch described above into deciduous forest reduced the total number of fledglings produced to independence on a landscape by 18% because of a decrease in the area available for nesting. Simulated reduction in the cumulative length of shrubby edge within a 50-m radius of any location in our landscapes from 0.6 to 0.3 km increased FSP by 5%. Our models demonstrated that the effects of any single management

  11. Application of parameters space analysis tools for empirical model validation

    Energy Technology Data Exchange (ETDEWEB)

    Paloma del Barrio, E. [LEPT-ENSAM UMR 8508, Talence (France); Guyon, G. [Electricite de France, Moret-sur-Loing (France)

    2004-01-01

    A new methodology for empirical model validation has been proposed in the framework of the Task 22 (Building Energy Analysis Tools) of the International Energy Agency. It involves two main steps: checking model validity and diagnosis. Both steps, as well as the underlying methods, have been presented in the first part of the paper. In this part, they are applied for testing modelling hypothesis in the framework of the thermal analysis of an actual building. Sensitivity analysis tools have been first used to identify the parts of the model that can be really tested on the available data. A preliminary diagnosis is then supplied by principal components analysis. Useful information for model behaviour improvement has been finally obtained by optimisation techniques. This example of application shows how model parameters space analysis is a powerful tool for empirical validation. In particular, diagnosis possibilities are largely increased in comparison with residuals analysis techniques. (author)

  12. More explicit communication after classroom-based crew resource management training: results of a pragmatic trial.

    Science.gov (United States)

    Verbeek-van Noord, Inge; de Bruijne, Martine C; Twisk, Jos W R; van Dyck, Cathy; Wagner, Cordula

    2015-02-01

    Aviation-based crew resource management trainings to optimize non-technical skills among professionals are often suggested for health care as a way to increase patient safety. Our aim was to evaluate the effect of a 2-day classroom-based crew resource management (CRM) training at emergency departments (EDs) on explicit professional oral communication (EPOC; non-technical skills). A pragmatic controlled before-after trial was conducted. Four EDs of general teaching hospitals were recruited (two intervention and two control departments). ED nurses and ED doctors were observed on their non-technical skills by means of a validated observation tool (EPOC). Our main outcome measure was the amount of EPOC observed per interaction in 30 minutes direct observations. Three outcome measures from EPOC were analysed: human interaction, anticipation on environment and an overall EPOC score. Linear and logistic mixed model analyses were performed. Models were corrected for the outcome measurement at baseline, days between training and observation, patient safety culture and error management culture at baseline. A statistically significant increase after the training was found on human interaction (β=0.27, 95% CI 0.08-0.49) and the overall EPOC score (β=0.25, 95% CI 0.06-0.43), but not for anticipation on environment (OR=1.19, 95% CI .45-3.15). This means that approximately 25% more explicit communication was shown after CRM training. We found an increase in the use of CRM skills after classroom-based crew resource management training. This study adds to the body of evidence that CRM trainings have the potential to increase patient safety by reducing communication flaws, which play an important role in health care-related adverse events. © 2014 John Wiley & Sons, Ltd.

  13. Limitations Of The Current State Space Modelling Approach In Multistage Machining Processes Due To Operation Variations

    Science.gov (United States)

    Abellán-Nebot, J. V.; Liu, J.; Romero, F.

    2009-11-01

    The State Space modelling approach has been recently proposed as an engineering-driven technique for part quality prediction in Multistage Machining Processes (MMP). Current State Space models incorporate fixture and datum variations in the multi-stage variation propagation, without explicitly considering common operation variations such as machine-tool thermal distortions, cutting-tool wear, cutting-tool deflections, etc. This paper shows the limitations of the current State Space model through an experimental case study where the effect of the spindle thermal expansion, cutting-tool flank wear and locator errors are introduced. The paper also discusses the extension of the current State Space model to include operation variations and its potential benefits.

  14. Measuring Explicit Word Learning of Preschool Children: A Development Study.

    Science.gov (United States)

    Kelley, Elizabeth Spencer

    2017-08-15

    The purpose of this article is to present preliminary results related to the development of a new measure of explicit word learning. The measure incorporated elements of explicit vocabulary instruction and dynamic assessment and was designed to be sensitive to differences in word learning skill and to be feasible for use in clinical settings. The explicit word learning measure included brief teaching trials and repeated fine-grained measurement of semantic knowledge and production of 3 novel words (2 verbs and 1 adjective). Preschool children (N = 23) completed the measure of explicit word learning; standardized, norm-referenced measures of expressive and receptive vocabulary; and an incidental word learning task. The measure of explicit word learning provided meaningful information about word learning. Performance on the explicit measure was related to existing vocabulary knowledge and incidental word learning. Findings from this development study indicate that further examination of the measure of explicit word learning is warranted. The measure may have the potential to identify children who are poor word learners. https://doi.org/10.23641/asha.5170738.

  15. Towards anatomic scale agent-based modeling with a massively parallel spatially explicit general-purpose model of enteric tissue (SEGMEnT_HPC).

    Science.gov (United States)

    Cockrell, Robert Chase; Christley, Scott; Chang, Eugene; An, Gary

    2015-01-01

    Perhaps the greatest challenge currently facing the biomedical research community is the ability to integrate highly detailed cellular and molecular mechanisms to represent clinical disease states as a pathway to engineer effective therapeutics. This is particularly evident in the representation of organ-level pathophysiology in terms of abnormal tissue structure, which, through histology, remains a mainstay in disease diagnosis and staging. As such, being able to generate anatomic scale simulations is a highly desirable goal. While computational limitations have previously constrained the size and scope of multi-scale computational models, advances in the capacity and availability of high-performance computing (HPC) resources have greatly expanded the ability of computational models of biological systems to achieve anatomic, clinically relevant scale. Diseases of the intestinal tract are exemplary examples of pathophysiological processes that manifest at multiple scales of spatial resolution, with structural abnormalities present at the microscopic, macroscopic and organ-levels. In this paper, we describe a novel, massively parallel computational model of the gut, the Spatially Explicitly General-purpose Model of Enteric Tissue_HPC (SEGMEnT_HPC), which extends an existing model of the gut epithelium, SEGMEnT, in order to create cell-for-cell anatomic scale simulations. We present an example implementation of SEGMEnT_HPC that simulates the pathogenesis of ileal pouchitis, and important clinical entity that affects patients following remedial surgery for ulcerative colitis.

  16. Analyzing key constraints to biogas production from crop residues and manure in the EU—A spatially explicit model

    Science.gov (United States)

    Persson, U. Martin

    2017-01-01

    This paper presents a spatially explicit method for making regional estimates of the potential for biogas production from crop residues and manure, accounting for key technical, biochemical, environmental and economic constraints. Methods for making such estimates are important as biofuels from agricultural residues are receiving increasing policy support from the EU and major biogas producers, such as Germany and Italy, in response to concerns over unintended negative environmental and social impacts of conventional biofuels. This analysis comprises a spatially explicit estimate of crop residue and manure production for the EU at 250 m resolution, and a biogas production model accounting for local constraints such as the sustainable removal of residues, transportation of substrates, and the substrates’ biochemical suitability for anaerobic digestion. In our base scenario, the EU biogas production potential from crop residues and manure is about 0.7 EJ/year, nearly double the current EU production of biogas from agricultural substrates, most of which does not come from residues or manure. An extensive sensitivity analysis of the model shows that the potential could easily be 50% higher or lower, depending on the stringency of economic, technical and biochemical constraints. We find that the potential is particularly sensitive to constraints on the substrate mixtures’ carbon-to-nitrogen ratio and dry matter concentration. Hence, the potential to produce biogas from crop residues and manure in the EU depends to large extent on the possibility to overcome the challenges associated with these substrates, either by complementing them with suitable co-substrates (e.g. household waste and energy crops), or through further development of biogas technology (e.g. pretreatment of substrates and recirculation of effluent). PMID:28141827

  17. Analyzing key constraints to biogas production from crop residues and manure in the EU-A spatially explicit model.

    Science.gov (United States)

    Einarsson, Rasmus; Persson, U Martin

    2017-01-01

    This paper presents a spatially explicit method for making regional estimates of the potential for biogas production from crop residues and manure, accounting for key technical, biochemical, environmental and economic constraints. Methods for making such estimates are important as biofuels from agricultural residues are receiving increasing policy support from the EU and major biogas producers, such as Germany and Italy, in response to concerns over unintended negative environmental and social impacts of conventional biofuels. This analysis comprises a spatially explicit estimate of crop residue and manure production for the EU at 250 m resolution, and a biogas production model accounting for local constraints such as the sustainable removal of residues, transportation of substrates, and the substrates' biochemical suitability for anaerobic digestion. In our base scenario, the EU biogas production potential from crop residues and manure is about 0.7 EJ/year, nearly double the current EU production of biogas from agricultural substrates, most of which does not come from residues or manure. An extensive sensitivity analysis of the model shows that the potential could easily be 50% higher or lower, depending on the stringency of economic, technical and biochemical constraints. We find that the potential is particularly sensitive to constraints on the substrate mixtures' carbon-to-nitrogen ratio and dry matter concentration. Hence, the potential to produce biogas from crop residues and manure in the EU depends to large extent on the possibility to overcome the challenges associated with these substrates, either by complementing them with suitable co-substrates (e.g. household waste and energy crops), or through further development of biogas technology (e.g. pretreatment of substrates and recirculation of effluent).

  18. Analyzing key constraints to biogas production from crop residues and manure in the EU-A spatially explicit model.

    Directory of Open Access Journals (Sweden)

    Rasmus Einarsson

    Full Text Available This paper presents a spatially explicit method for making regional estimates of the potential for biogas production from crop residues and manure, accounting for key technical, biochemical, environmental and economic constraints. Methods for making such estimates are important as biofuels from agricultural residues are receiving increasing policy support from the EU and major biogas producers, such as Germany and Italy, in response to concerns over unintended negative environmental and social impacts of conventional biofuels. This analysis comprises a spatially explicit estimate of crop residue and manure production for the EU at 250 m resolution, and a biogas production model accounting for local constraints such as the sustainable removal of residues, transportation of substrates, and the substrates' biochemical suitability for anaerobic digestion. In our base scenario, the EU biogas production potential from crop residues and manure is about 0.7 EJ/year, nearly double the current EU production of biogas from agricultural substrates, most of which does not come from residues or manure. An extensive sensitivity analysis of the model shows that the potential could easily be 50% higher or lower, depending on the stringency of economic, technical and biochemical constraints. We find that the potential is particularly sensitive to constraints on the substrate mixtures' carbon-to-nitrogen ratio and dry matter concentration. Hence, the potential to produce biogas from crop residues and manure in the EU depends to large extent on the possibility to overcome the challenges associated with these substrates, either by complementing them with suitable co-substrates (e.g. household waste and energy crops, or through further development of biogas technology (e.g. pretreatment of substrates and recirculation of effluent.

  19. Explicit mentalizing mechanisms and their adaptive role in memory conformity.

    Directory of Open Access Journals (Sweden)

    Rebecca Wheeler

    Full Text Available Memory conformity occurs when an individual endorses what other individuals remember about past events. Research on memory conformity is currently dominated by a 'forensic' perspective, which views the phenomenon as inherently undesirable. This is because conformity not only distorts the accuracy of an individual's memory, but also produces false corroboration between individuals, effects that act to undermine criminal justice systems. There is growing awareness, however, that memory conformity may be interpreted more generally as an adaptive social behavior regulated by explicit mentalizing mechanisms. Here, we provide novel evidence in support of this emerging alternative theoretical perspective. We carried out a memory conformity experiment which revealed that explicit belief-simulation (i.e. using one's own beliefs to model what other people believe systematically biases conformity towards like-minded individuals, even when there is no objective evidence that they have a more accurate memory than dissimilar individuals. We suggest that this bias is functional, i.e. adaptive, to the extent that it fosters trust, and hence cooperation, between in-group versus out-group individuals. We conclude that memory conformity is, in more fundamental terms, a highly desirable product of explicit mentalizing mechanisms that promote adaptive forms of social learning and cooperation.

  20. Explicit Supersymmetry Breaking on Boundaries of Warped Extra Dimensions

    Energy Technology Data Exchange (ETDEWEB)

    Hall, Lawrence J.; Nomura, Yasunori; Okui, Takemichi; Oliver, Steven J.

    2003-02-25

    Explicit supersymmetry breaking is studied in higher dimensional theories by having boundaries respect only a subgroup of the bulk symmetry. If the boundary symmetry is the maximal subgroup allowed by the boundary conditions imposed on the fields, then the symmetry can be consistently gauged; otherwise gauging leads to an inconsistent theory. In a warped fifth dimension, an explicit breaking of all bulk supersymmetries by the boundaries is found to be inconsistent with gauging; unlike the case of flat 5D, complete supersymmetry breaking by boundary conditions is not consistent with supergravity. Despite this result, the low energy effective theory resulting from boundary supersymmetry breaking becomes consistent in the limit where gravity decouples, and such models are explored in the hope that some way of successfully incorporating gravity can be found. A warped constrained standard model leads to a theory with one Higgs boson with mass expected close to the experimental limit. A unified theory in a warped fifth dimension is studied with boundary breaking of both SU(5) gauge symmetry and supersymmetry. The usual supersymmetric predictionfor gauge coupling unification holds even though the TeV spectrum is quite unlike the MSSM. Such a theory may unify matter and Higgs in the same SU(5) hypermultiplet.

  1. Modeling, methodologies and tools for molecular and nano-scale communications modeling, methodologies and tools

    CERN Document Server

    Nakano, Tadashi; Moore, Michael

    2017-01-01

    (Preliminary) The book presents the state of art in the emerging field of molecular and nanoscale communication. It gives special attention to fundamental models, and advanced methodologies and tools used in the field. It covers a wide range of applications, e.g. nanomedicine, nanorobot communication, bioremediation and environmental managements. It addresses advanced graduate students, academics and professionals working at the forefront in their fields and at the interfaces between different areas of research, such as engineering, computer science, biology and nanotechnology.

  2. Aligning building information model tools and construction management methods

    NARCIS (Netherlands)

    Hartmann, Timo; van Meerveld, H.J.; Vossebeld, N.; Adriaanse, Adriaan Maria

    2012-01-01

    Few empirical studies exist that can explain how different Building Information Model (BIM) based tool implementation strategies work in practical contexts. To help overcoming this gap, this paper describes the implementation of two BIM based tools, the first, to support the activities at an

  3. WMT: The CSDMS Web Modeling Tool

    Science.gov (United States)

    Piper, M.; Hutton, E. W. H.; Overeem, I.; Syvitski, J. P.

    2015-12-01

    The Community Surface Dynamics Modeling System (CSDMS) has a mission to enable model use and development for research in earth surface processes. CSDMS strives to expand the use of quantitative modeling techniques, promotes best practices in coding, and advocates for the use of open-source software. To streamline and standardize access to models, CSDMS has developed the Web Modeling Tool (WMT), a RESTful web application with a client-side graphical interface and a server-side database and API that allows users to build coupled surface dynamics models in a web browser on a personal computer or a mobile device, and run them in a high-performance computing (HPC) environment. With WMT, users can: Design a model from a set of components Edit component parameters Save models to a web-accessible server Share saved models with the community Submit runs to an HPC system Download simulation results The WMT client is an Ajax application written in Java with GWT, which allows developers to employ object-oriented design principles and development tools such as Ant, Eclipse and JUnit. For deployment on the web, the GWT compiler translates Java code to optimized and obfuscated JavaScript. The WMT client is supported on Firefox, Chrome, Safari, and Internet Explorer. The WMT server, written in Python and SQLite, is a layered system, with each layer exposing a web service API: wmt-db: database of component, model, and simulation metadata and output wmt-api: configure and connect components wmt-exe: launch simulations on remote execution servers The database server provides, as JSON-encoded messages, the metadata for users to couple model components, including descriptions of component exchange items, uses and provides ports, and input parameters. Execution servers are network-accessible computational resources, ranging from HPC systems to desktop computers, containing the CSDMS software stack for running a simulation. Once a simulation completes, its output, in NetCDF, is packaged

  4. Introduction to genetic algorithms as a modeling tool

    International Nuclear Information System (INIS)

    Wildberger, A.M.; Hickok, K.A.

    1990-01-01

    Genetic algorithms are search and classification techniques modeled on natural adaptive systems. This is an introduction to their use as a modeling tool with emphasis on prospects for their application in the power industry. It is intended to provide enough background information for its audience to begin to follow technical developments in genetic algorithms and to recognize those which might impact on electric power engineering. Beginning with a discussion of genetic algorithms and their origin as a model of biological adaptation, their advantages and disadvantages are described in comparison with other modeling tools such as simulation and neural networks in order to provide guidance in selecting appropriate applications. In particular, their use is described for improving expert systems from actual data and they are suggested as an aid in building mathematical models. Using the Thermal Performance Advisor as an example, it is suggested how genetic algorithms might be used to make a conventional expert system and mathematical model of a power plant adapt automatically to changes in the plant's characteristics

  5. Decadal shifts of East Asian summer monsoon in a climate model free of explicit GHGs and aerosols

    Science.gov (United States)

    Lin, Renping; Zhu, Jiang; Zheng, Fei

    2016-12-01

    The East Asian summer monsoon (EASM) experienced decadal transitions over the past few decades, and the associated "wetter-South-drier-North" shifts in rainfall patterns in China significantly affected the social and economic development in China. Two viewpoints stand out to explain these decadal shifts, regarding the shifts either a result of internal variability of climate system or that of external forcings (e.g. greenhouse gases (GHGs) and anthropogenic aerosols). However, most climate models, for example, the Atmospheric Model Intercomparison Project (AMIP)-type simulations and the Coupled Model Intercomparison Project (CMIP)-type simulations, fail to simulate the variation patterns, leaving the mechanisms responsible for these shifts still open to dispute. In this study, we conducted a successful simulation of these decadal transitions in a coupled model where we applied ocean data assimilation in the model free of explicit aerosols and GHGs forcing. The associated decadal shifts of the three-dimensional spatial structure in the 1990s, including the eastward retreat, the northward shift of the western Pacific subtropical high (WPSH), and the south-cool-north-warm pattern of the upper-level tropospheric temperature, were all well captured. Our simulation supports the argument that the variations of the oceanic fields are the dominant factor responsible for the EASM decadal transitions.

  6. Adaptation in integrated assessment modeling: where do we stand?

    NARCIS (Netherlands)

    Patt, A.; van Vuuren, D.P.; Berkhout, F.G.H.; Aaheim, A.; Hof, A.F.; Isaac, M.; Mechler, R.

    2010-01-01

    Adaptation is an important element on the climate change policy agenda. Integrated assessment models, which are key tools to assess climate change policies, have begun to address adaptation, either by including it implicitly in damage cost estimates, or by making it an explicit control variable. We

  7. Surviving the present: Modeling tools for organizational change

    International Nuclear Information System (INIS)

    Pangaro, P.

    1992-01-01

    The nuclear industry, like the rest of modern American business, is beset by a confluence of economic, technological, competitive, regulatory, and political pressures. For better or worse, business schools and management consultants have leapt to the rescue, offering the most modern conveniences that they can purvey. Recent advances in the study of organizations have led to new tools for their analysis, revision, and repair. There are two complementary tools that do not impose values or injunctions in themselves. One, called the organization modeler, captures the hierarchy of purposes that organizations and their subparts carry out. Any deficiency or pathology is quickly illuminated, and requirements for repair are made clear. The second, called THOUGHTSTICKER, is used to capture the semantic content of the conversations that occur across the interactions of parts of an organization. The distinctions and vocabulary in the language of an organization, and the relations within that domain, are elicited from the participants so that all three are available for debate and refinement. The product of the applications of these modeling tools is not the resulting models but rather the enhancement of the organization as a consequence of the process of constructing them

  8. Implicit vs. Explicit Trust in Social Matrix Factorization

    NARCIS (Netherlands)

    Fazeli, Soude; Loni, Babak; Bellogin, Alejandro; Drachsler, Hendrik; Sloep, Peter

    2014-01-01

    Incorporating social trust in Matrix Factorization (MF) methods demonstrably improves accuracy of rating prediction. Such approaches mainly use the trust scores explicitly expressed by users. However, it is often challenging to have users provide explicit trust scores of each other. There exist

  9. Using the IEA ETSAP modelling tools for Denmark

    Energy Technology Data Exchange (ETDEWEB)

    Grohnheit, Poul Erik

    2008-12-15

    An important part of the cooperation within the IEA (International Energy Agency) is organised through national contributions to 'Implementation Agreements' on energy technology and energy analyses. One of them is ETSAP (Energy Technology Systems Analysis Programme), started in 1976. Denmark has signed the agreement and contributed to some early annexes. This project is motivated by an invitation to participate in ETSAP Annex X, 'Global Energy Systems and Common Analyses: Climate friendly, Secure and Productive Energy Systems' for the period 2005 to 2007. The main activity is semi-annual workshops focusing on presentations of model analyses and use of the ETSAP tools (the MARKAL/TIMES family of models). The project was also planned to benefit from the EU project 'NEEDS - New Energy Externalities Developments for Sustainability'. ETSAP is contributing to a part of NEEDS that develops the TIMES model for 29 European countries with assessment of future technologies. An additional project 'Monitoring and Evaluation of the RES directives: implementation in EU27 and policy recommendations for 2020' (RES2020) under Intelligent Energy Europe was added, as well as the Danish 'Centre for Energy, Environment and Health (CEEH), starting from January 2007. This report summarises the activities under ETSAP Annex X and related project, emphasising the development of modelling tools that will be useful for modelling the Danish energy system. It is also a status report for the development of a model for Denmark, focusing on the tools and features that allow comparison with other countries and, particularly, to evaluate assumptions and results in international models covering Denmark. (au)

  10. The impact of convection in the West African monsoon region on global weather forecasts - explicit vs. parameterised convection simulations using the ICON model

    Science.gov (United States)

    Pante, Gregor; Knippertz, Peter

    2017-04-01

    The West African monsoon is the driving element of weather and climate during summer in the Sahel region. It interacts with mesoscale convective systems (MCSs) and the African easterly jet and African easterly waves. Poor representation of convection in numerical models, particularly its organisation on the mesoscale, can result in unrealistic forecasts of the monsoon dynamics. Arguably, the parameterisation of convection is one of the main deficiencies in models over this region. Overall, this has negative impacts on forecasts over West Africa itself but may also affect remote regions, as waves originating from convective heating are badly represented. Here we investigate those remote forecast impacts based on daily initialised 10-day forecasts for July 2016 using the ICON model. One set of simulations employs the default setup of the global model with a horizontal grid spacing of 13 km. It is compared with simulations using the 2-way nesting capability of ICON. A second model domain over West Africa (the nest) with 6.5 km grid spacing is sufficient to explicitly resolve MCSs in this region. In the 2-way nested simulations, the prognostic variables of the global model are influenced by the results of the nest through relaxation. The nest with explicit convection is able to reproduce single MCSs much more realistically compared to the stand-alone global simulation with parameterised convection. Explicit convection leads to cooler temperatures in the lower troposphere (below 500 hPa) over the northern Sahel due to stronger evaporational cooling. Overall, the feedback of dynamic variables from the nest to the global model shows clear positive effects when evaluating the output of the global domain of the 2-way nesting simulation and the output of the stand-alone global model with ERA-Interim re-analyses. Averaged over the 2-way nested region, bias and root mean squared error (RMSE) of temperature, geopotential, wind and relative humidity are significantly reduced in

  11. Attitudinal effects of degrading themes and sexual explicitness in video materials.

    Science.gov (United States)

    Golde, J A; Strassberg, D S; Turner, C M; Lowe, K

    2000-07-01

    This study examined the independent and interactive effects of sexual explicitness and degrading themes toward women on mens' attitudes following exposure to video presentations of male-female interactions. Subjects were 83 male college students who viewed video vignettes under one of four stimulus conditions: (a) sexually explicit/degrading, (b) sexually explicit/nondegrading, (c) nonexplicit/degrading, and (d) nonexplicit/nondegrading. Results revealed that men exposed to degrading material, regardless of explicitness, were significantly more likely to express attitudes supportive of rape, while explicitness had no significant main or interactive effect on these attitudes. Further, the interaction of explicitness with degradation was found to impact scores on a measure of sexual callousness. Theoretical and clinical implications of these findings are discussed.

  12. Implicit versus explicit : An ACT-R learning perspective

    NARCIS (Netherlands)

    Taatgen, N.A.

    1999-01-01

    Dienes & Perner propose a theory of implicit and explicit knowledge that is not entirely complete. It does not address many of the empirical issues, nor does it explain the difference between implicit and explicit learning. It does, however, provide a possible unified explanation, as opposed to the

  13. Springback Simulation and Tool Surface Compensation Algorithm for Sheet Metal Forming

    International Nuclear Information System (INIS)

    Shen Guozhe; Hu Ping; Zhang Xiangkui; Chen Xiaobin; Li Xiaoda

    2005-01-01

    Springback is an unquenchable forming defect in the sheet metal forming process. How to calculate springback accurately is a big challenge for a lot of FEA software. Springback compensation makes the stamped final part accordant with the designed part shape by modifying tool surface, which depends on the accurate springback amount. How ever, the meshing data based on numerical simulation is expressed by nodes and elements, such data can not be supplied directly to tool surface CAD data. In this paper, a tool surface compensation algorithm based on numerical simulation technique of springback process is proposed in which the independently developed dynamic explicit springback algorithm (DESA) is used to simulate springback amount. When doing the tool surface compensation, the springback amount of the projected point can be obtained by interpolation of the springback amount of the projected element nodes. So the modified values of tool surface can be calculated reversely. After repeating the springback and compensation calculations for 1∼3 times, the reasonable tool surface mesh is gained. Finally, the FEM data on the compensated tool surface is fitted into the surface by CAD modeling software. The examination of a real industrial part shows the validity of the present method

  14. Explicit modelling of SOA formation from α-pinene photooxidation: sensitivity to vapour pressure estimation

    Directory of Open Access Journals (Sweden)

    R. Valorso

    2011-07-01

    Full Text Available The sensitivity of the formation of secondary organic aerosol (SOA to the estimated vapour pressures of the condensable oxidation products is explored. A highly detailed reaction scheme was generated for α-pinene photooxidation using the Generator for Explicit Chemistry and Kinetics of Organics in the Atmosphere (GECKO-A. Vapour pressures (Pvap were estimated with three commonly used structure activity relationships. The values of Pvap were compared for the set of secondary species generated by GECKO-A to describe α-pinene oxidation. Discrepancies in the predicted vapour pressures were found to increase with the number of functional groups borne by the species. For semi-volatile organic compounds (i.e. organic species of interest for SOA formation, differences in the predicted Pvap range between a factor of 5 to 200 on average. The simulated SOA concentrations were compared to SOA observations in the Caltech chamber during three experiments performed under a range of NOx conditions. While the model captures the qualitative features of SOA formation for the chamber experiments, SOA concentrations are systematically overestimated. For the conditions simulated, the modelled SOA speciation appears to be rather insensitive to the Pvap estimation method.

  15. Open source Modeling and optimization tools for Planning

    Energy Technology Data Exchange (ETDEWEB)

    Peles, S. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2017-02-10

    Open source modeling and optimization tools for planning The existing tools and software used for planning and analysis in California are either expensive, difficult to use, or not generally accessible to a large number of participants. These limitations restrict the availability of participants for larger scale energy and grid studies in the state. The proposed initiative would build upon federal and state investments in open source software, and create and improve open source tools for use in the state planning and analysis activities. Computational analysis and simulation frameworks in development at national labs and universities can be brought forward to complement existing tools. An open source platform would provide a path for novel techniques and strategies to be brought into the larger community and reviewed by a broad set of stakeholders.

  16. Agent-based modeling as a tool for program design and evaluation.

    Science.gov (United States)

    Lawlor, Jennifer A; McGirr, Sara

    2017-12-01

    Recently, systems thinking and systems science approaches have gained popularity in the field of evaluation; however, there has been relatively little exploration of how evaluators could use quantitative tools to assist in the implementation of systems approaches therein. The purpose of this paper is to explore potential uses of one such quantitative tool, agent-based modeling, in evaluation practice. To this end, we define agent-based modeling and offer potential uses for it in typical evaluation activities, including: engaging stakeholders, selecting an intervention, modeling program theory, setting performance targets, and interpreting evaluation results. We provide demonstrative examples from published agent-based modeling efforts both inside and outside the field of evaluation for each of the evaluative activities discussed. We further describe potential pitfalls of this tool and offer cautions for evaluators who may chose to implement it in their practice. Finally, the article concludes with a discussion of the future of agent-based modeling in evaluation practice and a call for more formal exploration of this tool as well as other approaches to simulation modeling in the field. Copyright © 2017 Elsevier Ltd. All rights reserved.

  17. Systematic Methods and Tools for Computer Aided Modelling

    DEFF Research Database (Denmark)

    Fedorova, Marina

    and processes can be faster, cheaper and very efficient. The developed modelling framework involves five main elements: 1) a modelling tool, that includes algorithms for model generation; 2) a template library, which provides building blocks for the templates (generic models previously developed); 3) computer......-format and COM-objects, are incorporated to allow the export and import of mathematical models; 5) a user interface that provides the work-flow and data-flow to guide the user through the different modelling tasks....

  18. Scenario Evaluator for Electrical Resistivity survey pre-modeling tool

    Science.gov (United States)

    Terry, Neil; Day-Lewis, Frederick D.; Robinson, Judith L.; Slater, Lee D.; Halford, Keith J.; Binley, Andrew; Lane, John W.; Werkema, Dale D.

    2017-01-01

    Geophysical tools have much to offer users in environmental, water resource, and geotechnical fields; however, techniques such as electrical resistivity imaging (ERI) are often oversold and/or overinterpreted due to a lack of understanding of the limitations of the techniques, such as the appropriate depth intervals or resolution of the methods. The relationship between ERI data and resistivity is nonlinear; therefore, these limitations depend on site conditions and survey design and are best assessed through forward and inverse modeling exercises prior to field investigations. In this approach, proposed field surveys are first numerically simulated given the expected electrical properties of the site, and the resulting hypothetical data are then analyzed using inverse models. Performing ERI forward/inverse modeling, however, requires substantial expertise and can take many hours to implement. We present a new spreadsheet-based tool, the Scenario Evaluator for Electrical Resistivity (SEER), which features a graphical user interface that allows users to manipulate a resistivity model and instantly view how that model would likely be interpreted by an ERI survey. The SEER tool is intended for use by those who wish to determine the value of including ERI to achieve project goals, and is designed to have broad utility in industry, teaching, and research.

  19. Multiresolution and Explicit Methods for Vector Field Analysis and Visualization

    Science.gov (United States)

    Nielson, Gregory M.

    1997-01-01

    This is a request for a second renewal (3d year of funding) of a research project on the topic of multiresolution and explicit methods for vector field analysis and visualization. In this report, we describe the progress made on this research project during the second year and give a statement of the planned research for the third year. There are two aspects to this research project. The first is concerned with the development of techniques for computing tangent curves for use in visualizing flow fields. The second aspect of the research project is concerned with the development of multiresolution methods for curvilinear grids and their use as tools for visualization, analysis and archiving of flow data. We report on our work on the development of numerical methods for tangent curve computation first.

  20. Piecewise quadratic Lyapunov functions for stability verification of approximate explicit MPC

    Directory of Open Access Journals (Sweden)

    Morten Hovd

    2010-04-01

    Full Text Available Explicit MPC of constrained linear systems is known to result in a piecewise affine controller and therefore also piecewise affine closed loop dynamics. The complexity of such analytic formulations of the control law can grow exponentially with the prediction horizon. The suboptimal solutions offer a trade-off in terms of complexity and several approaches can be found in the literature for the construction of approximate MPC laws. In the present paper a piecewise quadratic (PWQ Lyapunov function is used for the stability verification of an of approximate explicit Model Predictive Control (MPC. A novel relaxation method is proposed for the LMI criteria on the Lyapunov function design. This relaxation is applicable to the design of PWQ Lyapunov functions for discrete-time piecewise affine systems in general.

  1. Guidelines for Reproducibly Building and Simulating Systems Biology Models.

    Science.gov (United States)

    Medley, J Kyle; Goldberg, Arthur P; Karr, Jonathan R

    2016-10-01

    Reproducibility is the cornerstone of the scientific method. However, currently, many systems biology models cannot easily be reproduced. This paper presents methods that address this problem. We analyzed the recent Mycoplasma genitalium whole-cell (WC) model to determine the requirements for reproducible modeling. We determined that reproducible modeling requires both repeatable model building and repeatable simulation. New standards and simulation software tools are needed to enhance and verify the reproducibility of modeling. New standards are needed to explicitly document every data source and assumption, and new deterministic parallel simulation tools are needed to quickly simulate large, complex models. We anticipate that these new standards and software will enable researchers to reproducibly build and simulate more complex models, including WC models.

  2. SEARCH: Spatially Explicit Animal Response to Composition of Habitat.

    Science.gov (United States)

    Pauli, Benjamin P; McCann, Nicholas P; Zollner, Patrick A; Cummings, Robert; Gilbert, Jonathan H; Gustafson, Eric J

    2013-01-01

    Complex decisions dramatically affect animal dispersal and space use. Dispersing individuals respond to a combination of fine-scale environmental stimuli and internal attributes. Individual-based modeling offers a valuable approach for the investigation of such interactions because it combines the heterogeneity of animal behaviors with spatial detail. Most individual-based models (IBMs), however, vastly oversimplify animal behavior and such behavioral minimalism diminishes the value of these models. We present program SEARCH (Spatially Explicit Animal Response to Composition of Habitat), a spatially explicit, individual-based, population model of animal dispersal through realistic landscapes. SEARCH uses values in Geographic Information System (GIS) maps to apply rules that animals follow during dispersal, thus allowing virtual animals to respond to fine-scale features of the landscape and maintain a detailed memory of areas sensed during movement. SEARCH also incorporates temporally dynamic landscapes so that the environment to which virtual animals respond can change during the course of a simulation. Animals in SEARCH are behaviorally dynamic and able to respond to stimuli based upon their individual experiences. Therefore, SEARCH is able to model behavioral traits of dispersing animals at fine scales and with many dynamic aspects. Such added complexity allows investigation of unique ecological questions. To illustrate SEARCH's capabilities, we simulated case studies using three mammals. We examined the impact of seasonally variable food resources on the weight distribution of dispersing raccoons (Procyon lotor), the effect of temporally dynamic mortality pressure in combination with various levels of behavioral responsiveness in eastern chipmunks (Tamias striatus), and the impact of behavioral plasticity and home range selection on disperser mortality and weight change in virtual American martens (Martes americana). These simulations highlight the relevance of

  3. Assessment of wear dependence parameters in complex model of cutting tool wear

    Science.gov (United States)

    Antsev, A. V.; Pasko, N. I.; Antseva, N. V.

    2018-03-01

    This paper addresses wear dependence of the generic efficient life period of cutting tools taken as an aggregate of the law of tool wear rate distribution and dependence of parameters of this law's on the cutting mode, factoring in the random factor as exemplified by the complex model of wear. The complex model of wear takes into account the variance of cutting properties within one batch of tools, variance in machinability within one batch of workpieces, and the stochastic nature of the wear process itself. A technique of assessment of wear dependence parameters in a complex model of cutting tool wear is provided. The technique is supported by a numerical example.

  4. Implicit and explicit memory for spatial information in Alzheimer's disease.

    Science.gov (United States)

    Kessels, R P C; Feijen, J; Postma, A

    2005-01-01

    There is abundant evidence that memory impairment in dementia in patients with Alzheimer's disease (AD) is related to explicit, conscious forms of memory, whereas implicit, unconscious forms of memory function remain relatively intact or are less severely affected. Only a few studies have been performed on spatial memory function in AD, showing that AD patients' explicit spatial memory is impaired, possibly related to hippocampal dysfunction. However, studies on implicit spatial memory in AD are lacking. The current study set out to investigate implicit and explicit spatial memory in AD patients (n=18) using an ecologically valid computer task, in which participants had to remember the locations of various objects in common rooms. The contribution of implicit and explicit memory functions was estimated by means of the process dissociation procedure. The results show that explicit spatial memory is impaired in AD patients compared with a control group (n=21). However, no group difference was found on implicit spatial function. This indicates that spared implicit memory in AD extends to the spatial domain, while the explicit spatial memory function deteriorates. Clinically, this finding might be relevant, in that an intact implicit memory function might be helpful in overcoming problems in explicit processing. Copyright (c) 2005 S. Karger AG, Basel.

  5. Evaluation of clinical information modeling tools.

    Science.gov (United States)

    Moreno-Conde, Alberto; Austin, Tony; Moreno-Conde, Jesús; Parra-Calderón, Carlos L; Kalra, Dipak

    2016-11-01

    Clinical information models are formal specifications for representing the structure and semantics of the clinical content within electronic health record systems. This research aims to define, test, and validate evaluation metrics for software tools designed to support the processes associated with the definition, management, and implementation of these models. The proposed framework builds on previous research that focused on obtaining agreement on the essential requirements in this area. A set of 50 conformance criteria were defined based on the 20 functional requirements agreed by that consensus and applied to evaluate the currently available tools. Of the 11 initiative developing tools for clinical information modeling identified, 9 were evaluated according to their performance on the evaluation metrics. Results show that functionalities related to management of data types, specifications, metadata, and terminology or ontology bindings have a good level of adoption. Improvements can be made in other areas focused on information modeling and associated processes. Other criteria related to displaying semantic relationships between concepts and communication with terminology servers had low levels of adoption. The proposed evaluation metrics were successfully tested and validated against a representative sample of existing tools. The results identify the need to improve tool support for information modeling and software development processes, especially in those areas related to governance, clinician involvement, and optimizing the technical validation of testing processes. This research confirmed the potential of these evaluation metrics to support decision makers in identifying the most appropriate tool for their organization. Los Modelos de Información Clínica son especificaciones para representar la estructura y características semánticas del contenido clínico en los sistemas de Historia Clínica Electrónica. Esta investigación define, prueba y valida

  6. Thermomechanical modelling of laser surface glazing for H13 tool steel

    Science.gov (United States)

    Kabir, I. R.; Yin, D.; Tamanna, N.; Naher, S.

    2018-03-01

    A two-dimensional thermomechanical finite element (FE) model of laser surface glazing (LSG) has been developed for H13 tool steel. The direct coupling technique of ANSYS 17.2 (APDL) has been utilised to solve the transient thermomechanical process. A H13 tool steel cylindrical cross-section has been modelled for laser power 200 W and 300 W at constant 0.2 mm beam width and 0.15 ms residence time. The model can predict temperature distribution, stress-strain increments in elastic and plastic region with time and space. The crack formation tendency also can be assumed by analysing the von Mises stress in the heat-concentrated zone. Isotropic and kinematic hardening models have been applied separately to predict the after-yield phenomena. At 200 W laser power, the peak surface temperature achieved is 1520 K which is below the melting point (1727 K) of H13 tool steel. For laser power 300 W, the peak surface temperature is 2523 K. Tensile residual stresses on surface have been found after cooling, which are in agreement with literature. Isotropic model shows higher residual stress that increases with laser power. Conversely, kinematic model gives lower residual stress which decreases with laser power. Therefore, both plasticity models could work in LSG for H13 tool steel.

  7. Explicit strong stability preserving multistep Runge–Kutta methods

    KAUST Repository

    Bresten, Christopher; Gottlieb, Sigal; Grant, Zachary; Higgs, Daniel; Ketcheson, David I.; Né meth, Adrian

    2015-01-01

    High-order spatial discretizations of hyperbolic PDEs are often designed to have strong stability properties, such as monotonicity. We study explicit multistep Runge-Kutta strong stability preserving (SSP) time integration methods for use with such discretizations. We prove an upper bound on the SSP coefficient of explicit multistep Runge-Kutta methods of order two and above. Numerical optimization is used to find optimized explicit methods of up to five steps, eight stages, and tenth order. These methods are tested on the linear advection and nonlinear Buckley-Leverett equations, and the results for the observed total variation diminishing and/or positivity preserving time-step are presented.

  8. Explicit strong stability preserving multistep Runge–Kutta methods

    KAUST Repository

    Bresten, Christopher

    2015-10-15

    High-order spatial discretizations of hyperbolic PDEs are often designed to have strong stability properties, such as monotonicity. We study explicit multistep Runge-Kutta strong stability preserving (SSP) time integration methods for use with such discretizations. We prove an upper bound on the SSP coefficient of explicit multistep Runge-Kutta methods of order two and above. Numerical optimization is used to find optimized explicit methods of up to five steps, eight stages, and tenth order. These methods are tested on the linear advection and nonlinear Buckley-Leverett equations, and the results for the observed total variation diminishing and/or positivity preserving time-step are presented.

  9. A cross-cultural study of explicit and implicit motivation for long-term volunteering

    NARCIS (Netherlands)

    Aydinli, A.; Bender, M.; Chasiotis, A.; van de Vijver, F.J.R.; Cemalcilar, Z.; Chong, A.; Yue, X.

    2016-01-01

    We propose a model of volunteering and test its validity across four cultural groups. We hypothesize that individuals’ explicit prosocial motivation relates positively to sustained volunteering, which is conceptualized as a latent factor comprising activity as a volunteer, service length, service

  10. Investigating the of Explicit Instruction of Apology Speech Act on Pragmatic Development of Iranian EFL Learners

    Directory of Open Access Journals (Sweden)

    Shima Rajabi

    2015-08-01

    Full Text Available Learning a foreign language requires students to acquire both grammatical knowledge and socio-pragmatic rules of a language. Pragmatic competence as one of the most difficult aspects of language provides several challenges to L2 learners in the process of learning a foreign language. To overcome this problem, EFL teachers should find the most effective way of teaching pragmatic knowledge to their students. Accordingly, the present study investigated the effect of explicit teaching of apology speech act, as an aspect of pragmatic competence, on the Iranian EFL learners’ appropriate use of the mentioned speech act. In so doing, a total of 73 EFL students at intermediate and advanced levels participated in a pre-posttest design research with experimental and control group. Data were collected using a Discourse Completion Test (DCT. The selection of apologetic situations in DCT was based on two variables of social status and social distance. The results revealed that explicit instruction was a facilitative tool that helped students use the proper apology strategies in different situations. Moreover, it was found that L2 proficiency had a significant influence on overall appropriateness of speech act production. Keywords: Explicit instruction; Apology speech act;   Pragmatic competence; Iranian EFL learners

  11. Simulation Tools for Electrical Machines Modelling: Teaching and ...

    African Journals Online (AJOL)

    Simulation tools are used both for research and teaching to allow a good comprehension of the systems under study before practical implementations. This paper illustrates the way MATLAB is used to model non-linearites in synchronous machine. The machine is modeled in rotor reference frame with currents as state ...

  12. Visual Basic, Excel-based fish population modeling tool - The pallid sturgeon example

    Science.gov (United States)

    Moran, Edward H.; Wildhaber, Mark L.; Green, Nicholas S.; Albers, Janice L.

    2016-02-10

    The model presented in this report is a spreadsheet-based model using Visual Basic for Applications within Microsoft Excel (http://dx.doi.org/10.5066/F7057D0Z) prepared in cooperation with the U.S. Army Corps of Engineers and U.S. Fish and Wildlife Service. It uses the same model structure and, initially, parameters as used by Wildhaber and others (2015) for pallid sturgeon. The difference between the model structure used for this report and that used by Wildhaber and others (2015) is that variance is not partitioned. For the model of this report, all variance is applied at the iteration and time-step levels of the model. Wildhaber and others (2015) partition variance into parameter variance (uncertainty about the value of a parameter itself) applied at the iteration level and temporal variance (uncertainty caused by random environmental fluctuations with time) applied at the time-step level. They included implicit individual variance (uncertainty caused by differences between individuals) within the time-step level.The interface developed for the model of this report is designed to allow the user the flexibility to change population model structure and parameter values and uncertainty separately for every component of the model. This flexibility makes the modeling tool potentially applicable to any fish species; however, the flexibility inherent in this modeling tool makes it possible for the user to obtain spurious outputs. The value and reliability of the model outputs are only as good as the model inputs. Using this modeling tool with improper or inaccurate parameter values, or for species for which the structure of the model is inappropriate, could lead to untenable management decisions. By facilitating fish population modeling, this modeling tool allows the user to evaluate a range of management options and implications. The goal of this modeling tool is to be a user-friendly modeling tool for developing fish population models useful to natural resource

  13. The fierce urgency of now: a proactive, pervasive content awareness tool

    Energy Technology Data Exchange (ETDEWEB)

    Powell, James E [Los Alamos National Laboratory; Collins, Linn M [Los Alamos National Laboratory; Martinez, Mark L B [Los Alamos National Laboratory

    2009-01-01

    Information awareness is distinct from explicit infonnation seeking, such as searching. We describe an information awareness tool that supports text composition by providing awareness of relevant content and references proactively and non-intrusively. As a user composes text, the tool automatically searches mUltiple sources, retrieves results, and displays links to the results. The tool has been implemented using Web 2.0 and Digital Library 2.0 technologies, and is flexible and highly configurable.

  14. Development of modeling tools for pin-by-pin precise reactor simulation

    International Nuclear Information System (INIS)

    Ma Yan; Li Shu; Li Gang; Zhang Baoyin; Deng Li; Fu Yuanguang

    2013-01-01

    In order to develop large-scale transport simulation and calculation method (such as simulation of whole reactor core pin-by-pin problem), the Institute of Applied Physics and Computational Mathematics developed the neutron-photon coupled transport code JMCT and the toolkit JCOGIN. Creating physical calculation model easily and efficiently can essentially reduce problem solving time. Currently, lots of visual modeling programs have been developed based on different CAD systems. In this article, the developing idea of a visual modeling tool based on field oriented development was introduced. Considering the feature of physical modeling, fast and convenient operation modules were developed. In order to solve the storage and conversion problems of large scale models, the data structure and conversional algorithm based on the hierarchical geometry tree were designed. The automatic conversion and generation of physical model input file for JMCT were realized. By using this modeling tool, the Dayawan reactor whole core physical model was created, and the transformed file was delivered to JMCT for transport calculation. The results validate the correctness of the visual modeling tool. (authors)

  15. Improving Power System Modeling. A Tool to Link Capacity Expansion and Production Cost Models

    Energy Technology Data Exchange (ETDEWEB)

    Diakov, Victor [National Renewable Energy Lab. (NREL), Golden, CO (United States); Cole, Wesley [National Renewable Energy Lab. (NREL), Golden, CO (United States); Sullivan, Patrick [National Renewable Energy Lab. (NREL), Golden, CO (United States); Brinkman, Gregory [National Renewable Energy Lab. (NREL), Golden, CO (United States); Margolis, Robert [National Renewable Energy Lab. (NREL), Golden, CO (United States)

    2015-11-01

    Capacity expansion models (CEM) provide a high-level long-term view at the prospects of the evolving power system. In simulating the possibilities of long-term capacity expansion, it is important to maintain the viability of power system operation in the short-term (daily, hourly and sub-hourly) scales. Production-cost models (PCM) simulate routine power system operation on these shorter time scales using detailed load, transmission and generation fleet data by minimizing production costs and following reliability requirements. When based on CEM 'predictions' about generating unit retirements and buildup, PCM provide more detailed simulation for the short-term system operation and, consequently, may confirm the validity of capacity expansion predictions. Further, production cost model simulations of a system that is based on capacity expansion model solution are 'evolutionary' sound: the generator mix is the result of logical sequence of unit retirement and buildup resulting from policy and incentives. The above has motivated us to bridge CEM with PCM by building a capacity expansion - to - production cost model Linking Tool (CEPCoLT). The Linking Tool is built to onset capacity expansion model prescriptions onto production cost model inputs. NREL's ReEDS and Energy Examplar's PLEXOS are the capacity expansion and the production cost models, respectively. Via the Linking Tool, PLEXOS provides details of operation for the regionally-defined ReEDS scenarios.

  16. Rapid State Space Modeling Tool for Rectangular Wing Aeroservoelastic Studies

    Science.gov (United States)

    Suh, Peter M.; Conyers, Howard Jason; Mavris, Dimitri N.

    2015-01-01

    This report introduces a modeling and simulation tool for aeroservoelastic analysis of rectangular wings with trailing-edge control surfaces. The inputs to the code are planform design parameters such as wing span, aspect ratio, and number of control surfaces. Using this information, the generalized forces are computed using the doublet-lattice method. Using Roger's approximation, a rational function approximation is computed. The output, computed in a few seconds, is a state space aeroservoelastic model which can be used for analysis and control design. The tool is fully parameterized with default information so there is little required interaction with the model developer. All parameters can be easily modified if desired. The focus of this report is on tool presentation, verification, and validation. These processes are carried out in stages throughout the report. The rational function approximation is verified against computed generalized forces for a plate model. A model composed of finite element plates is compared to a modal analysis from commercial software and an independently conducted experimental ground vibration test analysis. Aeroservoelastic analysis is the ultimate goal of this tool, therefore, the flutter speed and frequency for a clamped plate are computed using damping-versus-velocity and frequency-versus-velocity analysis. The computational results are compared to a previously published computational analysis and wind-tunnel results for the same structure. A case study of a generic wing model with a single control surface is presented. Verification of the state space model is presented in comparison to damping-versus-velocity and frequency-versus-velocity analysis, including the analysis of the model in response to a 1-cos gust.

  17. How Transparent About its Inflation Target Should a Central Bank be? An Agent-Based Model Assessment

    NARCIS (Netherlands)

    Salle, I.; Sénégas, M.A.; Yıldızoğlu, M.

    2013-01-01

    This paper revisits the benefits of explicitly announcing an inflation target for the con- duct of monetary policy in the framework of an agent-based model (ABM). This framework offers a flexible tool for modeling heterogeneity among individual agents and their bounded rationality, and to emphasize,

  18. Automation life-cycle cost model

    Science.gov (United States)

    Gathmann, Thomas P.; Reeves, Arlinda J.; Cline, Rick; Henrion, Max; Ruokangas, Corinne

    1992-01-01

    The problem domain being addressed by this contractual effort can be summarized by the following list: Automation and Robotics (A&R) technologies appear to be viable alternatives to current, manual operations; Life-cycle cost models are typically judged with suspicion due to implicit assumptions and little associated documentation; and Uncertainty is a reality for increasingly complex problems and few models explicitly account for its affect on the solution space. The objectives for this effort range from the near-term (1-2 years) to far-term (3-5 years). In the near-term, the envisioned capabilities of the modeling tool are annotated. In addition, a framework is defined and developed in the Decision Modelling System (DEMOS) environment. Our approach is summarized as follows: Assess desirable capabilities (structure into near- and far-term); Identify useful existing models/data; Identify parameters for utility analysis; Define tool framework; Encode scenario thread for model validation; and Provide transition path for tool development. This report contains all relevant, technical progress made on this contractual effort.

  19. Modelling stillbirth mortality reduction with the Lives Saved Tool

    Directory of Open Access Journals (Sweden)

    Hannah Blencowe

    2017-11-01

    Full Text Available Abstract Background The worldwide burden of stillbirths is large, with an estimated 2.6 million babies stillborn in 2015 including 1.3 million dying during labour. The Every Newborn Action Plan set a stillbirth target of ≤12 per 1000 in all countries by 2030. Planning tools will be essential as countries set policy and plan investment to scale up interventions to meet this target. This paper summarises the approach taken for modelling the impact of scaling-up health interventions on stillbirths in the Lives Saved tool (LiST, and potential future refinements. Methods The specific application to stillbirths of the general method for modelling the impact of interventions in LiST is described. The evidence for the effectiveness of potential interventions to reduce stillbirths are reviewed and the assumptions of the affected fraction of stillbirths who could potentially benefit from these interventions are presented. The current assumptions and their effects on stillbirth reduction are described and potential future improvements discussed. Results High quality evidence are not available for all parameters in the LiST stillbirth model. Cause-specific mortality data is not available for stillbirths, therefore stillbirths are modelled in LiST using an attributable fraction approach by timing of stillbirths (antepartum/ intrapartum. Of 35 potential interventions to reduce stillbirths identified, eight interventions are currently modelled in LiST. These include childbirth care, induction for prolonged pregnancy, multiple micronutrient and balanced energy supplementation, malaria prevention and detection and management of hypertensive disorders of pregnancy, diabetes and syphilis. For three of the interventions, childbirth care, detection and management of hypertensive disorders of pregnancy, and diabetes the estimate of effectiveness is based on expert opinion through a Delphi process. Only for malaria is coverage information available, with coverage

  20. Spin-orbit splitted excited states using explicitly-correlated equation-of-motion coupled-cluster singles and doubles eigenvectors

    Science.gov (United States)

    Bokhan, Denis; Trubnikov, Dmitrii N.; Perera, Ajith; Bartlett, Rodney J.

    2018-04-01

    An explicitly-correlated method of calculation of excited states with spin-orbit couplings, has been formulated and implemented. Developed approach utilizes left and right eigenvectors of equation-of-motion coupled-cluster model, which is based on the linearly approximated explicitly correlated coupled-cluster singles and doubles [CCSD(F12)] method. The spin-orbit interactions are introduced by using the spin-orbit mean field (SOMF) approximation of the Breit-Pauli Hamiltonian. Numerical tests for several atoms and molecules show good agreement between explicitly-correlated results and the corresponding values, calculated in complete basis set limit (CBS); the highly-accurate excitation energies can be obtained already at triple- ζ level.

  1. An advanced modelling tool for simulating complex river systems.

    Science.gov (United States)

    Trancoso, Ana Rosa; Braunschweig, Frank; Chambel Leitão, Pedro; Obermann, Matthias; Neves, Ramiro

    2009-04-01

    The present paper describes MOHID River Network (MRN), a 1D hydrodynamic model for river networks as part of MOHID Water Modelling System, which is a modular system for the simulation of water bodies (hydrodynamics and water constituents). MRN is capable of simulating water quality in the aquatic and benthic phase and its development was especially focused on the reproduction of processes occurring in temporary river networks (flush events, pools formation, and transmission losses). Further, unlike many other models, it allows the quantification of settled materials at the channel bed also over periods when the river falls dry. These features are very important to secure mass conservation in highly varying flows of temporary rivers. The water quality models existing in MOHID are base on well-known ecological models, such as WASP and ERSEM, the latter allowing explicit parameterization of C, N, P, Si, and O cycles. MRN can be coupled to the basin model, MOHID Land, with computes runoff and porous media transport, allowing for the dynamic exchange of water and materials between the river and surroundings, or it can be used as a standalone model, receiving discharges at any specified nodes (ASCII files of time series with arbitrary time step). These features account for spatial gradients in precipitation which can be significant in Mediterranean-like basins. An interface has been already developed for SWAT basin model.

  2. Drinkers’ memory bias for alcohol picture cues in explicit and implicit memory tasks

    Science.gov (United States)

    Nguyen-Louie, Tam T.; Buckman, Jennifer F.; Ray, Suchismita

    2016-01-01

    Background Alcohol cues can bias attention and elicit emotional reactions, especially in drinkers. Yet, little is known about how alcohol cues affect explicit and implicit memory processes, and how memory for alcohol cues is affected by acute alcohol intoxication. Methods Young adult participants (N=161) were randomly assigned to alcohol, placebo, or control beverage conditions. Following beverage consumption, they were shown neutral, emotional and alcohol-related pictures cues. Participants then completed free recall and repetition priming tasks to test explicit and implicit memory, respectively, for picture cues. Average blood alcohol concentration for the alcohol group was 74 ± 13 mg/dl when memory testing began. Two mixed linear model analyses were conducted to examine the effects of beverage condition, picture cue type, and their interaction on explicit and implicit memory. Results Picture cue type and beverage condition each significantly affected explicit recall of picture cues, whereas only picture cue type significantly influenced repetition priming. Individuals in the alcohol condition recalled significantly fewer pictures than those in other conditions, regardless of cue type. Both free recall and repetition priming were greater for emotional and alcohol-related cues compared to neutral picture cues. No interaction effects were detected. Conclusions Young adult drinkers showed enhanced explicit and implicit memory processing of alcohol cues compared to emotionally neutral cues. This enhanced processing for alcohol cues was on par with that seen for positive emotional cues. Acute alcohol intoxication did not alter this preferential memory processing for alcohol cues over neutral cues. PMID:26811126

  3. Drinkers' memory bias for alcohol picture cues in explicit and implicit memory tasks.

    Science.gov (United States)

    Nguyen-Louie, Tam T; Buckman, Jennifer F; Ray, Suchismita; Bates, Marsha E

    2016-03-01

    Alcohol cues can bias attention and elicit emotional reactions, especially in drinkers. Yet, little is known about how alcohol cues affect explicit and implicit memory processes, and how memory for alcohol cues is affected by acute alcohol intoxication. Young adult participants (N=161) were randomly assigned to alcohol, placebo, or control beverage conditions. Following beverage consumption, they were shown neutral, emotional and alcohol-related pictures cues. Participants then completed free recall and repetition priming tasks to test explicit and implicit memory, respectively, for picture cues. Average blood alcohol concentration for the alcohol group was 74±13mg/dl when memory testing began. Two mixed linear model analyses were conducted to examine the effects of beverage condition, picture cue type, and their interaction on explicit and implicit memory. Picture cue type and beverage condition each significantly affected explicit recall of picture cues, whereas only picture cue type significantly influenced repetition priming. Individuals in the alcohol condition recalled significantly fewer pictures than those in other conditions, regardless of cue type. Both free recall and repetition priming were greater for emotional and alcohol-related cues compared to neutral picture cues. No interaction effects were detected. Young adult drinkers showed enhanced explicit and implicit memory processing of alcohol cues compared to emotionally neutral cues. This enhanced processing for alcohol cues was on par with that seen for positive emotional cues. Acute alcohol intoxication did not alter this preferential memory processing for alcohol cues over neutral cues. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.

  4. Development Life Cycle and Tools for XML Content Models

    Energy Technology Data Exchange (ETDEWEB)

    Kulvatunyou, Boonserm [ORNL; Morris, Katherine [National Institute of Standards and Technology (NIST); Buhwan, Jeong [POSTECH University, South Korea; Goyal, Puja [National Institute of Standards and Technology (NIST)

    2004-11-01

    Many integration projects today rely on shared semantic models based on standards represented using Extensible Mark up Language (XML) technologies. Shared semantic models typically evolve and require maintenance. In addition, to promote interoperability and reduce integration costs, the shared semantics should be reused as much as possible. Semantic components must be consistent and valid in terms of agreed upon standards and guidelines. In this paper, we describe an activity model for creation, use, and maintenance of a shared semantic model that is coherent and supports efficient enterprise integration. We then use this activity model to frame our research and the development of tools to support those activities. We provide overviews of these tools primarily in the context of the W3C XML Schema. At the present, we focus our work on the W3C XML Schema as the representation of choice, due to its extensive adoption by industry.

  5. Multi-category micro-milling tool wear monitoring with continuous hidden Markov models

    Science.gov (United States)

    Zhu, Kunpeng; Wong, Yoke San; Hong, Geok Soon

    2009-02-01

    In-process monitoring of tool conditions is important in micro-machining due to the high precision requirement and high tool wear rate. Tool condition monitoring in micro-machining poses new challenges compared to conventional machining. In this paper, a multi-category classification approach is proposed for tool flank wear state identification in micro-milling. Continuous Hidden Markov models (HMMs) are adapted for modeling of the tool wear process in micro-milling, and estimation of the tool wear state given the cutting force features. For a noise-robust approach, the HMM outputs are connected via a medium filter to minimize the tool state before entry into the next state due to high noise level. A detailed study on the selection of HMM structures for tool condition monitoring (TCM) is presented. Case studies on the tool state estimation in the micro-milling of pure copper and steel demonstrate the effectiveness and potential of these methods.

  6. Scratch as a Computational Modelling Tool for Teaching Physics

    Science.gov (United States)

    Lopez, Victor; Hernandez, Maria Isabel

    2015-01-01

    The Scratch online authoring tool, which features a simple programming language that has been adapted to primary and secondary students, is being used more and more in schools as it offers students and teachers the opportunity to use a tool to build scientific models and evaluate their behaviour, just as can be done with computational modelling…

  7. Exploring patterns of explicit and implicit anti-gay attitudes in Muslims and Atheists

    NARCIS (Netherlands)

    Anderson, Joel; Koc, Yasin

    2015-01-01

    Research into the relationship between religion and anti-gay attitudes frequently focuses on Christianity. We explored the role of religiosity dimensions, previous contact, and factors in the dual-process motivation model as predictors of explicit and implicit anti-gay attitudes in samples of

  8. Individuals with fear of blushing explicitly and automatically associate blushing with social costs

    NARCIS (Netherlands)

    Glashouwer, K.A.; de Jong, P.J.; Dijk, C.; Buwalda, F.M.

    2011-01-01

    To explain fear of blushing, it has been proposed that individuals with fear of blushing overestimate the social costs of their blushing. Current information-processing models emphasize the relevance of differentiating between more automatic and more explicit cognitions, as both types of cognitions

  9. Individuals with Fear of Blushing Explicitly and Automatically Associate Blushing with Social Costs

    NARCIS (Netherlands)

    Glashouwer, Klaske A.; de Jong, Peter J.; Dijk, Corine; Buwalda, Femke M.

    2011-01-01

    To explain fear of blushing, it has been proposed that individuals with fear of blushing overestimate the social costs of their blushing. Current information-processing models emphasize the relevance of differentiating between more automatic and more explicit cognitions, as both types of cognitions

  10. The Ms. Stereotype Revisited: Implicit and Explicit Facets

    Science.gov (United States)

    Malcolmson, Kelly A.; Sinclair, Lisa

    2007-01-01

    Implicit and explicit stereotypes toward the title Ms. were examined. Participants read a short description of a target person whose title of address varied (Ms., Mrs., Miss, Mr.). They then rated the person on agentic and communal traits and completed an Implicit Association Test. Replicating earlier research (Dion, 1987), at an explicit level,…

  11. Modeling biochemical transformation processes and information processing with Narrator

    Directory of Open Access Journals (Sweden)

    Palfreyman Niall M

    2007-03-01

    Full Text Available Abstract Background Software tools that model and simulate the dynamics of biological processes and systems are becoming increasingly important. Some of these tools offer sophisticated graphical user interfaces (GUIs, which greatly enhance their acceptance by users. Such GUIs are based on symbolic or graphical notations used to describe, interact and communicate the developed models. Typically, these graphical notations are geared towards conventional biochemical pathway diagrams. They permit the user to represent the transport and transformation of chemical species and to define inhibitory and stimulatory dependencies. A critical weakness of existing tools is their lack of supporting an integrative representation of transport, transformation as well as biological information processing. Results Narrator is a software tool facilitating the development and simulation of biological systems as Co-dependence models. The Co-dependence Methodology complements the representation of species transport and transformation together with an explicit mechanism to express biological information processing. Thus, Co-dependence models explicitly capture, for instance, signal processing structures and the influence of exogenous factors or events affecting certain parts of a biological system or process. This combined set of features provides the system biologist with a powerful tool to describe and explore the dynamics of life phenomena. Narrator's GUI is based on an expressive graphical notation which forms an integral part of the Co-dependence Methodology. Behind the user-friendly GUI, Narrator hides a flexible feature which makes it relatively easy to map models defined via the graphical notation to mathematical formalisms and languages such as ordinary differential equations, the Systems Biology Markup Language or Gillespie's direct method. This powerful feature facilitates reuse, interoperability and conceptual model development. Conclusion Narrator is a

  12. A tool for model based diagnostics of the AGS Booster

    International Nuclear Information System (INIS)

    Luccio, A.

    1993-01-01

    A model-based algorithmic tool was developed to search for lattice errors by a systematic analysis of orbit data in the AGS Booster synchrotron. The algorithm employs transfer matrices calculated with MAD between points in the ring. Iterative model fitting of the data allows one to find and eventually correct magnet displacements and angles or field errors. The tool, implemented on a HP-Apollo workstation system, has proved very general and of immediate physical interpretation

  13. ADAS tools for collisional–radiative modelling of molecules

    Energy Technology Data Exchange (ETDEWEB)

    Guzmán, F., E-mail: francisco.guzman@cea.fr [Department of Physics, University of Strathclyde, Glasgow G4 0NG (United Kingdom); CEA, IRFM, Saint-Paul-lez-Durance 13108 (France); O’Mullane, M.; Summers, H.P. [Department of Physics, University of Strathclyde, Glasgow G4 0NG (United Kingdom)

    2013-07-15

    New theoretical and computational tools for molecular collisional–radiative models are presented. An application to the hydrogen molecule system has been made. At the same time, a structured database has been created where fundamental cross sections and rates for individual processes as well as derived data (effective coefficients) are stored. Relative populations for the vibrational states of the ground electronic state of H{sub 2} are presented and this vibronic resolution model is compared electronic resolution where vibronic transitions are summed over vibrational sub-states. Some new reaction rates are calculated by means of the impact parameter approximation. Computational tools have been developed to automate process and simplify the data assembly. Effective (collisional–radiative) rate coefficients versus temperature and density are presented.

  14. ADAS tools for collisional-radiative modelling of molecules

    Science.gov (United States)

    Guzmán, F.; O'Mullane, M.; Summers, H. P.

    2013-07-01

    New theoretical and computational tools for molecular collisional-radiative models are presented. An application to the hydrogen molecule system has been made. At the same time, a structured database has been created where fundamental cross sections and rates for individual processes as well as derived data (effective coefficients) are stored. Relative populations for the vibrational states of the ground electronic state of H2 are presented and this vibronic resolution model is compared electronic resolution where vibronic transitions are summed over vibrational sub-states. Some new reaction rates are calculated by means of the impact parameter approximation. Computational tools have been developed to automate process and simplify the data assembly. Effective (collisional-radiative) rate coefficients versus temperature and density are presented.

  15. Not explicit but implicit memory is influenced by individual perception style

    OpenAIRE

    Hine, Kyoko; Tsushima, Yoshiaki

    2018-01-01

    Not only explicit but also implicit memory has considerable influence on our daily life. However, it is still unclear whether explicit and implicit memories are sensitive to individual differences. Here, we investigated how individual perception style (global or local) correlates with implicit and explicit memory. As a result, we found that not explicit but implicit memory was affected by the perception style: local perception style people more greatly used implicit memory than global percept...

  16. DiVinE-CUDA - A Tool for GPU Accelerated LTL Model Checking

    Directory of Open Access Journals (Sweden)

    Jiří Barnat

    2009-12-01

    Full Text Available In this paper we present a tool that performs CUDA accelerated LTL Model Checking. The tool exploits parallel algorithm MAP adjusted to the NVIDIA CUDA architecture in order to efficiently detect the presence of accepting cycles in a directed graph. Accepting cycle detection is the core algorithmic procedure in automata-based LTL Model Checking. We demonstrate that the tool outperforms non-accelerated version of the algorithm and we discuss where the limits of the tool are and what we intend to do in the future to avoid them.

  17. Combination of process-oriented and pattern-oriented models of land-use change in a mountain area of Vietnam

    NARCIS (Netherlands)

    Castella, J.C.; Verburg, P.H.

    2007-01-01

    The tools and methods developed by different scientific communities to simulate the dynamics of land use have emphasised either processes or patterns of changes. Agent-based models (ABM) belong to the former category while many spatially explicit simulation models belong to the latter. These two

  18. Formulation of an explicit-multiple-time-step time integration method for use in a global primitive equation grid model

    Science.gov (United States)

    Chao, W. C.

    1982-01-01

    With appropriate modifications, a recently proposed explicit-multiple-time-step scheme (EMTSS) is incorporated into the UCLA model. In this scheme, the linearized terms in the governing equations that generate the gravity waves are split into different vertical modes. Each mode is integrated with an optimal time step, and at periodic intervals these modes are recombined. The other terms are integrated with a time step dictated by the CFL condition for low-frequency waves. This large time step requires a special modification of the advective terms in the polar region to maintain stability. Test runs for 72 h show that EMTSS is a stable, efficient and accurate scheme.

  19. "Tacit Knowledge" versus "Explicit Knowledge"

    DEFF Research Database (Denmark)

    Sanchez, Ron

    creators and carriers. By contrast, the explicit knowledge approach emphasizes processes for articulating knowledge held by individuals, the design of organizational approaches for creating new knowledge, and the development of systems (including information systems) to disseminate articulated knowledge...

  20. Business intelligence tools for radiology: creating a prototype model using open-source tools.

    Science.gov (United States)

    Prevedello, Luciano M; Andriole, Katherine P; Hanson, Richard; Kelly, Pauline; Khorasani, Ramin

    2010-04-01

    Digital radiology departments could benefit from the ability to integrate and visualize data (e.g. information reflecting complex workflow states) from all of their imaging and information management systems in one composite presentation view. Leveraging data warehousing tools developed in the business world may be one way to achieve this capability. In total, the concept of managing the information available in this data repository is known as Business Intelligence or BI. This paper describes the concepts used in Business Intelligence, their importance to modern Radiology, and the steps used in the creation of a prototype model of a data warehouse for BI using open-source tools.

  1. Tools for model-independent bounds in direct dark matter searches

    DEFF Research Database (Denmark)

    Cirelli, M.; Del Nobile, E.; Panci, P.

    2013-01-01

    We discuss a framework (based on non-relativistic operators) and a self-contained set of numerical tools to derive the bounds from some current direct detection experiments on virtually any arbitrary model of Dark Matter elastically scattering on nuclei.......We discuss a framework (based on non-relativistic operators) and a self-contained set of numerical tools to derive the bounds from some current direct detection experiments on virtually any arbitrary model of Dark Matter elastically scattering on nuclei....

  2. Representing climate, disturbance, and vegetation interactions in landscape models

    Science.gov (United States)

    Robert E. Keane; Donald McKenzie; Donald A. Falk; Erica A.H. Smithwick; Carol Miller; Lara-Karena B. Kellogg

    2015-01-01

    The prospect of rapidly changing climates over the next century calls for methods to predict their effects on myriad, interactive ecosystem processes. Spatially explicit models that simulate ecosystem dynamics at fine (plant, stand) to coarse (regional, global) scales are indispensable tools for meeting this challenge under a variety of possible futures. A special...

  3. Uncertainty Analysis of Coupled Socioeconomic-Cropping Models: Building Confidence in Climate Change Decision-Support Tools for Local Stakeholders

    Science.gov (United States)

    Malard, J. J.; Rojas, M.; Adamowski, J. F.; Gálvez, J.; Tuy, H. A.; Melgar-Quiñonez, H.

    2015-12-01

    While cropping models represent the biophysical aspects of agricultural systems, system dynamics modelling offers the possibility of representing the socioeconomic (including social and cultural) aspects of these systems. The two types of models can then be coupled in order to include the socioeconomic dimensions of climate change adaptation in the predictions of cropping models.We develop a dynamically coupled socioeconomic-biophysical model of agricultural production and its repercussions on food security in two case studies from Guatemala (a market-based, intensive agricultural system and a low-input, subsistence crop-based system). Through the specification of the climate inputs to the cropping model, the impacts of climate change on the entire system can be analysed, and the participatory nature of the system dynamics model-building process, in which stakeholders from NGOs to local governmental extension workers were included, helps ensure local trust in and use of the model.However, the analysis of climate variability's impacts on agroecosystems includes uncertainty, especially in the case of joint physical-socioeconomic modelling, and the explicit representation of this uncertainty in the participatory development of the models is important to ensure appropriate use of the models by the end users. In addition, standard model calibration, validation, and uncertainty interval estimation techniques used for physically-based models are impractical in the case of socioeconomic modelling. We present a methodology for the calibration and uncertainty analysis of coupled biophysical (cropping) and system dynamics (socioeconomic) agricultural models, using survey data and expert input to calibrate and evaluate the uncertainty of the system dynamics as well as of the overall coupled model. This approach offers an important tool for local decision makers to evaluate the potential impacts of climate change and their feedbacks through the associated socioeconomic system.

  4. Residual fatigue life evaluation of rail at squats seeds using 3D explicit finite element analysis

    NARCIS (Netherlands)

    Deng, X.; Naeimi, M.; Li, Z.; Qian, Z.

    2014-01-01

    A modeling procedure to predict the residual fatigue life of rail at squats seeds is developed in this article. Two models are involved: a 3D explicit Finite Element (FE) model to compute the stress and strain at squats in rail, and the J-S fatigue damage model to determine the residual fatigue life

  5. Explicit Bounds and Sharp Results for the Composition Operators Preserving the Exponential Class

    Directory of Open Access Journals (Sweden)

    Fernando Farroni

    2016-01-01

    Full Text Available Let f:Ω⊂Rn→Rn be a quasiconformal mapping whose Jacobian is denoted by Jf and let EXP(Ω be the space of exponentially integrable functions on Ω. We give an explicit bound for the norm of the composition operator Tf: u∈EXP(Ω↦u∘f-1∈EXP(f(Ω and, as a related question, we study the behaviour of the norm of log⁡Jf in the exponential class. The A∞ property of Jf is the counterpart in higher dimensions of the area distortion formula due to Astala in the plane and it is the key tool to prove the sharpness of our results.

  6. Advanced REACH tool: A Bayesian model for occupational exposure assessment

    NARCIS (Netherlands)

    McNally, K.; Warren, N.; Fransman, W.; Entink, R.K.; Schinkel, J.; Van Tongeren, M.; Cherrie, J.W.; Kromhout, H.; Schneider, T.; Tielemans, E.

    2014-01-01

    This paper describes a Bayesian model for the assessment of inhalation exposures in an occupational setting; the methodology underpins a freely available web-based application for exposure assessment, the Advanced REACH Tool (ART). The ART is a higher tier exposure tool that combines disparate

  7. Dynamic wind turbine models in power system simulation tool DIgSILENT

    OpenAIRE

    Hansen, A.D.; Jauch, C.; Sørensen, Poul Ejnar; Iov, F.; Blaabjerg, F.

    2004-01-01

    The present report describes the dynamic wind turbine models implemented in the power system simulation tool DIgSILENT (Version 12.0). The developed models are a part of the results of a national research project, whose overall objective is to create amodel database in different simulation tools. This model database should be able to support the analysis of the interaction between the mechanical structure of the wind turbine and the electrical grid during different operational modes. The repo...

  8. The time course of explicit and implicit categorization.

    Science.gov (United States)

    Smith, J David; Zakrzewski, Alexandria C; Herberger, Eric R; Boomer, Joseph; Roeder, Jessica L; Ashby, F Gregory; Church, Barbara A

    2015-10-01

    Contemporary theory in cognitive neuroscience distinguishes, among the processes and utilities that serve categorization, explicit and implicit systems of category learning that learn, respectively, category rules by active hypothesis testing or adaptive behaviors by association and reinforcement. Little is known about the time course of categorization within these systems. Accordingly, the present experiments contrasted tasks that fostered explicit categorization (because they had a one-dimensional, rule-based solution) or implicit categorization (because they had a two-dimensional, information-integration solution). In Experiment 1, participants learned categories under unspeeded or speeded conditions. In Experiment 2, they applied previously trained category knowledge under unspeeded or speeded conditions. Speeded conditions selectively impaired implicit category learning and implicit mature categorization. These results illuminate the processing dynamics of explicit/implicit categorization.

  9. Introducing Modeling Transition Diagrams as a Tool to Connect Mathematical Modeling to Mathematical Thinking

    Science.gov (United States)

    Czocher, Jennifer A.

    2016-01-01

    This study contributes a methodological tool to reconstruct the cognitive processes and mathematical activities carried out by mathematical modelers. Represented as Modeling Transition Diagrams (MTDs), individual modeling routes were constructed for four engineering undergraduate students. Findings stress the importance and limitations of using…

  10. Probabilistic language models in cognitive neuroscience: Promises and pitfalls.

    Science.gov (United States)

    Armeni, Kristijan; Willems, Roel M; Frank, Stefan L

    2017-12-01

    Cognitive neuroscientists of language comprehension study how neural computations relate to cognitive computations during comprehension. On the cognitive part of the equation, it is important that the computations and processing complexity are explicitly defined. Probabilistic language models can be used to give a computationally explicit account of language complexity during comprehension. Whereas such models have so far predominantly been evaluated against behavioral data, only recently have the models been used to explain neurobiological signals. Measures obtained from these models emphasize the probabilistic, information-processing view of language understanding and provide a set of tools that can be used for testing neural hypotheses about language comprehension. Here, we provide a cursory review of the theoretical foundations and example neuroimaging studies employing probabilistic language models. We highlight the advantages and potential pitfalls of this approach and indicate avenues for future research. Copyright © 2017 Elsevier Ltd. All rights reserved.

  11. Why explicit gluonic components are needed in multiquark wave functions

    International Nuclear Information System (INIS)

    Lipkin, H.J.

    1982-03-01

    Constituent quark models with nontrivial correlations between color and space and no explicit description of the gluon degrees of freedom are inconsistent with local color gauge invariance. The screening by gluons of local color density in the quark sector must be included in treatments of long range forces in multiquark states. Neglect of gauge invariance and screening leads to erroneous predictions like unobserved Van der Weals forces and unobserved ''baryonium'' states with localized color. (author)

  12. Universal correlators for multi-arc complex matrix models

    International Nuclear Information System (INIS)

    Akemann, G.

    1997-01-01

    The correlation functions of the multi-arc complex matrix model are shown to be universal for any finite number of arcs. The universality classes are characterized by the support of the eigenvalue density and are conjectured to fall into the same classes as the ones recently found for the Hermitian model. This is explicitly shown to be true for the case of two arcs, apart from the known result for one arc. The basic tool is the iterative solution of the loop equation for the complex matrix model with multiple arcs, which provides all multi-loop correlators up to an arbitrary genus. Explicit results for genus one are given for any number of arcs. The two-arc solution is investigated in detail, including the double-scaling limit. In addition universal expressions for the string susceptibility are given for both the complex and Hermitian model. (orig.)

  13. Spatially Explicit Estimation of Optimal Light Use Efficiency for Improved Satellite Data Driven Ecosystem Productivity Modeling

    Science.gov (United States)

    Madani, N.; Kimball, J. S.; Running, S. W.

    2014-12-01

    Remote sensing based light use efficiency (LUE) models, including the MODIS (MODerate resolution Imaging Spectroradiometer) MOD17 algorithm are commonly used for regional estimation and monitoring of vegetation gross primary production (GPP) and photosynthetic carbon (CO2) uptake. A common model assumption is that plants in a biome matrix operate at their photosynthetic capacity under optimal climatic conditions. A prescribed biome maximum light use efficiency parameter defines the maximum photosynthetic carbon conversion rate under prevailing climate conditions and is a large source of model uncertainty. Here, we used tower (FLUXNET) eddy covariance measurement based carbon flux data for estimating optimal LUE (LUEopt) over a North American domain. LUEopt was first estimated using tower observed daily carbon fluxes, meteorology and satellite (MODIS) observed fraction of photosynthetically active radiation (FPAR). LUEopt was then spatially interpolated over the domain using empirical models derived from independent geospatial data including global plant traits, surface soil moisture, terrain aspect, land cover type and percent tree cover. The derived LUEopt maps were then used as primary inputs to the MOD17 LUE algorithm for regional GPP estimation; these results were evaluated against tower observations and alternate MOD17 GPP estimates determined using Biome-specific LUEopt constants. Estimated LUEopt shows large spatial variability within and among different land cover classes indicated from a sparse North American tower network. Leaf nitrogen content and soil moisture are two important factors explaining LUEopt spatial variability. GPP estimated from spatially explicit LUEopt inputs shows significantly improved model accuracy against independent tower observations (R2 = 0.76; Mean RMSE plant trait information can explain spatial heterogeneity in LUEopt, leading to improved GPP estimates from satellite based LUE models.

  14. Spatially explicit modeling of blackbird abundance in the Prairie Pothole Region

    Science.gov (United States)

    Forcey, Greg M.; Thogmartin, Wayne E.; Linz, George M.; McKann, Patrick C.; Crimmins, Shawn M.

    2015-01-01

    Knowledge of factors influencing animal abundance is important to wildlife biologists developing management plans. This is especially true for economically important species such as blackbirds (Icteridae), which cause more than $100 million in crop damages annually in the United States. Using data from the North American Breeding Bird Survey, the National Land Cover Dataset, and the National Climatic Data Center, we modeled effects of regional environmental variables on relative abundance of 3 blackbird species (red-winged blackbird,Agelaius phoeniceus; yellow-headed blackbird, Xanthocephalus xanthocephalus; common grackle, Quiscalus quiscula) in the Prairie Pothole Region of the central United States. We evaluated landscape covariates at 3 logarithmically related spatial scales (1,000 ha, 10,000 ha, and 100,000 ha) and modeled weather variables at the 100,000-ha scale. We constructed models a priori using information from published habitat associations. We fit models with WinBUGS using Markov chain Monte Carlo techniques. Both landscape and weather variables contributed strongly to predicting blackbird relative abundance (95% credibility interval did not overlap 0). Variables with the strongest associations with blackbird relative abundance were the percentage of wetland area and precipitation amount from the year before bird surveys were conducted. The influence of spatial scale appeared small—models with the same variables expressed at different scales were often in the best model subset. This large-scale study elucidated regional effects of weather and landscape variables, suggesting that management strategies aimed at reducing damages caused by these species should consider the broader landscape, including weather effects, because such factors may outweigh the influence of localized conditions or site-specific management actions. The regional species distributional models we developed for blackbirds provide a tool for understanding these broader

  15. A mixed implicit/explicit procedure for soil-structure interaction

    International Nuclear Information System (INIS)

    Kunar, R.R.

    1982-01-01

    This paper describes an efficient method for the solution of dynamic soil-structure interaction problems. The method which combines implicit and explicit time integration procedures is ideally suited to problems in which the structure is considered linear and the soil non-linear. The equations relating to the linear structures are integrated using an unconditionally stable implicit scheme while the non-linear soil is treated explicitly. The explicit method is ideally suited to non-linear calculations as there is no need for iterative techniques. The structural equations can also be integrated explicitly, but this generally requires a time step that is much smaller than that for the soil. By using an unconditionally stable implicit algorithm for the structure, the complete analysis can be performed using the time step for the soil. The proposed procedure leads to economical solutions with the soil non-linearities handled accurately and efficiently. (orig.)

  16. The effect of encoding duration on implicit and explicit eyewitness memory.

    Science.gov (United States)

    Carol, Rolando N; Schreiber Compo, Nadja

    2018-05-01

    The present study investigated the effect of encoding duration on implicit and explicit eyewitness memory. Participants (N = 227) viewed a mock crime (brief, 15-s vs. long, 30-s vs. irrelevant/control) and were then tested with both implicit and explicit memory prompts or with explicit memory prompts only. Brief-encoding participants revealed more critical details implicitly than long-encoding or control participants. Further, the number and percentage of accurate details recalled explicitly were higher for long-encoding than for brief-encoding participants. Implicit testing prior to explicit recall-as compared to completing a filler task-was detrimental to free recall performance. Interestingly, brief-encoding participants were significantly more likely to remember critical details implicitly but not explicitly than long-encoding participants. This is the first study to investigate implicit eyewitness memory for a multimodal mock crime. Findings are theoretically consistent with prior research on cognition while expanding upon the extant eyewitness memory and investigative interviewing literature. Published by Elsevier Inc.

  17. Explicit symplectic algorithms based on generating functions for charged particle dynamics

    Science.gov (United States)

    Zhang, Ruili; Qin, Hong; Tang, Yifa; Liu, Jian; He, Yang; Xiao, Jianyuan

    2016-07-01

    Dynamics of a charged particle in the canonical coordinates is a Hamiltonian system, and the well-known symplectic algorithm has been regarded as the de facto method for numerical integration of Hamiltonian systems due to its long-term accuracy and fidelity. For long-term simulations with high efficiency, explicit symplectic algorithms are desirable. However, it is generally believed that explicit symplectic algorithms are only available for sum-separable Hamiltonians, and this restriction limits the application of explicit symplectic algorithms to charged particle dynamics. To overcome this difficulty, we combine the familiar sum-split method and a generating function method to construct second- and third-order explicit symplectic algorithms for dynamics of charged particle. The generating function method is designed to generate explicit symplectic algorithms for product-separable Hamiltonian with form of H (x ,p ) =pif (x ) or H (x ,p ) =xig (p ) . Applied to the simulations of charged particle dynamics, the explicit symplectic algorithms based on generating functions demonstrate superiorities in conservation and efficiency.

  18. Implicit and explicit processes in social cognition

    DEFF Research Database (Denmark)

    Frith, Christopher; Frith, Uta

    2008-01-01

    In this review we consider research on social cognition in which implicit processes can be compared and contrasted with explicit, conscious processes. In each case, their function is distinct, sometimes complementary and sometimes oppositional. We argue that implicit processes in social interaction...... are automatic and are often opposed to conscious strategies. While we are aware of explicit processes in social interaction, we cannot always use them to override implicit processes. Many studies show that implicit processes facilitate the sharing of knowledge, feelings, and actions, and hence, perhaps...

  19. Extrapolated stabilized explicit Runge-Kutta methods

    Science.gov (United States)

    Martín-Vaquero, J.; Kleefeld, B.

    2016-12-01

    Extrapolated Stabilized Explicit Runge-Kutta methods (ESERK) are proposed to solve multi-dimensional nonlinear partial differential equations (PDEs). In such methods it is necessary to evaluate the function nt times per step, but the stability region is O (nt2). Hence, the computational cost is O (nt) times lower than for a traditional explicit algorithm. In that way stiff problems can be integrated by the use of simple explicit evaluations in which case implicit methods usually had to be used. Therefore, they are especially well-suited for the method of lines (MOL) discretizations of parabolic nonlinear multi-dimensional PDEs. In this work, first s-stages first-order methods with extended stability along the negative real axis are obtained. They have slightly shorter stability regions than other traditional first-order stabilized explicit Runge-Kutta algorithms (also called Runge-Kutta-Chebyshev codes). Later, they are used to derive nt-stages second- and fourth-order schemes using Richardson extrapolation. The stability regions of these fourth-order codes include the interval [ - 0.01nt2, 0 ] (nt being the number of total functions evaluations), which are shorter than stability regions of ROCK4 methods, for example. However, the new algorithms neither suffer from propagation of errors (as other Runge-Kutta-Chebyshev codes as ROCK4 or DUMKA) nor internal instabilities. Additionally, many other types of higher-order (and also lower-order) methods can be obtained easily in a similar way. These methods also allow adaptation of the length step with no extra cost. Hence, the stability domain is adapted precisely to the spectrum of the problem at the current time of integration in an optimal way, i.e., with minimal number of additional stages. We compare the new techniques with other well-known algorithms with good results in very stiff diffusion or reaction-diffusion multi-dimensional nonlinear equations.

  20. Spatially explicit and stochastic simulation of forest landscape fire disturbance and succession

    Science.gov (United States)

    Hong S. He; David J. Mladenoff

    1999-01-01

    Understanding disturbance and recovery of forest landscapes is a challenge because of complex interactions over a range of temporal and spatial scales. Landscape simulation models offer an approach to studying such systems at broad scales. Fire can be simulated spatially using mechanistic or stochastic approaches. We describe the fire module in a spatially explicit,...

  1. Monte Carlo tools for Beyond the Standard Model Physics , April 14-16

    DEFF Research Database (Denmark)

    Badger...[], Simon; Christensen, Christian Holm; Dalsgaard, Hans Hjersing

    2011-01-01

    This workshop aims to gather together theorists and experimentalists interested in developing and using Monte Carlo tools for Beyond the Standard Model Physics in an attempt to be prepared for the analysis of data focusing on the Large Hadron Collider. Since a large number of excellent tools....... To identify promising models (or processes) for which the tools have not yet been constructed and start filling up these gaps. To propose ways to streamline the process of going from models to events, i.e. to make the process more user-friendly so that more people can get involved and perform serious collider...

  2. Age-dependent and coordinated shift in performance between implicit and explicit skill learning

    Directory of Open Access Journals (Sweden)

    Dezso eNemeth

    2013-10-01

    Full Text Available It has been reported recently that while general sequence learning across ages conforms to the typical inverted-U shape pattern, with best performance in early adulthood, surprisingly, the basic ability of picking up in an implicit manner triplets that occur with high vs. low probability in the sequence is best before 12 years of age and it significantly weakens afterwards. Based on these findings, it has been hypothesized that the cognitively controlled processes coming online at around 12 are useful for more targeted explicit learning at the cost of becoming relatively less sensitive to raw probabilities of events. To test this hypothesis, we collected data in a sequence learning task using probabilistic sequences in five age groups from 11 to 39 years of age (N=288, replicating the original implicit learning paradigm in an explicit task setting where subjects were guided to find repeating sequences. We found that in contrast to the implicit results, performance with the high- vs. low-probability triplets was at the same level in all age groups when subjects sought patterns in the sequence explicitly. Importantly, measurements of explicit knowledge about the identity of the sequences revealed a significant increase in ability to explicitly access the true sequences exactly around the age where the earlier study found the significant drop in ability to learn implicitly raw probabilities. These findings support the conjecture that the gradually increasing involvement of more complex internal models optimizes our skill learning abilities by compensating for the performance loss due to down-weighting the raw probabilities of the sensory input, while expanding our ability to acquire more sophisticated skills.

  3. Age-dependent and coordinated shift in performance between implicit and explicit skill learning.

    Science.gov (United States)

    Nemeth, Dezso; Janacsek, Karolina; Fiser, József

    2013-01-01

    It has been reported recently that while general sequence learning across ages conforms to the typical inverted-U shape pattern, with best performance in early adulthood, surprisingly, the basic ability of picking up in an implicit manner triplets that occur with high vs. low probability in the sequence is best before 12 years of age and it significantly weakens afterwards. Based on these findings, it has been hypothesized that the cognitively controlled processes coming online at around 12 are useful for more targeted explicit learning at the cost of becoming relatively less sensitive to raw probabilities of events. To test this hypothesis, we collected data in a sequence learning task using probabilistic sequences in five age groups from 11 to 39 years of age (N = 288), replicating the original implicit learning paradigm in an explicit task setting where subjects were guided to find repeating sequences. We found that in contrast to the implicit results, performance with the high- vs. low-probability triplets was at the same level in all age groups when subjects sought patterns in the sequence explicitly. Importantly, measurements of explicit knowledge about the identity of the sequences revealed a significant increase in ability to explicitly access the true sequences exactly around the age where the earlier study found the significant drop in ability to learn implicitly raw probabilities. These findings support the conjecture that the gradually increasing involvement of more complex internal models optimizes our skill learning abilities by compensating for the performance loss due to down-weighting the raw probabilities of the sensory input, while expanding our ability to acquire more sophisticated skills.

  4. Modeling the fate of nitrogen on the catchment scale using a spatially explicit hydro-biogeochemical simulation system

    Science.gov (United States)

    Klatt, S.; Butterbach-Bahl, K.; Kiese, R.; Haas, E.; Kraus, D.; Molina-Herrera, S. W.; Kraft, P.

    2015-12-01

    The continuous growth of the human population demands an equally growing supply for fresh water and food. As a result, available land for efficient agriculture is constantly diminishing which forces farmers to cultivate inferior croplands and intensify agricultural practices, e.g., increase the use of synthetic fertilizers. This intensification of marginal areas in particular will cause a dangerous rise in nitrate discharge into open waters or even drinking water resources. In order to reduce the amount of nitrate lost by surface runoff or lateral subsurface transport, bufferstrips have proved to be a valuable means. Current laws, however, promote rather static designs (i.e., width and usage) even though a multitude of factors, e.g., soil type, slope, vegetation and the nearby agricultural management, determines its effectiveness. We propose a spatially explicit modeling approach enabling to assess the effects of those factors on nitrate discharge from arable lands using the fully distributed hydrology model CMF coupled to the complex biogeochemical model LandscapeDNDC. Such a modeling scheme allows to observe the displacement of dissolved nutrients in both vertical and horizontal directions and serves to estimate both their uptake by the vegetated bufferstrip and loss to the environment. First results indicate a significant reduction of nitrate loss in the presence of a bufferstrip (2.5 m). We show effects induced by various buffer strip widths and plant cover on the nitrate retention.

  5. Not explicit but implicit memory is influenced by individual perception style.

    Science.gov (United States)

    Hine, Kyoko; Tsushima, Yoshiaki

    2018-01-01

    Not only explicit but also implicit memory has considerable influence on our daily life. However, it is still unclear whether explicit and implicit memories are sensitive to individual differences. Here, we investigated how individual perception style (global or local) correlates with implicit and explicit memory. As a result, we found that not explicit but implicit memory was affected by the perception style: local perception style people more greatly used implicit memory than global perception style people. These results help us to make the new effective application adapting to individual perception style and understand some clinical symptoms such as autistic spectrum disorder. Furthermore, this finding might give us new insight of memory involving consciousness and unconsciousness as well as relationship between implicit/explicit memory and individual perception style.

  6. Implicit and explicit memory in survivors of chronic interpersonal violence.

    Science.gov (United States)

    Minshew, Reese; D'Andrea, Wendy

    2015-01-01

    We investigated the relationship of implicit and explicit memory to a range of symptoms in a sample of 27 women with exposure to chronic interpersonal violence (IPV). Participants viewed the first 3 letters ("stems") of trauma-related, general threat, and neutral words; valenced words were matched with neutral words with the same stem. Free recall and a word-stem completion task were used to test explicit and implicit memory, respectively. Participants exhibited increased implicit memory for trauma-related words as compared with both general threat words and neutral "match" words. They also showed increased explicit memory for both general threat and trauma-related words. Finally, although neither implicit nor explicit memory was correlated with PTSD symptoms, implicit memory for trauma-related words was significantly correlated with symptoms associated with ongoing IPV. Interpersonal sensitivity, hostility, and alexithymia were significantly correlated with implicit, but not explicit, memory for trauma words. Somatization, dissociation, and alexithymia were negatively correlated with explicit, but not implicit, memory for general-threat words. These findings suggest that memory processes in survivors of IPV are closely related to the symptom profile associated with complex trauma. Exploring memory processes in survivors of IPV may lend unique insight into the development and maintenance of the symptom profile associated with IPV. (c) 2015 APA, all rights reserved).

  7. Continued development of modeling tools and theory for RF heating

    International Nuclear Information System (INIS)

    1998-01-01

    Mission Research Corporation (MRC) is pleased to present the Department of Energy (DOE) with its renewal proposal to the Continued Development of Modeling Tools and Theory for RF Heating program. The objective of the program is to continue and extend the earlier work done by the proposed principal investigator in the field of modeling (Radio Frequency) RF heating experiments in the large tokamak fusion experiments, particularly the Tokamak Fusion Test Reactor (TFTR) device located at Princeton Plasma Physics Laboratory (PPPL). An integral part of this work is the investigation and, in some cases, resolution of theoretical issues which pertain to accurate modeling. MRC is nearing the successful completion of the specified tasks of the Continued Development of Modeling Tools and Theory for RF Heating project. The following tasks are either completed or nearing completion. (1) Anisotropic temperature and rotation upgrades; (2) Modeling for relativistic ECRH; (3) Further documentation of SHOOT and SPRUCE. As a result of the progress achieved under this project, MRC has been urged to continue this effort. Specifically, during the performance of this project two topics were identified by PPPL personnel as new applications of the existing RF modeling tools. These two topics concern (a) future fast-wave current drive experiments on the large tokamaks including TFTR and (c) the interpretation of existing and future RF probe data from TFTR. To address each of these topics requires some modification or enhancement of the existing modeling tools, and the first topic requires resolution of certain theoretical issues to produce self-consistent results. This work falls within the scope of the original project and is more suited to the project's renewal than to the initiation of a new project

  8. Modeling with data tools and techniques for scientific computing

    CERN Document Server

    Klemens, Ben

    2009-01-01

    Modeling with Data fully explains how to execute computationally intensive analyses on very large data sets, showing readers how to determine the best methods for solving a variety of different problems, how to create and debug statistical models, and how to run an analysis and evaluate the results. Ben Klemens introduces a set of open and unlimited tools, and uses them to demonstrate data management, analysis, and simulation techniques essential for dealing with large data sets and computationally intensive procedures. He then demonstrates how to easily apply these tools to the many threads of statistical technique, including classical, Bayesian, maximum likelihood, and Monte Carlo methods

  9. Creating a spatially-explicit index: a method for assessing the global wildfire-water risk

    Science.gov (United States)

    Robinne, François-Nicolas; Parisien, Marc-André; Flannigan, Mike; Miller, Carol; Bladon, Kevin D.

    2017-04-01

    The wildfire-water risk (WWR) has been defined as the potential for wildfires to adversely affect water resources that are important for downstream ecosystems and human water needs for adequate water quantity and quality, therefore compromising the security of their water supply. While tools and methods are numerous for watershed-scale risk analysis, the development of a toolbox for the large-scale evaluation of the wildfire risk to water security has only started recently. In order to provide managers and policy-makers with an adequate tool, we implemented a method for the spatial analysis of the global WWR based on the Driving forces-Pressures-States-Impacts-Responses (DPSIR) framework. This framework relies on the cause-and-effect relationships existing between the five categories of the DPSIR chain. As this approach heavily relies on data, we gathered an extensive set of spatial indicators relevant to fire-induced hydrological hazards and water consumption patterns by human and natural communities. When appropriate, we applied a hydrological routing function to our indicators in order to simulate downstream accumulation of potentially harmful material. Each indicator was then assigned a DPSIR category. We collapsed the information in each category using a principal component analysis in order to extract the most relevant pixel-based information provided by each spatial indicator. Finally, we compiled our five categories using an additive indexation process to produce a spatially-explicit index of the WWR. A thorough sensitivity analysis has been performed in order to understand the relationship between the final risk values and the spatial pattern of each category used during the indexation. For comparison purposes, we aggregated index scores by global hydrological regions, or hydrobelts, to get a sense of regional DPSIR specificities. This rather simple method does not necessitate the use of complex physical models and provides a scalable and efficient tool

  10. Developing a cellular automata model of urban growth to inform spatial policy for flood mitigation : A case study in Kampala, Uganda

    NARCIS (Netherlands)

    Pérez-Molina, Eduardo; Sliuzas, R.V.; Flacke, J.; Jetten, V.G.

    2017-01-01

    Urban growth may intensify local flooding problems. Understanding the spatially explicit flood consequences of possible future land cover patterns contributes to inform policy for mitigating these impacts. A cellular automata model has been coupled with the openLISEM integrated flood modeling tool

  11. The South Florida Ecosystem Portfolio Model - A Map-Based Multicriteria Ecological, Economic, and Community Land-Use Planning Tool

    Science.gov (United States)

    Labiosa, William B.; Bernknopf, Richard; Hearn, Paul; Hogan, Dianna; Strong, David; Pearlstine, Leonard; Mathie, Amy M.; Wein, Anne M.; Gillen, Kevin; Wachter, Susan

    2009-01-01

    The South Florida Ecosystem Portfolio Model (EPM) prototype is a regional land-use planning Web tool that integrates ecological, economic, and social information and values of relevance to decision-makers and stakeholders. The EPM uses a multicriteria evaluation framework that builds on geographic information system-based (GIS) analysis and spatially-explicit models that characterize important ecological, economic, and societal endpoints and consequences that are sensitive to regional land-use/land-cover (LULC) change. The EPM uses both economics (monetized) and multiattribute utility (nonmonetized) approaches to valuing these endpoints and consequences. This hybrid approach represents a methodological middle ground between rigorous economic and ecological/ environmental scientific approaches. The EPM sacrifices some degree of economic- and ecological-forecasting precision to gain methodological transparency, spatial explicitness, and transferability, while maintaining credibility. After all, even small steps in the direction of including ecosystem services evaluation are an improvement over current land-use planning practice (Boyd and Wainger, 2003). There are many participants involved in land-use decision-making in South Florida, including local, regional, State, and Federal agencies, developers, environmental groups, agricultural groups, and other stakeholders (South Florida Regional Planning Council, 2003, 2004). The EPM's multicriteria evaluation framework is designed to cut across the objectives and knowledge bases of all of these participants. This approach places fundamental importance on social equity and stakeholder participation in land-use decision-making, but makes no attempt to determine normative socially 'optimal' land-use plans. The EPM is thus a map-based set of evaluation tools for planners and stakeholders to use in their deliberations of what is 'best', considering a balancing of disparate interests within a regional perspective. Although

  12. AgMIP Training in Multiple Crop Models and Tools

    Science.gov (United States)

    Boote, Kenneth J.; Porter, Cheryl H.; Hargreaves, John; Hoogenboom, Gerrit; Thornburn, Peter; Mutter, Carolyn

    2015-01-01

    The Agricultural Model Intercomparison and Improvement Project (AgMIP) has the goal of using multiple crop models to evaluate climate impacts on agricultural production and food security in developed and developing countries. There are several major limitations that must be overcome to achieve this goal, including the need to train AgMIP regional research team (RRT) crop modelers to use models other than the ones they are currently familiar with, plus the need to harmonize and interconvert the disparate input file formats used for the various models. Two activities were followed to address these shortcomings among AgMIP RRTs to enable them to use multiple models to evaluate climate impacts on crop production and food security. We designed and conducted courses in which participants trained on two different sets of crop models, with emphasis on the model of least experience. In a second activity, the AgMIP IT group created templates for inputting data on soils, management, weather, and crops into AgMIP harmonized databases, and developed translation tools for converting the harmonized data into files that are ready for multiple crop model simulations. The strategies for creating and conducting the multi-model course and developing entry and translation tools are reviewed in this chapter.

  13. Tools for macromolecular model building and refinement into electron cryo-microscopy reconstructions

    Energy Technology Data Exchange (ETDEWEB)

    Brown, Alan; Long, Fei; Nicholls, Robert A.; Toots, Jaan; Emsley, Paul; Murshudov, Garib, E-mail: garib@mrc-lmb.cam.ac.uk [MRC Laboratory of Molecular Biology, Francis Crick Avenue, Cambridge CB2 0QH (United Kingdom)

    2015-01-01

    A description is given of new tools to facilitate model building and refinement into electron cryo-microscopy reconstructions. The recent rapid development of single-particle electron cryo-microscopy (cryo-EM) now allows structures to be solved by this method at resolutions close to 3 Å. Here, a number of tools to facilitate the interpretation of EM reconstructions with stereochemically reasonable all-atom models are described. The BALBES database has been repurposed as a tool for identifying protein folds from density maps. Modifications to Coot, including new Jiggle Fit and morphing tools and improved handling of nucleic acids, enhance its functionality for interpreting EM maps. REFMAC has been modified for optimal fitting of atomic models into EM maps. As external structural information can enhance the reliability of the derived atomic models, stabilize refinement and reduce overfitting, ProSMART has been extended to generate interatomic distance restraints from nucleic acid reference structures, and a new tool, LIBG, has been developed to generate nucleic acid base-pair and parallel-plane restraints. Furthermore, restraint generation has been integrated with visualization and editing in Coot, and these restraints have been applied to both real-space refinement in Coot and reciprocal-space refinement in REFMAC.

  14. Tools for macromolecular model building and refinement into electron cryo-microscopy reconstructions

    International Nuclear Information System (INIS)

    Brown, Alan; Long, Fei; Nicholls, Robert A.; Toots, Jaan; Emsley, Paul; Murshudov, Garib

    2015-01-01

    A description is given of new tools to facilitate model building and refinement into electron cryo-microscopy reconstructions. The recent rapid development of single-particle electron cryo-microscopy (cryo-EM) now allows structures to be solved by this method at resolutions close to 3 Å. Here, a number of tools to facilitate the interpretation of EM reconstructions with stereochemically reasonable all-atom models are described. The BALBES database has been repurposed as a tool for identifying protein folds from density maps. Modifications to Coot, including new Jiggle Fit and morphing tools and improved handling of nucleic acids, enhance its functionality for interpreting EM maps. REFMAC has been modified for optimal fitting of atomic models into EM maps. As external structural information can enhance the reliability of the derived atomic models, stabilize refinement and reduce overfitting, ProSMART has been extended to generate interatomic distance restraints from nucleic acid reference structures, and a new tool, LIBG, has been developed to generate nucleic acid base-pair and parallel-plane restraints. Furthermore, restraint generation has been integrated with visualization and editing in Coot, and these restraints have been applied to both real-space refinement in Coot and reciprocal-space refinement in REFMAC

  15. Assessing Bioenergy Harvest Risks: Geospatially Explicit Tools for Maintaining Soil Productivity in Western US Forests

    Directory of Open Access Journals (Sweden)

    Deborah Page-Dumroese

    2011-09-01

    Full Text Available Biomass harvesting for energy production and forest health can impact the soil resource by altering inherent chemical, physical and biological properties. These impacts raise concern about damaging sensitive forest soils, even with the prospect of maintaining vigorous forest growth through biomass harvesting operations. Current forest biomass harvesting research concurs that harvest impacts to the soil resource are region- and site-specific, although generalized knowledge from decades of research can be incorporated into management activities. Based upon the most current forest harvesting research, we compiled information on harvest activities that decrease, maintain or increase soil-site productivity. We then developed a soil chemical and physical property risk assessment within a geographic information system for a timber producing region within the Northern Rocky Mountain ecoregion. Digital soil and geology databases were used to construct geospatially explicit best management practices to maintain or enhance soil-site productivity. The proposed risk assessments could aid in identifying resilient soils for forest land managers considering biomass operations, policy makers contemplating expansion of biomass harvesting and investors deliberating where to locate bioenergy conversion facilities.

  16. Large eddy simulations of round free jets using explicit filtering with/without dynamic Smagorinsky model

    International Nuclear Information System (INIS)

    Bogey, Christophe; Bailly, Christophe

    2006-01-01

    Large eddy simulations (LES) of round free jets at Mach number M = 0.9 with Reynolds numbers over the range 2.5 x 10 3 ≤ Re D ≤ 4 x 10 5 are performed using explicit selective/high-order filtering with or without dynamic Smagorinsky model (DSM). Features of the flows and of the turbulent kinetic energy budgets in the turbulent jets are reported. The contributions of molecular viscosity, filtering and DSM to energy dissipation are also presented. Using filtering alone, the results are independent of the filtering strength, and the effects of the Reynolds number on jet development are successfully calculated. Using DSM, the effective jet Reynolds number is found to be artificially decreased by the eddy viscosity. The results are also not appreciably modified when subgrid-scale kinetic energy is used. Moreover, unlike filtering which does not significantly affect the larger computed scales, the eddy viscosity is shown to dissipate energy through all the turbulent scales, in the same way as molecular viscosity at lower Reynolds numbers

  17. Customer Data Analysis Model using Business Intelligence Tools in Telecommunication Companies

    Directory of Open Access Journals (Sweden)

    Monica LIA

    2015-10-01

    Full Text Available This article presents a customer data analysis model in a telecommunication company and business intelligence tools for data modelling, transforming, data visualization and dynamic reports building . For a mature market, knowing the information inside the data and making forecast for strategic decision become more important in Romanian Market. Business Intelligence tools are used in business organization as support for decision making.

  18. Explicit Solutions for One-Dimensional Mean-Field Games

    KAUST Repository

    Prazeres, Mariana

    2017-04-05

    In this thesis, we consider stationary one-dimensional mean-field games (MFGs) with or without congestion. Our aim is to understand the qualitative features of these games through the analysis of explicit solutions. We are particularly interested in MFGs with a nonmonotonic behavior, which corresponds to situations where agents tend to aggregate. First, we derive the MFG equations from control theory. Then, we compute explicit solutions using the current formulation and examine their behavior. Finally, we represent the solutions and analyze the results. This thesis main contributions are the following: First, we develop the current method to solve MFG explicitly. Second, we analyze in detail non-monotonic MFGs and discover new phenomena: non-uniqueness, discontinuous solutions, empty regions and unhappiness traps. Finally, we address several regularization procedures and examine the stability of MFGs.

  19. Distributed MAP in the SpinJa Model Checker

    Directory of Open Access Journals (Sweden)

    Stefan Vijzelaar

    2011-10-01

    Full Text Available Spin in Java (SpinJa is an explicit state model checker for the Promela modelling language also used by the SPIN model checker. Designed to be extensible and reusable, the implementation of SpinJa follows a layered approach in which each new layer extends the functionality of the previous one. While SpinJa has preliminary support for shared-memory model checking, it did not yet support distributed-memory model checking. This tool paper presents a distributed implementation of a maximal accepting predecessors (MAP search algorithm on top of SpinJa.

  20. Programming Models and Tools for Intelligent Embedded Systems

    DEFF Research Database (Denmark)

    Sørensen, Peter Verner Bojsen

    Design automation and analysis tools targeting embedded platforms, developed using a component-based design approach, must be able to reason about the capabilities of the platforms. In the general case where nothing is assumed about the components comprising a platform or the platform topology...... is used for checking the consistency of a design with respect to the availablity of services and resources. In the second application, a tool for automatically implementing the communication infrastructure of a process network application, the Service Relation Model is used for analyzing the capabilities...

  1. Assessing Sustainability of Coral Reef Ecosystem Services using a Spatially-Explicit Decision Support Tool

    Science.gov (United States)

    Forecasting and communicating the potential outcomes of decision options requires support tools that aid in evaluating alternative scenarios in a user-friendly context and that highlight variables relevant to the decision options and valuable stakeholders. Envision is a GIS-base...

  2. Spatially Explicit Modelling of the Belgian Major Endurance Event 'The 100 km Dodentocht'.

    Directory of Open Access Journals (Sweden)

    Steffie Van Nieuland

    Full Text Available 'The 100 km Dodentocht', which takes place annually and has its start in Bornem, Belgium, is a long distance march where participants have to cover a 100 km trail in at most 24 hours. The approximately 11 000 marchers per edition are tracked by making use of passive radio-frequency-identification (RFID. These tracking data were analyzed to build a spatially explicit marching model that gives insights into the dynamics of the event and allows to evaluate the effect of changes in the starting procedure of the event. For building the model, the empirical distribution functions (edf of the marching speeds at every section of the trail in between two consecutive checkpoints and of the checkpoints where marchers retire, are determined, taking into account age, gender, and marching speeds at previous sections. These distribution functions are then used to sample the consecutive speeds and retirement, and as such to simulate the times when individual marchers pass by the consecutive checkpoints. We concluded that the data-driven model simulates the event reliably. Furthermore, we tested three scenarios to reduce the crowdiness along the first part of the trail and in this way were able to conclude that either the start should be moved to a location outside the town center where the streets are at least 25% wider, or that the marchers should start in two groups at two different locations, and that these groups should ideally merge at about 20 km after the start. The crowdiness at the start might also be reduced by installing a bottleneck at the start in order to limit the number of marchers that can pass per unit of time. Consequently, the operating hours of the consecutive checkpoints would be longer. The developed framework can likewise be used to analyze and improve the operation of other endurance events if sufficient tracking data are available.

  3. Age and time effects on implicit and explicit learning

    NARCIS (Netherlands)

    Verneau, M.; Kamp, J. van der; Savelsbergh, G.J.P.; Looze, M.P. de

    2014-01-01

    Study Context: It has been proposed that effects of aging are more pronounced for explicit than for implicit motor learning. The authors evaluated this claim by comparing the efficacy of explicit and implicit learning of a movement sequence in young and older adults, and by testing the resilience

  4. Age and Time Effects on Implicit and Explicit Learning

    NARCIS (Netherlands)

    Verneau, M.M.N.; van der Kamp, J.; Savelsbergh, G.J.P.; de Looze, M.P.

    2014-01-01

    Study Context: It has been proposed that effects of aging are more pronounced for explicit than for implicit motor learning. The authors evaluated this claim by comparing the efficacy of explicit and implicit learning of a movement sequence in young and older adults, and by testing the resilience

  5. Collidoscope: An Improved Tool for Computing Collisional Cross-Sections with the Trajectory Method

    Science.gov (United States)

    Ewing, Simon A.; Donor, Micah T.; Wilson, Jesse W.; Prell, James S.

    2017-04-01

    Ion mobility-mass spectrometry (IM-MS) can be a powerful tool for determining structural information about ions in the gas phase, from small covalent analytes to large, native-like or denatured proteins and complexes. For large biomolecular ions, which may have a wide variety of possible gas-phase conformations and multiple charge sites, quantitative, physically explicit modeling of collisional cross sections (CCSs) for comparison to IMS data can be challenging and time-consuming. We present a "trajectory method" (TM) based CCS calculator, named "Collidoscope," which utilizes parallel processing and optimized trajectory sampling, and implements both He and N2 as collision gas options. Also included is a charge-placement algorithm for determining probable charge site configurations for protonated protein ions given an input geometry in pdb file format. Results from Collidoscope are compared with those from the current state-of-the-art CCS simulation suite, IMoS. Collidoscope CCSs are within 4% of IMoS values for ions with masses from 18 Da to 800 kDa. Collidoscope CCSs using X-ray crystal geometries are typically within a few percent of IM-MS experimental values for ions with mass up to 3.5 kDa (melittin), and discrepancies for larger ions up to 800 kDa (GroEL) are attributed in large part to changes in ion structure during and after the electrospray process. Due to its physically explicit modeling of scattering, computational efficiency, and accuracy, Collidoscope can be a valuable tool for IM-MS research, especially for large biomolecular ions.

  6. 3D Finite Element Simulation of Micro End-Milling by Considering the Effect of Tool Run-Out

    DEFF Research Database (Denmark)

    Davoudinejad, Ali; Tosello, Guido; Parenti, Paolo

    2017-01-01

    Understanding the micro milling phenomena involved in the process is critical and difficult through physical experiments. This study presents a 3D finite element modeling (3D FEM) approach for the micro end-milling process on Al6082-T6. The proposed model employs a Lagrangian explicit finite...... element formulation to perform coupled thermo-mechanical transient analyses. FE simulations were performed at different cutting conditions to obtain realistic numerical predictions of chip formation, temperature distribution, and cutting forces by considering the effect of tool run-out in the model....... The predicted results of the model, involving the run-out influence, showed a good correlation with experimental chip formation and the signal shape of cutting forces....

  7. Teachers' Use of Computational Tools to Construct and Explore Dynamic Mathematical Models

    Science.gov (United States)

    Santos-Trigo, Manuel; Reyes-Rodriguez, Aaron

    2011-01-01

    To what extent does the use of computational tools offer teachers the possibility of constructing dynamic models to identify and explore diverse mathematical relations? What ways of reasoning or thinking about the problems emerge during the model construction process that involves the use of the tools? These research questions guided the…

  8. A masked negative self-esteem? : Implicit and explicit self-esteem in patients with Narcissistic Personality Disorder

    NARCIS (Netherlands)

    Marissen, Marlies A E; Brouwer, Marlies; Hiemstra, Annemarie M F; Deen, Mathijs L; Franken, Ingmar H A

    2016-01-01

    The mask model of narcissism states that the narcissistic traits of patients with NPD are the result of a compensatory reaction to underlying ego fragility. This model assumes that high explicit self-esteem masks low implicit self-esteem. However, research on narcissism has predominantly focused on

  9. Requirements for UML and OWL Integration Tool for User Data Consistency Modeling and Testing

    DEFF Research Database (Denmark)

    Nytun, J. P.; Jensen, Christian Søndergaard; Oleshchuk, V. A.

    2003-01-01

    The amount of data available on the Internet is continuously increasing, consequentially there is a growing need for tools that help to analyse the data. Testing of consistency among data received from different sources is made difficult by the number of different languages and schemas being used....... In this paper we analyze requirements for a tool that support integration of UML models and ontologies written in languages like the W3C Web Ontology Language (OWL). The tool can be used in the following way: after loading two legacy models into the tool, the tool user connects them by inserting modeling......, an important part of this technique is attaching of OCL expressions to special boolean class attributes that we call consistency attributes. The resulting integration model can be used for automatic consistency testing of two instances of the legacy models by automatically instantiate the whole integration...

  10. Explicit time integration of finite element models on a vectorized, concurrent computer with shared memory

    Science.gov (United States)

    Gilbertsen, Noreen D.; Belytschko, Ted

    1990-01-01

    The implementation of a nonlinear explicit program on a vectorized, concurrent computer with shared memory is described and studied. The conflict between vectorization and concurrency is described and some guidelines are given for optimal block sizes. Several example problems are summarized to illustrate the types of speed-ups which can be achieved by reprogramming as compared to compiler optimization.

  11. On the Reliability of Implicit and Explicit Memory Measures.

    Science.gov (United States)

    Buchner, Axel; Wippich, Werner

    2000-01-01

    Studied the reliability of implicit and explicit memory tests in experiments involving these tests. Results with 168, 84, 120, and 128 undergraduates show that methodological artifacts may cause implicit memory tests to have lower reliability than explicit memory tests, but that implicit tests need not necessarily be less reliable. (SLD)

  12. Exploring the Interaction of Implicit and Explicit Processes to Facilitate Individual Skill Learning

    National Research Council Canada - National Science Library

    Sun, Ron; Mathews, Robert C

    2005-01-01

    .... It helps us to explain (and eventually to predict) training and learning processes. The results of the experiments support the theory of the interactions of implicit and explicit learning processes during skill acquisition. The outcomes (data, models, and theories) provide a more detailed, clearer and more comprehensive perspective on skill learning.

  13. An ensemble model of QSAR tools for regulatory risk assessment.

    Science.gov (United States)

    Pradeep, Prachi; Povinelli, Richard J; White, Shannon; Merrill, Stephen J

    2016-01-01

    Quantitative structure activity relationships (QSARs) are theoretical models that relate a quantitative measure of chemical structure to a physical property or a biological effect. QSAR predictions can be used for chemical risk assessment for protection of human and environmental health, which makes them interesting to regulators, especially in the absence of experimental data. For compatibility with regulatory use, QSAR models should be transparent, reproducible and optimized to minimize the number of false negatives. In silico QSAR tools are gaining wide acceptance as a faster alternative to otherwise time-consuming clinical and animal testing methods. However, different QSAR tools often make conflicting predictions for a given chemical and may also vary in their predictive performance across different chemical datasets. In a regulatory context, conflicting predictions raise interpretation, validation and adequacy concerns. To address these concerns, ensemble learning techniques in the machine learning paradigm can be used to integrate predictions from multiple tools. By leveraging various underlying QSAR algorithms and training datasets, the resulting consensus prediction should yield better overall predictive ability. We present a novel ensemble QSAR model using Bayesian classification. The model allows for varying a cut-off parameter that allows for a selection in the desirable trade-off between model sensitivity and specificity. The predictive performance of the ensemble model is compared with four in silico tools (Toxtree, Lazar, OECD Toolbox, and Danish QSAR) to predict carcinogenicity for a dataset of air toxins (332 chemicals) and a subset of the gold carcinogenic potency database (480 chemicals). Leave-one-out cross validation results show that the ensemble model achieves the best trade-off between sensitivity and specificity (accuracy: 83.8 % and 80.4 %, and balanced accuracy: 80.6 % and 80.8 %) and highest inter-rater agreement [kappa ( κ ): 0

  14. Increasing the sampling efficiency of protein conformational transition using velocity-scaling optimized hybrid explicit/implicit solvent REMD simulation

    Energy Technology Data Exchange (ETDEWEB)

    Yu, Yuqi; Wang, Jinan; Shao, Qiang, E-mail: qshao@mail.shcnc.ac.cn, E-mail: Jiye.Shi@ucb.com, E-mail: wlzhu@mail.shcnc.ac.cn; Zhu, Weiliang, E-mail: qshao@mail.shcnc.ac.cn, E-mail: Jiye.Shi@ucb.com, E-mail: wlzhu@mail.shcnc.ac.cn [ACS Key Laboratory of Receptor Research, Drug Discovery and Design Center, Shanghai Institute of Materia Medica, Chinese Academy of Sciences, 555 Zuchongzhi Road, Shanghai 201203 (China); Shi, Jiye, E-mail: qshao@mail.shcnc.ac.cn, E-mail: Jiye.Shi@ucb.com, E-mail: wlzhu@mail.shcnc.ac.cn [UCB Pharma, 216 Bath Road, Slough SL1 4EN (United Kingdom)

    2015-03-28

    The application of temperature replica exchange molecular dynamics (REMD) simulation on protein motion is limited by its huge requirement of computational resource, particularly when explicit solvent model is implemented. In the previous study, we developed a velocity-scaling optimized hybrid explicit/implicit solvent REMD method with the hope to reduce the temperature (replica) number on the premise of maintaining high sampling efficiency. In this study, we utilized this method to characterize and energetically identify the conformational transition pathway of a protein model, the N-terminal domain of calmodulin. In comparison to the standard explicit solvent REMD simulation, the hybrid REMD is much less computationally expensive but, meanwhile, gives accurate evaluation of the structural and thermodynamic properties of the conformational transition which are in well agreement with the standard REMD simulation. Therefore, the hybrid REMD could highly increase the computational efficiency and thus expand the application of REMD simulation to larger-size protein systems.

  15. Biomass supply from alternative cellulosic crops and crop residues: A spatially explicit bioeconomic modeling approach

    International Nuclear Information System (INIS)

    Egbendewe-Mondzozo, Aklesso; Swinton, Scott M.; Izaurralde, César R.; Manowitz, David H.; Zhang, Xuesong

    2011-01-01

    This paper introduces a spatially-explicit bioeconomic model for the study of potential cellulosic biomass supply. For biomass crops to begin to replace current crops, farmers must earn more from them than from current crops. Using weather, topographic and soil data, the terrestrial ecosystem model, EPIC, dynamically simulates multiple cropping systems that vary by crop rotation, tillage, fertilization and residue removal rate. EPIC generates predicted crop yield and environmental outcomes over multiple watersheds. These EPIC results are used to parameterize a regional profit-maximization mathematical programming model that identifies profitable cropping system choices. The bioeconomic model is calibrated to 2007–09 crop production in a 9-county region of southwest Michigan. A simulation of biomass supply in response to rising biomass prices shows that cellulosic residues from corn stover and wheat straw begin to be supplied at minimum delivered biomass:corn grain price ratios of 0.15 and 0.18, respectively. At the mean corn price of $162.6/Mg ($4.13 per bushel) at commercial moisture content during 2007–2009, these ratios correspond to stover and straw prices of $24 and $29 per dry Mg. Perennial bioenergy crops begin to be supplied at price levels 2–3 times higher. Average biomass transport costs to the biorefinery plant range from $6 to $20/Mg compared to conventional crop production practices in the area, biomass supply from annual crop residues increased greenhouse gas emissions and reduced water quality through increased nutrient loss. By contrast, perennial cellulosic biomass crop production reduced greenhouse gas emissions and improved water quality. -- Highlights: ► A new bioeconomic model predicts biomass supply and its environmental impacts. ► The model captures the opportunity cost of switching to new cellulosic crops. ► Biomass from crop residues is supplied at lower biomass price than cellulosic crops. ► Biomass from cellulosic crops has

  16. An Excel®-based visualization tool of 2-D soil gas concentration profiles in petroleum vapor intrusion.

    Science.gov (United States)

    Verginelli, Iason; Yao, Yijun; Suuberg, Eric M

    2016-01-01

    In this study we present a petroleum vapor intrusion tool implemented in Microsoft ® Excel ® using Visual Basic for Applications (VBA) and integrated within a graphical interface. The latter helps users easily visualize two-dimensional soil gas concentration profiles and indoor concentrations as a function of site-specific conditions such as source strength and depth, biodegradation reaction rate constant, soil characteristics and building features. This tool is based on a two-dimensional explicit analytical model that combines steady-state diffusion-dominated vapor transport in a homogeneous soil with a piecewise first-order aerobic biodegradation model, in which rate is limited by oxygen availability. As recommended in the recently released United States Environmental Protection Agency's final Petroleum Vapor Intrusion guidance, a sensitivity analysis and a simplified Monte Carlo uncertainty analysis are also included in the spreadsheet.

  17. Reducing fertilizer-nitrogen losses from rowcrop landscapes: Insights and implications from a spatially explicit watershed model

    Science.gov (United States)

    McLellan, Eileen; Schilling, Keith; Robertson, Dale M.

    2015-01-01

    We present conceptual and quantitative models that predict changes in fertilizer-derived nitrogen delivery from rowcrop landscapes caused by agricultural conservation efforts implemented to reduce nutrient inputs and transport and increase nutrient retention in the landscape. To evaluate the relative importance of changes in the sources, transport, and sinks of fertilizer-derived nitrogen across a region, we use the spatially explicit SPAtially Referenced Regression On Watershed attributes watershed model to map the distribution, at the small watershed scale within the Upper Mississippi-Ohio River Basin (UMORB), of: (1) fertilizer inputs; (2) nutrient attenuation during delivery of those inputs to the UMORB outlet; and (3) nitrogen export from the UMORB outlet. Comparing these spatial distributions suggests that the amount of fertilizer input and degree of nutrient attenuation are both important in determining the extent of nitrogen export. From a management perspective, this means that agricultural conservation efforts to reduce nitrogen export would benefit by: (1) expanding their focus to include activities that restore and enhance nutrient processing in these highly altered landscapes; and (2) targeting specific types of best management practices to watersheds where they will be most valuable. Doing so successfully may result in a shift in current approaches to conservation planning, outreach, and funding.

  18. The Development of Explicit Memory for Basic Perceptual Features.

    Science.gov (United States)

    Gulya, Michelle; Rossi-George, Alba; Hartshorn, Kristen; Vieira, Aurora; Rovee-Collier, Carolyn; Johnson, Marcia K.; Chalfonte, Barbara L.

    2002-01-01

    Three experiments with 164 individuals between 4 and 80 years old examined age-related changes in explicit memory for three perceptual features: item identity, color, and location. Findings indicated that performance on explicit memory tests was not a consistent inverted U-shaped function of age across various features, but depended on the…

  19. Making things explicit using instructional materials: a case study of a Singapore teacher's practice

    Science.gov (United States)

    Leong, Yew Hoong; Cheng, Lu Pien; Toh, Wei Yeng Karen; Kaur, Berinderjeet; Toh, Tin Lam

    2018-04-01

    The phrase `make it explicit' is a common advice given to teachers. It is, however, not clear to us what this actually means when translated into classroom practice. Our review found that we are not alone: "explicit" is used in different ways in the education literature. This paper explores, through a case study of a teacher who stated "making things explicit" as an ostensible goal of his instructional practice, how the explicitation is realised in teaching mathematics. In particular, we examine how he used the instructional materials that he crafted to fulfil his goal of explicitation. We were able to uncover three strategies he used: explicit-from, explicit-within, and explicit-to.

  20. Information Theoretic Tools for Parameter Fitting in Coarse Grained Models

    KAUST Repository

    Kalligiannaki, Evangelia; Harmandaris, Vagelis; Katsoulakis, Markos A.; Plechac, Petr

    2015-01-01

    We study the application of information theoretic tools for model reduction in the case of systems driven by stochastic dynamics out of equilibrium. The model/dimension reduction is considered by proposing parametrized coarse grained dynamics

  1. Fluid Survival Tool: A Model Checker for Hybrid Petri Nets

    NARCIS (Netherlands)

    Postema, Björn Frits; Remke, Anne Katharina Ingrid; Haverkort, Boudewijn R.H.M.; Ghasemieh, Hamed

    2014-01-01

    Recently, algorithms for model checking Stochastic Time Logic (STL) on Hybrid Petri nets with a single general one-shot transition (HPNG) have been introduced. This paper presents a tool for model checking HPNG models against STL formulas. A graphical user interface (GUI) not only helps to

  2. Explicitly computing geodetic coordinates from Cartesian coordinates

    Science.gov (United States)

    Zeng, Huaien

    2013-04-01

    This paper presents a new form of quartic equation based on Lagrange's extremum law and a Groebner basis under the constraint that the geodetic height is the shortest distance between a given point and the reference ellipsoid. A very explicit and concise formulae of the quartic equation by Ferrari's line is found, which avoids the need of a good starting guess for iterative methods. A new explicit algorithm is then proposed to compute geodetic coordinates from Cartesian coordinates. The convergence region of the algorithm is investigated and the corresponding correct solution is given. Lastly, the algorithm is validated with numerical experiments.

  3. Content relatedness in the social web based on social explicit semantic analysis

    Science.gov (United States)

    Ntalianis, Klimis; Otterbacher, Jahna; Mastorakis, Nikolaos

    2017-06-01

    In this paper a novel content relatedness algorithm for social media content is proposed, based on the Explicit Semantic Analysis (ESA) technique. The proposed scheme takes into consideration social interactions. In particular starting from the vector space representation model, similarity is expressed by a summation of term weight products. In this paper, term weights are estimated by a social computing method, where the strength of each term is calculated by the attention the terms receives. For this reason each post is split into two parts, title and comments area, while attention is defined by the number of social interactions such as likes and shares. The overall approach is named Social Explicit Semantic Analysis. Experimental results on real data show the advantages and limitations of the proposed approach, while an initial comparison between ESA and S-ESA is very promising.

  4. Integration of orthographic, conceptual, and episodic information on implicit and explicit tests.

    Science.gov (United States)

    Weldon, M S; Massaro, D W

    1996-03-01

    An experiment was conducted to determine how orthographic and conceptual information are integrated during incidental and intentional retrieval. Subjects studied word lists with either a shallow (counting vowels) or deep (rating pleasantness) processing task, then received either an implicit or explicit word fragment completion (WFC) test. At test, word fragments contained 0, 1, 2, or 4 letters, and were accompanied by 0, 1, 2, or 3 semantically related words. On both the implicit and explicit tests, performance improved with increases in the numbers of letters and words. When semantic cues were presented with the word fragments, the implicit test became more conceptually drive. Still, conceptual processing had a larger effect in intentional than in incidental retrieval. The Fuzzy Logical Model of Perception (FLMP) provided a good description of how orthographic, semantic, and episodic information were combined during retrieval.

  5. A GIS Tool for evaluating and improving NEXRAD and its application in distributed hydrologic modeling

    Science.gov (United States)

    Zhang, X.; Srinivasan, R.

    2008-12-01

    In this study, a user friendly GIS tool was developed for evaluating and improving NEXRAD using raingauge data. This GIS tool can automatically read in raingauge and NEXRAD data, evaluate the accuracy of NEXRAD for each time unit, implement several geostatistical methods to improve the accuracy of NEXRAD through raingauge data, and output spatial precipitation map for distributed hydrologic model. The geostatistical methods incorporated in this tool include Simple Kriging with varying local means, Kriging with External Drift, Regression Kriging, Co-Kriging, and a new geostatistical method that was newly developed by Li et al. (2008). This tool was applied in two test watersheds at hourly and daily temporal scale. The preliminary cross-validation results show that incorporating raingauge data to calibrate NEXRAD can pronouncedly change the spatial pattern of NEXRAD and improve its accuracy. Using different geostatistical methods, the GIS tool was applied to produce long term precipitation input for a distributed hydrologic model - Soil and Water Assessment Tool (SWAT). Animated video was generated to vividly illustrate the effect of using different precipitation input data on distributed hydrologic modeling. Currently, this GIS tool is developed as an extension of SWAT, which is used as water quantity and quality modeling tool by USDA and EPA. The flexible module based design of this tool also makes it easy to be adapted for other hydrologic models for hydrological modeling and water resources management.

  6. A tool model for predicting atmospheric kinetics with sensitivity analysis

    Institute of Scientific and Technical Information of China (English)

    2001-01-01

    A package( a tool model) for program of predicting atmospheric chemical kinetics with sensitivity analysis is presented. The new direct method of calculating the first order sensitivity coefficients using sparse matrix technology to chemical kinetics is included in the tool model, it is only necessary to triangularize the matrix related to the Jacobian matrix of the model equation. The Gear type procedure is used to integrate amodel equation and its coupled auxiliary sensitivity coefficient equations. The FORTRAN subroutines of the model equation, the sensitivity coefficient equations, and their Jacobian analytical expressions are generated automatically from a chemical mechanism. The kinetic representation for the model equation and its sensitivity coefficient equations, and their Jacobian matrix is presented. Various FORTRAN subroutines in packages, such as SLODE, modified MA28, Gear package, with which the program runs in conjunction are recommended.The photo-oxidation of dimethyl disulfide is used for illustration.

  7. Automation based on knowledge modeling theory and its applications in engine diagnostic systems using Space Shuttle Main Engine vibrational data. M.S. Thesis

    Science.gov (United States)

    Kim, Jonnathan H.

    1995-01-01

    Humans can perform many complicated tasks without explicit rules. This inherent and advantageous capability becomes a hurdle when a task is to be automated. Modern computers and numerical calculations require explicit rules and discrete numerical values. In order to bridge the gap between human knowledge and automating tools, a knowledge model is proposed. Knowledge modeling techniques are discussed and utilized to automate a labor and time intensive task of detecting anomalous bearing wear patterns in the Space Shuttle Main Engine (SSME) High Pressure Oxygen Turbopump (HPOTP).

  8. Solving the Sea-Level Equation in an Explicit Time Differencing Scheme

    Science.gov (United States)

    Klemann, V.; Hagedoorn, J. M.; Thomas, M.

    2016-12-01

    In preparation of coupling the solid-earth to an ice-sheet compartment in an earth-system model, the dependency of initial topography on the ice-sheet history and viscosity structure has to be analysed. In this study, we discuss this dependency and how it influences the reconstruction of former sea level during a glacial cycle. The modelling is based on the VILMA code in which the field equations are solved in the time domain applying an explicit time-differencing scheme. The sea-level equation is solved simultaneously in the same explicit scheme as the viscoleastic field equations (Hagedoorn et al., 2007). With the assumption of only small changes, we neglect the iterative solution at each time step as suggested by e.g. Kendall et al. (2005). Nevertheless, the prediction of the initial paleo topography in case of moving coastlines remains to be iterated by repeated integration of the whole load history. The sensitivity study sketched at the beginning is accordingly motivated by the question if the iteration of the paleo topography can be replaced by a predefined one. This study is part of the German paleoclimate modelling initiative PalMod. Lit:Hagedoorn JM, Wolf D, Martinec Z, 2007. An estimate of global mean sea-level rise inferred from tide-gauge measurements using glacial-isostatic models consistent with the relative sea-level record. Pure appl. Geophys. 164: 791-818, doi:10.1007/s00024-007-0186-7Kendall RA, Mitrovica JX, Milne GA, 2005. On post-glacial sea level - II. Numerical formulation and comparative reesults on spherically symmetric models. Geophys. J. Int., 161: 679-706, doi:10.1111/j.365-246.X.2005.02553.x

  9. Analytical Modelling Of Milling For Tool Design And Selection

    International Nuclear Information System (INIS)

    Fontaine, M.; Devillez, A.; Dudzinski, D.

    2007-01-01

    This paper presents an efficient analytical model which allows to simulate a large panel of milling operations. A geometrical description of common end mills and of their engagement in the workpiece material is proposed. The internal radius of the rounded part of the tool envelope is used to define the considered type of mill. The cutting edge position is described for a constant lead helix and for a constant local helix angle. A thermomechanical approach of oblique cutting is applied to predict forces acting on the tool and these results are compared with experimental data obtained from milling tests on a 42CrMo4 steel for three classical types of mills. The influence of some tool's geometrical parameters on predicted cutting forces is presented in order to propose optimisation criteria for design and selection of cutting tools

  10. A Quasiphysics Intelligent Model for a Long Range Fast Tool Servo

    Science.gov (United States)

    Liu, Qiang; Zhou, Xiaoqin; Lin, Jieqiong; Xu, Pengzi; Zhu, Zhiwei

    2013-01-01

    Accurately modeling the dynamic behaviors of fast tool servo (FTS) is one of the key issues in the ultraprecision positioning of the cutting tool. Herein, a quasiphysics intelligent model (QPIM) integrating a linear physics model (LPM) and a radial basis function (RBF) based neural model (NM) is developed to accurately describe the dynamic behaviors of a voice coil motor (VCM) actuated long range fast tool servo (LFTS). To identify the parameters of the LPM, a novel Opposition-based Self-adaptive Replacement Differential Evolution (OSaRDE) algorithm is proposed which has been proved to have a faster convergence mechanism without compromising with the quality of solution and outperform than similar evolution algorithms taken for consideration. The modeling errors of the LPM and the QPIM are investigated by experiments. The modeling error of the LPM presents an obvious trend component which is about ±1.15% of the full span range verifying the efficiency of the proposed OSaRDE algorithm for system identification. As for the QPIM, the trend component in the residual error of LPM can be well suppressed, and the error of the QPIM maintains noise level. All the results verify the efficiency and superiority of the proposed modeling and identification approaches. PMID:24163627

  11. Modeling Behavior by Coastal River Otter (Lontra Canadensis in Response to Prey Availability in Prince William Sound, Alaska: A Spatially-Explicit Individual-Based Approach.

    Directory of Open Access Journals (Sweden)

    Shannon E Albeke

    Full Text Available Effects of climate change on animal behavior and cascading ecosystem responses are rarely evaluated. In coastal Alaska, social river otters (Lontra Canadensis, largely males, cooperatively forage on schooling fish and use latrine sites to communicate group associations and dominance. Conversely, solitary otters, mainly females, feed on intertidal-demersal fish and display mutual avoidance via scent marking. This behavioral variability creates "hotspots" of nutrient deposition and affects plant productivity and diversity on the terrestrial landscape. Because the abundance of schooling pelagic fish is predicted to decline with climate change, we developed a spatially-explicit individual-based model (IBM of otter behavior and tested six scenarios based on potential shifts to distribution patterns of schooling fish. Emergent patterns from the IBM closely mimicked observed otter behavior and landscape use in the absence of explicit rules of intraspecific attraction or repulsion. Model results were most sensitive to rules regarding spatial memory and activity state following an encounter with a fish school. With declining availability of schooling fish, the number of social groups and the time simulated otters spent in the company of conspecifics declined. Concurrently, model results suggested an elevation of defecation rate, a 25% increase in nitrogen transport to the terrestrial landscape, and significant changes to the spatial distribution of "hotspots" with declines in schooling fish availability. However, reductions in availability of schooling fish could lead to declines in otter density over time.

  12. Self-Dual Configurations in a Generalized Abelian Chern-Simons-Higgs Model with Explicit Breaking of the Lorentz Covariance

    International Nuclear Information System (INIS)

    Sourrouille, Lucas; Casana, Rodolfo

    2016-01-01

    We have studied the existence of self-dual solitonic solutions in a generalization of the Abelian Chern-Simons-Higgs model. Such a generalization introduces two different nonnegative functions, ω_1(|ϕ|) and ω(|ϕ|), which split the kinetic term of the Higgs field, |D_μϕ|"2→ω_1(|ϕ|)|D_0ϕ|"2-ω(|ϕ|)|D_kϕ|"2, breaking explicitly the Lorentz covariance. We have shown that a clean implementation of the Bogomolnyi procedure only can be implemented whether ω(|ϕ|)∝β|ϕ|"2"β"-"2 with β≥1. The self-dual or Bogomolnyi equations produce an infinity number of soliton solutions by choosing conveniently the generalizing function ω_1(|ϕ|) which must be able to provide a finite magnetic field. Also, we have shown that by properly choosing the generalizing functions it is possible to reproduce the Bogomolnyi equations of the Abelian Maxwell-Higgs and Chern-Simons-Higgs models. Finally, some new self-dual |ϕ|"6-vortex solutions have been analyzed from both theoretical and numerical point of view.

  13. Making context explicit for explanation and incremental knowledge acquisition

    Energy Technology Data Exchange (ETDEWEB)

    Brezillon, P. [Univ. Paris (France)

    1996-12-31

    Intelligent systems may be improved by making context explicit in problem solving. This is a lesson drawn from a study of the reasons why a number of knowledge-based systems (KBSs) failed. We discuss the interest to make context explicit in explanation generation and incremental knowledge acquisition, two important aspects of intelligent systems that aim to cooperate with users. We show how context can be used to better explain and incrementally acquire knowledge. The advantages of using context in explanation and incremental knowledge acquisition are discussed through SEPIT, an expert system for supporting diagnosis and explanation through simulation of power plants. We point out how the limitations of such systems may be overcome by making context explicit.

  14. Explicit solution of Calderon preconditioned time domain integral equations

    KAUST Repository

    Ulku, Huseyin Arda

    2013-07-01

    An explicit marching on-in-time (MOT) scheme for solving Calderon-preconditioned time domain integral equations is proposed. The scheme uses Rao-Wilton-Glisson and Buffa-Christiansen functions to discretize the domain and range of the integral operators and a PE(CE)m type linear multistep to march on in time. Unlike its implicit counterpart, the proposed explicit solver requires the solution of an MOT system with a Gram matrix that is sparse and well-conditioned independent of the time step size. Numerical results demonstrate that the explicit solver maintains its accuracy and stability even when the time step size is chosen as large as that typically used by an implicit solver. © 2013 IEEE.

  15. Parameter Extraction for PSpice Models by means of an Automated Optimization Tool – An IGBT model Study Case

    DEFF Research Database (Denmark)

    Suárez, Carlos Gómez; Reigosa, Paula Diaz; Iannuzzo, Francesco

    2016-01-01

    An original tool for parameter extraction of PSpice models has been released, enabling a simple parameter identification. A physics-based IGBT model is used to demonstrate that the optimization tool is capable of generating a set of parameters which predicts the steady-state and switching behavio...

  16. Solving the Little Hierarchy Problem with a Singlet and Explicit μ Terms

    International Nuclear Information System (INIS)

    Delgado, Antonio; Kolda, Christopher; Olson, J. Pocahontas; Puente, Alejandro de la

    2010-01-01

    We present a generalization of the next-to-minimal supersymmetric standard model, with an explicit μ term and a supersymmetric mass for the singlet superfield, as a route to alleviating the little hierarchy problem of the minimal supersymmetric standard model (MSSM). Though this model does not address the μ problem of the MSSM, we are able to generate masses for the lightest neutral Higgs boson up to 140 GeV with top squarks below the TeV scale, all couplings perturbative to the gauge unification scale, and with no need to fine-tune parameters in the scalar potential. This model more closely resembles the MSSM phenomenologically than the canonical next-to-minimal supersymmetric standard model.

  17. Knowledge modelling and reliability processing: presentation of the Figaro language and associated tools

    International Nuclear Information System (INIS)

    Bouissou, M.; Villatte, N.; Bouhadana, H.; Bannelier, M.

    1991-12-01

    EDF has been developing for several years an integrated set of knowledge-based and algorithmic tools for automation of reliability assessment of complex (especially sequential) systems. In this environment, the reliability expert has at his disposal all the powerful software tools for qualitative and quantitative processing, besides he gets various means to generate automatically the inputs for these tools, through the acquisition of graphical data. The development of these tools has been based on FIGARO, a specific language, which was built to get an homogeneous system modelling. Various compilers and interpreters get a FIGARO model into conventional models, such as fault-trees, Markov chains, Petri Networks. In this report, we introduce the main basics of FIGARO language, illustrating them with examples

  18. Implicit and explicit interethnic attitudes and ethnic discrimination in hiring

    NARCIS (Netherlands)

    Blommaert, E.C.C.A.; Tubergen, F.A. van; Coenders, M.T.A.

    2012-01-01

    We study effects of explicit and implicit interethnic attitudes on ethnic discrimination in hiring. Unlike explicit attitudes, implicit attitudes are characterised by reduced controllability, awareness or intention. Effects of implicit interethnic attitudes on ethnic discrimination in the labour

  19. Implicit but not explicit self-esteem predicts future depressive symptomatology.

    Science.gov (United States)

    Franck, Erik; De Raedt, Rudi; De Houwer, Jan

    2007-10-01

    To date, research on the predictive validity of implicit self-esteem for depressive relapse is very sparse. In the present study, we assessed implicit self-esteem using the Name Letter Preference Task and explicit self-esteem using the Rosenberg self-esteem scale in a group of currently depressed patients, formerly depressed individuals, and never depressed controls. In addition, we examined the predictive validity of explicit, implicit, and the interaction of explicit and implicit self-esteem in predicting future symptoms of depression in formerly depressed individuals and never depressed controls. The results showed that currently depressed individuals reported a lower explicit self-esteem as compared to formerly depressed individuals and never depressed controls. In line with previous research, all groups showed a positive implicit self-esteem not different from each other. Furthermore, after controlling for initial depressive symptomatology, implicit but not explicit self-esteem significantly predicted depressive symptoms at six months follow-up. Although implicit self-esteem assessed with the Name Letter Preference Test was not different between formerly depressed individuals and never depressed controls, the findings suggest it is an interesting variable in the study of vulnerability for depression relapse.

  20. Co-occurrence of social anxiety and depression symptoms in adolescence: differential links with implicit and explicit self-esteem?

    Science.gov (United States)

    de Jong, P J; Sportel, B E; de Hullu, E; Nauta, M H

    2012-03-01

    Social anxiety and depression often co-occur. As low self-esteem has been identified as a risk factor for both types of symptoms, it may help to explain their co-morbidity. Current dual process models of psychopathology differentiate between explicit and implicit self-esteem. Explicit self-esteem would reflect deliberate self-evaluative processes whereas implicit self-esteem would reflect simple associations in memory. Previous research suggests that low explicit self-esteem is involved in both social anxiety and depression whereas low implicit self-esteem is only involved in social anxiety. We tested whether the association between symptoms of social phobia and depression can indeed be explained by low explicit self-esteem, whereas low implicit self-esteem is only involved in social anxiety. Adolescents during the first stage of secondary education (n=1806) completed the Revised Child Anxiety and Depression Scale (RCADS) to measure symptoms of social anxiety and depression, the Rosenberg Self-Esteem Scale (RSES) to index explicit self-esteem and the Implicit Association Test (IAT) to measure implicit self-esteem. There was a strong association between symptoms of depression and social anxiety that could be largely explained by participants' explicit self-esteem. Only for girls did implicit self-esteem and the interaction between implicit and explicit self-esteem show small cumulative predictive validity for social anxiety, indicating that the association between low implicit self-esteem and social anxiety was most evident for girls with relatively low explicit self-esteem. Implicit self-esteem showed no significant predictive validity for depressive symptoms. The findings support the view that both shared and differential self-evaluative processes are involved in depression and social anxiety.

  1. Can Explaining Less Be More? Enhancing Vocabulary through Explicit versus Elaborative Storytelling

    Science.gov (United States)

    Vaahtoranta, Enni; Suggate, Sebastian; Jachmann, Cornelia; Lenhart, Jan; Lenhard, Wolfgang

    2018-01-01

    Shared reading represents an established practice to foster preschool vocabulary development, particularly when coupled with explicit instruction in word meanings. However, a question remains as to whether explicit word definitions detract from story delivery and hence language learning. Accordingly, this study compared explicit versus…

  2. Efficacy of an explicit handwriting program.

    Science.gov (United States)

    Kaiser, Marie-Laure; Albaret, Jean-Michel; Doudin, Pierre-André

    2011-04-01

    The aim of this study was to investigate the effects of an explicit handwriting program introduced during the first grade of elementary school. Grade 1 children (N=23) with an age range of 6.1 to 7.4 yr. (15 girls, 8 boys) were administered an additional handwriting program of two weekly sessions of 45 min. over six weeks. Another group of 19 Grade 1 children (11 girls, 8 boys) received only the regular handwriting program of one weekly session. The Concise Assessment Scale for Children's Handwriting was administered to measure the changes in quality and speed of handwriting. The children given the explicit program showed better quality and speed of handwriting than did the control group. Their handwriting was more regular, with fewer ambiguous letters and fewer incorrect relative heights.

  3. Review. Supporting problem structuring with computer-based tools in participatory forest planning

    Directory of Open Access Journals (Sweden)

    T. Hujala

    2013-07-01

    Full Text Available Aim of study: This review presents the state-of-art of using computerized techniques for problem structuring (PS in participatory forest planning. Frequency and modes of using different computerized tool types and their contribution for planning processes as well as critical observations are described, followed by recommendations on how to better integrate PS with the use of forest decision support systems.Area of study: The reviewed research cases are from Asia, Europe, North-America, Africa and Australia.Materials and methods: Via Scopus search and screening of abstracts, 32 research articles from years 2002–2011 were selected for review. Explicit and implicit evidence of using computerized tools for PS was recorded and assessed with content-driven qualitative analysis.Main results: GIS and forest-specific simulation tools were the most prevalent software types whereas cognitive modelling software and spreadsheet and calculation tools were less frequently used, followed by multi-criteria and interactive tools. The typical use type was to provide outputs of simulation–optimization or spatial analysis to negotiation situations or to compile summaries or illustrations afterwards; using software during group negotiation to foster interaction was observed only in a few cases.Research highlights: Expertise in both decision support systems and group learning is needed to better integrate PS and computerized decision analysis. From the knowledge management perspective, it is recommended to consider how the results of PS – e.g. conceptual models – could be stored into a problem perception database, and how PS and decision making could be streamlined by retrievals from such systems.Keywords: facilitated modeling; group negotiation; knowledge management; natural resource management; PSM; soft OR; stakeholders.

  4. DsixTools: the standard model effective field theory toolkit

    Energy Technology Data Exchange (ETDEWEB)

    Celis, Alejandro [Ludwig-Maximilians-Universitaet Muenchen, Fakultaet fuer Physik, Arnold Sommerfeld Center for Theoretical Physics, Munich (Germany); Fuentes-Martin, Javier; Vicente, Avelino [Universitat de Valencia-CSIC, Instituto de Fisica Corpuscular, Valencia (Spain); Virto, Javier [University of Bern, Albert Einstein Center for Fundamental Physics, Institute for Theoretical Physics, Bern (Switzerland)

    2017-06-15

    We present DsixTools, a Mathematica package for the handling of the dimension-six standard model effective field theory. Among other features, DsixTools allows the user to perform the full one-loop renormalization group evolution of the Wilson coefficients in the Warsaw basis. This is achieved thanks to the SMEFTrunner module, which implements the full one-loop anomalous dimension matrix previously derived in the literature. In addition, DsixTools also contains modules devoted to the matching to the ΔB = ΔS = 1, 2 and ΔB = ΔC = 1 operators of the Weak Effective Theory at the electroweak scale, and their QCD and QED Renormalization group evolution below the electroweak scale. (orig.)

  5. An integrated user-friendly ArcMAP tool for bivariate statistical modeling in geoscience applications

    Science.gov (United States)

    Jebur, M. N.; Pradhan, B.; Shafri, H. Z. M.; Yusof, Z.; Tehrany, M. S.

    2014-10-01

    Modeling and classification difficulties are fundamental issues in natural hazard assessment. A geographic information system (GIS) is a domain that requires users to use various tools to perform different types of spatial modeling. Bivariate statistical analysis (BSA) assists in hazard modeling. To perform this analysis, several calculations are required and the user has to transfer data from one format to another. Most researchers perform these calculations manually by using Microsoft Excel or other programs. This process is time consuming and carries a degree of uncertainty. The lack of proper tools to implement BSA in a GIS environment prompted this study. In this paper, a user-friendly tool, BSM (bivariate statistical modeler), for BSA technique is proposed. Three popular BSA techniques such as frequency ratio, weights-of-evidence, and evidential belief function models are applied in the newly proposed ArcMAP tool. This tool is programmed in Python and is created by a simple graphical user interface, which facilitates the improvement of model performance. The proposed tool implements BSA automatically, thus allowing numerous variables to be examined. To validate the capability and accuracy of this program, a pilot test area in Malaysia is selected and all three models are tested by using the proposed program. Area under curve is used to measure the success rate and prediction rate. Results demonstrate that the proposed program executes BSA with reasonable accuracy. The proposed BSA tool can be used in numerous applications, such as natural hazard, mineral potential, hydrological, and other engineering and environmental applications.

  6. An integrated user-friendly ArcMAP tool for bivariate statistical modelling in geoscience applications

    Science.gov (United States)

    Jebur, M. N.; Pradhan, B.; Shafri, H. Z. M.; Yusoff, Z. M.; Tehrany, M. S.

    2015-03-01

    Modelling and classification difficulties are fundamental issues in natural hazard assessment. A geographic information system (GIS) is a domain that requires users to use various tools to perform different types of spatial modelling. Bivariate statistical analysis (BSA) assists in hazard modelling. To perform this analysis, several calculations are required and the user has to transfer data from one format to another. Most researchers perform these calculations manually by using Microsoft Excel or other programs. This process is time-consuming and carries a degree of uncertainty. The lack of proper tools to implement BSA in a GIS environment prompted this study. In this paper, a user-friendly tool, bivariate statistical modeler (BSM), for BSA technique is proposed. Three popular BSA techniques, such as frequency ratio, weight-of-evidence (WoE), and evidential belief function (EBF) models, are applied in the newly proposed ArcMAP tool. This tool is programmed in Python and created by a simple graphical user interface (GUI), which facilitates the improvement of model performance. The proposed tool implements BSA automatically, thus allowing numerous variables to be examined. To validate the capability and accuracy of this program, a pilot test area in Malaysia is selected and all three models are tested by using the proposed program. Area under curve (AUC) is used to measure the success rate and prediction rate. Results demonstrate that the proposed program executes BSA with reasonable accuracy. The proposed BSA tool can be used in numerous applications, such as natural hazard, mineral potential, hydrological, and other engineering and environmental applications.

  7. Lightning-produced NOx in an explicit electrical scheme: a STERAO case study

    Science.gov (United States)

    Barthe, C.; Pinty, J.; Mari, C.

    2006-12-01

    An explicit lightning-produced nitrogen oxide scheme has been implemented in the French mesoscale model Meso-NH. The electrical scheme simulates explicitly the whole electric charge life cycle: charge separation, transfer, transport and neutralization by lightning flashes. The frequency and the 3D morphology of the lightning flashes are reproduced realistically. Therefore, fresh nitrogen oxide molecules can be added along the complex flash path as a function of the pressure, as suggested by results from laboratory experiments. No integral constraint on the total LNOx production at the cloud scale is added. The scheme is tested on the 10 July 1996, STERAO (Stratosphere-Troposphere Experiment-Radiation, Aerosols, and Ozone) storm. The model reproduces many features of the observed increase of electrical activity and LNOx flux through the anvil between the multicell and supercell stages. A large amount of LNOx is selectively produced in the upper part of the cells close to the updraft cores. Instantaneous peak concentrations exceed a few ppbv, as observed. The computed flux of NOx across the anvil compares favorably with the observations. The NOx production is estimated to 36 moles per lightning flash.

  8. A Meta-Analysis Suggests Different Neural Correlates for Implicit and Explicit Learning.

    Science.gov (United States)

    Loonis, Roman F; Brincat, Scott L; Antzoulatos, Evan G; Miller, Earl K

    2017-10-11

    A meta-analysis of non-human primates performing three different tasks (Object-Match, Category-Match, and Category-Saccade associations) revealed signatures of explicit and implicit learning. Performance improved equally following correct and error trials in the Match (explicit) tasks, but it improved more after correct trials in the Saccade (implicit) task, a signature of explicit versus implicit learning. Likewise, error-related negativity, a marker for error processing, was greater in the Match (explicit) tasks. All tasks showed an increase in alpha/beta (10-30 Hz) synchrony after correct choices. However, only the implicit task showed an increase in theta (3-7 Hz) synchrony after correct choices that decreased with learning. In contrast, in the explicit tasks, alpha/beta synchrony increased with learning and decreased thereafter. Our results suggest that explicit versus implicit learning engages different neural mechanisms that rely on different patterns of oscillatory synchrony. Copyright © 2017 Elsevier Inc. All rights reserved.

  9. Individual differences in explicit and implicit visuomotor learning and working memory capacity.

    Science.gov (United States)

    Christou, Antonios I; Miall, R Chris; McNab, Fiona; Galea, Joseph M

    2016-11-08

    The theoretical basis for the association between high working memory capacity (WMC) and enhanced visuomotor adaptation is unknown. Visuomotor adaptation involves interplay between explicit and implicit systems. We examined whether the positive association between adaptation and WMC is specific to the explicit component of adaptation. Experiment 1 replicated the positive correlation between WMC and adaptation, but revealed this was specific to the explicit component of adaptation, and apparently driven by a sub-group of participants who did not show any explicit adaptation in the correct direction. A negative correlation was observed between WMC and implicit learning. Experiments 2 and 3 showed that when the task restricted the development of an explicit strategy, high WMC was no longer associated with enhanced adaptation. This work reveals that the benefit of high WMC is specifically linked to an individual's capacity to use an explicit strategy. It also reveals an important contribution of individual differences in determining how adaptation is performed.

  10. Explicit formulation of second and third order optical nonlinearity in the FDTD framework

    Science.gov (United States)

    Varin, Charles; Emms, Rhys; Bart, Graeme; Fennel, Thomas; Brabec, Thomas

    2018-01-01

    The finite-difference time-domain (FDTD) method is a flexible and powerful technique for rigorously solving Maxwell's equations. However, three-dimensional optical nonlinearity in current commercial and research FDTD softwares requires solving iteratively an implicit form of Maxwell's equations over the entire numerical space and at each time step. Reaching numerical convergence demands significant computational resources and practical implementation often requires major modifications to the core FDTD engine. In this paper, we present an explicit method to include second and third order optical nonlinearity in the FDTD framework based on a nonlinear generalization of the Lorentz dispersion model. A formal derivation of the nonlinear Lorentz dispersion equation is equally provided, starting from the quantum mechanical equations describing nonlinear optics in the two-level approximation. With the proposed approach, numerical integration of optical nonlinearity and dispersion in FDTD is intuitive, transparent, and fully explicit. A strong-field formulation is also proposed, which opens an interesting avenue for FDTD-based modelling of the extreme nonlinear optics phenomena involved in laser filamentation and femtosecond micromachining of dielectrics.

  11. A High-Resolution Spatially Explicit Monte-Carlo Simulation Approach to Commercial and Residential Electricity and Water Demand Modeling

    Energy Technology Data Exchange (ETDEWEB)

    Morton, April M [ORNL; McManamay, Ryan A [ORNL; Nagle, Nicholas N [ORNL; Piburn, Jesse O [ORNL; Stewart, Robert N [ORNL; Surendran Nair, Sujithkumar [ORNL

    2016-01-01

    Abstract As urban areas continue to grow and evolve in a world of increasing environmental awareness, the need for high resolution spatially explicit estimates for energy and water demand has become increasingly important. Though current modeling efforts mark significant progress in the effort to better understand the spatial distribution of energy and water consumption, many are provided at a course spatial resolution or rely on techniques which depend on detailed region-specific data sources that are not publicly available for many parts of the U.S. Furthermore, many existing methods do not account for errors in input data sources and may therefore not accurately reflect inherent uncertainties in model outputs. We propose an alternative and more flexible Monte-Carlo simulation approach to high-resolution residential and commercial electricity and water consumption modeling that relies primarily on publicly available data sources. The method s flexible data requirement and statistical framework ensure that the model is both applicable to a wide range of regions and reflective of uncertainties in model results. Key words: Energy Modeling, Water Modeling, Monte-Carlo Simulation, Uncertainty Quantification Acknowledgment This manuscript has been authored by employees of UT-Battelle, LLC, under contract DE-AC05-00OR22725 with the U.S. Department of Energy. Accordingly, the United States Government retains and the publisher, by accepting the article for publication, acknowledges that the United States Government retains a non-exclusive, paid-up, irrevocable, world-wide license to publish or reproduce the published form of this manuscript, or allow others to do so, for United States Government purposes.

  12. Transformation of UML models to CSP : a case study for graph transformation tools

    NARCIS (Netherlands)

    Varró, D.; Asztalos, M.; Bisztray, D.; Boronat, A.; Dang, D.; Geiß, R.; Greenyer, J.; Van Gorp, P.M.E.; Kniemeyer, O.; Narayanan, A.; Rencis, E.; Weinell, E.; Schürr, A.; Nagl, M.; Zündorf, A.

    2008-01-01

    Graph transformation provides an intuitive mechanism for capturing model transformations. In the current paper, we investigate and compare various graph transformation tools using a compact practical model transformation case study carried out as part of the AGTIVE 2007 Tool Contest [22]. The aim of

  13. Integrating spatially explicit indices of abundance and habitat quality: an applied example for greater sage-grouse management.

    Science.gov (United States)

    Coates, Peter S; Casazza, Michael L; Ricca, Mark A; Brussee, Brianne E; Blomberg, Erik J; Gustafson, K Benjamin; Overton, Cory T; Davis, Dawn M; Niell, Lara E; Espinosa, Shawn P; Gardner, Scott C; Delehanty, David J

    2016-02-01

    Predictive species distributional models are a cornerstone of wildlife conservation planning. Constructing such models requires robust underpinning science that integrates formerly disparate data types to achieve effective species management.Greater sage-grouse Centrocercus urophasianus , hereafter 'sage-grouse' populations are declining throughout sagebrush-steppe ecosystems in North America, particularly within the Great Basin, which heightens the need for novel management tools that maximize the use of available information.Herein, we improve upon existing species distribution models by combining information about sage-grouse habitat quality, distribution and abundance from multiple data sources. To measure habitat, we created spatially explicit maps depicting habitat selection indices (HSI) informed by >35 500 independent telemetry locations from >1600 sage-grouse collected over 15 years across much of the Great Basin. These indices were derived from models that accounted for selection at different spatial scales and seasons. A region-wide HSI was calculated using the HSI surfaces modelled for 12 independent subregions and then demarcated into distinct habitat quality classes.We also employed a novel index to describe landscape patterns of sage-grouse abundance and space use (AUI). The AUI is a probabilistic composite of the following: (i) breeding density patterns based on the spatial configuration of breeding leks and associated trends in male attendance; and (ii) year-round patterns of space use indexed by the decreasing probability of use with increasing distance to leks. The continuous AUI surface was then reclassified into two classes representing high and low/no use and abundance. Synthesis and application s. Using the example of sage-grouse, we demonstrate how the joint application of indices of habitat selection, abundance and space use derived from multiple data sources yields a composite map that can guide effective allocation of management

  14. Integrating spatially explicit indices of abundance and habitat quality: an applied example for greater sage-grouse management

    Science.gov (United States)

    Coates, Peter S.; Casazza, Michael L.; Ricca, Mark A.; Brussee, Brianne E.; Blomberg, Erik J.; Gustafson, K. Benjamin; Overton, Cory T.; Davis, Dawn M.; Niell, Lara E.; Espinosa, Shawn P.; Gardner, Scott C.; Delehanty, David J.

    2016-01-01

    Predictive species distributional models are a cornerstone of wildlife conservation planning. Constructing such models requires robust underpinning science that integrates formerly disparate data types to achieve effective species management. Greater sage-grouse Centrocercus urophasianus, hereafter “sage-grouse” populations are declining throughout sagebrush-steppe ecosystems in North America, particularly within the Great Basin, which heightens the need for novel management tools that maximize use of available information. Herein, we improve upon existing species distribution models by combining information about sage-grouse habitat quality, distribution, and abundance from multiple data sources. To measure habitat, we created spatially explicit maps depicting habitat selection indices (HSI) informed by > 35 500 independent telemetry locations from > 1600 sage-grouse collected over 15 years across much of the Great Basin. These indices were derived from models that accounted for selection at different spatial scales and seasons. A region-wide HSI was calculated using the HSI surfaces modelled for 12 independent subregions and then demarcated into distinct habitat quality classes. We also employed a novel index to describe landscape patterns of sage-grouse abundance and space use (AUI). The AUI is a probabilistic composite of: (i) breeding density patterns based on the spatial configuration of breeding leks and associated trends in male attendance; and (ii) year-round patterns of space use indexed by the decreasing probability of use with increasing distance to leks. The continuous AUI surface was then reclassified into two classes representing high and low/no use and abundance. Synthesis and applications. Using the example of sage-grouse, we demonstrate how the joint application of indices of habitat selection, abundance, and space use derived from multiple data sources yields a composite map that can guide effective allocation of management intensity across

  15. Modelling rapid subsurface flow at the hillslope scale with explicit representation of preferential flow paths

    Science.gov (United States)

    Wienhöfer, J.; Zehe, E.

    2012-04-01

    Rapid lateral flow processes via preferential flow paths are widely accepted to play a key role for rainfall-runoff response in temperate humid headwater catchments. A quantitative description of these processes, however, is still a major challenge in hydrological research, not least because detailed information about the architecture of subsurface flow paths are often impossible to obtain at a natural site without disturbing the system. Our study combines physically based modelling and field observations with the objective to better understand how flow network configurations influence the hydrological response of hillslopes. The system under investigation is a forested hillslope with a small perennial spring at the study area Heumöser, a headwater catchment of the Dornbirnerach in Vorarlberg, Austria. In-situ points measurements of field-saturated hydraulic conductivity and dye staining experiments at the plot scale revealed that shrinkage cracks and biogenic macropores function as preferential flow paths in the fine-textured soils of the study area, and these preferential flow structures were active in fast subsurface transport of artificial tracers at the hillslope scale. For modelling of water and solute transport, we followed the approach of implementing preferential flow paths as spatially explicit structures of high hydraulic conductivity and low retention within the 2D process-based model CATFLOW. Many potential configurations of the flow path network were generated as realisations of a stochastic process informed by macropore characteristics derived from the plot scale observations. Together with different realisations of soil hydraulic parameters, this approach results in a Monte Carlo study. The model setups were used for short-term simulation of a sprinkling and tracer experiment, and the results were evaluated against measured discharges and tracer breakthrough curves. Although both criteria were taken for model evaluation, still several model setups

  16. Explicit formula for a fundamental class of functions

    OpenAIRE

    Avdispahić, Muharem; Smajlović, Lejla

    2005-01-01

    The purpose of this paper is to prove an analogue of A. Weil's explicit formula for a fundamental class of functions, i.e. the class of meromorphic functions that have an Euler sum representation and satisfy certain a functional equation. The advance of this explicit formula is that it enlarges the class of allowed test functions, from the class of functions with bounded Jordan variation to the class of functions of $\\phi $-bounded variation. A condition posed to the test fu...

  17. BPMNDiffViz : a tool for BPMN models comparison

    NARCIS (Netherlands)

    Ivanov, S.Y.; Kalenkova, A.A.; Aalst, van der W.M.P.; Daniel, F.; Zugal, S.

    2015-01-01

    Automatic comparison of business processes plays an important role in their analysis and optimization. In this paper we present the web-based tool BPMNDiffViz, that finds business processes discrepancies and visualizes them. BPMN (Business Process Model and Notation) 2.0 - one of the most commonly

  18. Managers' implicit and explicit risk-attitudes in managerial decision making

    NARCIS (Netherlands)

    Bittner, Jenny; Landwehr, Julia; Hertel, Guido; Binnewies, Carmen; Krumm, Stefan; Holling, Heinz; Kleinmar, Martin

    2013-01-01

    Purpose We examined the contribution of implicit and explicit risk-attitudes to the prediction of risky management decisions. Indirect methods allow for the measurement of implicit attitudes, while self-report is typically used to measure explicit, reflective attitudes. Indirect methods make it

  19. Explicit formulas for Clebsch-Gordan coefficients

    International Nuclear Information System (INIS)

    Rudnicki-Bujnowski, G.

    1975-01-01

    The problem is to obtain explicit algebraic formulas of Clebsch-Gordan coefficients for high values of angular momentum. The method of solution is an algebraic method based on the Racah formula using the FORMAC programming language. (Auth.)

  20. Modeling spatially explicit fire impact on gross primary production in interior Alaska using satellite images coupled with eddy covariance

    Science.gov (United States)

    Huang, Shengli; Liu, Heping; Dahal, Devendra; Jin, Suming; Welp, Lisa R.; Liu, Jinxun; Liu, Shuguang

    2013-01-01

    In interior Alaska, wildfires change gross primary production (GPP) after the initial disturbance. The impact of fires on GPP is spatially heterogeneous, which is difficult to evaluate by limited point-based comparisons or is insufficient to assess by satellite vegetation index. The direct prefire and postfire comparison is widely used, but the recovery identification may become biased due to interannual climate variability. The objective of this study is to propose a method to quantify the spatially explicit GPP change caused by fires and succession. We collected three Landsat images acquired on 13 July 2004, 5 August 2004, and 6 September 2004 to examine the GPP recovery of burned area from 1987 to 2004. A prefire Landsat image acquired in 1986 was used to reconstruct satellite images assuming that the fires of 1987–2004 had not occurred. We used a light-use efficiency model to estimate the GPP. This model was driven by maximum light-use efficiency (Emax) and fraction of photosynthetically active radiation absorbed by vegetation (FPAR). We applied this model to two scenarios (i.e., an actual postfire scenario and an assuming-no-fire scenario), where the changes in Emax and FPAR were taken into account. The changes in Emax were represented by the change in land cover of evergreen needleleaf forest, deciduous broadleaf forest, and shrub/grass mixed, whose Emax was determined from three fire chronosequence flux towers as 1.1556, 1.3336, and 0.5098 gC/MJ PAR. The changes in FPAR were inferred from NDVI change between the actual postfire NDVI and the reconstructed NDVI. After GPP quantification for July, August, and September 2004, we calculated the difference between the two scenarios in absolute and percent GPP changes. Our results showed rapid recovery of GPP post-fire with a 24% recovery immediately after burning and 43% one year later. For the fire scars with an age range of 2–17 years, the recovery rate ranged from 54% to 95%. In addition to the averaging